Java tutorial
/* * Licensed to Metamarkets Group Inc. (Metamarkets) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. Metamarkets licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, * software distributed under the License is distributed on an * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. */ package io.druid.segment.serde; import com.google.common.hash.HashFunction; import com.google.common.hash.Hashing; import com.google.common.primitives.Longs; import io.druid.hll.HyperLogLogCollector; import io.druid.java.util.common.io.smoosh.FileSmoosher; import io.druid.java.util.common.io.smoosh.Smoosh; import io.druid.java.util.common.io.smoosh.SmooshedFileMapper; import io.druid.java.util.common.io.smoosh.SmooshedWriter; import io.druid.segment.column.Column; import io.druid.segment.column.ColumnBuilder; import io.druid.segment.column.ComplexColumn; import io.druid.segment.column.ValueType; import io.druid.segment.data.IOPeon; import io.druid.segment.data.TmpFileIOPeon; import org.apache.commons.io.FileUtils; import org.junit.Assert; import org.junit.Test; import java.io.File; import java.io.IOException; public class LargeColumnSupportedComplexColumnSerializerTest { private final HashFunction fn = Hashing.murmur3_128(); @Test public void testSanity() throws IOException { HyperUniquesSerdeForTest serde = new HyperUniquesSerdeForTest(Hashing.murmur3_128()); int[] cases = { 1000, 5000, 10000, 20000 }; int[] columnSizes = { Integer.MAX_VALUE, Integer.MAX_VALUE / 2, Integer.MAX_VALUE / 4, 5000 * Longs.BYTES, 2500 * Longs.BYTES }; for (int columnSize : columnSizes) { for (int aCase : cases) { File tmpFile = FileUtils.getTempDirectory(); HyperLogLogCollector baseCollector = HyperLogLogCollector.makeLatestCollector(); try (IOPeon peon = new TmpFileIOPeon(); FileSmoosher v9Smoosher = new FileSmoosher(tmpFile)) { LargeColumnSupportedComplexColumnSerializer serializer = LargeColumnSupportedComplexColumnSerializer .createWithColumnSize(peon, "test", serde.getObjectStrategy(), columnSize); serializer.open(); for (int i = 0; i < aCase; i++) { HyperLogLogCollector collector = HyperLogLogCollector.makeLatestCollector(); byte[] hashBytes = fn.hashLong(i).asBytes(); collector.add(hashBytes); baseCollector.fold(collector); serializer.serialize(collector); } serializer.close(); try (final SmooshedWriter channel = v9Smoosher.addWithSmooshedWriter("test", serializer.getSerializedSize())) { serializer.writeToChannel(channel, v9Smoosher); } } SmooshedFileMapper mapper = Smoosh.map(tmpFile); final ColumnBuilder builder = new ColumnBuilder().setType(ValueType.COMPLEX) .setHasMultipleValues(false).setFileMapper(mapper); serde.deserializeColumn(mapper.mapFile("test"), builder); Column column = builder.build(); ComplexColumn complexColumn = column.getComplexColumn(); HyperLogLogCollector collector = HyperLogLogCollector.makeLatestCollector(); for (int i = 0; i < aCase; i++) { collector.fold((HyperLogLogCollector) complexColumn.getRowValue(i)); } Assert.assertEquals(baseCollector.estimateCardinality(), collector.estimateCardinality(), 0.0); } } } }