org.apache.drill.exec.fn.impl.TestNewMathFunctions.java Source code

Java tutorial

Introduction

Here is the source code for org.apache.drill.exec.fn.impl.TestNewMathFunctions.java

Source

/*
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 *
 * http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package org.apache.drill.exec.fn.impl;

import org.apache.drill.categories.OperatorTest;
import org.apache.drill.categories.UnlikelyTest;
import org.apache.drill.common.config.DrillConfig;
import org.apache.drill.exec.ExecConstants;
import org.apache.drill.exec.expr.fn.FunctionImplementationRegistry;
import org.apache.drill.exec.ops.FragmentContextImpl;
import org.apache.drill.exec.physical.PhysicalPlan;
import org.apache.drill.exec.physical.base.FragmentRoot;
import org.apache.drill.exec.physical.impl.ImplCreator;
import org.apache.drill.exec.physical.impl.SimpleRootExec;
import org.apache.drill.exec.planner.PhysicalPlanReader;
import org.apache.drill.exec.planner.PhysicalPlanReaderTestFactory;
import org.apache.drill.exec.proto.BitControl.PlanFragment;
import org.apache.drill.exec.rpc.UserClientConnection;
import org.apache.drill.exec.server.DrillbitContext;
import org.apache.drill.exec.vector.ValueVector;
import org.apache.drill.exec.vector.VarCharVector;
import org.apache.drill.test.BaseTestQuery;

import org.apache.hadoop.conf.Configuration;
import org.apache.parquet.column.ParquetProperties;
import org.apache.parquet.example.data.Group;
import org.apache.parquet.example.data.simple.SimpleGroupFactory;
import org.apache.parquet.hadoop.ParquetWriter;
import org.apache.parquet.hadoop.example.GroupWriteSupport;
import org.apache.parquet.hadoop.metadata.CompressionCodecName;
import org.apache.parquet.schema.MessageType;
import org.apache.parquet.schema.MessageTypeParser;
import org.apache.commons.io.FileUtils;

import org.junit.Test;
import org.junit.experimental.categories.Category;
import org.mockito.Mockito;

import com.google.common.base.Charsets;
import com.google.common.io.Resources;

import java.io.File;
import java.math.BigDecimal;

import static org.junit.Assert.assertEquals;
import static org.junit.Assert.assertTrue;

@Category({ UnlikelyTest.class, OperatorTest.class })
public class TestNewMathFunctions extends BaseTestQuery {
    private final DrillConfig c = DrillConfig.create();
    private PhysicalPlanReader reader;
    private FunctionImplementationRegistry registry;
    private FragmentContextImpl context;

    public Object[] getRunResult(SimpleRootExec exec) {
        int size = 0;
        for (final ValueVector v : exec) {
            size++;
        }

        final Object[] res = new Object[size];
        int i = 0;
        for (final ValueVector v : exec) {
            if (v instanceof VarCharVector) {
                res[i++] = new String(((VarCharVector) v).getAccessor().get(0));
            } else {
                res[i++] = v.getAccessor().getObject(0);
            }
        }
        return res;
    }

    public void runTest(Object[] expectedResults, String planPath) throws Throwable {

        final DrillbitContext bitContext = mockDrillbitContext();
        final UserClientConnection connection = Mockito.mock(UserClientConnection.class);

        final String planString = Resources.toString(Resources.getResource(planPath), Charsets.UTF_8);
        if (reader == null) {
            reader = PhysicalPlanReaderTestFactory.defaultPhysicalPlanReader(c);
        }
        if (registry == null) {
            registry = new FunctionImplementationRegistry(c);
        }
        if (context == null) {
            context = new FragmentContextImpl(bitContext, PlanFragment.getDefaultInstance(), connection, registry);
        }
        final PhysicalPlan plan = reader.readPhysicalPlan(planString);
        final SimpleRootExec exec = new SimpleRootExec(
                ImplCreator.getExec(context, (FragmentRoot) plan.getSortedOperators(false).iterator().next()));

        while (exec.next()) {
            final Object[] res = getRunResult(exec);
            assertEquals("return count does not match", expectedResults.length, res.length);

            for (int i = 0; i < res.length; i++) {
                assertEquals(String.format("column %s does not match", i), res[i], expectedResults[i]);
            }
        }

        if (context.getExecutorState().getFailureCause() != null) {
            throw context.getExecutorState().getFailureCause();
        }

        assertTrue(!context.getExecutorState().isFailed());
    }

    @Test
    public void testTrigoMathFunc() throws Throwable {
        final Object[] expected = new Object[] { Math.sin(45), Math.cos(45), Math.tan(45), Math.asin(45),
                Math.acos(45), Math.atan(45), Math.sinh(45), Math.cosh(45), Math.tanh(45) };
        runTest(expected, "functions/testTrigoMathFunctions.json");
    }

    @Test
    public void testExtendedMathFunc() throws Throwable {
        final BigDecimal d = new BigDecimal(
                "100111111111111111111111111111111111.00000000000000000000000000000000000000000000000000001");
        final Object[] expected = new Object[] { Math.cbrt(1000), Math.log(10), Math.log10(5),
                (Math.log(64.0) / Math.log(2.0)), Math.exp(10), Math.toDegrees(0.5), Math.toRadians(45.0), Math.PI,
                Math.cbrt(d.doubleValue()), Math.log(d.doubleValue()), (Math.log(d.doubleValue()) / Math.log(2)),
                Math.exp(d.doubleValue()), Math.toDegrees(d.doubleValue()), Math.toRadians(d.doubleValue()) };
        runTest(expected, "functions/testExtendedMathFunctions.json");
    }

    @Test
    public void testTruncDivMod() throws Throwable {
        final Object[] expected = new Object[] { 101.0, 0, 101, 1010.0, 101, 481.0, 0.001099999999931267 };
        runTest(expected, "functions/testDivModTruncFunctions.json");
    }

    @Test
    public void testIsNumeric() throws Throwable {
        final Object[] expected = new Object[] { 1, 1, 1, 0 };
        runTest(expected, "functions/testIsNumericFunction.json");
    }

    @Test
    public void testLog10WithDouble() throws Throwable {
        String json = "{" + "\"positive_infinity\" : Infinity," + "\"negative_infinity\" : -Infinity,"
                + "\"nan\" : NaN," + "\"num1\": 0.0," + "\"num2\": 0.1," + "\"num3\": 1.0," + "\"num4\": 1.5,"
                + "\"num5\": -1.5," + "\"num6\": 10.0" + "}";
        String query = "select " + "log10(positive_infinity) as pos_inf, " + "log10(negative_infinity) as neg_inf, "
                + "log10(nan) as nan, " + "log10(num1) as num1, " + "log10(num2) as num2, "
                + "log10(num3) as num3, " + "log10(num4) as num4, " + "log10(num5) as num5, "
                + "log10(num6) as num6 " + "from dfs.`data.json`";
        File file = new File(dirTestWatcher.getRootDir(), "data.json");
        try {
            FileUtils.writeStringToFile(file, json);
            setSessionOption(ExecConstants.JSON_READ_NUMBERS_AS_DOUBLE, true);
            setSessionOption(ExecConstants.JSON_READER_NAN_INF_NUMBERS, true);
            testBuilder().sqlQuery(query).ordered()
                    .baselineColumns("pos_inf", "neg_inf", "nan", "num1", "num2", "num3", "num4", "num5", "num6")
                    .baselineValues(Double.POSITIVE_INFINITY, Double.NaN, Double.NaN, Double.NEGATIVE_INFINITY,
                            -1.0d, 0d, 0.17609125905568124d, Double.NaN, 1.0d)
                    .go();
        } finally {
            resetSessionOption(ExecConstants.JSON_READ_NUMBERS_AS_DOUBLE);
            resetSessionOption(ExecConstants.JSON_READER_NAN_INF_NUMBERS);
            FileUtils.deleteQuietly(file);
        }
    }

    @Test
    public void testLog10WithFloat() throws Throwable {
        String json = "{" + "\"positive_infinity\" : Infinity," + "\"negative_infinity\" : -Infinity,"
                + "\"nan\" : NaN," + "\"num1\": 0.0," + "\"num2\": 0.1," + "\"num3\": 1.0," + "\"num4\": 1.5,"
                + "\"num5\": -1.5," + "\"num6\": 10.0" + "}";
        String query = "select " + "log10(cast(positive_infinity as float)) as pos_inf, "
                + "log10(cast(negative_infinity as float)) as neg_inf, " + "log10(cast(nan as float)) as nan, "
                + "log10(cast(num1 as float)) as num1, " + "log10(cast(num2 as float)) as num2, "
                + "log10(cast(num3 as float)) as num3, " + "log10(cast(num4 as float)) as num4, "
                + "log10(cast(num5 as float)) as num5, " + "log10(cast(num6 as float)) as num6 "
                + "from dfs.`data.json`";
        File file = new File(dirTestWatcher.getRootDir(), "data.json");
        try {
            FileUtils.writeStringToFile(file, json);
            setSessionOption(ExecConstants.JSON_READER_NAN_INF_NUMBERS, true);
            testBuilder().sqlQuery(query).ordered()
                    .baselineColumns("pos_inf", "neg_inf", "nan", "num1", "num2", "num3", "num4", "num5", "num6")
                    .baselineValues(Double.POSITIVE_INFINITY, Double.NaN, Double.NaN, Double.NEGATIVE_INFINITY,
                            -0.999999993528508d, 0d, 0.17609125905568124d, Double.NaN, 1.0d)
                    .go();
        } finally {
            resetSessionOption(ExecConstants.JSON_READER_NAN_INF_NUMBERS);
            FileUtils.deleteQuietly(file);
        }
    }

    @Test
    public void testLog10WithInt() throws Throwable {
        String json = "{" + "\"num1\": 0.0," + "\"num3\": 1.0," + "\"num5\": -1.0," + "\"num6\": 10.0" + "}";
        String query = "select " + "log10(cast(num1 as int)) as num1, " + "log10(cast(num3 as int)) as num3, "
                + "log10(cast(num5 as int)) as num5, " + "log10(cast(num6 as int)) as num6 "
                + "from dfs.`data.json`";
        File file = new File(dirTestWatcher.getRootDir(), "data.json");
        try {
            FileUtils.writeStringToFile(file, json);
            setSessionOption(ExecConstants.JSON_READER_NAN_INF_NUMBERS, true);
            testBuilder().sqlQuery(query).ordered().baselineColumns("num1", "num3", "num5", "num6")
                    .baselineValues(Double.NEGATIVE_INFINITY, 0d, Double.NaN, 1.0d).go();
        } finally {
            resetSessionOption(ExecConstants.JSON_READER_NAN_INF_NUMBERS);
            FileUtils.deleteQuietly(file);
        }
    }

    @Test
    public void testLog10WithBigInt() throws Throwable {
        String json = "{" + "\"num1\": 0," + "\"num3\": 1," + "\"num5\": -1," + "\"num6\": 10" + "}";
        String query = "select " + "log10(num1) as num1, " + "log10(num3) as num3, " + "log10(num5) as num5, "
                + "log10(num6) as num6 " + "from dfs.`data.json`";
        File file = new File(dirTestWatcher.getRootDir(), "data.json");
        try {
            FileUtils.writeStringToFile(file, json);

            setSessionOption(ExecConstants.JSON_READER_NAN_INF_NUMBERS, true);
            testBuilder().sqlQuery(query).ordered().baselineColumns("num1", "num3", "num5", "num6")
                    .baselineValues(Double.NEGATIVE_INFINITY, 0d, Double.NaN, 1.0d).go();
        } finally {
            resetSessionOption(ExecConstants.JSON_READER_NAN_INF_NUMBERS);
            FileUtils.deleteQuietly(file);
        }
    }

    @Test
    public void testLog10WithUint4() throws Exception {
        org.apache.hadoop.fs.Path file = new org.apache.hadoop.fs.Path(dirTestWatcher.getRootDir().toString(),
                "uint8.parquet");
        String schemaText = "message test { required int32 val(UINT_8); }";

        Configuration conf = new Configuration();
        MessageType schema = MessageTypeParser.parseMessageType(schemaText);
        GroupWriteSupport.setSchema(schema, conf);
        SimpleGroupFactory groupFactory = new SimpleGroupFactory(schema);

        try {
            try (ParquetWriter<Group> writer = new ParquetWriter<>(file, new GroupWriteSupport(),
                    CompressionCodecName.UNCOMPRESSED, 1024, 1024, 512, true, false,
                    ParquetProperties.WriterVersion.PARQUET_1_0, conf)) {

                writer.write(groupFactory.newGroup().append("val", 0));
                writer.write(groupFactory.newGroup().append("val", 1));
                writer.write(groupFactory.newGroup().append("val", -1));
                writer.write(groupFactory.newGroup().append("val", 10));
            }
            String query = "select log10(val) as col from dfs.`uint8.parquet`";
            testBuilder().sqlQuery(query).unOrdered().baselineColumns("col")
                    .baselineValues(Double.NEGATIVE_INFINITY).baselineValues(0d).baselineValues(Double.NaN)
                    .baselineValues(1.0d).build().run();
        } finally {
            FileUtils.deleteQuietly(new File(file.toString()));
        }
    }

}