You can download spark-hive_2.10-1.6.2.jar in this page.
Apache Open Source
spark-hive_2.10-1.6.2.jar file has the following types.
META-INF/DEPENDENCIES META-INF/LICENSE META-INF/MANIFEST.MF META-INF/NOTICE META-INF/maven/org.apache.spark/spark-hive_2.10/pom.properties META-INF/maven/org.apache.spark/spark-hive_2.10/pom.xml META-INF/maven/org.spark-project.spark/unused/pom.properties META-INF/maven/org.spark-project.spark/unused/pom.xml META-INF/services/org.apache.spark.sql.sources.DataSourceRegister org.apache.spark.sql.hive.CreateTableAsSelect.class org.apache.spark.sql.hive.CreateViewAsSelect.class org.apache.spark.sql.hive.CurrentDatabase.class org.apache.spark.sql.hive.DeferredObjectAdapter.class org.apache.spark.sql.hive.ExtendedHiveQlParser.class org.apache.spark.sql.hive.HadoopTableReader.class org.apache.spark.sql.hive.HiveContext.class org.apache.spark.sql.hive.HiveFunctionRegistry.class org.apache.spark.sql.hive.HiveGenericUDF.class org.apache.spark.sql.hive.HiveGenericUDTF.class org.apache.spark.sql.hive.HiveInspectors.class org.apache.spark.sql.hive.HiveMetastoreCatalog.class org.apache.spark.sql.hive.HiveMetastoreTypes.class org.apache.spark.sql.hive.HiveQLDialect.class org.apache.spark.sql.hive.HiveQl.class org.apache.spark.sql.hive.HiveSerDe.class org.apache.spark.sql.hive.HiveShim.class org.apache.spark.sql.hive.HiveSimpleUDF.class org.apache.spark.sql.hive.HiveStrategies.class org.apache.spark.sql.hive.HiveTableUtil.class org.apache.spark.sql.hive.HiveUDAFFunction.class org.apache.spark.sql.hive.HiveWindowFunction.class org.apache.spark.sql.hive.InsertIntoHiveTable.class org.apache.spark.sql.hive.MetastoreRelation.class org.apache.spark.sql.hive.NativePlaceholder.class org.apache.spark.sql.hive.ResolveHiveWindowFunction.class org.apache.spark.sql.hive.SparkHiveDynamicPartitionWriterContainer.class org.apache.spark.sql.hive.SparkHiveWriterContainer.class org.apache.spark.sql.hive.TableReader.class org.apache.spark.sql.hive.client.ClientInterface.class org.apache.spark.sql.hive.client.ClientWrapper.class org.apache.spark.sql.hive.client.ExternalTable.class org.apache.spark.sql.hive.client.HiveColumn.class org.apache.spark.sql.hive.client.HiveDatabase.class org.apache.spark.sql.hive.client.HivePartition.class org.apache.spark.sql.hive.client.HiveStorageDescriptor.class org.apache.spark.sql.hive.client.HiveTable.class org.apache.spark.sql.hive.client.IndexTable.class org.apache.spark.sql.hive.client.IsolatedClientLoader.class org.apache.spark.sql.hive.client.ManagedTable.class org.apache.spark.sql.hive.client.NonClosableMutableURLClassLoader.class org.apache.spark.sql.hive.client.Shim.class org.apache.spark.sql.hive.client.Shim_v0_12.class org.apache.spark.sql.hive.client.Shim_v0_13.class org.apache.spark.sql.hive.client.Shim_v0_14.class org.apache.spark.sql.hive.client.Shim_v1_0.class org.apache.spark.sql.hive.client.Shim_v1_1.class org.apache.spark.sql.hive.client.Shim_v1_2.class org.apache.spark.sql.hive.client.TableType.class org.apache.spark.sql.hive.client.VirtualView.class org.apache.spark.sql.hive.client.package.class org.apache.spark.sql.hive.execution.AddFile.class org.apache.spark.sql.hive.execution.AddJar.class org.apache.spark.sql.hive.execution.AnalyzeTable.class org.apache.spark.sql.hive.execution.CreateMetastoreDataSource.class org.apache.spark.sql.hive.execution.CreateMetastoreDataSourceAsSelect.class org.apache.spark.sql.hive.execution.CreateTableAsSelect.class org.apache.spark.sql.hive.execution.CreateViewAsSelect.class org.apache.spark.sql.hive.execution.DescribeHiveTableCommand.class org.apache.spark.sql.hive.execution.DropTable.class org.apache.spark.sql.hive.execution.HiveNativeCommand.class org.apache.spark.sql.hive.execution.HiveScriptIOSchema.class org.apache.spark.sql.hive.execution.HiveTableScan.class org.apache.spark.sql.hive.execution.InsertIntoHiveTable.class org.apache.spark.sql.hive.execution.ScriptTransformation.class org.apache.spark.sql.hive.execution.ScriptTransformationWriterThread.class org.apache.spark.sql.hive.orc.DefaultSource.class org.apache.spark.sql.hive.orc.OrcFileOperator.class org.apache.spark.sql.hive.orc.OrcFilters.class org.apache.spark.sql.hive.orc.OrcOutputWriter.class org.apache.spark.sql.hive.orc.OrcRelation.class org.apache.spark.sql.hive.orc.OrcTableScan.class org.apache.spark.sql.hive.package.class org.apache.spark.sql.hive.test.TestHive.class org.apache.spark.sql.hive.test.TestHiveContext.class org.apache.spark.sql.hive.test.TestHiveSingleton.class org.apache.spark.unused.UnusedStubClass.class
spark-hive_2.10-1.6.2.pom file content.
<?xml version="1.0" encoding="UTF-8"?> <project xmlns="http://maven.apache.org/POM/4.0.0" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://maven.apache.org/POM/4.0.0 http://maven.apache.org/maven-v4_0_0.xsd"> <parent> <artifactId>spark-parent_2.10</artifactId> <groupId>org.apache.spark</groupId> <version>1.6.2</version> <relativePath>../../pom.xml</relativePath> </parent> <modelVersion>4.0.0</modelVersion> <groupId>org.apache.spark</groupId> <artifactId>spark-hive_2.10</artifactId> <name>Spark Project Hive</name> <url>http://spark.apache.org/</url> <build> <outputDirectory>target/scala-${scala.binary.version}/classes</outputDirectory> <testOutputDirectory>target/scala-${scala.binary.version}/test-classes</testOutputDirectory> <plugins> <plugin> <groupId>org.scalatest</groupId> <artifactId>scalatest-maven-plugin</artifactId> <configuration> <argLine>-da -Xmx3g -XX:MaxPermSize=${MaxPermGen} -XX:ReservedCodeCacheSize=512m</argLine> </configuration> </plugin> <plugin> <groupId>org.codehaus.mojo</groupId> <artifactId>build-helper-maven-plugin</artifactId> <executions> <execution> <id>add-default-sources</id> <phase>generate-sources</phase> <goals> <goal>add-source</goal> </goals> <configuration> <sources> <source>v${hive.version.short}/src/main/scala</source> </sources> </configuration> </execution> </executions> </plugin> <plugin> <artifactId>maven-dependency-plugin</artifactId> <executions> <execution> <id>copy-dependencies</id> <phase>package</phase> <goals> <goal>copy-dependencies</goal> </goals> <configuration> <outputDirectory>${basedir}/../../lib_managed/jars</outputDirectory> <overWriteReleases>false</overWriteReleases> <overWriteSnapshots>false</overWriteSnapshots> <overWriteIfNewer>true</overWriteIfNewer> <includeGroupIds>org.datanucleus</includeGroupIds> </configuration> </execution> </executions> </plugin> </plugins> </build> <profiles> <profile> <id>hive</id> <build> <plugins> <plugin> <groupId>org.codehaus.mojo</groupId> <artifactId>build-helper-maven-plugin</artifactId> <executions> <execution> <id>add-scala-test-sources</id> <phase>generate-test-sources</phase> <goals> <goal>add-test-source</goal> </goals> <configuration> <sources> <source>compatibility/src/test/scala</source> </sources> </configuration> </execution> </executions> </plugin> </plugins> </build> </profile> </profiles> <dependencies> <dependency> <groupId>com.twitter</groupId> <artifactId>parquet-hadoop-bundle</artifactId> <version>1.6.0</version> <scope>compile</scope> </dependency> <dependency> <groupId>org.apache.spark</groupId> <artifactId>spark-core_2.10</artifactId> <version>1.6.2</version> <scope>compile</scope> <exclusions> <exclusion> <artifactId>jetty-plus</artifactId> <groupId>org.eclipse.jetty</groupId> </exclusion> <exclusion> <artifactId>jetty-security</artifactId> <groupId>org.eclipse.jetty</groupId> </exclusion> <exclusion> <artifactId>jetty-util</artifactId> <groupId>org.eclipse.jetty</groupId> </exclusion> <exclusion> <artifactId>jetty-server</artifactId> <groupId>org.eclipse.jetty</groupId> </exclusion> <exclusion> <artifactId>jetty-http</artifactId> <groupId>org.eclipse.jetty</groupId> </exclusion> <exclusion> <artifactId>jetty-continuation</artifactId> <groupId>org.eclipse.jetty</groupId> </exclusion> <exclusion> <artifactId>jetty-servlet</artifactId> <groupId>org.eclipse.jetty</groupId> </exclusion> </exclusions> </dependency> <dependency> <groupId>org.apache.spark</groupId> <artifactId>spark-core_2.10</artifactId> <version>1.6.2</version> <type>test-jar</type> <scope>test</scope> <exclusions> <exclusion> <artifactId>jetty-plus</artifactId> <groupId>org.eclipse.jetty</groupId> </exclusion> <exclusion> <artifactId>jetty-security</artifactId> <groupId>org.eclipse.jetty</groupId> </exclusion> <exclusion> <artifactId>jetty-util</artifactId> <groupId>org.eclipse.jetty</groupId> </exclusion> <exclusion> <artifactId>jetty-server</artifactId> <groupId>org.eclipse.jetty</groupId> </exclusion> <exclusion> <artifactId>jetty-http</artifactId> <groupId>org.eclipse.jetty</groupId> </exclusion> <exclusion> <artifactId>jetty-continuation</artifactId> <groupId>org.eclipse.jetty</groupId> </exclusion> <exclusion> <artifactId>jetty-servlet</artifactId> <groupId>org.eclipse.jetty</groupId> </exclusion> </exclusions> </dependency> <dependency> <groupId>org.apache.spark</groupId> <artifactId>spark-sql_2.10</artifactId> <version>1.6.2</version> <scope>compile</scope> </dependency> <dependency> <groupId>org.apache.spark</groupId> <artifactId>spark-test-tags_2.10</artifactId> <version>1.6.2</version> <scope>test</scope> </dependency> <dependency> <groupId>org.spark-project.hive</groupId> <artifactId>hive-cli</artifactId> <version>1.2.1.spark</version> <scope>compile</scope> <exclusions> <exclusion> <artifactId>hive-common</artifactId> <groupId>org.spark-project.hive</groupId> </exclusion> <exclusion> <artifactId>hive-exec</artifactId> <groupId>org.spark-project.hive</groupId> </exclusion> <exclusion> <artifactId>hive-jdbc</artifactId> <groupId>org.spark-project.hive</groupId> </exclusion> <exclusion> <artifactId>hive-metastore</artifactId> <groupId>org.spark-project.hive</groupId> </exclusion> <exclusion> <artifactId>hive-serde</artifactId> <groupId>org.spark-project.hive</groupId> </exclusion> <exclusion> <artifactId>hive-service</artifactId> <groupId>org.spark-project.hive</groupId> </exclusion> <exclusion> <artifactId>hive-shims</artifactId> <groupId>org.spark-project.hive</groupId> </exclusion> <exclusion> <artifactId>libthrift</artifactId> <groupId>org.apache.thrift</groupId> </exclusion> <exclusion> <artifactId>slf4j-api</artifactId> <groupId>org.slf4j</groupId> </exclusion> <exclusion> <artifactId>slf4j-log4j12</artifactId> <groupId>org.slf4j</groupId> </exclusion> <exclusion> <artifactId>log4j</artifactId> <groupId>log4j</groupId> </exclusion> <exclusion> <artifactId>commons-logging</artifactId> <groupId>commons-logging</groupId> </exclusion> </exclusions> </dependency> <dependency> <groupId>org.spark-project.hive</groupId> <artifactId>hive-exec</artifactId> <version>1.2.1.spark</version> <scope>compile</scope> <exclusions> <exclusion> <artifactId>hive-metastore</artifactId> <groupId>org.spark-project.hive</groupId> </exclusion> <exclusion> <artifactId>hive-shims</artifactId> <groupId>org.spark-project.hive</groupId> </exclusion> <exclusion> <artifactId>hive-ant</artifactId> <groupId>org.spark-project.hive</groupId> </exclusion> <exclusion> <artifactId>spark-client</artifactId> <groupId>org.spark-project.hive</groupId> </exclusion> <exclusion> <artifactId>ant</artifactId> <groupId>ant</groupId> </exclusion> <exclusion> <artifactId>ant</artifactId> <groupId>org.apache.ant</groupId> </exclusion> <exclusion> <artifactId>kryo</artifactId> <groupId>com.esotericsoftware.kryo</groupId> </exclusion> <exclusion> <artifactId>commons-codec</artifactId> <groupId>commons-codec</groupId> </exclusion> <exclusion> <artifactId>commons-httpclient</artifactId> <groupId>commons-httpclient</groupId> </exclusion> <exclusion> <artifactId>avro-mapred</artifactId> <groupId>org.apache.avro</groupId> </exclusion> <exclusion> <artifactId>calcite-core</artifactId> <groupId>org.apache.calcite</groupId> </exclusion> <exclusion> <artifactId>apache-curator</artifactId> <groupId>org.apache.curator</groupId> </exclusion> <exclusion> <artifactId>curator-client</artifactId> <groupId>org.apache.curator</groupId> </exclusion> <exclusion> <artifactId>curator-framework</artifactId> <groupId>org.apache.curator</groupId> </exclusion> <exclusion> <artifactId>libthrift</artifactId> <groupId>org.apache.thrift</groupId> </exclusion> <exclusion> <artifactId>libfb303</artifactId> <groupId>org.apache.thrift</groupId> </exclusion> <exclusion> <artifactId>zookeeper</artifactId> <groupId>org.apache.zookeeper</groupId> </exclusion> <exclusion> <artifactId>slf4j-api</artifactId> <groupId>org.slf4j</groupId> </exclusion> <exclusion> <artifactId>slf4j-log4j12</artifactId> <groupId>org.slf4j</groupId> </exclusion> <exclusion> <artifactId>log4j</artifactId> <groupId>log4j</groupId> </exclusion> <exclusion> <artifactId>commons-logging</artifactId> <groupId>commons-logging</groupId> </exclusion> <exclusion> <artifactId>groovy-all</artifactId> <groupId>org.codehaus.groovy</groupId> </exclusion> <exclusion> <artifactId>guava</artifactId> <groupId>com.google.guava</groupId> </exclusion> </exclusions> </dependency> <dependency> <groupId>org.spark-project.hive</groupId> <artifactId>hive-metastore</artifactId> <version>1.2.1.spark</version> <scope>compile</scope> <exclusions> <exclusion> <artifactId>hive-serde</artifactId> <groupId>org.spark-project.hive</groupId> </exclusion> <exclusion> <artifactId>hive-shims</artifactId> <groupId>org.spark-project.hive</groupId> </exclusion> <exclusion> <artifactId>libfb303</artifactId> <groupId>org.apache.thrift</groupId> </exclusion> <exclusion> <artifactId>libthrift</artifactId> <groupId>org.apache.thrift</groupId> </exclusion> <exclusion> <artifactId>servlet-api</artifactId> <groupId>org.mortbay.jetty</groupId> </exclusion> <exclusion> <artifactId>guava</artifactId> <groupId>com.google.guava</groupId> </exclusion> <exclusion> <artifactId>slf4j-api</artifactId> <groupId>org.slf4j</groupId> </exclusion> <exclusion> <artifactId>slf4j-log4j12</artifactId> <groupId>org.slf4j</groupId> </exclusion> </exclusions> </dependency> <dependency> <groupId>org.apache.avro</groupId> <artifactId>avro</artifactId> <version>1.7.7</version> <scope>compile</scope> </dependency> <dependency> <groupId>org.apache.avro</groupId> <artifactId>avro-mapred</artifactId> <version>1.7.7</version> <classifier>hadoop2</classifier> <scope>compile</scope> <exclusions> <exclusion> <artifactId>netty</artifactId> <groupId>io.netty</groupId> </exclusion> <exclusion> <artifactId>jetty</artifactId> <groupId>org.mortbay.jetty</groupId> </exclusion> <exclusion> <artifactId>jetty-util</artifactId> <groupId>org.mortbay.jetty</groupId> </exclusion> <exclusion> <artifactId>servlet-api</artifactId> <groupId>org.mortbay.jetty</groupId> </exclusion> <exclusion> <artifactId>velocity</artifactId> <groupId>org.apache.velocity</groupId> </exclusion> </exclusions> </dependency> <dependency> <groupId>commons-httpclient</groupId> <artifactId>commons-httpclient</artifactId> <version>3.1</version> <scope>compile</scope> </dependency> <dependency> <groupId>org.apache.calcite</groupId> <artifactId>calcite-avatica</artifactId> <version>1.2.0-incubating</version> <scope>compile</scope> <exclusions> <exclusion> <artifactId>jackson-annotations</artifactId> <groupId>com.fasterxml.jackson.core</groupId> </exclusion> <exclusion> <artifactId>jackson-core</artifactId> <groupId>com.fasterxml.jackson.core</groupId> </exclusion> <exclusion> <artifactId>jackson-databind</artifactId> <groupId>com.fasterxml.jackson.core</groupId> </exclusion> </exclusions> </dependency> <dependency> <groupId>org.apache.calcite</groupId> <artifactId>calcite-core</artifactId> <version>1.2.0-incubating</version> <scope>compile</scope> <exclusions> <exclusion> <artifactId>jackson-annotations</artifactId> <groupId>com.fasterxml.jackson.core</groupId> </exclusion> <exclusion> <artifactId>jackson-core</artifactId> <groupId>com.fasterxml.jackson.core</groupId> </exclusion> <exclusion> <artifactId>jackson-databind</artifactId> <groupId>com.fasterxml.jackson.core</groupId> </exclusion> <exclusion> <artifactId>guava</artifactId> <groupId>com.google.guava</groupId> </exclusion> <exclusion> <artifactId>jsr305</artifactId> <groupId>com.google.code.findbugs</groupId> </exclusion> <exclusion> <artifactId>janino</artifactId> <groupId>org.codehaus.janino</groupId> </exclusion> <exclusion> <artifactId>hsqldb</artifactId> <groupId>org.hsqldb</groupId> </exclusion> <exclusion> <artifactId>pentaho-aggdesigner-algorithm</artifactId> <groupId>org.pentaho</groupId> </exclusion> </exclusions> </dependency> <dependency> <groupId>org.apache.httpcomponents</groupId> <artifactId>httpclient</artifactId> <version>4.3.2</version> <scope>compile</scope> </dependency> <dependency> <groupId>org.codehaus.jackson</groupId> <artifactId>jackson-mapper-asl</artifactId> <version>1.9.13</version> <scope>compile</scope> </dependency> <dependency> <groupId>commons-codec</groupId> <artifactId>commons-codec</artifactId> <version>1.10</version> <scope>compile</scope> </dependency> <dependency> <groupId>joda-time</groupId> <artifactId>joda-time</artifactId> <version>2.9.3</version> <scope>compile</scope> </dependency> <dependency> <groupId>org.jodd</groupId> <artifactId>jodd-core</artifactId> <version>3.5.2</version> <scope>compile</scope> </dependency> <dependency> <groupId>com.google.code.findbugs</groupId> <artifactId>jsr305</artifactId> <version>1.3.9</version> <scope>compile</scope> </dependency> <dependency> <groupId>org.datanucleus</groupId> <artifactId>datanucleus-core</artifactId> <version>3.2.10</version> <scope>compile</scope> </dependency> <dependency> <groupId>org.apache.thrift</groupId> <artifactId>libthrift</artifactId> <version>0.9.2</version> <scope>compile</scope> <exclusions> <exclusion> <artifactId>httpclient</artifactId> <groupId>org.apache.httpcomponents</groupId> </exclusion> <exclusion> <artifactId>httpcore</artifactId> <groupId>org.apache.httpcomponents</groupId> </exclusion> <exclusion> <artifactId>slf4j-api</artifactId> <groupId>org.slf4j</groupId> </exclusion> </exclusions> </dependency> <dependency> <groupId>org.apache.thrift</groupId> <artifactId>libfb303</artifactId> <version>0.9.2</version> <scope>compile</scope> <exclusions> <exclusion> <artifactId>httpclient</artifactId> <groupId>org.apache.httpcomponents</groupId> </exclusion> <exclusion> <artifactId>httpcore</artifactId> <groupId>org.apache.httpcomponents</groupId> </exclusion> <exclusion> <artifactId>slf4j-api</artifactId> <groupId>org.slf4j</groupId> </exclusion> </exclusions> </dependency> <dependency> <groupId>org.scalacheck</groupId> <artifactId>scalacheck_2.10</artifactId> <version>1.11.3</version> <scope>test</scope> <exclusions> <exclusion> <artifactId>test-interface</artifactId> <groupId>org.scala-sbt</groupId> </exclusion> </exclusions> </dependency> <dependency> <groupId>org.apache.spark</groupId> <artifactId>spark-sql_2.10</artifactId> <version>1.6.2</version> <type>test-jar</type> <scope>test</scope> </dependency> <dependency> <groupId>org.apache.spark</groupId> <artifactId>spark-catalyst_2.10</artifactId> <version>1.6.2</version> <type>test-jar</type> <scope>test</scope> </dependency> <dependency> <groupId>org.scalatest</groupId> <artifactId>scalatest_2.10</artifactId> <version>2.2.1</version> <scope>test</scope> </dependency> <dependency> <groupId>junit</groupId> <artifactId>junit</artifactId> <version>4.11</version> <scope>test</scope> <exclusions> <exclusion> <artifactId>hamcrest-core</artifactId> <groupId>org.hamcrest</groupId> </exclusion> </exclusions> </dependency> <dependency> <groupId>com.novocode</groupId> <artifactId>junit-interface</artifactId> <version>0.11</version> <scope>test</scope> <exclusions> <exclusion> <artifactId>test-interface</artifactId> <groupId>org.scala-sbt</groupId> </exclusion> </exclusions> </dependency> </dependencies> <properties> <sbt.project.name>hive</sbt.project.name> </properties> </project>
<dependency> <groupId>org.apache.spark</groupId> <artifactId>spark-hive_2.10</artifactId> <version>1.6.2</version> </dependency>
If you think the following spark-hive_2.10-1.6.2.jar downloaded from Maven central repository is inappropriate, such as containing malicious code/tools or violating the copyright, please email , thanks.
Download spark-hive_2.10-1.6.2.jar file