List of usage examples for org.apache.hadoop.conf Configuration setClass
public void setClass(String name, Class<?> theClass, Class<?> xface)
name
property to the name of a theClass
implementing the given interface xface
. From source file:org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIOTest.java
License:Apache License
/** * This test validates functionality of {@link HadoopInputFormatIO.Read#withConfiguration() * withConfiguration()} function when Hadoop InputFormat class is not provided by the user in * configuration.//from www .ja v a 2 s .co m */ @Test public void testReadValidationFailsMissingInputFormatInConf() { Configuration configuration = new Configuration(); configuration.setClass("key.class", Text.class, Object.class); configuration.setClass("value.class", Employee.class, Object.class); thrown.expect(NullPointerException.class); HadoopInputFormatIO.<Text, Employee>read().withConfiguration(configuration); }
From source file:org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIOTest.java
License:Apache License
/** * This test validates functionality of {@link HadoopInputFormatIO.Read#withConfiguration() * withConfiguration()} function when key class is not provided by the user in configuration. *//* ww w. j a v a 2 s . com*/ @Test public void testReadValidationFailsMissingKeyClassInConf() { Configuration configuration = new Configuration(); configuration.setClass("mapreduce.job.inputformat.class", EmployeeInputFormat.class, InputFormat.class); configuration.setClass("value.class", Employee.class, Object.class); thrown.expect(NullPointerException.class); HadoopInputFormatIO.<Text, Employee>read().withConfiguration(configuration); }
From source file:org.apache.beam.sdk.io.hadoop.inputformat.HadoopInputFormatIOTest.java
License:Apache License
/** * This test validates functionality of {@link HadoopInputFormatIO.Read#withConfiguration() * withConfiguration()} function when value class is not provided by the user in configuration. *//*from w ww. jav a 2 s . c o m*/ @Test public void testReadValidationFailsMissingValueClassInConf() { Configuration configuration = new Configuration(); configuration.setClass("mapreduce.job.inputformat.class", EmployeeInputFormat.class, InputFormat.class); configuration.setClass("key.class", Text.class, Object.class); thrown.expect(NullPointerException.class); HadoopInputFormatIO.<Text, Employee>read().withConfiguration(configuration); }
From source file:org.apache.beam.sdk.io.hadoop.inputformat.HIFIOCassandraIT.java
License:Apache License
/** * Returns Hadoop configuration for reading data from Cassandra. To read data from Cassandra using * HadoopInputFormatIO, following properties must be set: InputFormat class, InputFormat key * class, InputFormat value class, Thrift address, Thrift port, partitioner class, keyspace and * columnfamily name./*ww w . j a v a 2s .c o m*/ */ private static Configuration getConfiguration(HIFITestOptions options) { Configuration conf = new Configuration(); conf.set(CASSANDRA_THRIFT_PORT_PROPERTY, options.getCassandraServerPort().toString()); conf.set(CASSANDRA_THRIFT_ADDRESS_PROPERTY, options.getCassandraServerIp()); conf.set(CASSANDRA_PARTITIONER_CLASS_PROPERTY, CASSANDRA_PARTITIONER_CLASS_VALUE); conf.set(CASSANDRA_KEYSPACE_PROPERTY, CASSANDRA_KEYSPACE); conf.set(CASSANDRA_COLUMNFAMILY_PROPERTY, CASSANDRA_TABLE); // Set user name and password if Cassandra instance has security configured. conf.set(USERNAME, options.getCassandraUserName()); conf.set(PASSWORD, options.getCassandraPassword()); conf.set(INPUT_KEYSPACE_USERNAME_CONFIG, options.getCassandraUserName()); conf.set(INPUT_KEYSPACE_PASSWD_CONFIG, options.getCassandraPassword()); conf.setClass("mapreduce.job.inputformat.class", org.apache.cassandra.hadoop.cql3.CqlInputFormat.class, InputFormat.class); conf.setClass("key.class", java.lang.Long.class, Object.class); conf.setClass("value.class", Row.class, Object.class); return conf; }
From source file:org.apache.beam.sdk.io.hadoop.inputformat.HIFIOElasticIT.java
License:Apache License
/** * Returns Hadoop configuration for reading data from Elasticsearch. Configuration object should * have InputFormat class, key class and value class to be set. Mandatory fields for ESInputFormat * to be set are es.resource, es.nodes, es.port, es.internal.es.version, es.nodes.wan.only. Please * refer <a href="https://www.elastic.co/guide/en/elasticsearch/hadoop/current/configuration.html" * >Elasticsearch Configuration</a> for more details. *///from w w w . j a v a 2 s . co m private static Configuration getConfiguration(HIFITestOptions options) { Configuration conf = new Configuration(); conf.set(ConfigurationOptions.ES_NODES, options.getElasticServerIp()); conf.set(ConfigurationOptions.ES_PORT, options.getElasticServerPort().toString()); conf.set(ConfigurationOptions.ES_NODES_WAN_ONLY, TRUE); // Set username and password if Elasticsearch is configured with security. conf.set(ConfigurationOptions.ES_NET_HTTP_AUTH_USER, options.getElasticUserName()); conf.set(ConfigurationOptions.ES_NET_HTTP_AUTH_PASS, options.getElasticPassword()); conf.set(ConfigurationOptions.ES_RESOURCE, ELASTIC_RESOURCE); conf.set("es.internal.es.version", ELASTIC_INTERNAL_VERSION); conf.set(ConfigurationOptions.ES_INDEX_AUTO_CREATE, TRUE); conf.setClass("mapreduce.job.inputformat.class", org.elasticsearch.hadoop.mr.EsInputFormat.class, InputFormat.class); conf.setClass("key.class", Text.class, Object.class); conf.setClass("value.class", LinkedMapWritable.class, Object.class); // Optimizations added to change the max docs per partition, scroll size and batch size of // bytes to improve the test time for large data conf.set("es.input.max.docs.per.partition", "50000"); conf.set("es.scroll.size", "400"); conf.set("es.batch.size.bytes", "8mb"); return conf; }
From source file:org.apache.beam.sdk.io.hadoop.inputformat.HIFIOWithCassandraTest.java
License:Apache License
/** * Returns configuration of CqlInutFormat. Mandatory parameters required apart from inputformat * class name, key class, value class are thrift port, thrift address, partitioner class, keyspace * and columnfamily name/*from www. jav a2s . co m*/ */ public Configuration getConfiguration() { Configuration conf = new Configuration(); conf.set(CASSANDRA_THRIFT_PORT_PROPERTY, CASSANDRA_PORT); conf.set(CASSANDRA_THRIFT_ADDRESS_PROPERTY, CASSANDRA_HOST); conf.set(CASSANDRA_PARTITIONER_CLASS_PROPERTY, CASSANDRA_PARTITIONER_CLASS_VALUE); conf.set(CASSANDRA_KEYSPACE_PROPERTY, CASSANDRA_KEYSPACE); conf.set(CASSANDRA_COLUMNFAMILY_PROPERTY, CASSANDRA_TABLE); conf.setClass(HadoopInputFormatIOConstants.INPUTFORMAT_CLASSNAME, org.apache.cassandra.hadoop.cql3.CqlInputFormat.class, InputFormat.class); conf.setClass(HadoopInputFormatIOConstants.KEY_CLASS, java.lang.Long.class, Object.class); conf.setClass(HadoopInputFormatIOConstants.VALUE_CLASS, com.datastax.driver.core.Row.class, Object.class); return conf; }
From source file:org.apache.beam.sdk.io.hadoop.inputformat.HIFIOWithElasticTest.java
License:Apache License
/** * Set the Elasticsearch configuration parameters in the Hadoop configuration object. * Configuration object should have InputFormat class, key class and value class set. Mandatory * fields for ESInputFormat to be set are es.resource, es.nodes, es.port, es.internal.es.version. * Please refer to//from ww w . j a v a 2 s . com * <a href="https://www.elastic.co/guide/en/elasticsearch/hadoop/current/configuration.html" * >Elasticsearch Configuration</a> for more details. */ public Configuration getConfiguration() { Configuration conf = new Configuration(); conf.set(ConfigurationOptions.ES_NODES, ELASTIC_IN_MEM_HOSTNAME); conf.set(ConfigurationOptions.ES_PORT, String.format("%s", ELASTIC_IN_MEM_PORT)); conf.set(ConfigurationOptions.ES_RESOURCE, ELASTIC_RESOURCE); conf.set("es.internal.es.version", ELASTIC_INTERNAL_VERSION); conf.set(ConfigurationOptions.ES_NODES_DISCOVERY, TRUE); conf.set(ConfigurationOptions.ES_INDEX_AUTO_CREATE, TRUE); conf.setClass("mapreduce.job.inputformat.class", org.elasticsearch.hadoop.mr.EsInputFormat.class, InputFormat.class); conf.setClass("key.class", Text.class, Object.class); conf.setClass("value.class", LinkedMapWritable.class, Object.class); return conf; }
From source file:org.apache.beam.sdk.io.hadoop.inputformat.HIFIOWithEmbeddedCassandraTest.java
License:Apache License
/** * Returns configuration of CqlInutFormat. Mandatory parameters required apart from inputformat * class name, key class, value class are thrift port, thrift address, partitioner class, keyspace * and columnfamily name// w w w. j a v a 2 s . com */ private Configuration getConfiguration() { Configuration conf = new Configuration(); conf.set(CASSANDRA_THRIFT_PORT_PROPERTY, CASSANDRA_PORT); conf.set(CASSANDRA_THRIFT_ADDRESS_PROPERTY, CASSANDRA_HOST); conf.set(CASSANDRA_PARTITIONER_CLASS_PROPERTY, CASSANDRA_PARTITIONER_CLASS_VALUE); conf.set(CASSANDRA_KEYSPACE_PROPERTY, CASSANDRA_KEYSPACE); conf.set(CASSANDRA_COLUMNFAMILY_PROPERTY, CASSANDRA_TABLE); conf.setClass("mapreduce.job.inputformat.class", org.apache.cassandra.hadoop.cql3.CqlInputFormat.class, InputFormat.class); conf.setClass("key.class", java.lang.Long.class, Object.class); conf.setClass("value.class", Row.class, Object.class); return conf; }
From source file:org.apache.beam.sdk.io.hadoop.inputformat.integration.tests.HIFIOCassandraIT.java
License:Apache License
/** * Returns Hadoop configuration for reading data from Cassandra. To read data from Cassandra using * HadoopInputFormatIO, following properties must be set: InputFormat class, InputFormat key * class, InputFormat value class, Thrift address, Thrift port, partitioner class, keyspace and * columnfamily name.//from w w w . ja va 2 s . com */ private static Configuration getConfiguration(HIFTestOptions options) { Configuration conf = new Configuration(); conf.set(CASSANDRA_THRIFT_PORT_PROPERTY, options.getCassandraServerPort().toString()); conf.set(CASSANDRA_THRIFT_ADDRESS_PROPERTY, options.getCassandraServerIp()); conf.set(CASSANDRA_PARTITIONER_CLASS_PROPERTY, CASSANDRA_PARTITIONER_CLASS_VALUE); conf.set(CASSANDRA_KEYSPACE_PROPERTY, CASSANDRA_KEYSPACE); conf.set(CASSANDRA_COLUMNFAMILY_PROPERTY, CASSANDRA_TABLE); // Set user name and password if Cassandra instance has security configured. conf.set(USERNAME, options.getCassandraUserName()); conf.set(PASSWORD, options.getCassandraPassword()); conf.set(INPUT_KEYSPACE_USERNAME_CONFIG, options.getCassandraUserName()); conf.set(INPUT_KEYSPACE_PASSWD_CONFIG, options.getCassandraPassword()); conf.setClass("mapreduce.job.inputformat.class", org.apache.cassandra.hadoop.cql3.CqlInputFormat.class, InputFormat.class); conf.setClass("key.class", java.lang.Long.class, Object.class); conf.setClass("value.class", com.datastax.driver.core.Row.class, Object.class); return conf; }
From source file:org.apache.beam.sdk.io.hadoop.inputformat.integration.tests.HIFIOElasticIT.java
License:Apache License
/** * Returns Hadoop configuration for reading data from Elasticsearch. Configuration object should * have InputFormat class, key class and value class to be set. Mandatory fields for ESInputFormat * to be set are es.resource, es.nodes, es.port, es.internal.es.version, es.nodes.wan.only. Please * refer <a href="https://www.elastic.co/guide/en/elasticsearch/hadoop/current/configuration.html" * >Elasticsearch Configuration</a> for more details. *//*from w w w. jav a2 s . c om*/ private static Configuration getConfiguration(HIFTestOptions options) { Configuration conf = new Configuration(); conf.set(ConfigurationOptions.ES_NODES, options.getElasticServerIp()); conf.set(ConfigurationOptions.ES_PORT, options.getElasticServerPort().toString()); conf.set(ConfigurationOptions.ES_NODES_WAN_ONLY, TRUE); // Set username and password if Elasticsearch is configured with security. conf.set(ConfigurationOptions.ES_NET_HTTP_AUTH_USER, options.getElasticUserName()); conf.set(ConfigurationOptions.ES_NET_HTTP_AUTH_PASS, options.getElasticPassword()); conf.set(ConfigurationOptions.ES_RESOURCE, ELASTIC_RESOURCE); conf.set("es.internal.es.version", ELASTIC_INTERNAL_VERSION); conf.set(ConfigurationOptions.ES_INDEX_AUTO_CREATE, TRUE); conf.setClass("mapreduce.job.inputformat.class", org.elasticsearch.hadoop.mr.EsInputFormat.class, InputFormat.class); conf.setClass("key.class", Text.class, Object.class); conf.setClass("value.class", LinkedMapWritable.class, Object.class); // Optimizations added to change the max docs per partition, scroll size and batch size of // bytes to improve the test time for large data conf.set("es.input.max.docs.per.partition", "50000"); conf.set("es.scroll.size", "400"); conf.set("es.batch.size.bytes", "8mb"); return conf; }