List of usage examples for org.apache.hadoop.conf Configuration setInt
public void setInt(String name, int value)
name
property to an int
. From source file:basic.PartitionGraph.java
License:Apache License
/** * Runs this tool./*from www. j a va2 s . c om*/ */ @SuppressWarnings({ "static-access" }) public int run(String[] args) throws Exception { Options options = new Options(); options.addOption(new Option(RANGE, "use range partitioner")); options.addOption(OptionBuilder.withArgName("path").hasArg().withDescription("input path").create(INPUT)); options.addOption(OptionBuilder.withArgName("path").hasArg().withDescription("output path").create(OUTPUT)); options.addOption( OptionBuilder.withArgName("num").hasArg().withDescription("number of nodes").create(NUM_NODES)); options.addOption(OptionBuilder.withArgName("num").hasArg().withDescription("number of partitions") .create(NUM_PARTITIONS)); CommandLine cmdline; CommandLineParser parser = new GnuParser(); try { cmdline = parser.parse(options, args); } catch (ParseException exp) { System.err.println("Error parsing command line: " + exp.getMessage()); return -1; } if (!cmdline.hasOption(INPUT) || !cmdline.hasOption(OUTPUT) || !cmdline.hasOption(NUM_NODES) || !cmdline.hasOption(NUM_PARTITIONS)) { System.out.println("args: " + Arrays.toString(args)); HelpFormatter formatter = new HelpFormatter(); formatter.setWidth(120); formatter.printHelp(this.getClass().getName(), options); ToolRunner.printGenericCommandUsage(System.out); return -1; } String inPath = cmdline.getOptionValue(INPUT); String outPath = cmdline.getOptionValue(OUTPUT); int nodeCount = Integer.parseInt(cmdline.getOptionValue(NUM_NODES)); int numParts = Integer.parseInt(cmdline.getOptionValue(NUM_PARTITIONS)); boolean useRange = cmdline.hasOption(RANGE); LOG.info("Tool name: " + PartitionGraph.class.getSimpleName()); LOG.info(" - input dir: " + inPath); LOG.info(" - output dir: " + outPath); LOG.info(" - num partitions: " + numParts); LOG.info(" - node cnt: " + nodeCount); LOG.info(" - use range partitioner: " + useRange); Configuration conf = getConf(); conf.setInt("NodeCount", nodeCount); Job job = Job.getInstance(conf); job.setJobName(PartitionGraph.class.getSimpleName() + ":" + inPath); job.setJarByClass(PartitionGraph.class); job.setNumReduceTasks(numParts); FileInputFormat.setInputPaths(job, new Path(inPath)); FileOutputFormat.setOutputPath(job, new Path(outPath)); job.setInputFormatClass(NonSplitableSequenceFileInputFormat.class); job.setOutputFormatClass(SequenceFileOutputFormat.class); job.setMapOutputKeyClass(IntWritable.class); job.setMapOutputValueClass(PageRankNode.class); job.setOutputKeyClass(IntWritable.class); job.setOutputValueClass(PageRankNode.class); if (useRange) { job.setPartitionerClass(RangePartitioner.class); } FileSystem.get(conf).delete(new Path(outPath), true); job.waitForCompletion(true); return 0; }
From source file:be.uantwerpen.adrem.bigfim.BigFIMDriver.java
License:Apache License
private static void setConfigurationValues(Configuration conf, FIMOptions opt) { conf.set(DELIMITER_KEY, opt.delimiter); conf.setInt(MIN_SUP_KEY, opt.minSup); conf.setInt(NUMBER_OF_MAPPERS_KEY, opt.nrMappers); conf.setInt(NUMBER_OF_CHUNKS, opt.nrMappers); conf.setInt(PREFIX_LENGTH_KEY, opt.prefixLength); conf.setStrings(OUTPUT_DIR_KEY, opt.outputDir); }
From source file:be.uantwerpen.adrem.bigfim.BigFIMDriver.java
License:Apache License
private void startCreatePrefixGroups(FIMOptions opt, int phase) throws IOException, ClassNotFoundException, InterruptedException, URISyntaxException { Path path = new Path(opt.outputDir + separator + "tg" + phase); for (FileStatus status : path.getFileSystem(new Configuration()).listStatus(path)) { String cacheFile = status.getPath().toString(); String trieGroupCount = cacheFile.substring(cacheFile.lastIndexOf('/'), cacheFile.length()); trieGroupCount = trieGroupCount.split("-")[1]; String outputFile = opt.outputDir + separator + "pg-trieGroup" + trieGroupCount; System.out.println("[CreatePrefixGroups]: input: " + opt.inputFile + ", output: " + opt.outputDir + ", cache: " + cacheFile); Job job = prepareJob(new Path(opt.inputFile), new Path(outputFile), SplitByKTextInputFormat.class, ComputeTidListMapper.class, Text.class, IntArrayWritable.class, ComputeTidListReducer.class, IntArrayWritable.class, IntMatrixWritable.class, SequenceFileOutputFormat.class); job.setJobName("Create Prefix Groups"); job.setJarByClass(BigFIMDriver.class); job.setNumReduceTasks(1);//from w w w .j a v a2 s .c o m Configuration conf = job.getConfiguration(); setConfigurationValues(conf, opt); conf.setInt(PREFIX_LENGTH_KEY, phase); addCacheFile(new URI(cacheFile.replace(" ", "%20")), job.getConfiguration()); runJob(job, "Prefix Creation"); } }
From source file:be.uantwerpen.adrem.bigfim.ComputeTidListReducerTest.java
License:Apache License
private Configuration createConfiguration() throws Exception { Configuration conf = new Configuration(); conf.setInt(MIN_SUP_KEY, 1); conf.setInt(NUMBER_OF_MAPPERS_KEY, 2); conf.setInt(SUBDB_SIZE, 10);/*from w w w . j a v a 2s . c om*/ conf.setStrings(OUTPUT_DIR_KEY, "file:///out"); return conf; }
From source file:be.uantwerpen.adrem.disteclat.DistEclatDriver.java
License:Apache License
/** * Passes all configuration flags to the Hadoop Configuration framework * //from w w w. j a va2 s. com * @param conf * the Hadoop configuration * @param config * the configuration that has user-defined flags */ private static void setConfigurationValues(Configuration conf, FIMOptions opt) { conf.set(DELIMITER_KEY, opt.delimiter); conf.setInt(MIN_SUP_KEY, opt.minSup); conf.setInt(NUMBER_OF_MAPPERS_KEY, opt.nrMappers); conf.setInt(NUMBER_OF_CHUNKS, opt.nrMappers); conf.setInt(PREFIX_LENGTH_KEY, opt.prefixLength); conf.setStrings(OUTPUT_DIR_KEY, opt.outputDir); }
From source file:be.ugent.intec.halvade.utils.HalvadeConf.java
License:Open Source License
public static void setVcores(Configuration conf, int val) { conf.setInt(vCores, val); }
From source file:be.ugent.intec.halvade.utils.HalvadeConf.java
License:Open Source License
public static void setMapThreads(Configuration conf, int val) { conf.setInt(mapThreads, val); }
From source file:be.ugent.intec.halvade.utils.HalvadeConf.java
License:Open Source License
public static void setReducerThreads(Configuration conf, int val) { conf.setInt(reduceThreads, val); }
From source file:be.ugent.intec.halvade.utils.HalvadeConf.java
License:Open Source License
public static void setKnownSitesOnHDFS(Configuration conf, String[] val) throws IOException, URISyntaxException { conf.setInt(numberOfSites, val.length); FileSystem fs;/*from ww w .j a v a2s . c o m*/ for (int i = 0; i < val.length; i++) { // check if dir add all files! fs = FileSystem.get(new URI(val[i]), conf); if (fs.isFile(new Path(val[i]))) { conf.set(sitesOnHDFSName + i, val[i]); } else { FileStatus[] files = fs.listStatus(new Path(val[i])); for (FileStatus file : files) { if (!file.isDir()) { conf.set(sitesOnHDFSName + i, file.getPath().toString()); } } } } }
From source file:be.ugent.intec.halvade.utils.HalvadeConf.java
License:Open Source License
public static void setMapContainerCount(Configuration conf, int val) { conf.setInt(totalContainers, val); }