List of usage examples for org.apache.hadoop.conf Configuration setInt
public void setInt(String name, int value)
name
property to an int
. From source file:edu.iu.daal_ridgereg.RidgeRegDaalLauncher.java
License:Apache License
/** * Launches all the tasks in order.// www. j a v a 2 s . c o m */ @Override public int run(String[] args) throws Exception { /* Put shared libraries into the distributed cache */ Configuration conf = this.getConf(); Initialize init = new Initialize(conf, args); /* Put shared libraries into the distributed cache */ init.loadDistributedLibs(); // load args init.loadSysArgs(); //load app args conf.setInt(HarpDAALConstants.FILE_DIM, Integer.parseInt(args[init.getSysArgNum()])); conf.setInt(HarpDAALConstants.FEATURE_DIM, Integer.parseInt(args[init.getSysArgNum() + 1])); conf.setInt(HarpDAALConstants.NUM_DEPVAR, Integer.parseInt(args[init.getSysArgNum() + 2])); conf.set(HarpDAALConstants.TEST_FILE_PATH, args[init.getSysArgNum() + 3]); conf.set(HarpDAALConstants.TEST_TRUTH_PATH, args[init.getSysArgNum() + 4]); // launch job System.out.println("Starting Job"); long perJobSubmitTime = System.currentTimeMillis(); System.out.println( "Start Job#" + " " + new SimpleDateFormat("HH:mm:ss.SSS").format(Calendar.getInstance().getTime())); Job rregJob = init.createJob("rregJob", RidgeRegDaalLauncher.class, RidgeRegDaalCollectiveMapper.class); // finish job boolean jobSuccess = rregJob.waitForCompletion(true); System.out.println( "End Job#" + " " + new SimpleDateFormat("HH:mm:ss.SSS").format(Calendar.getInstance().getTime())); System.out.println( "| Job#" + " Finished in " + (System.currentTimeMillis() - perJobSubmitTime) + " miliseconds |"); if (!jobSuccess) { rregJob.killJob(); System.out.println("rregJob failed"); } return 0; }
From source file:edu.iu.daal_sgd.SGDDaalLauncher.java
License:Apache License
/** * Launches SGD workers./*from ww w.j a v a 2 s . c o m*/ */ @Override public int run(String[] args) throws Exception { /* Put shared libraries into the distributed cache */ Configuration conf = this.getConf(); Initialize init = new Initialize(conf, args); /* Put shared libraries into the distributed cache */ init.loadDistributedLibs(); init.loadDistributedLibsExp(); // load args init.loadSysArgs(); //load app args conf.setInt(HarpDAALConstants.FILE_DIM, Integer.parseInt(args[init.getSysArgNum()])); conf.setInt(HarpDAALConstants.FEATURE_DIM, Integer.parseInt(args[init.getSysArgNum() + 1])); conf.setDouble(Constants.LAMBDA, Double.parseDouble(args[init.getSysArgNum() + 2])); conf.setDouble(Constants.EPSILON, Double.parseDouble(args[init.getSysArgNum() + 3])); conf.setBoolean(Constants.ENABLE_TUNING, Boolean.parseBoolean(args[init.getSysArgNum() + 4])); conf.set(HarpDAALConstants.TEST_FILE_PATH, args[init.getSysArgNum() + 5]); // launch job System.out.println("Starting Job"); long perJobSubmitTime = System.currentTimeMillis(); System.out.println( "Start Job#" + " " + new SimpleDateFormat("HH:mm:ss.SSS").format(Calendar.getInstance().getTime())); Job mfsgdJob = init.createJob("mfsgdJob", SGDDaalLauncher.class, SGDDaalCollectiveMapper.class); // finish job boolean jobSuccess = mfsgdJob.waitForCompletion(true); System.out.println( "End Job#" + " " + new SimpleDateFormat("HH:mm:ss.SSS").format(Calendar.getInstance().getTime())); System.out.println( "| Job#" + " Finished in " + (System.currentTimeMillis() - perJobSubmitTime) + " miliseconds |"); if (!jobSuccess) { mfsgdJob.killJob(); System.out.println("mfsgdJob failed"); } return 0; }
From source file:edu.iu.daal_sorting.SRTDaalLauncher.java
License:Apache License
/** * Launches all the tasks in order.// www. j a va2 s . c o m */ @Override public int run(String[] args) throws Exception { /* Put shared libraries into the distributed cache */ Configuration conf = this.getConf(); Initialize init = new Initialize(conf, args); /* Put shared libraries into the distributed cache */ init.loadDistributedLibs(); // load args init.loadSysArgs(); conf.setInt(HarpDAALConstants.FILE_DIM, Integer.parseInt(args[init.getSysArgNum()])); // launch job System.out.println("Starting Job"); long perJobSubmitTime = System.currentTimeMillis(); System.out.println( "Start Job#" + " " + new SimpleDateFormat("HH:mm:ss.SSS").format(Calendar.getInstance().getTime())); Job sortingJob = init.createJob("sortingJob", SRTDaalLauncher.class, SRTDaalCollectiveMapper.class); // finish job boolean jobSuccess = sortingJob.waitForCompletion(true); System.out.println( "End Job#" + " " + new SimpleDateFormat("HH:mm:ss.SSS").format(Calendar.getInstance().getTime())); System.out.println( "| Job#" + " Finished in " + (System.currentTimeMillis() - perJobSubmitTime) + " miliseconds |"); if (!jobSuccess) { sortingJob.killJob(); System.out.println("sortingJob failed"); } return 0; }
From source file:edu.iu.daal_stump.STUMPDaalLauncher.java
License:Apache License
/** * Launches all the tasks in order.//from www .j av a 2 s .co m */ @Override public int run(String[] args) throws Exception { /* Put shared libraries into the distributed cache */ Configuration conf = this.getConf(); Initialize init = new Initialize(conf, args); /* Put shared libraries into the distributed cache */ init.loadDistributedLibs(); // load args init.loadSysArgs(); conf.setInt(HarpDAALConstants.FILE_DIM, Integer.parseInt(args[init.getSysArgNum()])); conf.setInt(HarpDAALConstants.FEATURE_DIM, Integer.parseInt(args[init.getSysArgNum() + 1])); conf.set(HarpDAALConstants.TEST_FILE_PATH, args[init.getSysArgNum() + 2]); // launch job System.out.println("Starting Job"); long perJobSubmitTime = System.currentTimeMillis(); System.out.println( "Start Job#" + " " + new SimpleDateFormat("HH:mm:ss.SSS").format(Calendar.getInstance().getTime())); Job stumpJob = init.createJob("stumpJob", STUMPDaalLauncher.class, STUMPDaalCollectiveMapper.class); // finish job boolean jobSuccess = stumpJob.waitForCompletion(true); System.out.println( "End Job#" + " " + new SimpleDateFormat("HH:mm:ss.SSS").format(Calendar.getInstance().getTime())); System.out.println( "| Job#" + " Finished in " + (System.currentTimeMillis() - perJobSubmitTime) + " miliseconds |"); if (!jobSuccess) { stumpJob.killJob(); System.out.println("stumpJob failed"); } return 0; }
From source file:edu.iu.daal_subgraph.SCDaalLauncher.java
License:Apache License
private Job configureSCJob(String graphDir, String template, String outDir, int numMapTasks, boolean useLocalMultiThread, int numThreads, int numCores, String affinity, String omp_opt, int tpc, int mem, double memjavaratio, int send_array_limit, int nbr_split_len, boolean rotation_pipeline, int numIteration) throws IOException { Configuration configuration = getConf(); Job job = Job.getInstance(configuration, "subgraph counting"); Configuration jobConfig = job.getConfiguration(); Path jobOutDir = new Path(outDir); FileSystem fs = FileSystem.get(configuration); if (fs.exists(jobOutDir)) { fs.delete(jobOutDir, true);/*www . j a va 2 s . co m*/ } FileInputFormat.setInputPaths(job, graphDir); FileOutputFormat.setOutputPath(job, jobOutDir); //job.setInputFormatClass(KeyValueTextInputFormat.class); //use harp multifile input format to have a better control on num of map tasks job.setInputFormatClass(MultiFileInputFormat.class); job.setJarByClass(SCDaalLauncher.class); job.setMapperClass(SCDaalCollectiveMapper.class); JobConf jobConf = (JobConf) job.getConfiguration(); jobConf.set("mapreduce.framework.name", "map-collective"); // mapreduce.map.collective.memory.mb // 125000 jobConf.setInt("mapreduce.map.collective.memory.mb", mem); // mapreduce.map.collective.java.opts // -Xmx120000m -Xms120000m // int xmx = (mem - 5000) > (mem * 0.9) // ? (mem - 5000) : (int) Math.ceil(mem * 0.5); // int xmx = (int) Math.ceil((mem - 5000)*0.2); int xmx = (int) Math.ceil((mem - 5000) * memjavaratio); int xmn = (int) Math.ceil(0.25 * xmx); jobConf.set("mapreduce.map.collective.java.opts", "-Xmx" + xmx + "m -Xms" + xmx + "m" + " -Xmn" + xmn + "m"); jobConf.setNumMapTasks(numMapTasks); jobConf.setInt("mapreduce.job.max.split.locations", 10000); jobConf.setInt("mapreduce.task.timeout", 60000000); job.setNumReduceTasks(0); jobConfig.setInt(SCConstants.NUM_MAPPERS, numMapTasks); jobConfig.set(SCConstants.TEMPLATE_PATH, template); jobConfig.set(SCConstants.OUTPUT_PATH, outDir); jobConfig.setBoolean(SCConstants.USE_LOCAL_MULTITHREAD, useLocalMultiThread); jobConfig.setInt(SCConstants.NUM_THREADS_PER_NODE, numThreads); jobConfig.setInt(SCConstants.THREAD_NUM, numThreads); jobConfig.setInt(SCConstants.CORE_NUM, numCores); jobConfig.set(SCConstants.THD_AFFINITY, affinity); jobConfig.set(SCConstants.OMPSCHEDULE, omp_opt); jobConfig.setInt(SCConstants.TPC, tpc); jobConfig.setInt(SCConstants.SENDLIMIT, send_array_limit); jobConfig.setInt(SCConstants.NBRTASKLEN, nbr_split_len); jobConfig.setBoolean(SCConstants.ROTATION_PIPELINE, rotation_pipeline); jobConfig.setInt(SCConstants.NUM_ITERATION, numIteration); return job; }
From source file:edu.iu.daal_svd.SVDDaalLauncher.java
License:Apache License
/** * Launches all the tasks in order.// w w w. j av a 2 s . c om */ @Override public int run(String[] args) throws Exception { /* Put shared libraries into the distributed cache */ Configuration conf = this.getConf(); Initialize init = new Initialize(conf, args); /* Put shared libraries into the distributed cache */ init.loadDistributedLibs(); // load args init.loadSysArgs(); //load app args conf.setInt(HarpDAALConstants.FILE_DIM, Integer.parseInt(args[init.getSysArgNum()])); conf.setInt(HarpDAALConstants.FEATURE_DIM, Integer.parseInt(args[init.getSysArgNum() + 1])); // config job System.out.println("Starting Job"); long perJobSubmitTime = System.currentTimeMillis(); System.out.println( "Start Job#" + " " + new SimpleDateFormat("HH:mm:ss.SSS").format(Calendar.getInstance().getTime())); Job svdJob = init.createJob("svdJob", SVDDaalLauncher.class, SVDDaalCollectiveMapper.class); // initialize centroids data JobConf thisjobConf = (JobConf) svdJob.getConfiguration(); FileSystem fs = FileSystem.get(conf); int nFeatures = Integer.parseInt(args[init.getSysArgNum() + 1]); Path workPath = init.getWorkPath(); //generate Data if required boolean generateData = Boolean.parseBoolean(args[init.getSysArgNum() + 2]); if (generateData) { Path inputPath = init.getInputPath(); int total_points = Integer.parseInt(args[init.getSysArgNum() + 3]); int total_files = Integer.parseInt(args[init.getSysArgNum() + 4]); String tmpDirPathName = args[init.getSysArgNum() + 5]; DataGenerator.generateDenseDataMulti(total_points, nFeatures, total_files, 2, 1, ",", inputPath, tmpDirPathName, fs); } // finish job boolean jobSuccess = svdJob.waitForCompletion(true); System.out.println( "End Job#" + " " + new SimpleDateFormat("HH:mm:ss.SSS").format(Calendar.getInstance().getTime())); System.out.println( "| Job#" + " Finished in " + (System.currentTimeMillis() - perJobSubmitTime) + " miliseconds |"); if (!jobSuccess) { svdJob.killJob(); System.out.println("svdJob failed"); } return 0; }
From source file:edu.iu.daal_svm.MultiClassCSRBatch.SVMDaalLauncher.java
License:Apache License
/** * Launches all the tasks in order.//from w w w. j a v a2s .com */ @Override public int run(String[] args) throws Exception { /* Put shared libraries into the distributed cache */ Configuration conf = this.getConf(); Initialize init = new Initialize(conf, args); /* Put shared libraries into the distributed cache */ init.loadDistributedLibs(); // load args init.loadSysArgs(); conf.setInt(HarpDAALConstants.NUM_CLASS, Integer.parseInt(args[init.getSysArgNum()])); conf.set(HarpDAALConstants.TRAIN_TRUTH_PATH, args[init.getSysArgNum() + 1]); conf.set(HarpDAALConstants.TEST_FILE_PATH, args[init.getSysArgNum() + 2]); conf.set(HarpDAALConstants.TEST_TRUTH_PATH, args[init.getSysArgNum() + 3]); // launch job System.out.println("Starting Job"); long perJobSubmitTime = System.currentTimeMillis(); System.out.println( "Start Job#" + " " + new SimpleDateFormat("HH:mm:ss.SSS").format(Calendar.getInstance().getTime())); Job svmJob = init.createJob("svmJob", SVMDaalLauncher.class, SVMDaalCollectiveMapper.class); // finish job boolean jobSuccess = svmJob.waitForCompletion(true); System.out.println( "End Job#" + " " + new SimpleDateFormat("HH:mm:ss.SSS").format(Calendar.getInstance().getTime())); System.out.println( "| Job#" + " Finished in " + (System.currentTimeMillis() - perJobSubmitTime) + " miliseconds |"); if (!jobSuccess) { svmJob.killJob(); System.out.println("svmJob failed"); } return 0; }
From source file:edu.iu.daal_svm.MultiClassDenseBatch.SVMDaalLauncher.java
License:Apache License
/** * Launches all the tasks in order./* w w w . j a va 2s .c o m*/ */ @Override public int run(String[] args) throws Exception { /* Put shared libraries into the distributed cache */ Configuration conf = this.getConf(); Initialize init = new Initialize(conf, args); /* Put shared libraries into the distributed cache */ init.loadDistributedLibs(); // load args init.loadSysArgs(); conf.setInt(HarpDAALConstants.FILE_DIM, Integer.parseInt(args[init.getSysArgNum()])); conf.setInt(HarpDAALConstants.FEATURE_DIM, Integer.parseInt(args[init.getSysArgNum() + 1])); conf.setInt(HarpDAALConstants.NUM_CLASS, Integer.parseInt(args[init.getSysArgNum() + 2])); conf.set(HarpDAALConstants.TEST_FILE_PATH, args[init.getSysArgNum() + 3]); // launch job System.out.println("Starting Job"); long perJobSubmitTime = System.currentTimeMillis(); System.out.println( "Start Job#" + " " + new SimpleDateFormat("HH:mm:ss.SSS").format(Calendar.getInstance().getTime())); Job svmJob = init.createJob("svmJob", SVMDaalLauncher.class, SVMDaalCollectiveMapper.class); // finish job boolean jobSuccess = svmJob.waitForCompletion(true); System.out.println( "End Job#" + " " + new SimpleDateFormat("HH:mm:ss.SSS").format(Calendar.getInstance().getTime())); System.out.println( "| Job#" + " Finished in " + (System.currentTimeMillis() - perJobSubmitTime) + " miliseconds |"); if (!jobSuccess) { svmJob.killJob(); System.out.println("svmJob failed"); } return 0; }
From source file:edu.iu.frlayout.FRJobLauncher.java
License:Apache License
private Job prepareJob(String inputDir, String layoutFile, int totalVtx, int iterationCount, int jobCount, int numMapTasks, int partiitonPerWorker, String outputDirPath, Configuration configuration) throws IOException, URISyntaxException, InterruptedException, ClassNotFoundException { Job job = new Job(configuration, "fr-" + jobCount); Configuration jobConfig = job.getConfiguration(); Path outputDir = new Path(outputDirPath); FileInputFormat.setInputPaths(job, inputDir); FileOutputFormat.setOutputPath(job, outputDir); jobConfig.set(FRConstants.LAYOUT_FILE, layoutFile); jobConfig.setInt(FRConstants.ITERATION, iterationCount); jobConfig.setInt(FRConstants.TOTAL_VTX, totalVtx); jobConfig.setInt(FRConstants.NUM_MAPS, numMapTasks); jobConfig.setInt(FRConstants.PARTITION_PER_WORKER, partiitonPerWorker); jobConfig.setInt("mapreduce.job.max.split.locations", 10000); // input class to file-based class // job.setInputFormatClass(DataFileInputFormat.class); job.setInputFormatClass(MultiFileInputFormat.class); // job.setOutputFormatClass(SequenceFileOutputFormat.class); job.setJarByClass(FRJobLauncher.class); job.setMapperClass(FRCollectiveMapper.class); org.apache.hadoop.mapred.JobConf jobConf = (JobConf) job.getConfiguration(); jobConf.set("mapreduce.framework.name", "map-collective"); jobConf.setNumMapTasks(numMapTasks); job.setNumReduceTasks(0);//from w ww . java 2 s.c o m return job; }
From source file:edu.iu.kmeans.KMeansMapCollective.java
License:Apache License
private Job configureKMeansJob(int numOfDataPoints, int numCentroids, int vectorSize, int numPointFiles, int numMapTasks, Configuration configuration, Path workDirPath, Path dataDir, Path cDir, Path outDir, int jobID, int iterationCount) throws IOException, URISyntaxException { Job job = new Job(configuration, "kmeans_job_" + jobID); Configuration jobConfig = job.getConfiguration(); Path jobOutDir = new Path(outDir, "kmeans_out_" + jobID); FileSystem fs = FileSystem.get(configuration); if (fs.exists(jobOutDir)) { fs.delete(jobOutDir, true);//from w ww.j ava2 s. c o m } FileInputFormat.setInputPaths(job, dataDir); FileOutputFormat.setOutputPath(job, jobOutDir); // The first centroid file with ID 0, // which should match with the centroid file name in data generation Path cFile = new Path(cDir, KMeansConstants.CENTROID_FILE_PREFIX + jobID); System.out.println("Centroid File Path: " + cFile.toString()); jobConfig.set(KMeansConstants.CFILE, cFile.toString()); jobConfig.setInt(KMeansConstants.JOB_ID, jobID); jobConfig.setInt(KMeansConstants.ITERATION_COUNT, iterationCount); // input class to file-based class // job.setInputFormatClass(DataFileInputFormat.class); job.setInputFormatClass(MultiFileInputFormat.class); // job.setOutputKeyClass(IntWritable.class); // job.setOutputValueClass(V2DDataWritable.class); // job.setOutputFormatClass(SequenceFileOutputFormat.class); job.setJarByClass(KMeansMapCollective.class); job.setMapperClass(KMeansCollectiveMapper.class); org.apache.hadoop.mapred.JobConf jobConf = (JobConf) job.getConfiguration(); jobConf.set("mapreduce.framework.name", "map-collective"); jobConf.setNumMapTasks(numMapTasks); jobConf.setInt("mapreduce.job.max.split.locations", 10000); job.setNumReduceTasks(0); jobConfig.setInt(KMeansConstants.VECTOR_SIZE, vectorSize); jobConfig.setInt(KMeansConstants.NUM_CENTROIDS, numCentroids); jobConfig.setInt(KMeansConstants.POINTS_PER_FILE, numOfDataPoints / numPointFiles); jobConfig.set(KMeansConstants.WORK_DIR, workDirPath.toString()); jobConfig.setInt(KMeansConstants.NUM_MAPPERS, numMapTasks); return job; }