Java tutorial
/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package com.yourcompany.hadoop.mapreduce; import org.apache.hadoop.fs.Path; import org.apache.hadoop.io.IntWritable; import org.apache.hadoop.io.Text; import org.apache.hadoop.mapreduce.Job; import org.apache.hadoop.mapreduce.lib.input.FileInputFormat; import org.apache.hadoop.mapreduce.lib.output.FileOutputFormat; import org.apache.hadoop.util.ToolRunner; import java.io.IOException; /** * ? Wordcount MapReduce Job Driver. * <p/> * <pre> * #hadoop jar JAR_FILE wordcount -input IN -output OUT -minSupport 0 -reducer 2 -exactMatch true * -bigrammable false -hasOrigin false -originCNoun false * </pre> * * @author Edward KIM * @version 0.1 * @see <a href="http://cafe.naver.com/korlucene.cafe?iframe_url=/MyCafeIntro.nhn%3Fclubid=17291730">Lucene ?</a> */ public class KoreanWordcountDriver extends org.apache.hadoop.conf.Configured implements org.apache.hadoop.util.Tool { public static void main(String[] args) throws Exception { int res = ToolRunner.run(new KoreanWordcountDriver(), args); System.exit(res); } public int run(String[] args) throws Exception { Job job = new Job(); parseArguements(args, job); job.setJarByClass(KoreanWordcountDriver.class); job.setMapperClass(KoreanWordcountMapper.class); job.setReducerClass(KoreanWordcountReducer.class); job.setMapOutputKeyClass(Text.class); job.setMapOutputValueClass(IntWritable.class); job.setOutputKeyClass(Text.class); job.setOutputValueClass(IntWritable.class); return job.waitForCompletion(true) ? 0 : 1; } private void parseArguements(String[] args, Job job) throws IOException { for (int i = 0; i < args.length; ++i) { if ("-input".equals(args[i])) { FileInputFormat.addInputPaths(job, args[++i]); } else if ("-output".equals(args[i])) { FileOutputFormat.setOutputPath(job, new Path(args[++i])); } else if ("-exactMatch".equals(args[i])) { job.getConfiguration().set("exactMatch", args[++i]); } else if ("-bigrammable".equals(args[i])) { job.getConfiguration().set("bigrammable", args[++i]); } else if ("-hasOrigin".equals(args[i])) { job.getConfiguration().set("hasOrigin", args[++i]); } else if ("-originCNoun".equals(args[i])) { job.getConfiguration().set("originCNoun", args[++i]); } else if ("-reducer".equals(args[i])) { job.setNumReduceTasks(Integer.parseInt(args[++i])); } else if ("-minSupport".equals(args[i])) { job.getConfiguration().set("minSupport", args[++i]); } } } }