Java tutorial
/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.apache.hadoop.chukwa.dataloader; import java.io.File; import java.io.IOException; import java.util.ArrayList; import java.util.HashSet; import java.util.concurrent.CompletionService; import java.util.concurrent.ExecutorCompletionService; import java.util.concurrent.ExecutorService; import java.util.concurrent.Executors; import java.util.concurrent.TimeUnit; import org.apache.commons.logging.Log; import org.apache.commons.logging.LogFactory; import org.apache.hadoop.fs.FileStatus; import org.apache.hadoop.fs.FileSystem; import org.apache.hadoop.fs.FileUtil; import org.apache.hadoop.fs.Path; import org.apache.hadoop.util.ToolRunner; import org.apache.hadoop.chukwa.analysis.salsa.fsm.FSMBuilder; import org.apache.hadoop.chukwa.conf.ChukwaConfiguration; import org.apache.hadoop.chukwa.util.ExceptionUtil; import org.apache.hadoop.conf.Configuration; public class FSMDataLoader extends DataLoaderFactory { private static Log log = LogFactory.getLog(FSMDataLoader.class); protected MetricDataLoader threads[] = null; private static String DATA_LOADER_THREAD_LIMIT = "chukwa.data.loader.threads.limit"; private int size = 1; private static CompletionService completion = null; private static ExecutorService executor = null; private static String[] mappers = { "org.apache.hadoop.chukwa.analysis.salsa.fsm.DataNodeClientTraceMapper", "org.apache.hadoop.chukwa.analysis.salsa.fsm.TaskTrackerClientTraceMapper", "org.apache.hadoop.chukwa.analysis.salsa.fsm.JobHistoryTaskDataMapper" }; public FSMDataLoader() { } public void load(ChukwaConfiguration conf, FileSystem fs, FileStatus[] fileList) throws IOException { if (executor == null) { try { this.size = Integer.parseInt(conf.get(DATA_LOADER_THREAD_LIMIT)); } catch (Exception e) { this.size = 1; } executor = Executors.newFixedThreadPool(size); } if (completion == null) { completion = new ExecutorCompletionService(executor); } try { // Locate directory output directories of the current demux, and create a unique directory list. HashSet<Path> inputPaths = new HashSet<Path>(); HashSet<Path> outputPaths = new HashSet<Path>(); int counter = 0; for (int i = 0; i < fileList.length; i++) { Path temp = fileList[i].getPath().getParent(); if (!inputPaths.contains(temp)) { inputPaths.add(temp); } } String outputDir = conf.get("chukwa.tmp.data.dir") + File.separator + "fsm_" + System.currentTimeMillis() + "_"; if (inputPaths.size() > 0) { Configuration fsmConf = new Configuration(); // Run fsm map reduce job for dn, tt, and jobhist. for (String mapper : mappers) { String[] args = new String[inputPaths.size() + 3]; args[0] = "-in"; int k = 2; boolean hasData = false; for (Path temp : inputPaths) { String tempPath = temp.toUri().toString(); if ((mapper.intern() == mappers[0].intern() && tempPath.indexOf("ClientTraceDetailed") > 0) || (mapper.intern() == mappers[1].intern() && tempPath.indexOf("ClientTraceDetailed") > 0) || (mapper.intern() == mappers[2].intern() && tempPath.indexOf("TaskData") > 0) || (mapper.intern() == mappers[2].intern() && tempPath.indexOf("JobData") > 0)) { args[k] = tempPath; k++; hasData = true; } } args[1] = k - 2 + ""; fsmConf.set("chukwa.salsa.fsm.mapclass", mapper); args[k] = outputDir + mapper; Path outputPath = new Path(args[k]); outputPaths.add(outputPath); if (hasData) { int res = ToolRunner.run(fsmConf, new FSMBuilder(), args); } } } // Find the mapreduce output and load to MDL. for (Path outputPath : outputPaths) { Path searchDir = new Path(outputPath.toUri().toString() + "/*/*/*.evt"); log.info("Search dir:" + searchDir.toUri().toString()); FileStatus[] outputList = fs.globStatus(searchDir); if (outputList != null) { for (int j = 0; j < outputList.length; j++) { String outputFile = outputList[j].getPath().toUri().toString(); log.info("FSM -> MDL loading: " + outputFile); completion.submit(new MetricDataLoader(conf, fs, outputFile)); counter++; } } else { log.warn("No output to load."); } } for (int i = 0; i < counter; i++) { completion.take().get(); } // Clean up mapreduce output of fsm. for (Path dir : outputPaths) { fs.delete(dir, true); } } catch (Exception e) { log.error(ExceptionUtil.getStackTrace(e)); throw new IOException(); } finally { } } public void shutdown() throws InterruptedException { executor.shutdown(); executor.awaitTermination(30, TimeUnit.SECONDS); executor.shutdownNow(); } }