gobblin.runtime.mapreduce.GobblinOutputCommitter.java Source code

Java tutorial

Introduction

Here is the source code for gobblin.runtime.mapreduce.GobblinOutputCommitter.java

Source

/*
 * Licensed to the Apache Software Foundation (ASF) under one or more
 * contributor license agreements.  See the NOTICE file distributed with
 * this work for additional information regarding copyright ownership.
 * The ASF licenses this file to You under the Apache License, Version 2.0
 * (the "License"); you may not use this file except in compliance with
 * the License.  You may obtain a copy of the License at
 *
 *    http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */

package gobblin.runtime.mapreduce;

import java.io.DataInputStream;
import java.io.IOException;
import java.net.URI;
import java.util.Map;
import java.util.concurrent.ConcurrentHashMap;

import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.FileStatus;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.PathFilter;
import org.apache.hadoop.mapreduce.JobContext;
import org.apache.hadoop.mapreduce.JobStatus;
import org.apache.hadoop.mapreduce.OutputCommitter;
import org.apache.hadoop.mapreduce.TaskAttemptContext;

import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

import com.google.common.io.Closer;

import gobblin.configuration.ConfigurationKeys;
import gobblin.configuration.WorkUnitState;
import gobblin.runtime.AbstractJobLauncher;
import gobblin.runtime.GobblinMultiTaskAttempt;
import gobblin.runtime.listeners.JobListener;
import gobblin.source.workunit.MultiWorkUnit;
import gobblin.source.workunit.WorkUnit;
import gobblin.util.JobLauncherUtils;

import lombok.Getter;

/**
 * Hadoop {@link OutputCommitter} implementation that overrides the default
 * {@link #abortJob(JobContext, org.apache.hadoop.mapreduce.JobStatus.State)} behavior.
 * This is necessary to add functionality for cleaning up staging data when the
 * {@link gobblin.runtime.JobLauncher#cancelJob(JobListener)} method is
 * called via Azkaban. Azkaban only allows the cancel method run to for 5 ms until it
 * does a hard kill on the process. In order to make sure the staging data still gets
 * cleaned-up, the cleanup will take place in the AM.
 */
public class GobblinOutputCommitter extends OutputCommitter {

    private static final Logger LOG = LoggerFactory.getLogger(GobblinOutputFormat.class);

    @Getter
    private Map<String, GobblinMultiTaskAttempt> attemptIdToMultiTaskAttempt = new ConcurrentHashMap<>();

    @Override
    public void abortJob(JobContext jobContext, JobStatus.State state) throws IOException {
        LOG.info("Aborting Job: " + jobContext.getJobID() + " with state: " + state);

        Configuration conf = jobContext.getConfiguration();

        URI fsUri = URI.create(conf.get(ConfigurationKeys.FS_URI_KEY, ConfigurationKeys.LOCAL_FS_URI));
        FileSystem fs = FileSystem.get(fsUri, conf);

        Path mrJobDir = new Path(conf.get(ConfigurationKeys.MR_JOB_ROOT_DIR_KEY),
                conf.get(ConfigurationKeys.JOB_NAME_KEY));
        Path jobInputDir = new Path(mrJobDir, MRJobLauncher.INPUT_DIR_NAME);

        if (!fs.exists(jobInputDir) || !fs.isDirectory(jobInputDir)) {
            LOG.warn(String.format("%s either does not exist or is not a directory. No data to cleanup.",
                    jobInputDir));
            return;
        }

        // Iterate through all files in the jobInputDir, each file should correspond to a serialized wu or mwu
        try {
            for (FileStatus status : fs.listStatus(jobInputDir, new WorkUnitFilter())) {

                Closer workUnitFileCloser = Closer.create();

                // If the file ends with ".wu" de-serialize it into a WorkUnit
                if (status.getPath().getName().endsWith(AbstractJobLauncher.WORK_UNIT_FILE_EXTENSION)) {
                    WorkUnit wu = WorkUnit.createEmpty();
                    try {
                        wu.readFields(workUnitFileCloser.register(new DataInputStream(fs.open(status.getPath()))));
                    } finally {
                        workUnitFileCloser.close();
                    }
                    JobLauncherUtils.cleanTaskStagingData(new WorkUnitState(wu), LOG);
                }

                // If the file ends with ".mwu" de-serialize it into a MultiWorkUnit
                if (status.getPath().getName().endsWith(AbstractJobLauncher.MULTI_WORK_UNIT_FILE_EXTENSION)) {
                    MultiWorkUnit mwu = MultiWorkUnit.createEmpty();
                    try {
                        mwu.readFields(workUnitFileCloser.register(new DataInputStream(fs.open(status.getPath()))));
                    } finally {
                        workUnitFileCloser.close();
                    }
                    for (WorkUnit wu : mwu.getWorkUnits()) {
                        JobLauncherUtils.cleanTaskStagingData(new WorkUnitState(wu), LOG);
                    }
                }
            }
        } finally {
            try {
                cleanUpWorkingDirectory(mrJobDir, fs);
            } finally {
                super.abortJob(jobContext, state);
            }
        }
    }

    @Override
    public void abortTask(TaskAttemptContext arg0) throws IOException {
    }

    @Override
    public void commitTask(TaskAttemptContext arg0) throws IOException {
        String taskAttemptId = arg0.getTaskAttemptID().toString();
        LOG.info("Committing task attempt: " + taskAttemptId);
        this.attemptIdToMultiTaskAttempt.get(taskAttemptId).commit();
    }

    @Override
    public boolean needsTaskCommit(TaskAttemptContext arg0) throws IOException {
        return this.attemptIdToMultiTaskAttempt.containsKey(arg0.getTaskAttemptID().toString());
    }

    @Override
    public void setupJob(JobContext arg0) throws IOException {
    }

    @Override
    public void setupTask(TaskAttemptContext arg0) throws IOException {
    }

    /**
     * Replicates the default behavior of the {@link OutputCommitter} used by
     * {@link org.apache.hadoop.mapreduce.lib.output.NullOutputFormat}.
     * @return true
     */
    @Override
    public boolean isRecoverySupported() {
        return true;
    }

    /**
     * Replicates the default behavior of the {@link OutputCommitter} used by
     * {@link org.apache.hadoop.mapreduce.lib.output.NullOutputFormat}.
     */
    @Override
    public void recoverTask(TaskAttemptContext taskContext) throws IOException {
    }

    /**
     * Cleanup the Hadoop MR working directory.
     */
    private static void cleanUpWorkingDirectory(Path mrJobDir, FileSystem fs) throws IOException {
        if (fs.exists(mrJobDir)) {
            fs.delete(mrJobDir, true);
            LOG.info("Deleted working directory " + mrJobDir);
        }
    }

    private static class WorkUnitFilter implements PathFilter {
        @Override
        public boolean accept(Path path) {
            return path.getName().endsWith(AbstractJobLauncher.WORK_UNIT_FILE_EXTENSION)
                    || path.getName().endsWith(AbstractJobLauncher.MULTI_WORK_UNIT_FILE_EXTENSION);
        }
    }
}