org.apache.hadoop.mapred.TestNodeDecommissioning.java Source code

Java tutorial

Introduction

Here is the source code for org.apache.hadoop.mapred.TestNodeDecommissioning.java

Source

/**
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */

package org.apache.hadoop.mapred;

import java.util.Hashtable;

import org.apache.commons.logging.LogFactory;
import org.apache.commons.logging.Log;
import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.mapreduce.test.system.JTProtocol;
import org.apache.hadoop.mapreduce.test.system.JTClient;
import org.apache.hadoop.mapreduce.test.system.TTClient;
import org.apache.hadoop.mapreduce.test.system.MRCluster;
import org.apache.hadoop.test.system.process.HadoopDaemonRemoteCluster;
import java.util.List;

import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.util.RemoteExecution;
import org.apache.hadoop.util.SSHRemoteExecution;
import org.apache.hadoop.examples.SleepJob;
import org.apache.hadoop.fs.Path;
import java.net.InetAddress;

import org.junit.Assert;
import org.junit.BeforeClass;
import org.junit.AfterClass;
import org.junit.Test;

/**
 * Verify the task tracker node Decommission.
 */

public class TestNodeDecommissioning {

    private static MRCluster cluster = null;
    private static FileSystem dfs = null;
    private static JobClient jobClient = null;
    private static Configuration conf = null;
    private static Path excludeHostPath = null;

    static final Log LOG = LogFactory.getLog(TestNodeDecommissioning.class);

    public TestNodeDecommissioning() throws Exception {
    }

    @BeforeClass
    public static void setUp() throws Exception {
        cluster = MRCluster.createCluster(new Configuration());
        String[] expExcludeList = { "java.net.ConnectException", "java.io.IOException" };
        cluster.setExcludeExpList(expExcludeList);
        cluster.setUp();
        jobClient = cluster.getJTClient().getClient();
        conf = cluster.getJTClient().getProxy().getDaemonConf();
        String confFile = "mapred-site.xml";
        Hashtable<String, String> prop = new Hashtable<String, String>();
        prop.put("mapred.hosts.exclude", "/tmp/mapred.exclude");
        prop.put("mapreduce.cluster.administrators", " gridadmin,hadoop,users");
        cluster.restartClusterWithNewConfig(prop, confFile);
        UtilsForTests.waitFor(1000);
    }

    @AfterClass
    public static void tearDown() throws Exception {
        cluster.restart();
        cluster.tearDown();
    }

    /**
     * This tests whether a node is successfully able to be
     * decommissioned or not.
     * First a node is decommissioned and verified.
     * Then it is removed from decommissoned and verified again.
     * At last the node is started.
     */
    @Test
    public void TestNodeDecommissioning() throws Exception {

        JTProtocol remoteJTClientProxy = cluster.getJTClient().getProxy();

        JTClient remoteJTClient = cluster.getJTClient();
        String jtClientHostName = remoteJTClient.getHostName();
        InetAddress localMachine = java.net.InetAddress.getLocalHost();
        String testRunningHostName = localMachine.getHostName();
        LOG.info("Hostname of local machine: " + testRunningHostName);

        List<TTClient> ttClients = cluster.getTTClients();

        //One slave is got
        TTClient ttClient = (TTClient) ttClients.get(0);
        String ttClientHostName = ttClient.getHostName();

        //Hadoop Conf directory is got
        String hadoopConfDir = cluster.getConf().get(HadoopDaemonRemoteCluster.CONF_HADOOPCONFDIR);

        LOG.info("hadoopConfDir is:" + hadoopConfDir);

        //Hadoop Home is got
        String hadoopHomeDir = cluster.getConf().get(HadoopDaemonRemoteCluster.CONF_HADOOPHOME);

        LOG.info("hadoopHomeDir is:" + hadoopHomeDir);

        conf = cluster.getJTClient().getProxy().getDaemonConf();
        //"mapred.hosts.exclude" path is got
        String excludeHostPathString = (String) conf.get("mapred.hosts.exclude");
        String keytabForHadoopqaUser = "/homes/hadoopqa/hadoopqa.dev.headless.keytab hadoopqa";
        excludeHostPath = new Path(excludeHostPathString);
        LOG.info("exclude Host pathString is :" + excludeHostPathString);

        //One sleep job is submitted
        SleepJob job = new SleepJob();
        job.setConf(conf);
        conf = job.setupJobConf(1, 0, 100, 100, 100, 100);
        JobConf jconf = new JobConf(conf);
        RunningJob rJob = cluster.getJTClient().getClient().submitJob(jconf);

        //username who submitted the job is got.
        String userName = null;
        try {
            JobStatus[] jobStatus = cluster.getJTClient().getClient().getAllJobs();
            userName = jobStatus[0].getUsername();
        } catch (Exception ex) {
            LOG.error("Failed to get user name");
            boolean status = false;
            Assert.assertTrue("Failed to get the userName", status);
        }

        //The client which needs to be decommissioned is put in the exclude path. 
        String command = "echo " + ttClientHostName + " > " + excludeHostPath;

        LOG.info("command is : " + command);
        RemoteExecution rExec = new SSHRemoteExecution();
        rExec.executeCommand(jtClientHostName, userName, command);

        //The refreshNode command is created and execute in Job Tracker Client.
        String refreshNodeCommand = "export HADOOP_CONF_DIR=" + hadoopConfDir + "; export HADOOP_HOME="
                + hadoopHomeDir + ";cd " + hadoopHomeDir + ";kinit -k -t " + keytabForHadoopqaUser
                + ";bin/hadoop mradmin -refreshNodes;";
        LOG.info("refreshNodeCommand is : " + refreshNodeCommand);
        try {
            rExec.executeCommand(testRunningHostName, userName, refreshNodeCommand);
        } catch (Exception e) {
            e.printStackTrace();
        }

        //Checked whether the node is really decommissioned.
        boolean nodeDecommissionedOrNot = false;
        nodeDecommissionedOrNot = remoteJTClientProxy.isNodeDecommissioned(ttClientHostName);

        //The TTClient host is removed from the exclude path
        command = "rm " + excludeHostPath;

        LOG.info("command is : " + command);
        rExec.executeCommand(jtClientHostName, userName, command);

        Assert.assertTrue("Node should be decommissioned", nodeDecommissionedOrNot);

        //The refreshNode command is created and execute in Job Tracker Client.
        rExec.executeCommand(jtClientHostName, userName, refreshNodeCommand);

        //Checked whether the node is out of decommission.
        nodeDecommissionedOrNot = false;
        nodeDecommissionedOrNot = remoteJTClientProxy.isNodeDecommissioned(ttClientHostName);
        Assert.assertFalse("present of not is", nodeDecommissionedOrNot);

        //Starting that node
        String ttClientStart = "export HADOOP_CONF_DIR=" + hadoopConfDir + "; export HADOOP_HOME=" + hadoopHomeDir
                + ";cd " + hadoopHomeDir + ";kinit -k -t " + keytabForHadoopqaUser
                + ";bin/hadoop-daemons.sh start tasktracker;";
        LOG.info("ttClientStart is : " + ttClientStart);
        rExec.executeCommand(jtClientHostName, userName, ttClientStart);
    }
}