org.apache.hadoop.hdfs.nfs.mount.RpcProgramMountd.java Source code

Java tutorial

Introduction

Here is the source code for org.apache.hadoop.hdfs.nfs.mount.RpcProgramMountd.java

Source

/**
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package org.apache.hadoop.hdfs.nfs.mount;

import com.google.common.annotations.VisibleForTesting;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hdfs.DFSClient;
import org.apache.hadoop.hdfs.protocol.HdfsFileStatus;
import org.apache.hadoop.hdfs.server.namenode.NameNode;
import org.apache.hadoop.mount.MountEntry;
import org.apache.hadoop.mount.MountInterface;
import org.apache.hadoop.mount.MountResponse;
import org.apache.hadoop.nfs.AccessPrivilege;
import org.apache.hadoop.nfs.NfsExports;
import org.apache.hadoop.nfs.nfs3.FileHandle;
import org.apache.hadoop.nfs.nfs3.Nfs3Constant;
import org.apache.hadoop.nfs.nfs3.Nfs3Status;
import org.apache.hadoop.oncrpc.RpcAcceptedReply;
import org.apache.hadoop.oncrpc.RpcCall;
import org.apache.hadoop.oncrpc.RpcInfo;
import org.apache.hadoop.oncrpc.RpcProgram;
import org.apache.hadoop.oncrpc.RpcResponse;
import org.apache.hadoop.oncrpc.RpcUtil;
import org.apache.hadoop.oncrpc.XDR;
import org.apache.hadoop.oncrpc.security.VerifierNone;
import org.apache.hadoop.security.SecurityUtil;
import org.apache.hadoop.security.UserGroupInformation;
import org.jboss.netty.buffer.ChannelBuffer;
import org.jboss.netty.buffer.ChannelBuffers;
import org.jboss.netty.channel.ChannelHandlerContext;

import java.io.IOException;
import java.net.InetAddress;
import java.net.InetSocketAddress;
import java.util.ArrayList;
import java.util.Collections;
import java.util.List;

import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NFS_KEYTAB_FILE_KEY;
import static org.apache.hadoop.hdfs.DFSConfigKeys.DFS_NFS_USER_NAME_KEY;

/**
 * RPC program corresponding to mountd daemon. See {@link Mountd}.
 */
public class RpcProgramMountd extends RpcProgram implements MountInterface {
    private static final Log LOG = LogFactory.getLog(RpcProgramMountd.class);
    public static final int PROGRAM = 100005;
    public static final int VERSION_1 = 1;
    public static final int VERSION_2 = 2;
    public static final int VERSION_3 = 3;
    public static final int PORT = 4242;

    // Need DFSClient for branch-1 to get ExtendedHdfsFileStatus
    private final DFSClient dfsClient;

    /**
     * Synchronized list
     */
    private final List<MountEntry> mounts;

    /**
     * List that is unmodifiable
     */
    private final List<String> exports;

    private final NfsExports hostsMatcher;

    public RpcProgramMountd(Configuration config) throws IOException {
        // Note that RPC cache is not enabled
        super("mountd", "localhost", config.getInt("nfs3.mountd.port", PORT), PROGRAM, VERSION_1, VERSION_3);
        exports = new ArrayList<String>();
        exports.add(config.get(Nfs3Constant.EXPORT_POINT, Nfs3Constant.EXPORT_POINT_DEFAULT));
        this.hostsMatcher = NfsExports.getInstance(config);
        this.mounts = Collections.synchronizedList(new ArrayList<MountEntry>());
        UserGroupInformation.setConfiguration(config);
        SecurityUtil.login(config, DFS_NFS_KEYTAB_FILE_KEY, DFS_NFS_USER_NAME_KEY);
        this.dfsClient = new DFSClient(NameNode.getAddress(config), config);
    }

    @Override
    public XDR nullOp(XDR out, int xid, InetAddress client) {
        if (LOG.isDebugEnabled()) {
            LOG.debug("MOUNT NULLOP : " + " client: " + client);
        }
        return RpcAcceptedReply.getAcceptInstance(xid, new VerifierNone()).write(out);
    }

    @Override
    public XDR mnt(XDR xdr, XDR out, int xid, InetAddress client) {
        AccessPrivilege accessPrivilege = hostsMatcher.getAccessPrivilege(client);
        if (accessPrivilege == AccessPrivilege.NONE) {
            return MountResponse.writeMNTResponse(Nfs3Status.NFS3ERR_ACCES, out, xid, null);
        }

        String path = xdr.readString();
        if (LOG.isDebugEnabled()) {
            LOG.debug("MOUNT MNT path: " + path + " client: " + client);
        }

        String host = client.getHostName();
        if (LOG.isDebugEnabled()) {
            LOG.debug("Got host: " + host + " path: " + path);
        }
        if (!exports.contains(path)) {
            LOG.info("Path " + path + " is not shared.");
            MountResponse.writeMNTResponse(Nfs3Status.NFS3ERR_NOENT, out, xid, null);
            return out;
        }

        FileHandle handle = null;
        try {
            HdfsFileStatus exFileStatus = dfsClient.getFileInfo(path);

            handle = new FileHandle(exFileStatus.getFileId());
        } catch (IOException e) {
            LOG.error("Can't get handle for export:" + path, e);
            MountResponse.writeMNTResponse(Nfs3Status.NFS3ERR_NOENT, out, xid, null);
            return out;
        }

        assert (handle != null);
        LOG.info("Giving handle (fileId:" + handle.getFileId() + ") to client for export " + path);
        mounts.add(new MountEntry(host, path));

        MountResponse.writeMNTResponse(Nfs3Status.NFS3_OK, out, xid, handle.getContent());
        return out;
    }

    @Override
    public XDR dump(XDR out, int xid, InetAddress client) {
        if (LOG.isDebugEnabled()) {
            LOG.debug("MOUNT NULLOP : " + " client: " + client);
        }

        List<MountEntry> copy = new ArrayList<MountEntry>(mounts);
        MountResponse.writeMountList(out, xid, copy);
        return out;
    }

    @Override
    public XDR umnt(XDR xdr, XDR out, int xid, InetAddress client) {
        String path = xdr.readString();
        if (LOG.isDebugEnabled()) {
            LOG.debug("MOUNT UMNT path: " + path + " client: " + client);
        }

        String host = client.getHostName();
        mounts.remove(new MountEntry(host, path));
        RpcAcceptedReply.getAcceptInstance(xid, new VerifierNone()).write(out);
        return out;
    }

    @Override
    public XDR umntall(XDR out, int xid, InetAddress client) {
        if (LOG.isDebugEnabled()) {
            LOG.debug("MOUNT UMNTALL : " + " client: " + client);
        }
        mounts.clear();
        return RpcAcceptedReply.getAcceptInstance(xid, new VerifierNone()).write(out);
    }

    @Override
    public void handleInternal(ChannelHandlerContext ctx, RpcInfo info) {
        RpcCall rpcCall = (RpcCall) info.header();
        final MNTPROC mntproc = MNTPROC.fromValue(rpcCall.getProcedure());
        int xid = rpcCall.getXid();
        byte[] data = new byte[info.data().readableBytes()];
        info.data().readBytes(data);
        XDR xdr = new XDR(data);
        XDR out = new XDR();
        InetAddress client = ((InetSocketAddress) info.remoteAddress()).getAddress();

        if (mntproc == MNTPROC.NULL) {
            out = nullOp(out, xid, client);
        } else if (mntproc == MNTPROC.MNT) {
            out = mnt(xdr, out, xid, client);
        } else if (mntproc == MNTPROC.DUMP) {
            out = dump(out, xid, client);
        } else if (mntproc == MNTPROC.UMNT) {
            out = umnt(xdr, out, xid, client);
        } else if (mntproc == MNTPROC.UMNTALL) {
            umntall(out, xid, client);
        } else if (mntproc == MNTPROC.EXPORT) {
            // Currently only support one NFS export 
            List<NfsExports> hostsMatchers = new ArrayList<NfsExports>();
            hostsMatchers.add(hostsMatcher);
            out = MountResponse.writeExportList(out, xid, exports, hostsMatchers);
        } else {
            // Invalid procedure
            RpcAcceptedReply.getInstance(xid, RpcAcceptedReply.AcceptState.PROC_UNAVAIL, new VerifierNone())
                    .write(out);
        }
        ChannelBuffer buf = ChannelBuffers.wrappedBuffer(out.asReadOnlyWrap().buffer());
        RpcResponse rsp = new RpcResponse(buf, info.remoteAddress());
        RpcUtil.sendRpcResponse(ctx, rsp);
    }

    @Override
    protected boolean isIdempotent(RpcCall call) {
        // Not required, because cache is turned off
        return false;
    }

    @VisibleForTesting
    public List<String> getExports() {
        return this.exports;
    }
}