org.apache.hadoop.hdfs.server.namenode.FSEditLog.java Source code

Java tutorial

Introduction

Here is the source code for org.apache.hadoop.hdfs.server.namenode.FSEditLog.java

Source

/**
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package org.apache.hadoop.hdfs.server.namenode;

import static org.apache.hadoop.hdfs.server.common.Util.now;

import java.io.IOException;
import java.lang.reflect.Constructor;
import java.net.URI;
import java.util.ArrayList;
import java.util.Collection;
import java.util.List;

import org.apache.commons.lang.StringUtils;
import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.classification.InterfaceAudience;
import org.apache.hadoop.classification.InterfaceStability;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.fs.Options;
import org.apache.hadoop.fs.permission.FsPermission;
import org.apache.hadoop.hdfs.DFSConfigKeys;
import org.apache.hadoop.hdfs.protocol.HdfsConstants;
import org.apache.hadoop.hdfs.security.token.delegation.DelegationTokenIdentifier;
import org.apache.hadoop.hdfs.server.common.HdfsServerConstants.NamenodeRole;
import org.apache.hadoop.hdfs.server.common.Storage.StorageDirectory;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.AddOp;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.CancelDelegationTokenOp;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.CloseOp;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.ConcatDeleteOp;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.DeleteOp;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.GetDelegationTokenOp;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.LogSegmentOp;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.MkdirOp;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.OpInstanceCache;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.ReassignLeaseOp;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.RenameOldOp;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.RenameOp;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.RenewDelegationTokenOp;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.SetGenstampOp;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.SetOwnerOp;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.SetPermissionsOp;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.SetQuotaOp;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.SetReplicationOp;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.SymlinkOp;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.TimesOp;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.UpdateBlocksOp;
import org.apache.hadoop.hdfs.server.namenode.FSEditLogOp.UpdateMasterKeyOp;
import org.apache.hadoop.hdfs.server.namenode.JournalSet.JournalAndStream;
import org.apache.hadoop.hdfs.server.namenode.metrics.NameNodeMetrics;
import org.apache.hadoop.hdfs.server.protocol.NamenodeRegistration;
import org.apache.hadoop.hdfs.server.protocol.RemoteEditLogManifest;
import org.apache.hadoop.io.IOUtils;
import org.apache.hadoop.security.token.delegation.DelegationKey;

import com.google.common.annotations.VisibleForTesting;
import com.google.common.base.Preconditions;
import com.google.common.collect.Lists;

/**
 * FSEditLog maintains a log of the namespace modifications.
 * 
 */
@InterfaceAudience.Private
@InterfaceStability.Evolving
public class FSEditLog {

    static final Log LOG = LogFactory.getLog(FSEditLog.class);

    /**
     * State machine for edit log.
     * 
     * In a non-HA setup:
     * 
     * The log starts in UNITIALIZED state upon construction. Once it's
     * initialized, it is usually in IN_SEGMENT state, indicating that edits may
     * be written. In the middle of a roll, or while saving the namespace, it
     * briefly enters the BETWEEN_LOG_SEGMENTS state, indicating that the previous
     * segment has been closed, but the new one has not yet been opened.
     * 
     * In an HA setup:
     * 
     * The log starts in UNINITIALIZED state upon construction. Once it's
     * initialized, it sits in the OPEN_FOR_READING state the entire time that the
     * NN is in standby. Upon the NN transition to active, the log will be CLOSED,
     * and then move to being BETWEEN_LOG_SEGMENTS, much as if the NN had just
     * started up, and then will move to IN_SEGMENT so it can begin writing to the
     * log. The log states will then revert to behaving as they do in a non-HA
     * setup.
     */
    private enum State {
        UNINITIALIZED, BETWEEN_LOG_SEGMENTS, IN_SEGMENT, OPEN_FOR_READING, CLOSED;
    }

    private State state = State.UNINITIALIZED;

    //initialize
    private JournalSet journalSet = null;
    private EditLogOutputStream editLogStream = null;

    // a monotonically increasing counter that represents transactionIds.
    private long txid = 0;

    // stores the last synced transactionId.
    private long synctxid = 0;

    // the first txid of the log that's currently open for writing.
    // If this value is N, we are currently writing to edits_inprogress_N
    private long curSegmentTxId = HdfsConstants.INVALID_TXID;

    // the time of printing the statistics to the log file.
    private long lastPrintTime;

    // is a sync currently running?
    private volatile boolean isSyncRunning;

    // is an automatic sync scheduled?
    private volatile boolean isAutoSyncScheduled = false;

    // Used to exit in the event of a failure to sync to all journals. It's a
    // member variable so it can be swapped out for testing.
    private Runtime runtime = Runtime.getRuntime();

    // these are statistics counters.
    private long numTransactions; // number of transactions
    private long numTransactionsBatchedInSync;
    private long totalTimeTransactions; // total time for all transactions
    private NameNodeMetrics metrics;

    private NNStorage storage;
    private Configuration conf;

    private List<URI> editsDirs;

    private ThreadLocal<OpInstanceCache> cache = new ThreadLocal<OpInstanceCache>() {
        @Override
        protected OpInstanceCache initialValue() {
            return new OpInstanceCache();
        }
    };

    /**
     * The edit directories that are shared between primary and secondary.
     */
    private List<URI> sharedEditsDirs;

    private static class TransactionId {
        public long txid;

        TransactionId(long value) {
            this.txid = value;
        }
    }

    // stores the most current transactionId of this thread.
    private static final ThreadLocal<TransactionId> myTransactionId = new ThreadLocal<TransactionId>() {
        protected synchronized TransactionId initialValue() {
            return new TransactionId(Long.MAX_VALUE);
        }
    };

    /**
     * Construct FSEditLog with default configuration, taking editDirs from NNStorage
     * 
     * @param storage Storage object used by namenode
     */
    @VisibleForTesting
    FSEditLog(NNStorage storage) throws IOException {
        Configuration conf = new Configuration();
        // Make sure the edits dirs are set in the provided configuration object.
        conf.set(DFSConfigKeys.DFS_NAMENODE_EDITS_DIR_KEY, StringUtils.join(storage.getEditsDirectories(), ","));
        init(conf, storage, FSNamesystem.getNamespaceEditsDirs(conf));
    }

    /**
     * Constructor for FSEditLog. Underlying journals are constructed, but 
     * no streams are opened until open() is called.
     * 
     * @param conf The namenode configuration
     * @param storage Storage object used by namenode
     * @param editsDirs List of journals to use
     */
    FSEditLog(Configuration conf, NNStorage storage, List<URI> editsDirs) {
        init(conf, storage, editsDirs);
    }

    private void init(Configuration conf, NNStorage storage, List<URI> editsDirs) {
        isSyncRunning = false;
        this.conf = conf;
        this.storage = storage;
        metrics = NameNode.getNameNodeMetrics();
        lastPrintTime = now();

        // If this list is empty, an error will be thrown on first use
        // of the editlog, as no journals will exist
        this.editsDirs = Lists.newArrayList(editsDirs);

        this.sharedEditsDirs = FSNamesystem.getSharedEditsDirs(conf);
    }

    public synchronized void initJournalsForWrite() {
        Preconditions.checkState(state == State.UNINITIALIZED || state == State.CLOSED, "Unexpected state: %s",
                state);

        initJournals(this.editsDirs);
        state = State.BETWEEN_LOG_SEGMENTS;
    }

    public synchronized void initSharedJournalsForRead() {
        if (state == State.OPEN_FOR_READING) {
            LOG.warn("Initializing shared journals for READ, already open for READ", new Exception());
            return;
        }
        Preconditions.checkState(state == State.UNINITIALIZED || state == State.CLOSED);

        initJournals(this.sharedEditsDirs);
        state = State.OPEN_FOR_READING;
    }

    private synchronized void initJournals(List<URI> dirs) {
        int minimumRedundantJournals = conf.getInt(DFSConfigKeys.DFS_NAMENODE_EDITS_DIR_MINIMUM_KEY,
                DFSConfigKeys.DFS_NAMENODE_EDITS_DIR_MINIMUM_DEFAULT);

        journalSet = new JournalSet(minimumRedundantJournals);
        for (URI u : dirs) {
            boolean required = FSNamesystem.getRequiredNamespaceEditsDirs(conf).contains(u);
            if (u.getScheme().equals(NNStorage.LOCAL_URI_SCHEME)) {
                StorageDirectory sd = storage.getStorageDirectory(u);
                if (sd != null) {
                    journalSet.add(new FileJournalManager(sd, storage), required);
                }
            } else {
                journalSet.add(createJournal(u), required);
            }
        }

        if (journalSet.isEmpty()) {
            LOG.error("No edits directories configured!");
        }
    }

    /**
     * Get the list of URIs the editlog is using for storage
     * @return collection of URIs in use by the edit log
     */
    Collection<URI> getEditURIs() {
        return editsDirs;
    }

    /**
     * Initialize the output stream for logging, opening the first
     * log segment.
     */
    synchronized void openForWrite() throws IOException {
        Preconditions.checkState(state == State.BETWEEN_LOG_SEGMENTS, "Bad state: %s", state);

        long segmentTxId = getLastWrittenTxId() + 1;
        // Safety check: we should never start a segment if there are
        // newer txids readable.
        EditLogInputStream s = journalSet.getInputStream(segmentTxId, true);
        try {
            Preconditions.checkState(s == null,
                    "Cannot start writing at txid %s when there is a stream " + "available for read: %s",
                    segmentTxId, s);
        } finally {
            IOUtils.closeStream(s);
        }

        startLogSegmentAndWriteHeaderTxn(segmentTxId);
        assert state == State.IN_SEGMENT : "Bad state: " + state;
    }

    /**
     * @return true if the log is currently open in write mode, regardless
     * of whether it actually has an open segment.
     */
    synchronized boolean isOpenForWrite() {
        return state == State.IN_SEGMENT || state == State.BETWEEN_LOG_SEGMENTS;
    }

    /**
     * @return true if the log is open in write mode and has a segment open
     * ready to take edits.
     */
    synchronized boolean isSegmentOpen() {
        return state == State.IN_SEGMENT;
    }

    /**
     * @return true if the log is open in read mode.
     */
    public synchronized boolean isOpenForRead() {
        return state == State.OPEN_FOR_READING;
    }

    /**
     * Shutdown the file store.
     */
    synchronized void close() {
        if (state == State.CLOSED) {
            LOG.debug("Closing log when already closed");
            return;
        }
        if (state == State.IN_SEGMENT) {
            assert editLogStream != null;
            waitForSyncToFinish();
            endCurrentLogSegment(true);
        }

        if (!journalSet.isEmpty()) {
            try {
                journalSet.close();
            } catch (IOException ioe) {
                LOG.warn("Error closing journalSet", ioe);
            }
        }

        state = State.CLOSED;
    }

    /**
     * Write an operation to the edit log. Do not sync to persistent
     * store yet.
     */
    void logEdit(final FSEditLogOp op) {
        synchronized (this) {
            assert isOpenForWrite() : "bad state: " + state;

            // wait if an automatic sync is scheduled
            waitIfAutoSyncScheduled();

            long start = beginTransaction();
            op.setTransactionId(txid);

            try {
                editLogStream.write(op);
            } catch (IOException ex) {
                // All journals failed, it is handled in logSync.
            }

            endTransaction(start);

            // check if it is time to schedule an automatic sync
            if (!shouldForceSync()) {
                return;
            }
            isAutoSyncScheduled = true;
        }

        // sync buffered edit log entries to persistent store
        logSync();
    }

    /**
     * Wait if an automatic sync is scheduled
     * @throws InterruptedException
     */
    synchronized void waitIfAutoSyncScheduled() {
        try {
            while (isAutoSyncScheduled) {
                this.wait(1000);
            }
        } catch (InterruptedException e) {
        }
    }

    /**
     * Signal that an automatic sync scheduling is done if it is scheduled
     */
    synchronized void doneWithAutoSyncScheduling() {
        if (isAutoSyncScheduled) {
            isAutoSyncScheduled = false;
            notifyAll();
        }
    }

    /**
     * Check if should automatically sync buffered edits to 
     * persistent store
     * 
     * @return true if any of the edit stream says that it should sync
     */
    private boolean shouldForceSync() {
        return editLogStream.shouldForceSync();
    }

    private long beginTransaction() {
        assert Thread.holdsLock(this);
        // get a new transactionId
        txid++;

        //
        // record the transactionId when new data was written to the edits log
        //
        TransactionId id = myTransactionId.get();
        id.txid = txid;
        return now();
    }

    private void endTransaction(long start) {
        assert Thread.holdsLock(this);

        // update statistics
        long end = now();
        numTransactions++;
        totalTimeTransactions += (end - start);
        if (metrics != null) // Metrics is non-null only when used inside name node
            metrics.addTransaction(end - start);
    }

    /**
     * Return the transaction ID of the last transaction written to the log.
     */
    public synchronized long getLastWrittenTxId() {
        return txid;
    }

    /**
     * @return the first transaction ID in the current log segment
     */
    synchronized long getCurSegmentTxId() {
        Preconditions.checkState(isSegmentOpen(), "Bad state: %s", state);
        return curSegmentTxId;
    }

    /**
     * Set the transaction ID to use for the next transaction written.
     */
    synchronized void setNextTxId(long nextTxId) {
        Preconditions.checkArgument(synctxid <= txid && nextTxId >= txid,
                "May not decrease txid." + " synctxid=%s txid=%s nextTxId=%s", synctxid, txid, nextTxId);

        txid = nextTxId - 1;
    }

    /**
     * Blocks until all ongoing edits have been synced to disk.
     * This differs from logSync in that it waits for edits that have been
     * written by other threads, not just edits from the calling thread.
     *
     * NOTE: this should be done while holding the FSNamesystem lock, or
     * else more operations can start writing while this is in progress.
     */
    void logSyncAll() {
        // Record the most recent transaction ID as our own id
        synchronized (this) {
            TransactionId id = myTransactionId.get();
            id.txid = txid;
        }
        // Then make sure we're synced up to this point
        logSync();
    }

    /**
     * Sync all modifications done by this thread.
     *
     * The internal concurrency design of this class is as follows:
     *   - Log items are written synchronized into an in-memory buffer,
     *     and each assigned a transaction ID.
     *   - When a thread (client) would like to sync all of its edits, logSync()
     *     uses a ThreadLocal transaction ID to determine what edit number must
     *     be synced to.
     *   - The isSyncRunning volatile boolean tracks whether a sync is currently
     *     under progress.
     *
     * The data is double-buffered within each edit log implementation so that
     * in-memory writing can occur in parallel with the on-disk writing.
     *
     * Each sync occurs in three steps:
     *   1. synchronized, it swaps the double buffer and sets the isSyncRunning
     *      flag.
     *   2. unsynchronized, it flushes the data to storage
     *   3. synchronized, it resets the flag and notifies anyone waiting on the
     *      sync.
     *
     * The lack of synchronization on step 2 allows other threads to continue
     * to write into the memory buffer while the sync is in progress.
     * Because this step is unsynchronized, actions that need to avoid
     * concurrency with sync() should be synchronized and also call
     * waitForSyncToFinish() before assuming they are running alone.
     */
    public void logSync() {
        long syncStart = 0;

        // Fetch the transactionId of this thread. 
        long mytxid = myTransactionId.get().txid;

        boolean sync = false;
        try {
            EditLogOutputStream logStream = null;
            synchronized (this) {
                try {
                    printStatistics(false);

                    // if somebody is already syncing, then wait
                    while (mytxid > synctxid && isSyncRunning) {
                        try {
                            wait(1000);
                        } catch (InterruptedException ie) {
                        }
                    }

                    //
                    // If this transaction was already flushed, then nothing to do
                    //
                    if (mytxid <= synctxid) {
                        numTransactionsBatchedInSync++;
                        if (metrics != null) {
                            // Metrics is non-null only when used inside name node
                            metrics.incrTransactionsBatchedInSync();
                        }
                        return;
                    }

                    // now, this thread will do the sync
                    syncStart = txid;
                    isSyncRunning = true;
                    sync = true;

                    // swap buffers
                    try {
                        if (journalSet.isEmpty()) {
                            throw new IOException("No journals available to flush");
                        }
                        editLogStream.setReadyToFlush();
                    } catch (IOException e) {
                        LOG.fatal("Could not sync enough journals to persistent storage. "
                                + "Unsynced transactions: " + (txid - synctxid), new Exception());
                        runtime.exit(1);
                    }
                } finally {
                    // Prevent RuntimeException from blocking other log edit write 
                    doneWithAutoSyncScheduling();
                }
                //editLogStream may become null,
                //so store a local variable for flush.
                logStream = editLogStream;
            }

            // do the sync
            long start = now();
            try {
                if (logStream != null) {
                    logStream.flush();
                }
            } catch (IOException ex) {
                synchronized (this) {
                    LOG.fatal("Could not sync enough journals to persistent storage. " + "Unsynced transactions: "
                            + (txid - synctxid), new Exception());
                    runtime.exit(1);
                }
            }
            long elapsed = now() - start;

            if (metrics != null) { // Metrics non-null only when used inside name node
                metrics.addSync(elapsed);
            }

        } finally {
            // Prevent RuntimeException from blocking other log edit sync 
            synchronized (this) {
                if (sync) {
                    synctxid = syncStart;
                    isSyncRunning = false;
                }
                this.notifyAll();
            }
        }
    }

    //
    // print statistics every 1 minute.
    //
    private void printStatistics(boolean force) {
        long now = now();
        if (lastPrintTime + 60000 > now && !force) {
            return;
        }
        lastPrintTime = now;
        StringBuilder buf = new StringBuilder();
        buf.append("Number of transactions: ");
        buf.append(numTransactions);
        buf.append(" Total time for transactions(ms): ");
        buf.append(totalTimeTransactions);
        buf.append("Number of transactions batched in Syncs: ");
        buf.append(numTransactionsBatchedInSync);
        buf.append(" Number of syncs: ");
        buf.append(editLogStream.getNumSync());
        buf.append(" SyncTimes(ms): ");
        buf.append(journalSet.getSyncTimes());
        LOG.info(buf);
    }

    /** 
     * Add open lease record to edit log. 
     * Records the block locations of the last block.
     */
    public void logOpenFile(String path, INodeFileUnderConstruction newNode) {
        AddOp op = AddOp.getInstance(cache.get()).setPath(path).setReplication(newNode.getReplication())
                .setModificationTime(newNode.getModificationTime()).setAccessTime(newNode.getAccessTime())
                .setBlockSize(newNode.getPreferredBlockSize()).setBlocks(newNode.getBlocks())
                .setPermissionStatus(newNode.getPermissionStatus()).setClientName(newNode.getClientName())
                .setClientMachine(newNode.getClientMachine());

        logEdit(op);
    }

    /** 
     * Add close lease record to edit log.
     */
    public void logCloseFile(String path, INodeFile newNode) {
        CloseOp op = CloseOp.getInstance(cache.get()).setPath(path).setReplication(newNode.getReplication())
                .setModificationTime(newNode.getModificationTime()).setAccessTime(newNode.getAccessTime())
                .setBlockSize(newNode.getPreferredBlockSize()).setBlocks(newNode.getBlocks())
                .setPermissionStatus(newNode.getPermissionStatus());

        logEdit(op);
    }

    public void logUpdateBlocks(String path, INodeFileUnderConstruction file) {
        UpdateBlocksOp op = UpdateBlocksOp.getInstance(cache.get()).setPath(path).setBlocks(file.getBlocks());
        logEdit(op);
    }

    /** 
     * Add create directory record to edit log
     */
    public void logMkDir(String path, INode newNode) {
        MkdirOp op = MkdirOp.getInstance(cache.get()).setPath(path).setTimestamp(newNode.getModificationTime())
                .setPermissionStatus(newNode.getPermissionStatus());
        logEdit(op);
    }

    /** 
     * Add rename record to edit log
     * TODO: use String parameters until just before writing to disk
     */
    void logRename(String src, String dst, long timestamp) {
        RenameOldOp op = RenameOldOp.getInstance(cache.get()).setSource(src).setDestination(dst)
                .setTimestamp(timestamp);
        logEdit(op);
    }

    /** 
     * Add rename record to edit log
     */
    void logRename(String src, String dst, long timestamp, Options.Rename... options) {
        RenameOp op = RenameOp.getInstance(cache.get()).setSource(src).setDestination(dst).setTimestamp(timestamp)
                .setOptions(options);
        logEdit(op);
    }

    /** 
     * Add set replication record to edit log
     */
    void logSetReplication(String src, short replication) {
        SetReplicationOp op = SetReplicationOp.getInstance(cache.get()).setPath(src).setReplication(replication);
        logEdit(op);
    }

    /** Add set namespace quota record to edit log
     * 
     * @param src the string representation of the path to a directory
     * @param quota the directory size limit
     */
    void logSetQuota(String src, long nsQuota, long dsQuota) {
        SetQuotaOp op = SetQuotaOp.getInstance(cache.get()).setSource(src).setNSQuota(nsQuota).setDSQuota(dsQuota);
        logEdit(op);
    }

    /**  Add set permissions record to edit log */
    void logSetPermissions(String src, FsPermission permissions) {
        SetPermissionsOp op = SetPermissionsOp.getInstance(cache.get()).setSource(src).setPermissions(permissions);
        logEdit(op);
    }

    /**  Add set owner record to edit log */
    void logSetOwner(String src, String username, String groupname) {
        SetOwnerOp op = SetOwnerOp.getInstance(cache.get()).setSource(src).setUser(username).setGroup(groupname);
        logEdit(op);
    }

    /**
     * concat(trg,src..) log
     */
    void logConcat(String trg, String[] srcs, long timestamp) {
        ConcatDeleteOp op = ConcatDeleteOp.getInstance(cache.get()).setTarget(trg).setSources(srcs)
                .setTimestamp(timestamp);
        logEdit(op);
    }

    /** 
     * Add delete file record to edit log
     */
    void logDelete(String src, long timestamp) {
        DeleteOp op = DeleteOp.getInstance(cache.get()).setPath(src).setTimestamp(timestamp);
        logEdit(op);
    }

    /** 
     * Add generation stamp record to edit log
     */
    void logGenerationStamp(long genstamp) {
        SetGenstampOp op = SetGenstampOp.getInstance(cache.get()).setGenerationStamp(genstamp);
        logEdit(op);
    }

    /** 
     * Add access time record to edit log
     */
    void logTimes(String src, long mtime, long atime) {
        TimesOp op = TimesOp.getInstance(cache.get()).setPath(src).setModificationTime(mtime).setAccessTime(atime);
        logEdit(op);
    }

    /** 
     * Add a create symlink record.
     */
    void logSymlink(String path, String value, long mtime, long atime, INodeSymlink node) {
        SymlinkOp op = SymlinkOp.getInstance(cache.get()).setPath(path).setValue(value).setModificationTime(mtime)
                .setAccessTime(atime).setPermissionStatus(node.getPermissionStatus());
        logEdit(op);
    }

    /**
     * log delegation token to edit log
     * @param id DelegationTokenIdentifier
     * @param expiryTime of the token
     */
    void logGetDelegationToken(DelegationTokenIdentifier id, long expiryTime) {
        GetDelegationTokenOp op = GetDelegationTokenOp.getInstance(cache.get()).setDelegationTokenIdentifier(id)
                .setExpiryTime(expiryTime);
        logEdit(op);
    }

    void logRenewDelegationToken(DelegationTokenIdentifier id, long expiryTime) {
        RenewDelegationTokenOp op = RenewDelegationTokenOp.getInstance(cache.get()).setDelegationTokenIdentifier(id)
                .setExpiryTime(expiryTime);
        logEdit(op);
    }

    void logCancelDelegationToken(DelegationTokenIdentifier id) {
        CancelDelegationTokenOp op = CancelDelegationTokenOp.getInstance(cache.get())
                .setDelegationTokenIdentifier(id);
        logEdit(op);
    }

    void logUpdateMasterKey(DelegationKey key) {
        UpdateMasterKeyOp op = UpdateMasterKeyOp.getInstance(cache.get()).setDelegationKey(key);
        logEdit(op);
    }

    void logReassignLease(String leaseHolder, String src, String newHolder) {
        ReassignLeaseOp op = ReassignLeaseOp.getInstance(cache.get()).setLeaseHolder(leaseHolder).setPath(src)
                .setNewHolder(newHolder);
        logEdit(op);
    }

    /**
     * Get all the journals this edit log is currently operating on.
     */
    synchronized List<JournalAndStream> getJournals() {
        return journalSet.getAllJournalStreams();
    }

    /**
     * Used only by tests.
     */
    @VisibleForTesting
    synchronized public JournalSet getJournalSet() {
        return journalSet;
    }

    /**
     * Used only by unit tests.
     */
    @VisibleForTesting
    synchronized void setRuntimeForTesting(Runtime runtime) {
        this.runtime = runtime;
        this.journalSet.setRuntimeForTesting(runtime);
    }

    /**
     * Used only by tests.
     */
    @VisibleForTesting
    void setMetricsForTests(NameNodeMetrics metrics) {
        this.metrics = metrics;
    }

    /**
     * Return a manifest of what finalized edit logs are available
     */
    public synchronized RemoteEditLogManifest getEditLogManifest(long fromTxId) throws IOException {
        return journalSet.getEditLogManifest(fromTxId);
    }

    /**
     * Finalizes the current edit log and opens a new log segment.
     * @return the transaction id of the BEGIN_LOG_SEGMENT transaction
     * in the new log.
     */
    synchronized long rollEditLog() throws IOException {
        LOG.info("Rolling edit logs.");
        endCurrentLogSegment(true);

        long nextTxId = getLastWrittenTxId() + 1;
        startLogSegmentAndWriteHeaderTxn(nextTxId);

        assert curSegmentTxId == nextTxId;
        return nextTxId;
    }

    /**
     * Remote namenode just has started a log segment, start log segment locally.
     */
    public synchronized void startLogSegment(long txid, boolean abortCurrentLogSegment) throws IOException {
        LOG.info("Namenode started a new log segment at txid " + txid);
        if (isSegmentOpen()) {
            if (getLastWrittenTxId() == txid - 1) {
                //In sync with the NN, so end and finalize the current segment`
                endCurrentLogSegment(false);
            } else {
                //Missed some transactions: probably lost contact with NN temporarily.
                final String mess = "Cannot start a new log segment at txid " + txid + " since only up to txid "
                        + getLastWrittenTxId() + " have been written in the log segment starting at "
                        + getCurSegmentTxId() + ".";
                if (abortCurrentLogSegment) {
                    //Mark the current segment as aborted.
                    LOG.warn(mess);
                    abortCurrentLogSegment();
                } else {
                    throw new IOException(mess);
                }
            }
        }
        setNextTxId(txid);
        startLogSegment(txid);
    }

    /**
     * Start writing to the log segment with the given txid.
     * Transitions from BETWEEN_LOG_SEGMENTS state to IN_LOG_SEGMENT state. 
     */
    private void startLogSegment(final long segmentTxId) throws IOException {
        assert Thread.holdsLock(this);

        LOG.info("Starting log segment at " + segmentTxId);
        Preconditions.checkArgument(segmentTxId > 0, "Bad txid: %s", segmentTxId);
        Preconditions.checkState(state == State.BETWEEN_LOG_SEGMENTS, "Bad state: %s", state);
        Preconditions.checkState(segmentTxId > curSegmentTxId, "Cannot start writing to log segment " + segmentTxId
                + " when previous log segment started at " + curSegmentTxId);
        Preconditions.checkArgument(segmentTxId == txid + 1,
                "Cannot start log segment at txid %s when next expected " + "txid is %s", segmentTxId, txid + 1);

        numTransactions = totalTimeTransactions = numTransactionsBatchedInSync = 0;

        // TODO no need to link this back to storage anymore!
        // See HDFS-2174.
        storage.attemptRestoreRemovedStorage();

        try {
            editLogStream = journalSet.startLogSegment(segmentTxId);
        } catch (IOException ex) {
            throw new IOException(
                    "Unable to start log segment " + segmentTxId + ": too few journals successfully started.", ex);
        }

        curSegmentTxId = segmentTxId;
        state = State.IN_SEGMENT;
    }

    synchronized void startLogSegmentAndWriteHeaderTxn(final long segmentTxId) throws IOException {
        startLogSegment(segmentTxId);

        logEdit(LogSegmentOp.getInstance(cache.get(), FSEditLogOpCodes.OP_START_LOG_SEGMENT));
        logSync();
    }

    /**
     * Finalize the current log segment.
     * Transitions from IN_SEGMENT state to BETWEEN_LOG_SEGMENTS state.
     */
    synchronized void endCurrentLogSegment(boolean writeEndTxn) {
        LOG.info("Ending log segment " + curSegmentTxId);
        Preconditions.checkState(isSegmentOpen(), "Bad state: %s", state);

        if (writeEndTxn) {
            logEdit(LogSegmentOp.getInstance(cache.get(), FSEditLogOpCodes.OP_END_LOG_SEGMENT));
            logSync();
        }

        printStatistics(true);

        final long lastTxId = getLastWrittenTxId();

        try {
            journalSet.finalizeLogSegment(curSegmentTxId, lastTxId);
            editLogStream = null;
        } catch (IOException e) {
            //All journals have failed, it will be handled in logSync.
        }

        state = State.BETWEEN_LOG_SEGMENTS;
    }

    /**
     * Abort all current logs. Called from the backup node.
     */
    synchronized void abortCurrentLogSegment() {
        try {
            //Check for null, as abort can be called any time.
            if (editLogStream != null) {
                editLogStream.abort();
                editLogStream = null;
                state = State.BETWEEN_LOG_SEGMENTS;
            }
        } catch (IOException e) {
            LOG.warn("All journals failed to abort", e);
        }
    }

    /**
     * Archive any log files that are older than the given txid.
     */
    public synchronized void purgeLogsOlderThan(final long minTxIdToKeep) {
        assert curSegmentTxId == HdfsConstants.INVALID_TXID || // on format this is no-op
                minTxIdToKeep <= curSegmentTxId : "cannot purge logs older than txid " + minTxIdToKeep
                        + " when current segment starts at " + curSegmentTxId;

        // This could be improved to not need synchronization. But currently,
        // journalSet is not threadsafe, so we need to synchronize this method.
        try {
            journalSet.purgeLogsOlderThan(minTxIdToKeep);
        } catch (IOException ex) {
            //All journals have failed, it will be handled in logSync.
        }
    }

    /**
     * The actual sync activity happens while not synchronized on this object.
     * Thus, synchronized activities that require that they are not concurrent
     * with file operations should wait for any running sync to finish.
     */
    synchronized void waitForSyncToFinish() {
        while (isSyncRunning) {
            try {
                wait(1000);
            } catch (InterruptedException ie) {
            }
        }
    }

    /**
     * Return the txid of the last synced transaction.
     * For test use only
     */
    synchronized long getSyncTxId() {
        return synctxid;
    }

    // sets the initial capacity of the flush buffer.
    synchronized void setOutputBufferCapacity(int size) {
        journalSet.setOutputBufferCapacity(size);
    }

    /**
     * Create (or find if already exists) an edit output stream, which
     * streams journal records (edits) to the specified backup node.<br>
     * 
     * The new BackupNode will start receiving edits the next time this
     * NameNode's logs roll.
     * 
     * @param bnReg the backup node registration information.
     * @param nnReg this (active) name-node registration.
     * @throws IOException
     */
    synchronized void registerBackupNode(NamenodeRegistration bnReg, // backup node
            NamenodeRegistration nnReg) // active name-node
            throws IOException {
        if (bnReg.isRole(NamenodeRole.CHECKPOINT))
            return; // checkpoint node does not stream edits

        JournalManager jas = findBackupJournal(bnReg);
        if (jas != null) {
            // already registered
            LOG.info("Backup node " + bnReg + " re-registers");
            return;
        }

        LOG.info("Registering new backup node: " + bnReg);
        BackupJournalManager bjm = new BackupJournalManager(bnReg, nnReg);
        journalSet.add(bjm, false);
    }

    synchronized void releaseBackupStream(NamenodeRegistration registration) throws IOException {
        BackupJournalManager bjm = this.findBackupJournal(registration);
        if (bjm != null) {
            LOG.info("Removing backup journal " + bjm);
            journalSet.remove(bjm);
        }
    }

    /**
     * Find the JournalAndStream associated with this BackupNode.
     * 
     * @return null if it cannot be found
     */
    private synchronized BackupJournalManager findBackupJournal(NamenodeRegistration bnReg) {
        for (JournalManager bjm : journalSet.getJournalManagers()) {
            if ((bjm instanceof BackupJournalManager) && ((BackupJournalManager) bjm).matchesRegistration(bnReg)) {
                return (BackupJournalManager) bjm;
            }
        }
        return null;
    }

    /** Write the batch of edits to edit log. */
    public synchronized void journal(long firstTxId, int numTxns, byte[] data) {
        final long expectedTxId = getLastWrittenTxId() + 1;
        Preconditions.checkState(firstTxId == expectedTxId,
                "received txid batch starting at %s but expected txid %s", firstTxId, expectedTxId);
        setNextTxId(firstTxId + numTxns - 1);
        logEdit(data.length, data);
        logSync();
    }

    /**
     * Write an operation to the edit log. Do not sync to persistent
     * store yet.
     */
    synchronized void logEdit(final int length, final byte[] data) {
        long start = beginTransaction();

        try {
            editLogStream.writeRaw(data, 0, length);
        } catch (IOException ex) {
            // All journals have failed, it will be handled in logSync.
        }
        endTransaction(start);
    }

    /**
     * Run recovery on all journals to recover any unclosed segments
     */
    synchronized void recoverUnclosedStreams() {
        Preconditions.checkState(state == State.BETWEEN_LOG_SEGMENTS, "May not recover segments - wrong state: %s",
                state);
        try {
            journalSet.recoverUnfinalizedSegments();
        } catch (IOException ex) {
            // All journals have failed, it is handled in logSync.
        }
    }

    Collection<EditLogInputStream> selectInputStreams(long fromTxId, long toAtLeastTxId) throws IOException {
        return selectInputStreams(fromTxId, toAtLeastTxId, true);
    }

    /**
     * Select a list of input streams to load.
     * 
     * @param fromTxId first transaction in the selected streams
     * @param toAtLeast the selected streams must contain this transaction
     * @param inProgessOk set to true if in-progress streams are OK
     */
    public synchronized Collection<EditLogInputStream> selectInputStreams(long fromTxId, long toAtLeastTxId,
            boolean inProgressOk) throws IOException {
        List<EditLogInputStream> streams = new ArrayList<EditLogInputStream>();
        EditLogInputStream stream = journalSet.getInputStream(fromTxId, inProgressOk);
        while (stream != null) {
            streams.add(stream);
            // We're now looking for a higher range, so reset the fromTxId
            fromTxId = stream.getLastTxId() + 1;
            stream = journalSet.getInputStream(fromTxId, inProgressOk);
        }

        if (fromTxId <= toAtLeastTxId) {
            closeAllStreams(streams);
            throw new IOException(String.format("Gap in transactions. Expected to "
                    + "be able to read up until at least txid %d but unable to find any "
                    + "edit logs containing txid %d", toAtLeastTxId, fromTxId));
        }
        return streams;
    }

    /** 
     * Close all the streams in a collection
     * @param streams The list of streams to close
     */
    static void closeAllStreams(Iterable<EditLogInputStream> streams) {
        for (EditLogInputStream s : streams) {
            IOUtils.closeStream(s);
        }
    }

    /**
     * Retrieve the implementation class for a Journal scheme.
     * @param conf The configuration to retrieve the information from
     * @param uriScheme The uri scheme to look up.
     * @return the class of the journal implementation
     * @throws IllegalArgumentException if no class is configured for uri
     */
    static Class<? extends JournalManager> getJournalClass(Configuration conf, String uriScheme) {
        String key = DFSConfigKeys.DFS_NAMENODE_EDITS_PLUGIN_PREFIX + "." + uriScheme;
        Class<? extends JournalManager> clazz = null;
        try {
            clazz = conf.getClass(key, null, JournalManager.class);
        } catch (RuntimeException re) {
            throw new IllegalArgumentException("Invalid class specified for " + uriScheme, re);
        }

        if (clazz == null) {
            LOG.warn("No class configured for " + uriScheme + ", " + key + " is empty");
            throw new IllegalArgumentException("No class configured for " + uriScheme);
        }
        return clazz;
    }

    /**
     * Construct a custom journal manager.
     * The class to construct is taken from the configuration.
     * @param uri Uri to construct
     * @return The constructed journal manager
     * @throws IllegalArgumentException if no class is configured for uri
     */
    private JournalManager createJournal(URI uri) {
        Class<? extends JournalManager> clazz = getJournalClass(conf, uri.getScheme());

        try {
            Constructor<? extends JournalManager> cons = clazz.getConstructor(Configuration.class, URI.class);
            return cons.newInstance(conf, uri);
        } catch (Exception e) {
            throw new IllegalArgumentException("Unable to construct journal, " + uri, e);
        }
    }
}