org.apache.hadoop.hbase.index.Indexer.java Source code

Java tutorial

Introduction

Here is the source code for org.apache.hadoop.hbase.index.Indexer.java

Source

/*
 * Copyright 2014 The Apache Software Foundation
 *
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 *
 * http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package org.apache.hadoop.hbase.index;

import static org.apache.hadoop.hbase.index.util.IndexManagementUtil.rethrowIndexingException;

import java.io.DataInput;
import java.io.DataOutput;
import java.io.IOException;
import java.util.ArrayList;
import java.util.Collection;
import java.util.HashMap;
import java.util.List;
import java.util.Map;
import java.util.Map.Entry;
import java.util.concurrent.locks.ReentrantReadWriteLock;
import java.util.concurrent.locks.ReentrantReadWriteLock.ReadLock;

import org.apache.commons.logging.Log;
import org.apache.commons.logging.LogFactory;
import org.apache.hadoop.conf.Configuration;
import org.apache.hadoop.hbase.Coprocessor;
import org.apache.hadoop.hbase.CoprocessorEnvironment;
import org.apache.hadoop.hbase.HConstants;
import org.apache.hadoop.hbase.HRegionInfo;
import org.apache.hadoop.hbase.HTableDescriptor;
import org.apache.hadoop.hbase.KeyValue;
import org.apache.hadoop.hbase.client.Delete;
import org.apache.hadoop.hbase.client.Durability;
import org.apache.hadoop.hbase.client.Mutation;
import org.apache.hadoop.hbase.client.Put;
import org.apache.hadoop.hbase.coprocessor.BaseRegionObserver;
import org.apache.hadoop.hbase.coprocessor.ObserverContext;
import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment;
import org.apache.hadoop.hbase.regionserver.InternalScanner;
import org.apache.hadoop.hbase.regionserver.KeyValueScanner;
import org.apache.hadoop.hbase.regionserver.MiniBatchOperationInProgress;
import org.apache.hadoop.hbase.regionserver.ScanType;
import org.apache.hadoop.hbase.regionserver.Store;
import org.apache.hadoop.hbase.regionserver.wal.HLog;
import org.apache.hadoop.hbase.regionserver.wal.HLogKey;
import org.apache.hadoop.hbase.regionserver.wal.WALEdit;
import org.apache.hadoop.hbase.util.Pair;

import com.google.common.collect.Multimap;
import org.apache.hadoop.hbase.index.builder.IndexBuildManager;
import org.apache.hadoop.hbase.index.builder.IndexBuilder;
import org.apache.hadoop.hbase.index.builder.IndexBuildingFailureException;
import org.apache.hadoop.hbase.index.table.HTableInterfaceReference;
import org.apache.hadoop.hbase.index.util.ImmutableBytesPtr;
import org.apache.hadoop.hbase.index.util.IndexManagementUtil;
import org.apache.hadoop.hbase.index.wal.IndexedKeyValue;
import org.apache.hadoop.hbase.index.write.IndexFailurePolicy;
import org.apache.hadoop.hbase.index.write.IndexWriter;
import org.apache.hadoop.hbase.index.write.recovery.PerRegionIndexWriteCache;
import org.apache.hadoop.hbase.index.write.recovery.StoreFailuresInCachePolicy;
import org.apache.hadoop.hbase.index.write.recovery.TrackingParallelWriterIndexCommitter;
import org.apache.phoenix.util.MetaDataUtil;

/**
 * Do all the work of managing index updates from a single coprocessor. All Puts/Delets are passed
 * to an {@link IndexBuilder} to determine the actual updates to make.
 * <p>
 * If the WAL is enabled, these updates are then added to the WALEdit and attempted to be written to
 * the WAL after the WALEdit has been saved. If any of the index updates fail, this server is
 * immediately terminated and we rely on WAL replay to attempt the index updates again (see
 * {@link #preWALRestore(ObserverContext, HRegionInfo, HLogKey, WALEdit)}).
 * <p>
 * If the WAL is disabled, the updates are attempted immediately. No consistency guarantees are made
 * if the WAL is disabled - some or none of the index updates may be successful. All updates in a
 * single batch must have the same durability level - either everything gets written to the WAL or
 * nothing does. Currently, we do not support mixed-durability updates within a single batch. If you
 * want to have different durability levels, you only need to split the updates into two different
 * batches.
 */
public class Indexer extends BaseRegionObserver {

    private static final Log LOG = LogFactory.getLog(Indexer.class);

    /** WAL on this server */
    private HLog log;
    protected IndexWriter writer;
    protected IndexBuildManager builder;

    /** Configuration key for the {@link IndexBuilder} to use */
    public static final String INDEX_BUILDER_CONF_KEY = "index.builder";

    // Setup out locking on the index edits/WAL so we can be sure that we don't lose a roll a WAL edit
    // before an edit is applied to the index tables
    private static final ReentrantReadWriteLock INDEX_READ_WRITE_LOCK = new ReentrantReadWriteLock(true);
    public static final ReadLock INDEX_UPDATE_LOCK = INDEX_READ_WRITE_LOCK.readLock();

    /**
     * Configuration key for if the indexer should check the version of HBase is running. Generally,
     * you only want to ignore this for testing or for custom versions of HBase.
     */
    public static final String CHECK_VERSION_CONF_KEY = "com.saleforce.hbase.index.checkversion";

    private static final String INDEX_RECOVERY_FAILURE_POLICY_KEY = "org.apache.hadoop.hbase.index.recovery.failurepolicy";

    /**
     * Marker {@link KeyValue} to indicate that we are doing a batch operation. Needed because the
     * coprocessor framework throws away the WALEdit from the prePut/preDelete hooks when checking a
     * batch if there were no {@link KeyValue}s attached to the {@link WALEdit}. When you get down to
     * the preBatch hook, there won't be any WALEdits to which to add the index updates.
     */
    private static KeyValue BATCH_MARKER = new KeyValue();

    /**
     * cache the failed updates to the various regions. Used for making the WAL recovery mechanisms
     * more robust in the face of recoverying index regions that were on the same server as the
     * primary table region
     */
    private PerRegionIndexWriteCache failedIndexEdits = new PerRegionIndexWriteCache();

    /**
     * IndexWriter for writing the recovered index edits. Separate from the main indexer since we need
     * different write/failure policies
     */
    private IndexWriter recoveryWriter;

    private boolean stopped;
    private boolean disabled;

    public static final String RecoveryFailurePolicyKeyForTesting = INDEX_RECOVERY_FAILURE_POLICY_KEY;

    public static final int INDEXING_SUPPORTED_MAJOR_VERSION = MetaDataUtil.encodeMaxPatchVersion(0, 94);
    public static final int INDEXING_SUPPORTED__MIN_MAJOR_VERSION = MetaDataUtil.encodeVersion("0.94.0");
    private static final int INDEX_WAL_COMPRESSION_MINIMUM_SUPPORTED_VERSION = MetaDataUtil.encodeVersion("0.94.9");

    @Override
    public void start(CoprocessorEnvironment e) throws IOException {
        try {
            final RegionCoprocessorEnvironment env = (RegionCoprocessorEnvironment) e;
            String serverName = env.getRegionServerServices().getServerName().getServerName();
            if (env.getConfiguration().getBoolean(CHECK_VERSION_CONF_KEY, true)) {
                // make sure the right version <-> combinations are allowed.
                String errormsg = Indexer.validateVersion(env.getHBaseVersion(), env.getConfiguration());
                if (errormsg != null) {
                    IOException ioe = new IOException(errormsg);
                    env.getRegionServerServices().abort(errormsg, ioe);
                    throw ioe;
                }
            }

            this.builder = new IndexBuildManager(env);

            // get a reference to the WAL
            log = env.getRegionServerServices().getWAL();
            // add a synchronizer so we don't archive a WAL that we need
            log.registerWALActionsListener(new IndexLogRollSynchronizer(INDEX_READ_WRITE_LOCK.writeLock()));

            // setup the actual index writer
            this.writer = new IndexWriter(env, serverName + "-index-writer");

            // setup the recovery writer that does retries on the failed edits
            TrackingParallelWriterIndexCommitter recoveryCommmiter = new TrackingParallelWriterIndexCommitter();

            try {
                // get the specified failure policy. We only ever override it in tests, but we need to do it
                // here
                Class<? extends IndexFailurePolicy> policyClass = env.getConfiguration().getClass(
                        INDEX_RECOVERY_FAILURE_POLICY_KEY, StoreFailuresInCachePolicy.class,
                        IndexFailurePolicy.class);
                IndexFailurePolicy policy = policyClass.getConstructor(PerRegionIndexWriteCache.class)
                        .newInstance(failedIndexEdits);
                LOG.debug("Setting up recovery writter with committer: " + recoveryCommmiter.getClass()
                        + " and failure policy: " + policy.getClass());
                recoveryWriter = new IndexWriter(recoveryCommmiter, policy, env, serverName + "-recovery-writer");
            } catch (Exception ex) {
                throw new IOException("Could not instantiate recovery failure policy!", ex);
            }
        } catch (NoSuchMethodError ex) {
            disabled = true;
            super.start(e);
            LOG.error("Must be too early a version of HBase. Disabled coprocessor ", ex);
        }
    }

    @Override
    public void stop(CoprocessorEnvironment e) throws IOException {
        if (this.stopped) {
            return;
        }
        if (this.disabled) {
            super.stop(e);
            return;
        }
        this.stopped = true;
        String msg = "Indexer is being stopped";
        this.builder.stop(msg);
        this.writer.stop(msg);
        this.recoveryWriter.stop(msg);
    }

    @Override
    public void prePut(final ObserverContext<RegionCoprocessorEnvironment> c, final Put put, final WALEdit edit,
            final boolean writeToWAL) throws IOException {
        if (this.disabled) {
            super.prePut(c, put, edit, writeToWAL);
            return;
        }
        // just have to add a batch marker to the WALEdit so we get the edit again in the batch
        // processing step. We let it throw an exception here because something terrible has happened.
        edit.add(BATCH_MARKER);
    }

    @Override
    public void preDelete(ObserverContext<RegionCoprocessorEnvironment> e, Delete delete, WALEdit edit,
            boolean writeToWAL) throws IOException {
        if (this.disabled) {
            super.preDelete(e, delete, edit, writeToWAL);
            return;
        }
        try {
            preDeleteWithExceptions(e, delete, edit, writeToWAL);
            return;
        } catch (Throwable t) {
            rethrowIndexingException(t);
        }
        throw new RuntimeException(
                "Somehow didn't return an index update but also didn't propagate the failure to the client!");
    }

    public void preDeleteWithExceptions(ObserverContext<RegionCoprocessorEnvironment> e, Delete delete,
            WALEdit edit, boolean writeToWAL) throws Exception {
        // if we are making the update as part of a batch, we need to add in a batch marker so the WAL
        // is retained
        if (this.builder.getBatchId(delete) != null) {
            edit.add(BATCH_MARKER);
            return;
        }

        // get the mapping for index column -> target index table
        Collection<Pair<Mutation, byte[]>> indexUpdates = this.builder.getIndexUpdate(delete);

        if (doPre(indexUpdates, edit, writeToWAL)) {
            takeUpdateLock("delete");
        }
    }

    @Override
    public void preBatchMutate(ObserverContext<RegionCoprocessorEnvironment> c,
            MiniBatchOperationInProgress<Pair<Mutation, Integer>> miniBatchOp) throws IOException {
        if (this.disabled) {
            super.preBatchMutate(c, miniBatchOp);
            return;
        }
        try {
            preBatchMutateWithExceptions(c, miniBatchOp);
            return;
        } catch (Throwable t) {
            rethrowIndexingException(t);
        }
        throw new RuntimeException(
                "Somehow didn't return an index update but also didn't propagate the failure to the client!");
    }

    @SuppressWarnings("deprecation")
    public void preBatchMutateWithExceptions(ObserverContext<RegionCoprocessorEnvironment> c,
            MiniBatchOperationInProgress<Pair<Mutation, Integer>> miniBatchOp) throws Throwable {

        // first group all the updates for a single row into a single update to be processed
        Map<ImmutableBytesPtr, MultiMutation> mutations = new HashMap<ImmutableBytesPtr, MultiMutation>();
        boolean durable = false;
        for (int i = 0; i < miniBatchOp.size(); i++) {
            // remove the batch keyvalue marker - its added for all puts
            WALEdit edit = miniBatchOp.getWalEdit(i);
            // we don't have a WALEdit for immutable index cases, which still see this path
            // we could check is indexing is enable for the mutation in prePut and then just skip this
            // after checking here, but this saves us the checking again.
            if (edit != null) {
                KeyValue kv = edit.getKeyValues().remove(0);
                assert kv == BATCH_MARKER : "Expected batch marker from the WALEdit, but got: " + kv;
            }
            Pair<Mutation, Integer> op = miniBatchOp.getOperation(i);
            Mutation m = op.getFirst();
            // skip this mutation if we aren't enabling indexing
            // unfortunately, we really should ask if the raw mutation (rather than the combined mutation)
            // should be indexed, which means we need to expose another method on the builder. Such is the
            // way optimization go though.
            if (!this.builder.isEnabled(m)) {
                continue;
            }

            // figure out if this is batch is durable or not
            if (!durable) {
                durable = m.getDurability() != Durability.SKIP_WAL;
            }

            // add the mutation to the batch set
            ImmutableBytesPtr row = new ImmutableBytesPtr(m.getRow());
            MultiMutation stored = mutations.get(row);
            // we haven't seen this row before, so add it
            if (stored == null) {
                stored = new MultiMutation(row, m.getWriteToWAL());
                mutations.put(row, stored);
            }
            stored.addAll(m);
        }

        // early exit if it turns out we don't have any edits
        if (mutations.entrySet().size() == 0) {
            return;
        }

        // dump all the index updates into a single WAL. They will get combined in the end anyways, so
        // don't worry which one we get
        WALEdit edit = miniBatchOp.getWalEdit(0);

        // get the index updates for all elements in this batch
        Collection<Pair<Mutation, byte[]>> indexUpdates = this.builder.getIndexUpdate(miniBatchOp,
                mutations.values());
        // write them
        if (doPre(indexUpdates, edit, durable)) {
            takeUpdateLock("batch mutation");
        }
    }

    private void takeUpdateLock(String opDesc) throws IndexBuildingFailureException {
        boolean interrupted = false;
        // lock the log, so we are sure that index write gets atomically committed
        LOG.debug("Taking INDEX_UPDATE readlock for " + opDesc);
        // wait for the update lock
        while (!this.stopped) {
            try {
                INDEX_UPDATE_LOCK.lockInterruptibly();
                LOG.debug("Got the INDEX_UPDATE readlock for " + opDesc);
                // unlock the lock so the server can shutdown, if we find that we have stopped since getting
                // the lock
                if (this.stopped) {
                    INDEX_UPDATE_LOCK.unlock();
                    throw new IndexBuildingFailureException(
                            "Found server stop after obtaining the update lock, killing update attempt");
                }
                break;
            } catch (InterruptedException e) {
                LOG.info("Interrupted while waiting for update lock. Ignoring unless stopped");
                interrupted = true;
            }
        }
        if (interrupted) {
            Thread.currentThread().interrupt();
        }
    }

    private class MultiMutation extends Mutation {

        private ImmutableBytesPtr rowKey;

        public MultiMutation(ImmutableBytesPtr rowkey, boolean writeToWal) {
            this.rowKey = rowkey;
            this.writeToWAL = writeToWal;
        }

        /**
         * @param stored
         */
        @SuppressWarnings("deprecation")
        public void addAll(Mutation stored) {
            // add all the kvs
            for (Entry<byte[], List<KeyValue>> kvs : stored.getFamilyMap().entrySet()) {
                byte[] family = kvs.getKey();
                List<KeyValue> list = getKeyValueList(family, kvs.getValue().size());
                list.addAll(kvs.getValue());
                familyMap.put(family, list);
            }

            // add all the attributes, not overriding already stored ones
            for (Entry<String, byte[]> attrib : stored.getAttributesMap().entrySet()) {
                if (this.getAttribute(attrib.getKey()) == null) {
                    this.setAttribute(attrib.getKey(), attrib.getValue());
                }
            }
            if (stored.getWriteToWAL()) {
                this.writeToWAL = true;
            }
        }

        private List<KeyValue> getKeyValueList(byte[] family, int hint) {
            List<KeyValue> list = familyMap.get(family);
            if (list == null) {
                list = new ArrayList<KeyValue>(hint);
            }
            return list;
        }

        @Override
        public byte[] getRow() {
            return this.rowKey.copyBytesIfNecessary();
        }

        @Override
        public int hashCode() {
            return this.rowKey.hashCode();
        }

        @Override
        public boolean equals(Object o) {
            return o == null ? false : o.hashCode() == this.hashCode();
        }

        @Override
        public void readFields(DataInput arg0) throws IOException {
            throw new UnsupportedOperationException("MultiMutations cannot be read/written");
        }

        @Override
        public void write(DataOutput arg0) throws IOException {
            throw new UnsupportedOperationException("MultiMutations cannot be read/written");
        }
    }

    /**
     * Add the index updates to the WAL, or write to the index table, if the WAL has been disabled
     * @return <tt>true</tt> if the WAL has been updated.
     * @throws IOException
     */
    private boolean doPre(Collection<Pair<Mutation, byte[]>> indexUpdates, final WALEdit edit,
            final boolean writeToWAL) throws IOException {
        // no index updates, so we are done
        if (indexUpdates == null || indexUpdates.size() == 0) {
            return false;
        }

        // if writing to wal is disabled, we never see the WALEdit updates down the way, so do the index
        // update right away
        if (!writeToWAL) {
            try {
                this.writer.write(indexUpdates);
                return false;
            } catch (Throwable e) {
                LOG.error("Failed to update index with entries:" + indexUpdates, e);
                IndexManagementUtil.rethrowIndexingException(e);
            }
        }

        // we have all the WAL durability, so we just update the WAL entry and move on
        for (Pair<Mutation, byte[]> entry : indexUpdates) {
            edit.add(new IndexedKeyValue(entry.getSecond(), entry.getFirst()));
        }

        return true;
    }

    @Override
    public void postPut(ObserverContext<RegionCoprocessorEnvironment> e, Put put, WALEdit edit, boolean writeToWAL)
            throws IOException {
        if (this.disabled) {
            super.postPut(e, put, edit, writeToWAL);
            return;
        }
        doPost(edit, put, writeToWAL);
    }

    @Override
    public void postDelete(ObserverContext<RegionCoprocessorEnvironment> e, Delete delete, WALEdit edit,
            boolean writeToWAL) throws IOException {
        if (this.disabled) {
            super.postDelete(e, delete, edit, writeToWAL);
            return;
        }
        doPost(edit, delete, writeToWAL);
    }

    @Override
    public void postBatchMutate(ObserverContext<RegionCoprocessorEnvironment> c,
            MiniBatchOperationInProgress<Pair<Mutation, Integer>> miniBatchOp) throws IOException {
        if (this.disabled) {
            super.postBatchMutate(c, miniBatchOp);
            return;
        }
        this.builder.batchCompleted(miniBatchOp);
        // noop for the rest of the indexer - its handled by the first call to put/delete
    }

    private void doPost(WALEdit edit, Mutation m, boolean writeToWAL) throws IOException {
        try {
            doPostWithExceptions(edit, m, writeToWAL);
            return;
        } catch (Throwable e) {
            rethrowIndexingException(e);
        }
        throw new RuntimeException(
                "Somehow didn't complete the index update, but didn't return succesfully either!");
    }

    private void doPostWithExceptions(WALEdit edit, Mutation m, boolean writeToWAL) throws Exception {
        //short circuit, if we don't need to do any work
        if (!writeToWAL || !this.builder.isEnabled(m)) {
            // already did the index update in prePut, so we are done
            return;
        }

        // there is a little bit of excess here- we iterate all the non-indexed kvs for this check first
        // and then do it again later when getting out the index updates. This should be pretty minor
        // though, compared to the rest of the runtime
        IndexedKeyValue ikv = getFirstIndexedKeyValue(edit);
        /*
         * early exit - we have nothing to write, so we don't need to do anything else. NOTE: we don't
         * release the WAL Rolling lock (INDEX_UPDATE_LOCK) since we never take it in doPre if there are
         * no index updates.
         */
        if (ikv == null) {
            return;
        }

        /*
         * only write the update if we haven't already seen this batch. We only want to write the batch
         * once (this hook gets called with the same WALEdit for each Put/Delete in a batch, which can
         * lead to writing all the index updates for each Put/Delete).
         */
        if (!ikv.getBatchFinished()) {
            Collection<Pair<Mutation, byte[]>> indexUpdates = extractIndexUpdate(edit);

            // the WAL edit is kept in memory and we already specified the factory when we created the
            // references originally - therefore, we just pass in a null factory here and use the ones
            // already specified on each reference
            try {
                writer.writeAndKillYourselfOnFailure(indexUpdates);
            } finally {
                // With a custom kill policy, we may throw instead of kill the server.
                // Without doing this in a finally block (at least with the mini cluster),
                // the region server never goes down.

                // mark the batch as having been written. In the single-update case, this never gets check
                // again, but in the batch case, we will check it again (see above).
                ikv.markBatchFinished();

                // release the lock on the index, we wrote everything properly
                // we took the lock for each Put/Delete, so we have to release it a matching number of times
                // batch cases only take the lock once, so we need to make sure we don't over-release the
                // lock.
                LOG.debug("Releasing INDEX_UPDATE readlock");
                INDEX_UPDATE_LOCK.unlock();
            }
        }
    }

    /**
     * Search the {@link WALEdit} for the first {@link IndexedKeyValue} present
     * @param edit {@link WALEdit}
     * @return the first {@link IndexedKeyValue} in the {@link WALEdit} or <tt>null</tt> if not
     *         present
     */
    private IndexedKeyValue getFirstIndexedKeyValue(WALEdit edit) {
        for (KeyValue kv : edit.getKeyValues()) {
            if (kv instanceof IndexedKeyValue) {
                return (IndexedKeyValue) kv;
            }
        }
        return null;
    }

    /**
     * Extract the index updates from the WAL Edit
     * @param edit to search for index updates
     * @return the mutations to apply to the index tables
     */
    private Collection<Pair<Mutation, byte[]>> extractIndexUpdate(WALEdit edit) {
        Collection<Pair<Mutation, byte[]>> indexUpdates = new ArrayList<Pair<Mutation, byte[]>>();
        for (KeyValue kv : edit.getKeyValues()) {
            if (kv instanceof IndexedKeyValue) {
                IndexedKeyValue ikv = (IndexedKeyValue) kv;
                indexUpdates.add(new Pair<Mutation, byte[]>(ikv.getMutation(), ikv.getIndexTable()));
            }
        }

        return indexUpdates;
    }

    @Override
    public void postOpen(final ObserverContext<RegionCoprocessorEnvironment> c) {
        Multimap<HTableInterfaceReference, Mutation> updates = failedIndexEdits
                .getEdits(c.getEnvironment().getRegion());

        if (this.disabled) {
            super.postOpen(c);
            return;
        }
        LOG.info("Found some outstanding index updates that didn't succeed during"
                + " WAL replay - attempting to replay now.");
        //if we have no pending edits to complete, then we are done
        if (updates == null || updates.size() == 0) {
            return;
        }

        // do the usual writer stuff, killing the server again, if we can't manage to make the index
        // writes succeed again
        try {
            writer.writeAndKillYourselfOnFailure(updates);
        } catch (IOException e) {
            LOG.error("Exception thrown instead of killing server during index writing", e);
        }
    }

    @Override
    public void preWALRestore(ObserverContext<RegionCoprocessorEnvironment> env, HRegionInfo info, HLogKey logKey,
            WALEdit logEdit) throws IOException {
        if (this.disabled) {
            super.preWALRestore(env, info, logKey, logEdit);
            return;
        }
        // TODO check the regions in transition. If the server on which the region lives is this one,
        // then we should rety that write later in postOpen.
        // we might be able to get even smarter here and pre-split the edits that are server-local
        // into their own recovered.edits file. This then lets us do a straightforward recovery of each
        // region (and more efficiently as we aren't writing quite as hectically from this one place).

        /*
         * Basically, we let the index regions recover for a little while long before retrying in the
         * hopes they come up before the primary table finishes.
         */
        Collection<Pair<Mutation, byte[]>> indexUpdates = extractIndexUpdate(logEdit);
        recoveryWriter.writeAndKillYourselfOnFailure(indexUpdates);
    }

    /**
     * Create a custom {@link InternalScanner} for a compaction that tracks the versions of rows that
     * are removed so we can clean then up from the the index table(s).
     * <p>
     * This is not yet implemented - its not clear if we should even mess around with the Index table
     * for these rows as those points still existed. TODO: v2 of indexing
     */
    @Override
    public InternalScanner preCompactScannerOpen(ObserverContext<RegionCoprocessorEnvironment> c, Store store,
            List<? extends KeyValueScanner> scanners, ScanType scanType, long earliestPutTs, InternalScanner s)
            throws IOException {
        return super.preCompactScannerOpen(c, store, scanners, scanType, earliestPutTs, s);
    }

    /**
     * Exposed for testing!
     * @return the currently instantiated index builder
     */
    public IndexBuilder getBuilderForTesting() {
        return this.builder.getBuilderForTesting();
    }

    /**
     * Validate that the version and configuration parameters are supported
     * @param hbaseVersion current version of HBase on which <tt>this</tt> coprocessor is installed
     * @param conf configuration to check for allowed parameters (e.g. WAL Compression only if >=
     *            0.94.9)
     * @return <tt>null</tt> if the version is supported, the error message to display otherwise
     */
    public static String validateVersion(String hbaseVersion, Configuration conf) {
        int encodedVersion = MetaDataUtil.encodeVersion(hbaseVersion);
        // above 0.94 everything should be supported
        if (encodedVersion > INDEXING_SUPPORTED_MAJOR_VERSION) {
            return null;
        }
        // check to see if its at least 0.94
        if (encodedVersion < INDEXING_SUPPORTED__MIN_MAJOR_VERSION) {
            return "Indexing not supported for versions older than 0.94.X";
        }
        // if less than 0.94.9, we need to check if WAL Compression is enabled
        if (encodedVersion < INDEX_WAL_COMPRESSION_MINIMUM_SUPPORTED_VERSION) {
            if (conf.getBoolean(HConstants.ENABLE_WAL_COMPRESSION, false)) {
                return "Indexing not supported with WAL Compression for versions of HBase older than 0.94.9 - found version:"
                        + hbaseVersion;
            }
        }
        return null;
    }

    /**
     * Enable indexing on the given table
     * @param desc {@link HTableDescriptor} for the table on which indexing should be enabled
     * @param builder class to use when building the index for this table
     * @param properties map of custom configuration options to make available to your
     *          {@link IndexBuilder} on the server-side
     * @throws IOException the Indexer coprocessor cannot be added
     */
    public static void enableIndexing(HTableDescriptor desc, Class<? extends IndexBuilder> builder,
            Map<String, String> properties) throws IOException {
        if (properties == null) {
            properties = new HashMap<String, String>();
        }
        properties.put(Indexer.INDEX_BUILDER_CONF_KEY, builder.getName());
        desc.addCoprocessor(Indexer.class.getName(), null, Coprocessor.PRIORITY_USER, properties);
    }
}