org.apache.solr.hadoop.MorphlineGoLiveMiniMRTest.java Source code

Java tutorial

Introduction

Here is the source code for org.apache.solr.hadoop.MorphlineGoLiveMiniMRTest.java

Source

/*
 * Licensed to the Apache Software Foundation (ASF) under one or more
 * contributor license agreements.  See the NOTICE file distributed with
 * this work for additional information regarding copyright ownership.
 * The ASF licenses this file to You under the Apache License, Version 2.0
 * (the "License"); you may not use this file except in compliance with
 * the License.  You may obtain a copy of the License at
 *
 *     http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing, software
 * distributed under the License is distributed on an "AS IS" BASIS,
 * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
 * See the License for the specific language governing permissions and
 * limitations under the License.
 */
package org.apache.solr.hadoop;

import java.io.File;
import java.io.IOException;
import java.io.OutputStream;
import java.io.OutputStreamWriter;
import java.io.UnsupportedEncodingException;
import java.io.Writer;
import java.lang.reflect.Array;
import java.net.URI;
import java.util.ArrayList;
import java.util.Collection;
import java.util.HashSet;
import java.util.List;
import java.util.Set;

import org.apache.hadoop.fs.FileSystem;
import org.apache.hadoop.fs.Path;
import org.apache.hadoop.fs.permission.FsPermission;
import org.apache.hadoop.hdfs.MiniDFSCluster;
import org.apache.hadoop.mapred.JobConf;
import org.apache.hadoop.mapred.MiniMRCluster;
import org.apache.hadoop.mapreduce.Job;
import org.apache.hadoop.security.authorize.ProxyUsers;
import org.apache.hadoop.util.JarFinder;
import org.apache.hadoop.util.ToolRunner;
import org.apache.lucene.util.LuceneTestCase.SuppressCodecs;
import org.apache.solr.client.solrj.SolrQuery;
import org.apache.solr.client.solrj.SolrServerException;
import org.apache.solr.client.solrj.embedded.JettySolrRunner;
import org.apache.solr.client.solrj.impl.HttpSolrServer;
import org.apache.solr.client.solrj.request.QueryRequest;
import org.apache.solr.client.solrj.response.QueryResponse;
import org.apache.solr.cloud.AbstractFullDistribZkTestBase;
import org.apache.solr.common.cloud.Replica;
import org.apache.solr.common.cloud.Slice;
import org.apache.solr.common.cloud.SolrZkClient;
import org.apache.solr.common.cloud.ZkCoreNodeProps;
import org.apache.solr.common.params.ModifiableSolrParams;
import org.apache.solr.common.params.CollectionParams.CollectionAction;
import org.apache.solr.common.util.NamedList;
import org.apache.solr.handler.extraction.ExtractingParams;
import org.junit.After;
import org.junit.AfterClass;
import org.junit.Before;
import org.junit.BeforeClass;
import org.junit.Test;

import com.carrotsearch.randomizedtesting.annotations.ThreadLeakAction;
import com.carrotsearch.randomizedtesting.annotations.ThreadLeakAction.Action;
import com.carrotsearch.randomizedtesting.annotations.ThreadLeakLingering;
import com.carrotsearch.randomizedtesting.annotations.ThreadLeakScope;
import com.carrotsearch.randomizedtesting.annotations.ThreadLeakScope.Scope;
import com.carrotsearch.randomizedtesting.annotations.ThreadLeakZombies;
import com.carrotsearch.randomizedtesting.annotations.ThreadLeakZombies.Consequence;

@ThreadLeakAction({ Action.WARN })
@ThreadLeakLingering(linger = 0)
@ThreadLeakZombies(Consequence.CONTINUE)
@ThreadLeakScope(Scope.NONE)
@SuppressCodecs({ "Lucene3x", "Lucene40" })
public class MorphlineGoLiveMiniMRTest extends AbstractFullDistribZkTestBase {

    private static final String RESOURCES_DIR = "target/test-classes";
    private static final String DOCUMENTS_DIR = RESOURCES_DIR + "/test-documents";
    private static final File MINIMR_CONF_DIR = new File(RESOURCES_DIR + "/solr/minimr");

    private static final String SEARCH_ARCHIVES_JAR = JarFinder.getJar(MapReduceIndexerTool.class);

    private static MiniDFSCluster dfsCluster = null;
    private static MiniMRCluster mrCluster = null;
    private static int numRuns = 0;

    private final String inputAvroFile1;
    private final String inputAvroFile2;
    private final String inputAvroFile3;

    @Override
    public String getSolrHome() {
        return MINIMR_CONF_DIR.getAbsolutePath();
    }

    public MorphlineGoLiveMiniMRTest() {
        this.inputAvroFile1 = "sample-statuses-20120521-100919.avro";
        this.inputAvroFile2 = "sample-statuses-20120906-141433.avro";
        this.inputAvroFile3 = "sample-statuses-20120906-141433-medium.avro";

        fixShardCount = true;
        sliceCount = 3;
        shardCount = 3;
    }

    private static boolean isYarn() {
        try {
            Job.class.getMethod("getCluster");
            return true;
        } catch (NoSuchMethodException e) {
            return false;
        }
    }

    @BeforeClass
    public static void setupClass() throws Exception {
        //    if (isYarn()) {
        //      org.junit.Assume.assumeTrue(false); // ignore test on Yarn until CDH-10420 is fixed
        //    }
        if (System.getProperty("hadoop.log.dir") == null) {
            System.setProperty("hadoop.log.dir", "target");
        }
        int taskTrackers = 2;
        int dataNodes = 2;

        System.setProperty("solr.hdfs.blockcache.enabled", "false");

        JobConf conf = new JobConf();
        conf.set("dfs.block.access.token.enable", "false");
        conf.set("dfs.permissions", "true");
        conf.set("hadoop.security.authentication", "simple");

        createTempDir();
        System.setProperty("test.build.data", dataDir + File.separator + "hdfs" + File.separator + "build");
        System.setProperty("test.cache.data", dataDir + File.separator + "hdfs" + File.separator + "cache");

        dfsCluster = new MiniDFSCluster(conf, dataNodes, true, null);
        FileSystem fileSystem = dfsCluster.getFileSystem();
        fileSystem.mkdirs(new Path("/tmp"));
        fileSystem.mkdirs(new Path("/user"));
        fileSystem.mkdirs(new Path("/hadoop/mapred/system"));
        fileSystem.setPermission(new Path("/tmp"), FsPermission.valueOf("-rwxrwxrwx"));
        fileSystem.setPermission(new Path("/user"), FsPermission.valueOf("-rwxrwxrwx"));
        fileSystem.setPermission(new Path("/hadoop/mapred/system"), FsPermission.valueOf("-rwx------"));
        String nnURI = fileSystem.getUri().toString();
        int numDirs = 1;
        String[] racks = null;
        String[] hosts = null;

        mrCluster = new MiniMRCluster(0, 0, taskTrackers, nnURI, numDirs, racks, hosts, null, conf);
        ProxyUsers.refreshSuperUserGroupsConfiguration(conf);
    }

    @Override
    @Before
    public void setUp() throws Exception {
        super.setUp();
        System.setProperty("host", "127.0.0.1");
        System.setProperty("numShards", Integer.toString(sliceCount));
        URI uri = dfsCluster.getFileSystem().getUri();
        System.setProperty("solr.hdfs.home", uri.toString() + "/" + this.getClass().getName());
        uploadConfFiles();
    }

    @Override
    @After
    public void tearDown() throws Exception {
        super.tearDown();
        System.clearProperty("host");
        System.clearProperty("numShards");
        System.clearProperty("solr.hdfs.home");
    }

    @AfterClass
    public static void teardownClass() throws Exception {
        System.clearProperty("solr.hdfs.blockcache.enabled");
        if (mrCluster != null) {
            mrCluster.shutdown();
            mrCluster = null;
        }
        if (dfsCluster != null) {
            dfsCluster.shutdown();
            dfsCluster = null;
        }
        FileSystem.closeAll();
    }

    private JobConf getJobConf() {
        return mrCluster.createJobConf();
    }

    @Test
    @Override
    public void testDistribSearch() throws Exception {
        super.testDistribSearch();
    }

    @Test
    public void testBuildShardUrls() throws Exception {
        // 2x3
        Integer numShards = 2;
        List<Object> urls = new ArrayList<Object>();
        urls.add("shard1");
        urls.add("shard2");
        urls.add("shard3");
        urls.add("shard4");
        urls.add("shard5");
        urls.add("shard6");
        List<List<String>> shardUrls = MapReduceIndexerTool.buildShardUrls(urls, numShards);

        assertEquals(shardUrls.toString(), 2, shardUrls.size());

        for (List<String> u : shardUrls) {
            assertEquals(3, u.size());
        }

        // 1x6
        numShards = 1;
        shardUrls = MapReduceIndexerTool.buildShardUrls(urls, numShards);

        assertEquals(shardUrls.toString(), 1, shardUrls.size());

        for (List<String> u : shardUrls) {
            assertEquals(6, u.size());
        }

        // 6x1
        numShards = 6;
        shardUrls = MapReduceIndexerTool.buildShardUrls(urls, numShards);

        assertEquals(shardUrls.toString(), 6, shardUrls.size());

        for (List<String> u : shardUrls) {
            assertEquals(1, u.size());
        }

        // 3x2
        numShards = 3;
        shardUrls = MapReduceIndexerTool.buildShardUrls(urls, numShards);

        assertEquals(shardUrls.toString(), 3, shardUrls.size());

        for (List<String> u : shardUrls) {
            assertEquals(2, u.size());
        }

        // null shards, 6x1
        numShards = null;
        shardUrls = MapReduceIndexerTool.buildShardUrls(urls, numShards);

        assertEquals(shardUrls.toString(), 6, shardUrls.size());

        for (List<String> u : shardUrls) {
            assertEquals(1, u.size());
        }

        // null shards 3x1
        numShards = null;

        urls = new ArrayList<Object>();
        urls.add("shard1");
        urls.add("shard2");
        urls.add("shard3");

        shardUrls = MapReduceIndexerTool.buildShardUrls(urls, numShards);

        assertEquals(shardUrls.toString(), 3, shardUrls.size());

        for (List<String> u : shardUrls) {
            assertEquals(1, u.size());
        }

        // 2x(2,3) off balance
        numShards = 2;
        urls = new ArrayList<Object>();
        urls.add("shard1");
        urls.add("shard2");
        urls.add("shard3");
        urls.add("shard4");
        urls.add("shard5");
        shardUrls = MapReduceIndexerTool.buildShardUrls(urls, numShards);

        assertEquals(shardUrls.toString(), 2, shardUrls.size());

        Set<Integer> counts = new HashSet<Integer>();
        counts.add(shardUrls.get(0).size());
        counts.add(shardUrls.get(1).size());

        assertTrue(counts.contains(2));
        assertTrue(counts.contains(3));
    }

    private String[] prependInitialArgs(String[] args) {
        String[] head = new String[] {
                "--morphline-file=" + RESOURCES_DIR + "/test-morphlines/solrCellDocumentTypes.conf",
                "--morphline-id=morphline1", };
        return concat(head, args);
    }

    @Override
    public void doTest() throws Exception {

        waitForRecoveriesToFinish(false);

        FileSystem fs = dfsCluster.getFileSystem();
        Path inDir = fs.makeQualified(new Path("/user/testing/testMapperReducer/input"));
        fs.delete(inDir, true);
        String DATADIR = "/user/testing/testMapperReducer/data";
        Path dataDir = fs.makeQualified(new Path(DATADIR));
        fs.delete(dataDir, true);
        Path outDir = fs.makeQualified(new Path("/user/testing/testMapperReducer/output"));
        fs.delete(outDir, true);

        assertTrue(fs.mkdirs(inDir));
        Path INPATH = upAvroFile(fs, inDir, DATADIR, dataDir, inputAvroFile1);

        JobConf jobConf = getJobConf();
        // enable mapred.job.tracker = local to run in debugger and set breakpoints
        // jobConf.set("mapred.job.tracker", "local");
        jobConf.setMaxMapAttempts(1);
        jobConf.setMaxReduceAttempts(1);
        jobConf.setJar(SEARCH_ARCHIVES_JAR);
        jobConf.setBoolean(ExtractingParams.IGNORE_TIKA_EXCEPTION, false);

        MapReduceIndexerTool tool;
        int res;
        QueryResponse results;
        HttpSolrServer server = new HttpSolrServer(cloudJettys.get(0).url);

        String[] args = new String[] { "--solr-home-dir=" + MINIMR_CONF_DIR.getAbsolutePath(),
                "--output-dir=" + outDir.toString(), "--mappers=3",
                ++numRuns % 2 == 0 ? "--input-list=" + INPATH.toString() : dataDir.toString(), "--shard-url",
                cloudJettys.get(0).url, "--shard-url", cloudJettys.get(1).url, "--shard-url",
                cloudJettys.get(2).url, "--go-live-threads", Integer.toString(random().nextInt(15) + 1),
                "--verbose", "--go-live" };
        args = prependInitialArgs(args);

        if (true) {
            tool = new MapReduceIndexerTool();
            res = ToolRunner.run(jobConf, tool, args);
            assertEquals(0, res);
            assertTrue(tool.job.isComplete());
            assertTrue(tool.job.isSuccessful());
            results = server.query(new SolrQuery("*:*"));
            assertEquals(20, results.getResults().getNumFound());
        }

        fs.delete(inDir, true);
        fs.delete(outDir, true);
        fs.delete(dataDir, true);
        assertTrue(fs.mkdirs(inDir));
        INPATH = upAvroFile(fs, inDir, DATADIR, dataDir, inputAvroFile2);

        args = new String[] { "--solr-home-dir=" + MINIMR_CONF_DIR.getAbsolutePath(),
                "--output-dir=" + outDir.toString(), "--mappers=3", "--verbose", "--go-live",
                ++numRuns % 2 == 0 ? "--input-list=" + INPATH.toString() : dataDir.toString(), "--shard-url",
                cloudJettys.get(0).url, "--shard-url", cloudJettys.get(1).url, "--shard-url",
                cloudJettys.get(2).url, "--go-live-threads", Integer.toString(random().nextInt(15) + 1) };
        args = prependInitialArgs(args);

        if (true) {
            tool = new MapReduceIndexerTool();
            res = ToolRunner.run(jobConf, tool, args);
            assertEquals(0, res);
            assertTrue(tool.job.isComplete());
            assertTrue(tool.job.isSuccessful());
            results = server.query(new SolrQuery("*:*"));

            assertEquals(22, results.getResults().getNumFound());
        }

        // try using zookeeper
        String collection = "collection1";
        if (random().nextBoolean()) {
            // sometimes, use an alias
            createAlias("updatealias", "collection1");
            collection = "updatealias";
        }

        fs.delete(inDir, true);
        fs.delete(outDir, true);
        fs.delete(dataDir, true);
        INPATH = upAvroFile(fs, inDir, DATADIR, dataDir, inputAvroFile3);

        args = new String[] { "--output-dir=" + outDir.toString(), "--mappers=3", "--reducers=6", "--verbose",
                "--go-live", ++numRuns % 2 == 0 ? "--input-list=" + INPATH.toString() : dataDir.toString(),
                "--zk-host", zkServer.getZkAddress(), "--collection", collection };
        args = prependInitialArgs(args);

        if (true) {
            tool = new MapReduceIndexerTool();
            res = ToolRunner.run(jobConf, tool, args);
            assertEquals(0, res);
            assertTrue(tool.job.isComplete());
            assertTrue(tool.job.isSuccessful());

            results = server.query(new SolrQuery("*:*"));
            assertEquals(2126, results.getResults().getNumFound());
        }

        server.shutdown();

        // try using zookeeper with replication
        String replicatedCollection = "replicated_collection";
        createCollection(replicatedCollection, 2, 3, 2);
        waitForRecoveriesToFinish(false);
        cloudClient.setDefaultCollection(replicatedCollection);
        fs.delete(inDir, true);
        fs.delete(outDir, true);
        fs.delete(dataDir, true);
        assertTrue(fs.mkdirs(dataDir));
        INPATH = upAvroFile(fs, inDir, DATADIR, dataDir, inputAvroFile3);

        args = new String[] { "--solr-home-dir=" + MINIMR_CONF_DIR.getAbsolutePath(),
                "--output-dir=" + outDir.toString(), "--mappers=3", "--reducers=6", "--verbose", "--go-live",
                "--zk-host", zkServer.getZkAddress(), "--collection", replicatedCollection, dataDir.toString() };
        args = prependInitialArgs(args);

        if (true) {
            tool = new MapReduceIndexerTool();
            res = ToolRunner.run(jobConf, tool, args);
            assertEquals(0, res);
            assertTrue(tool.job.isComplete());
            assertTrue(tool.job.isSuccessful());

            results = cloudClient.query(new SolrQuery("*:*"));
            assertEquals(2104, results.getResults().getNumFound());

            checkConsistency(replicatedCollection);
        }

        // try using solr_url with replication
        cloudClient.deleteByQuery("*:*");
        cloudClient.commit();
        fs.delete(inDir, true);
        fs.delete(dataDir, true);
        assertTrue(fs.mkdirs(dataDir));
        INPATH = upAvroFile(fs, inDir, DATADIR, dataDir, inputAvroFile3);

        args = new String[] { "--solr-home-dir=" + MINIMR_CONF_DIR.getAbsolutePath(),
                "--output-dir=" + outDir.toString(), "--shards", "2", "--mappers=3", "--verbose", "--go-live",
                "--go-live-threads", Integer.toString(random().nextInt(15) + 1), dataDir.toString() };
        args = prependInitialArgs(args);

        List<String> argList = new ArrayList<String>();
        getShardUrlArgs(argList, replicatedCollection);
        args = concat(args, argList.toArray(new String[0]));

        if (true) {
            tool = new MapReduceIndexerTool();
            res = ToolRunner.run(jobConf, tool, args);
            assertEquals(0, res);
            assertTrue(tool.job.isComplete());
            assertTrue(tool.job.isSuccessful());

            checkConsistency(replicatedCollection);

            results = cloudClient.query(new SolrQuery("*:*"));
            assertEquals(2104, results.getResults().getNumFound());
        }

    }

    private void checkConsistency(String replicatedCollection) throws SolrServerException {
        Collection<Slice> slices = cloudClient.getZkStateReader().getClusterState().getSlices(replicatedCollection);
        for (Slice slice : slices) {
            Collection<Replica> replicas = slice.getReplicas();
            long found = -1;
            for (Replica replica : replicas) {
                HttpSolrServer client = new HttpSolrServer(new ZkCoreNodeProps(replica).getCoreUrl());
                SolrQuery query = new SolrQuery("*:*");
                query.set("distrib", false);
                QueryResponse replicaResults = client.query(query);
                long count = replicaResults.getResults().getNumFound();
                if (found != -1) {
                    assertEquals(slice.getName() + " is inconsistent " + new ZkCoreNodeProps(replica).getCoreUrl(),
                            found, count);
                }
                found = count;
            }
        }
    }

    private void getShardUrlArgs(List<String> args, String replicatedCollection) {
        Collection<Slice> slices = cloudClient.getZkStateReader().getClusterState().getSlices(replicatedCollection);
        for (Slice slice : slices) {
            Collection<Replica> replicas = slice.getReplicas();
            for (Replica replica : replicas) {
                args.add("--shard-url");
                args.add(new ZkCoreNodeProps(replica).getCoreUrl());
            }
        }
    }

    private Path upAvroFile(FileSystem fs, Path inDir, String DATADIR, Path dataDir, String localFile)
            throws IOException, UnsupportedEncodingException {
        Path INPATH = new Path(inDir, "input.txt");
        OutputStream os = fs.create(INPATH);
        Writer wr = new OutputStreamWriter(os, "UTF-8");
        wr.write(DATADIR + File.separator + localFile);
        wr.close();

        assertTrue(fs.mkdirs(dataDir));
        fs.copyFromLocalFile(new Path(DOCUMENTS_DIR, localFile), dataDir);
        return INPATH;
    }

    @Override
    public JettySolrRunner createJetty(File solrHome, String dataDir, String shardList, String solrConfigOverride,
            String schemaOverride) throws Exception {

        JettySolrRunner jetty = new JettySolrRunner(solrHome.getAbsolutePath(), context, 0, solrConfigOverride,
                schemaOverride);

        jetty.setShards(shardList);

        if (System.getProperty("collection") == null) {
            System.setProperty("collection", "collection1");
        }

        jetty.start();

        System.clearProperty("collection");

        return jetty;
    }

    private static void putConfig(SolrZkClient zkClient, File solrhome, String name) throws Exception {
        putConfig(zkClient, solrhome, name, name);
    }

    private static void putConfig(SolrZkClient zkClient, File solrhome, String srcName, String destName)
            throws Exception {

        File file = new File(solrhome, "conf" + File.separator + srcName);
        if (!file.exists()) {
            // LOG.info("skipping " + file.getAbsolutePath() +
            // " because it doesn't exist");
            return;
        }

        String destPath = "/configs/conf1/" + destName;
        // LOG.info("put " + file.getAbsolutePath() + " to " + destPath);
        zkClient.makePath(destPath, file, false, true);
    }

    private void uploadConfFiles() throws Exception {
        // upload our own config files
        SolrZkClient zkClient = new SolrZkClient(zkServer.getZkAddress(), 10000);
        putConfig(zkClient, new File(RESOURCES_DIR + "/solr/solrcloud"), "solrconfig.xml");
        putConfig(zkClient, MINIMR_CONF_DIR, "schema.xml");
        putConfig(zkClient, MINIMR_CONF_DIR, "elevate.xml");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_en.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_ar.txt");

        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_bg.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_ca.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_cz.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_da.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_el.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_es.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_eu.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_de.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_fa.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_fi.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_fr.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_ga.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_gl.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_hi.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_hu.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_hy.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_id.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_it.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_ja.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_lv.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_nl.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_no.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_pt.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_ro.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_ru.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_sv.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_th.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stopwords_tr.txt");

        putConfig(zkClient, MINIMR_CONF_DIR, "lang/contractions_ca.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/contractions_fr.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/contractions_ga.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "lang/contractions_it.txt");

        putConfig(zkClient, MINIMR_CONF_DIR, "lang/stemdict_nl.txt");

        putConfig(zkClient, MINIMR_CONF_DIR, "lang/hyphenations_ga.txt");

        putConfig(zkClient, MINIMR_CONF_DIR, "stopwords.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "protwords.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "currency.xml");
        putConfig(zkClient, MINIMR_CONF_DIR, "open-exchange-rates.json");
        putConfig(zkClient, MINIMR_CONF_DIR, "mapping-ISOLatin1Accent.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "old_synonyms.txt");
        putConfig(zkClient, MINIMR_CONF_DIR, "synonyms.txt");
        zkClient.close();
    }

    protected static <T> T[] concat(T[]... arrays) {
        if (arrays.length <= 0) {
            throw new IllegalArgumentException();
        }
        Class clazz = null;
        int length = 0;
        for (T[] array : arrays) {
            clazz = array.getClass();
            length += array.length;
        }
        T[] result = (T[]) Array.newInstance(clazz.getComponentType(), length);
        int pos = 0;
        for (T[] array : arrays) {
            System.arraycopy(array, 0, result, pos, array.length);
            pos += array.length;
        }
        return result;
    }

    private NamedList<Object> createAlias(String alias, String collections)
            throws SolrServerException, IOException {
        ModifiableSolrParams params = new ModifiableSolrParams();
        params.set("collections", collections);
        params.set("name", alias);
        params.set("action", CollectionAction.CREATEALIAS.toString());
        QueryRequest request = new QueryRequest(params);
        request.setPath("/admin/collections");
        return cloudClient.request(request);
    }

}