com.metamx.druid.loading.SingleSegmentLoader.java Source code

Java tutorial

Introduction

Here is the source code for com.metamx.druid.loading.SingleSegmentLoader.java

Source

/*
 * Druid - a distributed column store.
 * Copyright (C) 2012  Metamarkets Group Inc.
 *
 * This program is free software; you can redistribute it and/or
 * modify it under the terms of the GNU General Public License
 * as published by the Free Software Foundation; either version 2
 * of the License, or (at your option) any later version.
 *
 * This program is distributed in the hope that it will be useful,
 * but WITHOUT ANY WARRANTY; without even the implied warranty of
 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
 * GNU General Public License for more details.
 *
 * You should have received a copy of the GNU General Public License
 * along with this program; if not, write to the Free Software
 * Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA  02110-1301, USA.
 */

package com.metamx.druid.loading;

import com.google.common.base.Preconditions;
import com.google.common.collect.ImmutableList;
import com.google.common.collect.Sets;
import com.google.common.primitives.Longs;
import com.google.inject.Inject;
import com.metamx.common.IAE;
import com.metamx.common.ISE;
import com.metamx.common.logger.Logger;
import com.metamx.druid.client.DataSegment;
import com.metamx.druid.index.QueryableIndex;
import com.metamx.druid.index.QueryableIndexSegment;
import com.metamx.druid.index.Segment;
import org.apache.commons.io.FileUtils;

import java.io.File;
import java.io.IOException;
import java.util.Iterator;
import java.util.List;
import java.util.Set;

/**
 */
public class SingleSegmentLoader implements SegmentLoader {
    private static final Logger log = new Logger(SingleSegmentLoader.class);

    private final DataSegmentPuller dataSegmentPuller;
    private final QueryableIndexFactory factory;

    private final List<StorageLocation> locations;

    @Inject
    public SingleSegmentLoader(DataSegmentPuller dataSegmentPuller, QueryableIndexFactory factory,
            SegmentLoaderConfig config) {
        this.dataSegmentPuller = dataSegmentPuller;
        this.factory = factory;

        final ImmutableList.Builder<StorageLocation> locBuilder = ImmutableList.builder();

        // This is a really, really stupid way of getting this information.  Splitting on commas and bars is error-prone
        // We should instead switch it up to be a JSON Array of JSON Object or something and cool stuff like that
        // But, that'll have to wait for some other day.
        for (String dirSpec : config.getCacheDirectory().split(",")) {
            String[] dirSplit = dirSpec.split("\\|");
            if (dirSplit.length == 1) {
                locBuilder.add(new StorageLocation(new File(dirSplit[0]), config.getServerMaxSize()));
            } else if (dirSplit.length == 2) {
                final Long maxSize = Longs.tryParse(dirSplit[1]);
                if (maxSize == null) {
                    throw new IAE("Size of a local segment storage location must be an integral number, got[%s]",
                            dirSplit[1]);
                }
                locBuilder.add(new StorageLocation(new File(dirSplit[0]), maxSize));
            } else {
                throw new ISE("Unknown segment storage location[%s]=>[%s], config[%s].", dirSplit.length, dirSpec,
                        config.getCacheDirectory());
            }
        }
        locations = locBuilder.build();

        Preconditions.checkArgument(locations.size() > 0, "Must have at least one segment cache directory.");
        log.info("Using storage locations[%s]", locations);
    }

    @Override
    public boolean isSegmentLoaded(final DataSegment segment) {
        return findStorageLocationIfLoaded(segment) != null;
    }

    public StorageLocation findStorageLocationIfLoaded(final DataSegment segment) {
        for (StorageLocation location : locations) {
            File localStorageDir = new File(location.getPath(), DataSegmentPusherUtil.getStorageDir(segment));
            if (localStorageDir.exists()) {
                return location;
            }
        }
        return null;
    }

    @Override
    public Segment getSegment(DataSegment segment) throws SegmentLoadingException {
        File segmentFiles = getSegmentFiles(segment);
        final QueryableIndex index = factory.factorize(segmentFiles);

        return new QueryableIndexSegment(segment.getIdentifier(), index);
    }

    public File getSegmentFiles(DataSegment segment) throws SegmentLoadingException {
        StorageLocation loc = findStorageLocationIfLoaded(segment);

        final File retVal;

        if (loc == null) {
            Iterator<StorageLocation> locIter = locations.iterator();
            loc = locIter.next();
            while (locIter.hasNext()) {
                loc = loc.mostEmpty(locIter.next());
            }

            if (!loc.canHandle(segment.getSize())) {
                throw new ISE("Segment[%s:%,d] too large for storage[%s:%,d].", segment.getIdentifier(),
                        segment.getSize(), loc.getPath(), loc.available());
            }

            File storageDir = new File(loc.getPath(), DataSegmentPusherUtil.getStorageDir(segment));
            if (!storageDir.mkdirs()) {
                log.debug("Unable to make parent file[%s]", storageDir);
            }

            dataSegmentPuller.getSegmentFiles(segment, storageDir);
            loc.addSegment(segment);

            retVal = storageDir;
        } else {
            retVal = new File(loc.getPath(), DataSegmentPusherUtil.getStorageDir(segment));
        }

        loc.addSegment(segment);

        return retVal;
    }

    @Override
    public void cleanup(DataSegment segment) throws SegmentLoadingException {
        StorageLocation loc = findStorageLocationIfLoaded(segment);

        if (loc == null) {
            log.info("Asked to cleanup something[%s] that didn't exist.  Skipping.", segment);
            return;
        }

        try {
            File cacheFile = new File(loc.getPath(), DataSegmentPusherUtil.getStorageDir(segment));
            log.info("Deleting directory[%s]", cacheFile);
            FileUtils.deleteDirectory(cacheFile);
            loc.removeSegment(segment);
        } catch (IOException e) {
            throw new SegmentLoadingException(e, e.getMessage());
        }
    }

    static class StorageLocation {
        private final File path;
        private final long maxSize;
        private final Set<DataSegment> segments;

        private volatile long currSize = 0;

        StorageLocation(File path, long maxSize) {
            this.path = path;
            this.maxSize = maxSize;

            this.segments = Sets.newHashSet();
        }

        File getPath() {
            return path;
        }

        Long getMaxSize() {
            return maxSize;
        }

        synchronized void addSegment(DataSegment segment) {
            if (segments.add(segment)) {
                currSize += segment.getSize();
            }
        }

        synchronized void removeSegment(DataSegment segment) {
            if (segments.remove(segment)) {
                currSize -= segment.getSize();
            }
        }

        boolean canHandle(long size) {
            return available() >= size;
        }

        synchronized long available() {
            return maxSize - currSize;
        }

        StorageLocation mostEmpty(StorageLocation other) {
            return available() > other.available() ? this : other;
        }

        @Override
        public String toString() {
            return "StorageLocation{" + "path=" + path + ", maxSize=" + maxSize + '}';
        }
    }
}