com.spotify.helios.agent.QueueingHistoryWriter.java Source code

Java tutorial

Introduction

Here is the source code for com.spotify.helios.agent.QueueingHistoryWriter.java

Source

/*
 * Copyright (c) 2014 Spotify AB.
 *
 * Licensed to the Apache Software Foundation (ASF) under one
 * or more contributor license agreements.  See the NOTICE file
 * distributed with this work for additional information
 * regarding copyright ownership.  The ASF licenses this file
 * to you under the Apache License, Version 2.0 (the
 * "License"); you may not use this file except in compliance
 * with the License.  You may obtain a copy of the License at
 *
 *   http://www.apache.org/licenses/LICENSE-2.0
 *
 * Unless required by applicable law or agreed to in writing,
 * software distributed under the License is distributed on an
 * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
 * KIND, either express or implied.  See the License for the
 * specific language governing permissions and limitations
 * under the License.
 */

package com.spotify.helios.agent;

import com.google.common.base.Function;
import com.google.common.base.Supplier;
import com.google.common.collect.ImmutableSet;
import com.google.common.collect.Iterables;
import com.google.common.collect.Lists;
import com.google.common.collect.Maps;
import com.google.common.util.concurrent.AbstractIdleService;
import com.google.common.util.concurrent.MoreExecutors;

import com.fasterxml.jackson.core.type.TypeReference;
import com.spotify.helios.common.descriptors.JobId;
import com.spotify.helios.common.descriptors.TaskStatus;
import com.spotify.helios.common.descriptors.TaskStatusEvent;
import com.spotify.helios.servicescommon.PersistentAtomicReference;
import com.spotify.helios.servicescommon.coordination.Paths;
import com.spotify.helios.servicescommon.coordination.ZooKeeperClient;

import org.apache.zookeeper.KeeperException;
import org.apache.zookeeper.KeeperException.NodeExistsException;
import org.slf4j.Logger;
import org.slf4j.LoggerFactory;

import java.io.IOException;
import java.nio.channels.ClosedByInterruptException;
import java.nio.file.Path;
import java.util.Collections;
import java.util.Deque;
import java.util.List;
import java.util.concurrent.ConcurrentLinkedDeque;
import java.util.concurrent.ConcurrentMap;
import java.util.concurrent.Executors;
import java.util.concurrent.ScheduledExecutorService;
import java.util.concurrent.ScheduledThreadPoolExecutor;
import java.util.concurrent.TimeUnit;
import java.util.concurrent.atomic.AtomicInteger;

import static com.google.common.base.Preconditions.checkState;
import static java.util.concurrent.TimeUnit.SECONDS;
import static org.apache.zookeeper.KeeperException.ConnectionLossException;

/**
 * Just some breadcrumbs so next time, the person that follows me can understand why things are
 * the way they are.
 *
 * Theory of operation:
 * 1. saveHistoryItem should never block for any significant amount of time.  Specifically, it
 *    should not block on ZK being in any particular state, and ideally not while a file write is
 *    occurring, as the file may get large if ZK has been away for a long time.
 * 2. We limit each job to max 30 events in memory (and in ZK for that matter)
 * 3. Maximum of 600 total events, so as not to consume all available memory.
 */
public class QueueingHistoryWriter extends AbstractIdleService implements Runnable {
    private static final Logger log = LoggerFactory.getLogger(QueueingHistoryWriter.class);

    public static final int MAX_NUMBER_STATUS_EVENTS_TO_RETAIN = 30;
    private static final int MAX_QUEUE_SIZE = 30;
    private static final int MAX_TOTAL_SIZE = 600;

    private final ConcurrentMap<JobId, Deque<TaskStatusEvent>> items;
    private final ScheduledExecutorService zkWriterExecutor = MoreExecutors.getExitingScheduledExecutorService(
            (ScheduledThreadPoolExecutor) Executors.newScheduledThreadPool(1), 0, SECONDS);
    private final String hostname;
    private final AtomicInteger count;
    private final ZooKeeperClient client;
    private final PersistentAtomicReference<ConcurrentMap<JobId, Deque<TaskStatusEvent>>> backingStore;

    public QueueingHistoryWriter(final String hostname, final ZooKeeperClient client, final Path backingFile)
            throws IOException, InterruptedException {
        this.hostname = hostname;
        this.client = client;
        this.backingStore = PersistentAtomicReference.create(backingFile,
                new TypeReference<ConcurrentMap<JobId, Deque<TaskStatusEvent>>>() {
                }, new Supplier<ConcurrentMap<JobId, Deque<TaskStatusEvent>>>() {
                    @Override
                    public ConcurrentMap<JobId, Deque<TaskStatusEvent>> get() {
                        return Maps.newConcurrentMap();
                    }
                });
        this.items = backingStore.get();

        // Clean out any errant null values.  Normally shouldn't have any, but we did have a few
        // where it happened, and this will make sure we can get out of a bad state if we get into it.
        final ImmutableSet<JobId> curKeys = ImmutableSet.copyOf(this.items.keySet());
        for (JobId key : curKeys) {
            if (this.items.get(key) == null) {
                this.items.remove(key);
            }
        }

        int itemCount = 0;
        for (Deque<TaskStatusEvent> deque : items.values()) {
            itemCount += deque.size();
        }
        this.count = new AtomicInteger(itemCount);
    }

    @Override
    protected void startUp() throws Exception {
        zkWriterExecutor.scheduleAtFixedRate(this, 1, 1, TimeUnit.SECONDS);
    }

    @Override
    protected void shutDown() throws Exception {
        zkWriterExecutor.shutdownNow();
        zkWriterExecutor.awaitTermination(1, TimeUnit.MINUTES);
    }

    private void add(TaskStatusEvent item) throws InterruptedException {
        // If too many "globally", toss them
        while (count.get() >= MAX_TOTAL_SIZE) {
            getNext();
        }

        final JobId key = item.getStatus().getJob().getId();
        final Deque<TaskStatusEvent> deque = getDeque(key);

        synchronized (deque) {
            // if too many in the particular deque, toss them
            while (deque.size() >= MAX_QUEUE_SIZE) {
                deque.remove();
                count.decrementAndGet();
            }
            deque.add(item);
            count.incrementAndGet();
        }

        try {
            backingStore.set(items);
        } catch (ClosedByInterruptException e) {
            log.debug("Writing task status event to backing store was interrupted");
        } catch (IOException e) { // We are best effort after all...
            log.warn("Failed to write task status event to backing store", e);
        }
    }

    private Deque<TaskStatusEvent> getDeque(final JobId key) {
        synchronized (items) {
            final Deque<TaskStatusEvent> deque = items.get(key);
            if (deque == null) { // try more assertively to get a deque
                final ConcurrentLinkedDeque<TaskStatusEvent> newDeque = new ConcurrentLinkedDeque<TaskStatusEvent>();
                items.put(key, newDeque);
                return newDeque;
            }
            return deque;
        }
    }

    public void saveHistoryItem(final JobId jobId, final TaskStatus status) throws InterruptedException {
        saveHistoryItem(jobId, status, System.currentTimeMillis());
    }

    public void saveHistoryItem(final JobId jobId, final TaskStatus status, long timestamp)
            throws InterruptedException {
        add(new TaskStatusEvent(status, timestamp, hostname));
    }

    private TaskStatusEvent getNext() {
        // Some explanation: We first find the eldest event from amongst the queues (ok, they're
        // deques, but we really use it as a put back queue), and only then to we try to get
        // a lock on the relevant queue from whence we got the event.  Assuming that all worked
        // *and* that the event we have wasn't rolled off due to max-size limitations, we then
        // pull the item off the queue and return it.  We're basically doing optimistic concurrency,
        // and skewing things so that adding to this should be cheap.

        while (true) {
            final TaskStatusEvent current = findEldestEvent();

            // Didn't find anything that needed processing?
            if (current == null) {
                return null;
            }

            final JobId id = current.getStatus().getJob().getId();
            final Deque<TaskStatusEvent> deque = items.get(id);
            if (deque == null) {
                // shouldn't happen because we should be the only one pulling items off, but....
                continue;
            }

            synchronized (deque) {
                if (!deque.peek().equals(current)) {
                    // item got rolled off, try again
                    continue;
                }

                // Pull it off the queue and be paranoid.
                final TaskStatusEvent newCurrent = deque.poll();
                count.decrementAndGet();
                checkState(current.equals(newCurrent), "current should equal newCurrent");
                // Safe because this is the *only* place we hold these two locks at the same time.
                synchronized (items) {
                    // Extra paranoia: curDeque should always == deque
                    final Deque<TaskStatusEvent> curDeque = items.get(id);
                    if (curDeque != null && curDeque.isEmpty()) {
                        items.remove(id);
                    }
                }
                return current;
            }
        }
    }

    public boolean isEmpty() {
        return count.get() == 0;
    }

    private void putBack(TaskStatusEvent event) {
        final JobId key = event.getStatus().getJob().getId();
        final Deque<TaskStatusEvent> queue = getDeque(key);
        synchronized (queue) {
            if (queue.size() >= MAX_QUEUE_SIZE) {
                // already full, just toss the event
                return;
            }
            queue.push(event);
            count.incrementAndGet();
        }
    }

    private TaskStatusEvent findEldestEvent() {
        // We don't lock anything because in the worst case, we just put things in out of order which
        // while not perfect, won't cause any actual harm.  Out of order meaning between jobids, not
        // within the same job id.  Whether this is the best strategy (as opposed to fullest deque)
        // is arguable.
        TaskStatusEvent current = null;
        for (Deque<TaskStatusEvent> queue : items.values()) {
            if (queue == null) {
                continue;
            }
            final TaskStatusEvent item = queue.peek();
            if (current == null || (item.getTimestamp() < current.getTimestamp())) {
                current = item;
            }
        }
        return current;
    }

    @Override
    public void run() {
        while (true) {
            final TaskStatusEvent item = getNext();
            if (item == null) {
                return;
            }

            try {
                final JobId jobId = item.getStatus().getJob().getId();
                final String historyPath = Paths.historyJobHostEventsTimestamp(jobId, hostname,
                        item.getTimestamp());
                log.debug("writing queued item to zookeeper {} {}", item.getStatus().getJob().getId(),
                        item.getTimestamp());
                client.ensurePath(historyPath, true);
                client.createAndSetData(historyPath, item.getStatus().toJsonBytes());

                // See if too many
                final List<String> events = client.getChildren(Paths.historyJobHostEvents(jobId, hostname));
                if (events.size() > MAX_NUMBER_STATUS_EVENTS_TO_RETAIN) {
                    trimStatusEvents(events, jobId);
                }
            } catch (NodeExistsException e) {
                // Ahh, the two generals problem...  We handle by doing nothing since the thing
                // we wanted in, is in.
                log.debug("item we wanted in is already there");
            } catch (ConnectionLossException e) {
                log.warn("Connection lost while putting item into zookeeper, will retry");
                putBack(item);
                break;
            } catch (KeeperException e) {
                log.error("Error putting item into zookeeper, will retry", e);
                putBack(item);
                break;
            }
        }
    }

    private void trimStatusEvents(List<String> events, JobId jobId) {
        // CleanupExecutor only has one thread so can assume no others are fiddling as we do this.
        // All this to sort numerically instead of lexically....
        final List<Long> eventsAsLongs = Lists
                .newArrayList(Iterables.transform(events, new Function<String, Long>() {
                    @Override
                    public Long apply(String name) {
                        return Long.valueOf(name);
                    }
                }));
        Collections.sort(eventsAsLongs);

        for (int i = 0; i < (eventsAsLongs.size() - MAX_NUMBER_STATUS_EVENTS_TO_RETAIN); i++) {
            try {
                client.delete(Paths.historyJobHostEventsTimestamp(jobId, hostname, eventsAsLongs.get(i)));
            } catch (KeeperException e) {
                log.warn("failure deleting overflow of status items - we're hoping a later" + " execution will fix",
                        e);
            }
        }
    }
}