Java tutorial
/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, * software distributed under the License is distributed on an * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. */ package org.apache.usergrid.chop.webapp.coordinator; import java.io.File; import java.util.LinkedList; import java.util.concurrent.Callable; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.apache.commons.lang.builder.HashCodeBuilder; import org.apache.usergrid.chop.api.ProviderParams; import org.apache.usergrid.chop.api.store.amazon.AmazonFig; import org.apache.usergrid.chop.spi.InstanceManager; import org.apache.usergrid.chop.spi.IpRuleManager; import org.apache.usergrid.chop.spi.LaunchResult; import org.apache.usergrid.chop.stack.BasicInstanceSpec; import org.apache.usergrid.chop.stack.CoordinatedStack; import org.apache.usergrid.chop.stack.ICoordinatedCluster; import org.apache.usergrid.chop.stack.Instance; import org.apache.usergrid.chop.stack.SetupStackSignal; import org.apache.usergrid.chop.webapp.ChopUiFig; import org.apache.usergrid.chop.webapp.dao.ProviderParamsDao; import org.apache.usergrid.chop.webapp.service.InjectorFactory; import com.google.inject.Inject; /** Encapsulates a CoordinatedStack and sets it up asynchronously */ public class SetupStackThread implements Callable<CoordinatedStack> { private static final Logger LOG = LoggerFactory.getLogger(SetupStackThread.class); @Inject private ChopUiFig chopUiFig; @Inject private ProviderParamsDao providerParamsDao; private CoordinatedStack stack; private String errorMessage; public SetupStackThread(CoordinatedStack stack) { this.stack = stack; } public CoordinatedStack getStack() { return stack; } public String getErrorMessage() { return errorMessage; } @Override public CoordinatedStack call() { String keyFile; LinkedList<String> launchedInstances = new LinkedList<String>(); providerParamsDao = InjectorFactory.getInstance(ProviderParamsDao.class); chopUiFig = InjectorFactory.getInstance(ChopUiFig.class); ProviderParams providerParams = providerParamsDao.getByUser(stack.getUser().getUsername()); /** Bypass the keys in AmazonFig so that it uses the ones belonging to the user */ AmazonFig amazonFig = InjectorFactory.getInstance(AmazonFig.class); amazonFig.bypass(AmazonFig.AWS_ACCESS_KEY, providerParams.getAccessKey()); amazonFig.bypass(AmazonFig.AWS_SECRET_KEY, providerParams.getSecretKey()); InstanceManager instanceManager = InjectorFactory.getInstance(InstanceManager.class); IpRuleManager ipRuleManager = InjectorFactory.getInstance(IpRuleManager.class); File runnerJar = CoordinatorUtils.getRunnerJar(chopUiFig.getContextPath(), stack); ipRuleManager.setDataCenter(stack.getDataCenter()); ipRuleManager.applyIpRuleSet(stack.getIpRuleSet()); /** Setup clusters */ for (ICoordinatedCluster cluster : stack.getClusters()) { LOG.info("Starting setting up cluster {}...", cluster.getName()); keyFile = providerParams.getKeys().get(cluster.getInstanceSpec().getKeyName()); if (keyFile == null) { errorMessage = "No key found with name " + cluster.getInstanceSpec().getKeyName() + " for cluster " + cluster.getName(); LOG.warn(errorMessage + ", aborting and terminating launched instances..."); instanceManager.terminateInstances(launchedInstances); stack.setSetupState(SetupStackSignal.FAIL); stack.notifyAll(); return null; } if (!(new File(keyFile)).exists()) { errorMessage = "Key file " + keyFile + " for cluster " + cluster.getName() + " not found"; LOG.warn(errorMessage + ", aborting and terminating launched instances..."); instanceManager.terminateInstances(launchedInstances); stack.setSetupState(SetupStackSignal.FAIL); stack.notifyAll(); return null; } LaunchResult result = instanceManager.launchCluster(stack, cluster, chopUiFig.getLaunchClusterTimeout()); for (Instance instance : result.getInstances()) { launchedInstances.add(instance.getId()); cluster.add(instance); } /** Setup system properties, deploy the scripts and execute them on cluster instances */ boolean success = false; try { success = CoordinatorUtils.executeClusterSSHCommands(cluster, runnerJar, keyFile); } catch (Exception e) { LOG.warn("Error while executing SSH commands", e); } if (!success) { errorMessage = "SSH commands have failed, will not continue"; instanceManager.terminateInstances(launchedInstances); stack.setSetupState(SetupStackSignal.FAIL); stack.notifyAll(); return null; } LOG.info("Cluster {} is ready, moving on...", cluster.getName()); } /** Setup runners */ LOG.info("Starting setting up runner instances..."); keyFile = providerParams.getKeys().get(providerParams.getKeyName()); if (keyFile == null) { errorMessage = "No key found with name " + providerParams.getKeyName() + " for runners"; LOG.warn(errorMessage + ", aborting and terminating launched instances..."); instanceManager.terminateInstances(launchedInstances); stack.setSetupState(SetupStackSignal.FAIL); stack.notifyAll(); return null; } if (!(new File(keyFile)).exists()) { errorMessage = "Key file " + keyFile + " for runners not found"; LOG.warn(errorMessage + ", aborting and terminating launched instances..."); instanceManager.terminateInstances(launchedInstances); stack.setSetupState(SetupStackSignal.FAIL); stack.notifyAll(); return null; } BasicInstanceSpec runnerSpec = new BasicInstanceSpec(); runnerSpec.setImageId(providerParams.getImageId()); runnerSpec.setType(providerParams.getInstanceType()); runnerSpec.setKeyName(providerParams.getKeyName()); LaunchResult result = instanceManager.launchRunners(stack, runnerSpec, chopUiFig.getLaunchClusterTimeout()); for (Instance instance : result.getInstances()) { launchedInstances.add(instance.getId()); stack.addRunnerInstance(instance); } /** Deploy and start runner.jar on instances */ boolean success = false; try { success = CoordinatorUtils.executeRunnerSSHCommands(stack, runnerJar, keyFile); } catch (Exception e) { LOG.warn("Error while executing SSH commands", e); } if (!success) { errorMessage = "SSH commands have failed, will not continue"; instanceManager.terminateInstances(launchedInstances); stack.setSetupState(SetupStackSignal.FAIL); stack.notifyAll(); return null; } stack.setSetupState(SetupStackSignal.COMPLETE); LOG.info("Stack {} is set up and ready...", stack.getName()); stack.notifyAll(); return stack; } @Override public int hashCode() { if (errorMessage != null) { return new HashCodeBuilder(97, 71).append(errorMessage).toHashCode(); } return new HashCodeBuilder(97, 71).append(stack.getId().toString()).append(stack.getUser().getUsername()) .append(stack.getCommit().getId()).append(stack.getModule().getId()).append(stack.getRunnerCount()) .toHashCode(); } @Override public boolean equals(final Object obj) { if (this == obj) { return true; } return obj != null && obj instanceof SetupStackThread && obj.hashCode() == this.hashCode(); } }