Java tutorial
/* * Druid - a distributed column store. * Copyright 2012 - 2015 Metamarkets Group Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package io.druid.storage.s3; import com.google.common.base.Predicate; import com.google.common.base.Throwables; import com.google.common.collect.ImmutableMap; import com.google.common.collect.Maps; import com.google.inject.Inject; import com.metamx.common.ISE; import com.metamx.common.MapUtils; import com.metamx.common.logger.Logger; import io.druid.segment.loading.DataSegmentMover; import io.druid.segment.loading.SegmentLoadingException; import io.druid.timeline.DataSegment; import org.jets3t.service.ServiceException; import org.jets3t.service.acl.gs.GSAccessControlList; import org.jets3t.service.impl.rest.httpclient.RestS3Service; import org.jets3t.service.model.S3Object; import java.util.Map; import java.util.concurrent.Callable; public class S3DataSegmentMover implements DataSegmentMover { private static final Logger log = new Logger(S3DataSegmentMover.class); private final RestS3Service s3Client; private final S3DataSegmentPusherConfig config; @Inject public S3DataSegmentMover(RestS3Service s3Client, S3DataSegmentPusherConfig config) { this.s3Client = s3Client; this.config = config; } @Override public DataSegment move(DataSegment segment, Map<String, Object> targetLoadSpec) throws SegmentLoadingException { try { Map<String, Object> loadSpec = segment.getLoadSpec(); String s3Bucket = MapUtils.getString(loadSpec, "bucket"); String s3Path = MapUtils.getString(loadSpec, "key"); String s3DescriptorPath = S3Utils.descriptorPathForSegmentPath(s3Path); final String targetS3Bucket = MapUtils.getString(targetLoadSpec, "bucket"); final String targetS3BaseKey = MapUtils.getString(targetLoadSpec, "baseKey"); final String targetS3Path = S3Utils.constructSegmentPath(targetS3BaseKey, segment); String targetS3DescriptorPath = S3Utils.descriptorPathForSegmentPath(targetS3Path); if (targetS3Bucket.isEmpty()) { throw new SegmentLoadingException("Target S3 bucket is not specified"); } if (targetS3Path.isEmpty()) { throw new SegmentLoadingException("Target S3 baseKey is not specified"); } safeMove(s3Bucket, s3Path, targetS3Bucket, targetS3Path); safeMove(s3Bucket, s3DescriptorPath, targetS3Bucket, targetS3DescriptorPath); return segment.withLoadSpec(ImmutableMap.<String, Object>builder() .putAll(Maps.filterKeys(loadSpec, new Predicate<String>() { @Override public boolean apply(String input) { return !(input.equals("bucket") || input.equals("key")); } })).put("bucket", targetS3Bucket).put("key", targetS3Path).build()); } catch (ServiceException e) { throw new SegmentLoadingException(e, "Unable to move segment[%s]: [%s]", segment.getIdentifier(), e); } } private void safeMove(final String s3Bucket, final String s3Path, final String targetS3Bucket, final String targetS3Path) throws ServiceException, SegmentLoadingException { try { S3Utils.retryS3Operation(new Callable<Void>() { @Override public Void call() throws Exception { if (s3Client.isObjectInBucket(s3Bucket, s3Path)) { if (s3Bucket.equals(targetS3Bucket) && s3Path.equals(targetS3Path)) { log.info("No need to move file[s3://%s/%s] onto itself", s3Bucket, s3Path); } else { final S3Object[] list = s3Client.listObjects(s3Bucket, s3Path, ""); if (list.length == 0) { // should never happen throw new ISE("Unable to list object [s3://%s/%s]", s3Bucket, s3Path); } final S3Object s3Object = list[0]; if (s3Object.getStorageClass() != null && s3Object.getStorageClass().equals(S3Object.STORAGE_CLASS_GLACIER)) { log.warn("Cannot move file[s3://%s/%s] of storage class glacier, skipping.", s3Bucket, s3Path); } else { log.info("Moving file[s3://%s/%s] to [s3://%s/%s]", s3Bucket, s3Path, targetS3Bucket, targetS3Path); final S3Object target = new S3Object(targetS3Path); if (!config.getDisableAcl()) { target.setAcl(GSAccessControlList.REST_CANNED_BUCKET_OWNER_FULL_CONTROL); } s3Client.moveObject(s3Bucket, s3Path, targetS3Bucket, target, false); } } } else { // ensure object exists in target location if (s3Client.isObjectInBucket(targetS3Bucket, targetS3Path)) { log.info( "Not moving file [s3://%s/%s], already present in target location [s3://%s/%s]", s3Bucket, s3Path, targetS3Bucket, targetS3Path); } else { throw new SegmentLoadingException( "Unable to move file [s3://%s/%s] to [s3://%s/%s], not present in either source or target location", s3Bucket, s3Path, targetS3Bucket, targetS3Path); } } return null; } }); } catch (Exception e) { Throwables.propagateIfInstanceOf(e, ServiceException.class); Throwables.propagateIfInstanceOf(e, SegmentLoadingException.class); throw Throwables.propagate(e); } } }