Java tutorial
/* * Copyright 2009-2013 by The Regents of the University of California * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * you may obtain a copy of the License from * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package edu.uci.ics.pregelix.runtime.converter; import java.io.DataInput; import java.io.DataInputStream; import org.apache.asterix.builders.RecordBuilder; import org.apache.asterix.om.types.ARecordType; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.mapreduce.TaskAttemptContext; import org.apache.hyracks.api.context.IHyracksTaskContext; import org.apache.hyracks.api.exceptions.HyracksDataException; import org.apache.hyracks.dataflow.common.comm.io.ArrayTupleBuilder; import org.apache.hyracks.hdfs.ContextFactory; import edu.uci.ics.external.connector.asterixdb.api.IWriteConverter; import edu.uci.ics.external.connector.asterixdb.api.IWriteConverterFactory; import edu.uci.ics.pregelix.api.converter.VertexOutputConverter; import edu.uci.ics.pregelix.api.graph.Vertex; import edu.uci.ics.pregelix.api.util.BspUtils; import edu.uci.ics.pregelix.api.util.ResetableByteArrayInputStream; import edu.uci.ics.pregelix.dataflow.base.IConfigurationFactory; import edu.uci.ics.pregelix.dataflow.util.IterationUtils; public class WriteConverterFactory implements IWriteConverterFactory { private static final long serialVersionUID = 1L; private final IConfigurationFactory confFactory; public WriteConverterFactory(IConfigurationFactory confFactory) { this.confFactory = confFactory; } @SuppressWarnings("rawtypes") @Override public IWriteConverter getFieldWriteConverter(IHyracksTaskContext ctx, int partitionId) throws HyracksDataException { final Configuration conf = confFactory.createConfiguration(); // Set context properly ContextFactory ctxFactory = new ContextFactory(); TaskAttemptContext mapperContext = ctxFactory.createContext(conf, partitionId); mapperContext.getConfiguration().setClassLoader(ctx.getJobletContext().getClassLoader()); conf.setClassLoader(ctx.getJobletContext().getClassLoader()); IterationUtils.setJobContext(BspUtils.getJobId(conf), ctx, mapperContext); Vertex.taskContext = mapperContext; final Vertex vertex = BspUtils.createVertex(conf); vertex.setVertexContext(IterationUtils.getVertexContext(BspUtils.getJobId(conf), ctx)); final VertexOutputConverter outputConverter = BspUtils.createVertexOutputConverter(conf); final ResetableByteArrayInputStream inputStream = new ResetableByteArrayInputStream(); final DataInput dataInput = new DataInputStream(inputStream); final RecordBuilder recordBuilder = new RecordBuilder(); return new IWriteConverter() { @Override public void open(ARecordType recordType) throws HyracksDataException { recordBuilder.reset(recordType); outputConverter.open(recordType); } @Override public void convert(byte[] data, int start, int len, ArrayTupleBuilder outputTb) throws HyracksDataException { try { inputStream.setByteArray(data, start); vertex.readFields(dataInput); recordBuilder.init(); outputTb.reset(); outputConverter.convert(vertex.getVertexId(), outputTb.getDataOutput()); outputTb.addFieldEndOffset(); outputConverter.convert(vertex, recordBuilder); // By default, the record type tag is stored in AsterixDB. recordBuilder.write(outputTb.getDataOutput(), true); outputTb.addFieldEndOffset(); } catch (Exception e) { throw new HyracksDataException(e); } } @Override public void close() throws HyracksDataException { outputConverter.close(); } }; } }