List of usage examples for java.io DataOutput writeInt
void writeInt(int v) throws IOException;
int
value, which is comprised of four bytes, to the output stream. From source file:com.fiorano.openesb.application.aps.ApplicationHeader.java
/** * This method writes this <code>ApplicationHeader</code> object to the * specified output stream object./*from w ww . j a v a 2s. c o m*/ * * @param out DataOutput object * @param versionNo * @exception IOException if an error occurs while converting data and * writing it to a binary stream. * @see #fromStream(DataInput, int) * @since Tifosi2.0 */ public void toStream(DataOutput out, int versionNo) throws IOException { super.toStream(out, versionNo); out.writeBoolean(m_canBeSubGraphed); if (m_scope != null) UTFReaderWriter.writeUTF(out, m_scope); else UTFReaderWriter.writeUTF(out, ""); if (m_appName != null) UTFReaderWriter.writeUTF(out, m_appName); else UTFReaderWriter.writeUTF(out, ""); if (m_appGUID != null) UTFReaderWriter.writeUTF(out, m_appGUID); else UTFReaderWriter.writeUTF(out, ""); if (m_creationDate != null) UTFReaderWriter.writeUTF(out, m_creationDate); else UTFReaderWriter.writeUTF(out, ""); if (m_category != null) UTFReaderWriter.writeUTF(out, m_category); else UTFReaderWriter.writeUTF(out, ""); if (m_authorNames != null) { int length = m_authorNames.size(); out.writeInt(length); Enumeration names = m_authorNames.elements(); while (names.hasMoreElements()) { String name = (String) names.nextElement(); UTFReaderWriter.writeUTF(out, name); } } else out.writeInt(0); if (m_iconName != null) UTFReaderWriter.writeUTF(out, m_iconName); else UTFReaderWriter.writeUTF(out, ""); if (m_shortDescription != null) UTFReaderWriter.writeUTF(out, m_shortDescription); else UTFReaderWriter.writeUTF(out, ""); if (m_longDescription != null) UTFReaderWriter.writeUTF(out, m_longDescription); else UTFReaderWriter.writeUTF(out, ""); if (m_compatibleWith != null) { int length = m_compatibleWith.size(); out.writeInt(length); Enumeration names = m_compatibleWith.elements(); while (names.hasMoreElements()) { String name = (String) names.nextElement(); UTFReaderWriter.writeUTF(out, name); } } else out.writeInt(0); out.writeInt(m_maxInstLimit); if (m_params != null) { int length = m_params.size(); out.writeInt(length); Enumeration params = m_params.elements(); while (params.hasMoreElements()) { Param param = (Param) params.nextElement(); param.toStream(out, versionNo); } } else out.writeInt(0); if (m_version != null) UTFReaderWriter.writeUTF(out, m_version); else UTFReaderWriter.writeUTF(out, ""); out.writeBoolean(m_isVersionLocked); UTFReaderWriter.writeUTF(out, m_strProfile); if (m_appContext != null) { out.writeInt(1); m_appContext.toStream(out, versionNo); } else { out.writeInt(0); } }
From source file:org.mhisoft.wallet.service.AttachmentService.java
/** * @param transferStoreMode a new store will be created because the old one has t * @param oldStoreFileName provide along with the transferStoreMode. for read the contents to be transfered to the new store when doing compacting. * @param itemStartPos start position for the data output file. * @param dataOut data output file/stream * @param t the FileAccessTable contains the entries to write. * @param encryptor The encryptor * @throws IOException/*from ww w. j a v a 2 s . c o m*/ */ protected void writeFileEntries(WalletModel model, boolean transferStoreMode, String oldStoreFileName, final long itemStartPos, DataOutput dataOut, final FileAccessTable t, final PBEEncryptor oldEncryptorForRead, final PBEEncryptor encryptor) throws IOException { long pos = itemStartPos; //write it out //dataOut = new DataOutputStream(new FileOutputStream(new File(outoutFIleName))); //write the total number of entries first /*#0*/ //dataOut.writeInt(t.getEntries().size()); //itemStartPos = 4; String impAttachmentStoreName = null; if (model.getImpModel() != null) impAttachmentStoreName = getAttachmentFileName(model.getImpModel().getVaultFileName()); //write each entry for (int i = 0; i < t.getEntries().size(); i++) { FileAccessEntry fileAccessEntry = t.getEntries().get(i); //not handleing the deleted entries. if (fileAccessEntry.getAccessFlag() == FileAccessFlag.Delete) { continue; } logger.fine("------------------"); logger.fine("Write entry: " + fileAccessEntry.getGUID() + "-" + fileAccessEntry.getFileName()); logger.fine("\t access flag: " + fileAccessEntry.getAccessFlag()); logger.fine("\t start pos: " + pos); fileAccessEntry.setPosition(pos); /* UUID*/ int uuidSize = FileUtils.writeString(dataOut, fileAccessEntry.getGUID()); pos += 40; /* accessflag */ dataOut.writeInt(0); pos += 4; /* position */ dataOut.writeLong(fileAccessEntry.getPosition()); pos += 8; /* write filename encrypted */ String strFName = FileUtils.getFileNameWithoutPath(fileAccessEntry.getFileName()); logger.fine("\t write file name: " + strFName); byte[] _byteFileName = StringUtils.getBytes(strFName); pos = writeEncryptedContent(_byteFileName, encryptor, dataOut, pos); logger.fine("\t pos: " + pos); /* Attachment body */ byte[] attachmentBytes; if (fileAccessEntry.getAccessFlag() == FileAccessFlag.Merge && fileAccessEntry.getEncSize() > 0) { if (impAttachmentStoreName == null) throw new IOException("impAttachmentStoreName is not set."); //depends on the import model version, it might or might not be compressed //test case: import an old version data file with attachments. byte[] _impBytes = readCompressedFileContent(model.getImpModel().getCurrentDataFileVersion(), impAttachmentStoreName, fileAccessEntry, model.getImpModel().getEncryptor()); if (model.getImpModel().getCurrentDataFileVersion() >= WalletModel.LATEST_DATA_VERSION) { //it is compressed. attachmentBytes = _impBytes; } else { //not compressed, need to compress it. attachmentBytes = CompressionUtil.getCompressedBytes(new ByteArrayInputStream(_impBytes)); } } else if (transferStoreMode && fileAccessEntry.getAccessFlag() == FileAccessFlag.None && fileAccessEntry.getEncSize() > 0) { // no change, this is an transfer to the new store. need to read the filecontent from the old store. // could be upgrade scenario as well. // should be able to skip the compression-decompression process. byte[] _bytes = readCompressedFileContent(model.getCurrentDataFileVersion(), oldStoreFileName, fileAccessEntry, oldEncryptorForRead); if (model.getCurrentDataFileVersion() < WalletModel.LATEST_DATA_VERSION) { //it was not compressed. attachmentBytes = CompressionUtil.getCompressedBytes(new ByteArrayInputStream(_bytes)); logger.fine("\t transfer from old store. compress contennt. before " + _bytes.length + ", after:" + attachmentBytes.length); } else attachmentBytes = _bytes; } else { /* for CREATE OR UPDATE or a new store */ //read from file //since v14 , start to compress the contents. File f = fileAccessEntry.getFile(); FileInputStream sourceInputStream = new FileInputStream(f); if (model.getCurrentDataFileVersion() < WalletModel.LATEST_DATA_VERSION) { //for testing, write the old version format. un-compressed attachmentBytes = FileUtils.readFile(sourceInputStream); logger.fine("\t write entry by reading from file, not compressed: " + fileAccessEntry.getFile() + ", original size:" + fileAccessEntry.getFile().length()); } else { attachmentBytes = CompressionUtil.getCompressedBytes(sourceInputStream); logger.fine("\t write entry by reading from file: " + fileAccessEntry.getFile() + ", original size:" + fileAccessEntry.getFile().length() + ", compressed size:" + attachmentBytes.length); } } //encrypted the compressed bytes and write out. pos = writeEncryptedContent(attachmentBytes, encryptor, dataOut, pos); } }
From source file:com.ricemap.spateDB.core.RTree.java
/** * Builds the RTree given a serialized list of elements. It uses the given * stockObject to deserialize these elements and build the tree. Also writes * the created tree to the disk directly. * // ww w . j a v a 2s.c om * @param elements * - serialization of elements to be written * @param offset * - index of the first element to use in the elements array * @param len * - number of bytes to user from the elements array * @param bytesAvailable * - size available (in bytes) to store the tree structures * @param dataOut * - an output to use for writing the tree to * @param fast_sort * - setting this to <code>true</code> allows the method to run * faster by materializing the offset of each element in the list * which speeds up the comparison. However, this requires an * additional 16 bytes per element. So, for each 1M elements, the * method will require an additional 16 M bytes (approximately). */ public void bulkLoadWrite(final byte[] element_bytes, final int offset, final int len, final int degree, DataOutput dataOut, final boolean fast_sort, final boolean columnarStorage) { try { columnar = columnarStorage; //TODO: the order of fields should be stable under Oracle JVM, but not guaranteed Field[] fields = stockObject.getClass().getDeclaredFields(); // Count number of elements in the given text int i_start = offset; final Text line = new Text(); while (i_start < offset + len) { int i_end = skipToEOL(element_bytes, i_start); // Extract the line without end of line character line.set(element_bytes, i_start, i_end - i_start - 1); stockObject.fromText(line); elementCount++; i_start = i_end; } LOG.info("Bulk loading an RTree with " + elementCount + " elements"); // It turns out the findBestDegree returns the best degree when the // whole // tree is loaded to memory when processed. However, as current // algorithms // process the tree while it's on disk, a higher degree should be // selected // such that a node fits one file block (assumed to be 4K). // final int degree = findBestDegree(bytesAvailable, elementCount); LOG.info("Writing an RTree with degree " + degree); int height = Math.max(1, (int) Math.ceil(Math.log(elementCount) / Math.log(degree))); int leafNodeCount = (int) Math.pow(degree, height - 1); if (elementCount < 2 * leafNodeCount && height > 1) { height--; leafNodeCount = (int) Math.pow(degree, height - 1); } int nodeCount = (int) ((Math.pow(degree, height) - 1) / (degree - 1)); int nonLeafNodeCount = nodeCount - leafNodeCount; // Keep track of the offset of each element in the text final int[] offsets = new int[elementCount]; final int[] ids = new int[elementCount]; final double[] ts = fast_sort ? new double[elementCount] : null; final double[] xs = fast_sort ? new double[elementCount] : null; final double[] ys = fast_sort ? new double[elementCount] : null; //initialize columnar data output ByteArrayOutputStream index_bos = new ByteArrayOutputStream(); DataOutputStream index_dos = new DataOutputStream(index_bos); ByteArrayOutputStream[] bos = new ByteArrayOutputStream[fields.length]; DataOutputStream[] dos = new DataOutputStream[fields.length]; for (int i = 0; i < bos.length; i++) { bos[i] = new ByteArrayOutputStream(); dos[i] = new DataOutputStream(bos[i]); } i_start = offset; line.clear(); for (int i = 0; i < elementCount; i++) { offsets[i] = i_start; ids[i] = i; int i_end = skipToEOL(element_bytes, i_start); if (xs != null) { // Extract the line with end of line character line.set(element_bytes, i_start, i_end - i_start - 1); stockObject.fromText(line); // Sample center of the shape ts[i] = (stockObject.getMBR().t1 + stockObject.getMBR().t2) / 2; xs[i] = (stockObject.getMBR().x1 + stockObject.getMBR().x2) / 2; ys[i] = (stockObject.getMBR().y1 + stockObject.getMBR().y2) / 2; //build columnar storage if (stockObject instanceof Point3d) { index_dos.writeDouble(ts[i]); index_dos.writeDouble(xs[i]); index_dos.writeDouble(ys[i]); } else { throw new RuntimeException("Indexing non-point shape with RTREE is not supported yet"); } for (int j = 0; j < fields.length; j++) { if (fields[j].getType().equals(Integer.TYPE)) { dos[j].writeInt(fields[j].getInt(stockObject)); } else if (fields[j].getType().equals(Double.TYPE)) { dos[j].writeDouble(fields[j].getDouble(stockObject)); } else if (fields[j].getType().equals(Long.TYPE)) { dos[j].writeLong(fields[j].getLong(stockObject)); } else { continue; //throw new RuntimeException("Field type is not supported yet"); } } } i_start = i_end; } index_dos.close(); for (int i = 0; i < dos.length; i++) { dos[i].close(); } /** A struct to store information about a split */ class SplitStruct extends Prism { /** Start and end index for this split */ int index1, index2; /** Direction of this split */ byte direction; /** Index of first element on disk */ int offsetOfFirstElement; static final byte DIRECTION_T = 0; static final byte DIRECTION_X = 1; static final byte DIRECTION_Y = 2; SplitStruct(int index1, int index2, byte direction) { this.index1 = index1; this.index2 = index2; this.direction = direction; } @Override public void write(DataOutput out) throws IOException { // if (columnarStorage) out.writeInt(index1); else out.writeInt(offsetOfFirstElement); super.write(out); } void partition(Queue<SplitStruct> toBePartitioned) { IndexedSortable sortableT; IndexedSortable sortableX; IndexedSortable sortableY; if (fast_sort) { // Use materialized xs[] and ys[] to do the comparisons sortableT = new IndexedSortable() { @Override public void swap(int i, int j) { // Swap ts double tempt = ts[i]; ts[i] = ts[j]; ts[j] = tempt; // Swap xs double tempx = xs[i]; xs[i] = xs[j]; xs[j] = tempx; // Swap ys double tempY = ys[i]; ys[i] = ys[j]; ys[j] = tempY; // Swap id int tempid = offsets[i]; offsets[i] = offsets[j]; offsets[j] = tempid; tempid = ids[i]; ids[i] = ids[j]; ids[j] = tempid; } @Override public int compare(int i, int j) { if (ts[i] < ts[j]) return -1; if (ts[i] > ts[j]) return 1; return 0; } }; sortableX = new IndexedSortable() { @Override public void swap(int i, int j) { // Swap ts double tempt = ts[i]; ts[i] = ts[j]; ts[j] = tempt; // Swap xs double tempx = xs[i]; xs[i] = xs[j]; xs[j] = tempx; // Swap ys double tempY = ys[i]; ys[i] = ys[j]; ys[j] = tempY; // Swap id int tempid = offsets[i]; offsets[i] = offsets[j]; offsets[j] = tempid; tempid = ids[i]; ids[i] = ids[j]; ids[j] = tempid; } @Override public int compare(int i, int j) { if (ts[i] < ts[j]) return -1; if (xs[i] < xs[j]) return -1; if (xs[i] > xs[j]) return 1; return 0; } }; sortableY = new IndexedSortable() { @Override public void swap(int i, int j) { // Swap ts double tempt = ts[i]; ts[i] = ts[j]; ts[j] = tempt; // Swap xs double tempx = xs[i]; xs[i] = xs[j]; xs[j] = tempx; // Swap ys double tempY = ys[i]; ys[i] = ys[j]; ys[j] = tempY; // Swap id int tempid = offsets[i]; offsets[i] = offsets[j]; offsets[j] = tempid; tempid = ids[i]; ids[i] = ids[j]; ids[j] = tempid; } @Override public int compare(int i, int j) { if (ys[i] < ys[j]) return -1; if (ys[i] > ys[j]) return 1; return 0; } }; } else { // No materialized xs and ys. Always deserialize objects // to compare sortableT = new IndexedSortable() { @Override public void swap(int i, int j) { // Swap id int tempid = offsets[i]; offsets[i] = offsets[j]; offsets[j] = tempid; tempid = ids[i]; ids[i] = ids[j]; ids[j] = tempid; } @Override public int compare(int i, int j) { // Get end of line int eol = skipToEOL(element_bytes, offsets[i]); line.set(element_bytes, offsets[i], eol - offsets[i] - 1); stockObject.fromText(line); double ti = (stockObject.getMBR().t1 + stockObject.getMBR().t2) / 2; eol = skipToEOL(element_bytes, offsets[j]); line.set(element_bytes, offsets[j], eol - offsets[j] - 1); stockObject.fromText(line); double tj = (stockObject.getMBR().t1 + stockObject.getMBR().t2) / 2; if (ti < tj) return -1; if (ti > tj) return 1; return 0; } }; sortableX = new IndexedSortable() { @Override public void swap(int i, int j) { // Swap id int tempid = offsets[i]; offsets[i] = offsets[j]; offsets[j] = tempid; tempid = ids[i]; ids[i] = ids[j]; ids[j] = tempid; } @Override public int compare(int i, int j) { // Get end of line int eol = skipToEOL(element_bytes, offsets[i]); line.set(element_bytes, offsets[i], eol - offsets[i] - 1); stockObject.fromText(line); double xi = (stockObject.getMBR().x1 + stockObject.getMBR().x2) / 2; eol = skipToEOL(element_bytes, offsets[j]); line.set(element_bytes, offsets[j], eol - offsets[j] - 1); stockObject.fromText(line); double xj = (stockObject.getMBR().x1 + stockObject.getMBR().x2) / 2; if (xi < xj) return -1; if (xi > xj) return 1; return 0; } }; sortableY = new IndexedSortable() { @Override public void swap(int i, int j) { // Swap id int tempid = offsets[i]; offsets[i] = offsets[j]; offsets[j] = tempid; tempid = ids[i]; ids[i] = ids[j]; ids[j] = tempid; } @Override public int compare(int i, int j) { int eol = skipToEOL(element_bytes, offsets[i]); line.set(element_bytes, offsets[i], eol - offsets[i] - 1); stockObject.fromText(line); double yi = (stockObject.getMBR().y1 + stockObject.getMBR().y2) / 2; eol = skipToEOL(element_bytes, offsets[j]); line.set(element_bytes, offsets[j], eol - offsets[j] - 1); stockObject.fromText(line); double yj = (stockObject.getMBR().y1 + stockObject.getMBR().y2) / 2; if (yi < yj) return -1; if (yi > yj) return 1; return 0; } }; } final IndexedSorter sorter = new QuickSort(); final IndexedSortable[] sortables = new IndexedSortable[3]; sortables[SplitStruct.DIRECTION_T] = sortableT; sortables[SplitStruct.DIRECTION_X] = sortableX; sortables[SplitStruct.DIRECTION_Y] = sortableY; sorter.sort(sortables[direction], index1, index2); // Partition into maxEntries partitions (equally) and // create a SplitStruct for each partition int i1 = index1; for (int iSplit = 0; iSplit < degree; iSplit++) { int i2 = index1 + (index2 - index1) * (iSplit + 1) / degree; SplitStruct newSplit; if (direction == 0) { newSplit = new SplitStruct(i1, i2, (byte) 1); } else if (direction == 1) { newSplit = new SplitStruct(i1, i2, (byte) 2); } else { newSplit = new SplitStruct(i1, i2, (byte) 0); } toBePartitioned.add(newSplit); i1 = i2; } } } // All nodes stored in level-order traversal Vector<SplitStruct> nodes = new Vector<SplitStruct>(); final Queue<SplitStruct> toBePartitioned = new LinkedList<SplitStruct>(); toBePartitioned.add(new SplitStruct(0, elementCount, SplitStruct.DIRECTION_X)); while (!toBePartitioned.isEmpty()) { SplitStruct split = toBePartitioned.poll(); if (nodes.size() < nonLeafNodeCount) { // This is a non-leaf split.partition(toBePartitioned); } nodes.add(split); } if (nodes.size() != nodeCount) { throw new RuntimeException( "Expected node count: " + nodeCount + ". Real node count: " + nodes.size()); } // Now we have our data sorted in the required order. Start building // the tree. // Store the offset of each leaf node in the tree FSDataOutputStream fakeOut = new FSDataOutputStream(new java.io.OutputStream() { // Null output stream @Override public void write(int b) throws IOException { // Do nothing } @Override public void write(byte[] b, int off, int len) throws IOException { // Do nothing } @Override public void write(byte[] b) throws IOException { // Do nothing } }, null, TreeHeaderSize + nodes.size() * NodeSize); for (int i_leaf = nonLeafNodeCount, i = 0; i_leaf < nodes.size(); i_leaf++) { nodes.elementAt(i_leaf).offsetOfFirstElement = (int) fakeOut.getPos(); if (i != nodes.elementAt(i_leaf).index1) throw new RuntimeException(); double t1, x1, y1, t2, x2, y2; // Initialize MBR to first object int eol = skipToEOL(element_bytes, offsets[i]); fakeOut.write(element_bytes, offsets[i], eol - offsets[i]); line.set(element_bytes, offsets[i], eol - offsets[i] - 1); stockObject.fromText(line); Prism mbr = stockObject.getMBR(); t1 = mbr.t1; x1 = mbr.x1; y1 = mbr.y1; t2 = mbr.t2; x2 = mbr.x2; y2 = mbr.y2; i++; while (i < nodes.elementAt(i_leaf).index2) { eol = skipToEOL(element_bytes, offsets[i]); fakeOut.write(element_bytes, offsets[i], eol - offsets[i]); line.set(element_bytes, offsets[i], eol - offsets[i] - 1); stockObject.fromText(line); mbr = stockObject.getMBR(); if (mbr.t1 < t1) t1 = mbr.t1; if (mbr.x1 < x1) x1 = mbr.x1; if (mbr.y1 < y1) y1 = mbr.y1; if (mbr.t2 > t2) t2 = mbr.t2; if (mbr.x2 > x2) x2 = mbr.x2; if (mbr.y2 > y2) y2 = mbr.y2; i++; } nodes.elementAt(i_leaf).set(t1, x1, y1, t2, x2, y2); } fakeOut.close(); fakeOut = null; // Calculate MBR and offsetOfFirstElement for non-leaves for (int i_node = nonLeafNodeCount - 1; i_node >= 0; i_node--) { int i_first_child = i_node * degree + 1; nodes.elementAt(i_node).offsetOfFirstElement = nodes.elementAt(i_first_child).offsetOfFirstElement; int i_child = 0; Prism mbr; mbr = nodes.elementAt(i_first_child + i_child); double t1 = mbr.t1; double x1 = mbr.x1; double y1 = mbr.y1; double t2 = mbr.t2; double x2 = mbr.x2; double y2 = mbr.y2; i_child++; while (i_child < degree) { mbr = nodes.elementAt(i_first_child + i_child); if (mbr.t1 < t1) t1 = mbr.t1; if (mbr.x1 < x1) x1 = mbr.x1; if (mbr.y1 < y1) y1 = mbr.y1; if (mbr.t2 > t2) t2 = mbr.t2; if (mbr.x2 > x2) x2 = mbr.x2; if (mbr.y2 > y2) y2 = mbr.y2; i_child++; } nodes.elementAt(i_node).set(t1, x1, y1, t2, x2, y2); } // Start writing the tree // write tree header (including size) // Total tree size. (== Total bytes written - 8 bytes for the size // itself) dataOut.writeInt(TreeHeaderSize + NodeSize * nodeCount + len); // Tree height dataOut.writeInt(height); // Degree dataOut.writeInt(degree); dataOut.writeInt(elementCount); //isColumnar dataOut.writeInt(columnarStorage ? 1 : 0); // write nodes for (SplitStruct node : nodes) { node.write(dataOut); } // write elements if (columnarStorage) { byte[] index_bs = index_bos.toByteArray(); byte[][] bss = new byte[bos.length][]; for (int i = 0; i < bss.length; i++) { bss[i] = bos[i].toByteArray(); } for (int element_i = 0; element_i < elementCount; element_i++) { //int eol = skipToEOL(element_bytes, offsets[element_i]); //dataOut.write(element_bytes, offsets[element_i], eol - offsets[element_i]); dataOut.write(index_bs, ids[element_i] * IndexUnitSize, IndexUnitSize); } for (int i = 0; i < fields.length; i++) { int fieldSize = 0; if (fields[i].getType().equals(Integer.TYPE)) { fieldSize = 4; } else if (fields[i].getType().equals(Long.TYPE)) { fieldSize = 8; } else if (fields[i].getType().equals(Double.TYPE)) { fieldSize = 8; } else { //throw new RuntimeException("Unsupported field type: " + fields[i].getType().getName()); continue; } for (int element_i = 0; element_i < elementCount; element_i++) { //int eol = skipToEOL(element_bytes, offsets[element_i]); //dataOut.write(element_bytes, offsets[element_i], eol - offsets[element_i]); dataOut.write(bss[i], ids[element_i] * fieldSize, fieldSize); } } } else { for (int element_i = 0; element_i < elementCount; element_i++) { int eol = skipToEOL(element_bytes, offsets[element_i]); dataOut.write(element_bytes, offsets[element_i], eol - offsets[element_i]); } } } catch (IOException e) { e.printStackTrace(); } catch (IllegalArgumentException e) { // TODO Auto-generated catch block e.printStackTrace(); } catch (IllegalAccessException e) { // TODO Auto-generated catch block e.printStackTrace(); } }
From source file:org.apache.geode.internal.InternalDataSerializer.java
/** * write a set of Long objects/* ww w .j a va2 s. co m*/ * * @param set the set of Long objects * @param hasLongIDs if false, write only ints, not longs * @param out the output stream */ public static void writeSetOfLongs(Set set, boolean hasLongIDs, DataOutput out) throws IOException { if (set == null) { out.writeInt(-1); } else { out.writeInt(set.size()); out.writeBoolean(hasLongIDs); for (Object aSet : set) { Long l = (Long) aSet; if (hasLongIDs) { out.writeLong(l); } else { out.writeInt((int) l.longValue()); } } } }
From source file:org.apache.geode.internal.InternalDataSerializer.java
/** * write a set of Long objects TODO: writeListOfLongs is unused * // w w w. ja va2 s .c o m * @param list the set of Long objects * @param hasLongIDs if false, write only ints, not longs * @param out the output stream */ public static void writeListOfLongs(List list, boolean hasLongIDs, DataOutput out) throws IOException { if (list == null) { out.writeInt(-1); } else { out.writeInt(list.size()); out.writeBoolean(hasLongIDs); for (Object aList : list) { Long l = (Long) aList; if (hasLongIDs) { out.writeLong(l); } else { out.writeInt((int) l.longValue()); } } } }
From source file:org.apache.hadoop.mapred.ReduceTask.java
@Override public void write(DataOutput out) throws IOException { super.write(out); out.writeInt(numMaps); // write the number of maps }
From source file:org.apache.hadoop.mapred.RecoverReducerTask.java
@Override public void write(DataOutput out) throws IOException { super.write(out); out.writeInt(numMaps); // write the number of maps }
From source file:hybridewah.HybridBitmap.java
/** * Serialize.//from w ww. j a v a2 s . c o m * * @param out * the DataOutput stream * @throws IOException * Signals that an I/O exception has occurred. */ public void serialize(DataOutput out) throws IOException { out.writeInt(this.sizeinbits); out.writeInt(this.actualsizeinwords); for (int k = 0; k < this.actualsizeinwords; ++k) out.writeLong(this.buffer[k]); out.writeInt(this.rlw.position); }
From source file:org.apache.geode.internal.InternalDataSerializer.java
/** * Persist this class's map to out TODO: saveRegistrations is unused *//*from w w w. j a v a2 s. co m*/ public static void saveRegistrations(DataOutput out) throws IOException { for (Object v : idsToSerializers.values()) { if (v instanceof InitMarker) { v = ((Marker) v).getSerializer(); } if (v instanceof DataSerializer) { DataSerializer ds = (DataSerializer) v; out.writeInt(ds.getId()); // since 5.7 an int instead of a byte DataSerializer.writeClass(ds.getClass(), out); } } if (!dsClassesToHolders.isEmpty()) { Iterator<Entry<String, SerializerAttributesHolder>> iterator = dsClassesToHolders.entrySet().iterator(); Class dsClass = null; while (iterator.hasNext()) { try { dsClass = getCachedClass(iterator.next().getKey()); } catch (ClassNotFoundException ignored) { logger.info(LogMarker.SERIALIZER, LocalizedMessage.create( LocalizedStrings.InternalDataSerializer_COULD_NOT_LOAD_DATASERIALIZER_CLASS_0, dsClass)); continue; } DataSerializer ds = register(dsClass, false); iterator.remove(); idsToHolders.remove(ds.getId()); for (Class clazz : ds.getSupportedClasses()) { supportedClassesToHolders.remove(clazz.getName()); } out.writeInt(ds.getId()); // since 5.7 an int instead of a byte DataSerializer.writeClass(ds.getClass(), out); } } // We know that DataSerializer's id must be > 0 so write a zero // to mark the end of the ds list. out.writeInt(0); // since 5.7 an int instead of a byte }
From source file:org.apache.geode.internal.InternalDataSerializer.java
public static void writeArrayLength(int len, DataOutput out) throws IOException { if (len == -1) { out.writeByte(NULL_ARRAY);// w w w. j a v a2s . co m } else if (len <= MAX_BYTE_ARRAY_LEN) { out.writeByte(len); } else if (len <= 0xFFFF) { out.writeByte(SHORT_ARRAY_LEN); out.writeShort(len); } else { out.writeByte(INT_ARRAY_LEN); out.writeInt(len); } }