Java tutorial
/* * Licensed to the Apache Software Foundation (ASF) under one or more * contributor license agreements. See the NOTICE file distributed with * this work for additional information regarding copyright ownership. * The ASF licenses this file to You under the Apache License, Version 2.0 * (the "License"); you may not use this file except in compliance with * the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.apache.jackrabbit.oak.fixture; import java.io.File; import java.lang.management.ManagementFactory; import javax.sql.DataSource; import org.apache.commons.io.FileUtils; import org.apache.jackrabbit.oak.Oak; import org.apache.jackrabbit.oak.plugins.document.DocumentMK; import org.apache.jackrabbit.oak.plugins.document.rdb.RDBBlobStore; import org.apache.jackrabbit.oak.plugins.document.rdb.RDBDataSourceFactory; import org.apache.jackrabbit.oak.plugins.document.rdb.RDBDocumentStore; import org.apache.jackrabbit.oak.plugins.document.rdb.RDBOptions; import org.apache.jackrabbit.oak.plugins.document.util.MongoConnection; import org.apache.jackrabbit.oak.plugins.memory.EmptyNodeState; import org.apache.jackrabbit.oak.plugins.memory.MemoryNodeStore; import org.apache.jackrabbit.oak.plugins.segment.SegmentNodeStore; import org.apache.jackrabbit.oak.plugins.segment.SegmentStore; import org.apache.jackrabbit.oak.plugins.segment.file.FileStore; import org.apache.jackrabbit.oak.spi.blob.BlobStore; import org.apache.jackrabbit.oak.spi.state.NodeStore; public abstract class OakFixture { public static final String OAK_MEMORY = "Oak-Memory"; public static final String OAK_MEMORY_NS = "Oak-MemoryNS"; public static final String OAK_MONGO = "Oak-Mongo"; public static final String OAK_MONGO_FDS = "Oak-Mongo-FDS"; public static final String OAK_MONGO_NS = "Oak-MongoNS"; public static final String OAK_RDB = "Oak-RDB"; public static final String OAK_RDB_FDS = "Oak-RDB-FDS"; public static final String OAK_TAR = "Oak-Tar"; public static final String OAK_TAR_FDS = "Oak-Tar-FDS"; public static final String OAK_SEGMENT_TAR = "Oak-Segment-Tar"; public static final String OAK_SEGMENT_TAR_FDS = "Oak-Segment-Tar-FDS"; private final String name; protected final String unique; protected OakFixture(String name) { this.name = name; this.unique = getUniqueDatabaseName(name); } public static String getUniqueDatabaseName(String name) { return String.format("%s-%d", name, System.currentTimeMillis()); } public abstract Oak getOak(int clusterId) throws Exception; public abstract Oak[] setUpCluster(int n) throws Exception; public abstract void tearDownCluster(); @Override public String toString() { return name; } public static OakFixture getMemory(long cacheSize) { return getMemory(OAK_MEMORY, cacheSize); } public static OakFixture getMemoryNS(long cacheSize) { return getMemory(OAK_MEMORY_NS, cacheSize); } public static OakFixture getMemory(String name, final long cacheSize) { return new OakFixture(name) { @Override public Oak getOak(int clusterId) throws Exception { Oak oak; oak = newOak(new MemoryNodeStore()); return oak; } @Override public Oak[] setUpCluster(int n) throws Exception { Oak[] cluster = new Oak[n]; for (int i = 0; i < cluster.length; i++) { Oak oak; oak = newOak(new MemoryNodeStore()); cluster[i] = oak; } return cluster; } @Override public void tearDownCluster() { // nothing to do } }; } public static OakFixture getMongo(String uri, boolean dropDBAfterTest, long cacheSize) { return getMongo(OAK_MONGO, uri, dropDBAfterTest, cacheSize, false, null, 0); } public static OakFixture getMongo(String host, int port, String database, boolean dropDBAfterTest, long cacheSize) { return getMongo(OAK_MONGO, host, port, database, dropDBAfterTest, cacheSize, false, null, 0); } public static OakFixture getMongoNS(String uri, boolean dropDBAfterTest, long cacheSize) { return getMongo(OAK_MONGO_NS, uri, dropDBAfterTest, cacheSize, false, null, 0); } public static OakFixture getMongoNS(String host, int port, String database, boolean dropDBAfterTest, long cacheSize) { return getMongo(OAK_MONGO_NS, host, port, database, dropDBAfterTest, cacheSize, false, null, 0); } public static OakFixture getMongo(String name, final String host, final int port, String database, final boolean dropDBAfterTest, final long cacheSize, final boolean useFileDataStore, final File base, final int fdsCacheInMB) { if (database == null) { database = getUniqueDatabaseName(name); } String uri = "mongodb://" + host + ":" + port + "/" + database; return getMongo(name, uri, dropDBAfterTest, cacheSize, useFileDataStore, base, fdsCacheInMB); } public static OakFixture getMongo(final String name, final String uri, final boolean dropDBAfterTest, final long cacheSize, final boolean useFileDataStore, final File base, final int fdsCacheInMB) { return new OakFixture(name) { private DocumentMK[] kernels; private BlobStoreFixture blobStoreFixture; { if (useFileDataStore) { blobStoreFixture = BlobStoreFixture.getFileDataStore(base, fdsCacheInMB); } else { blobStoreFixture = BlobStoreFixture.create(base, false); } } @Override public Oak getOak(int clusterId) throws Exception { MongoConnection mongo = new MongoConnection(uri); DocumentMK.Builder mkBuilder = new DocumentMK.Builder().setMongoDB(mongo.getDB()) .memoryCacheSize(cacheSize).setPersistentCache("target/persistentCache,time") .setClusterId(clusterId).setLogging(false); setupBlobStore(mkBuilder); DocumentMK dmk = mkBuilder.open(); return newOak(dmk.getNodeStore()); } @Override public Oak[] setUpCluster(int n) throws Exception { Oak[] cluster = new Oak[n]; kernels = new DocumentMK[cluster.length]; for (int i = 0; i < cluster.length; i++) { MongoConnection mongo = new MongoConnection(uri); DocumentMK.Builder mkBuilder = new DocumentMK.Builder().setMongoDB(mongo.getDB()) .memoryCacheSize(cacheSize).setPersistentCache("target/persistentCache,time") .setClusterId(i + 1).setLogging(false); setupBlobStore(mkBuilder); kernels[i] = mkBuilder.open(); cluster[i] = newOak(kernels[i].getNodeStore()); } return cluster; } @Override public void tearDownCluster() { for (DocumentMK kernel : kernels) { kernel.dispose(); } if (dropDBAfterTest) { try { MongoConnection mongo = new MongoConnection(uri); mongo.getDB().dropDatabase(); mongo.close(); if (blobStoreFixture != null) { blobStoreFixture.tearDown(); } } catch (Exception e) { throw new RuntimeException(e); } } } private void setupBlobStore(DocumentMK.Builder mkBuilder) { if (blobStoreFixture != null) { mkBuilder.setBlobStore(blobStoreFixture.setUp()); } } }; } public static OakFixture getRDB(final String name, final String jdbcuri, final String jdbcuser, final String jdbcpasswd, final String tablePrefix, final boolean dropDBAfterTest, final long cacheSize) { return getRDB(name, jdbcuri, jdbcuser, jdbcpasswd, tablePrefix, dropDBAfterTest, cacheSize, false, null, 0); } public static OakFixture getRDB(final String name, final String jdbcuri, final String jdbcuser, final String jdbcpasswd, final String tablePrefix, final boolean dropDBAfterTest, final long cacheSize, final boolean useFileDataStore, final File base, final int fdsCacheInMB) { return new OakFixture(name) { private DocumentMK[] kernels; private BlobStoreFixture blobStoreFixture; { if (useFileDataStore) { blobStoreFixture = BlobStoreFixture.getFileDataStore(base, fdsCacheInMB); } } private RDBOptions getOptions(boolean dropDBAFterTest, String tablePrefix) { return new RDBOptions().dropTablesOnClose(dropDBAfterTest).tablePrefix(tablePrefix); } private BlobStore getBlobStore() { try { if (useFileDataStore) { return blobStoreFixture.setUp(); } else { DataSource ds = RDBDataSourceFactory.forJdbcUrl(jdbcuri, jdbcuser, jdbcpasswd); return new RDBBlobStore(ds, getOptions(dropDBAfterTest, tablePrefix)); } } catch (Exception e) { throw new RuntimeException(e); } } @Override public Oak getOak(int clusterId) throws Exception { DataSource ds = RDBDataSourceFactory.forJdbcUrl(jdbcuri, jdbcuser, jdbcpasswd); DocumentMK.Builder mkBuilder = new DocumentMK.Builder() .setRDBConnection(ds, getOptions(dropDBAfterTest, tablePrefix)).memoryCacheSize(cacheSize) .setClusterId(clusterId).setLogging(false); BlobStore blobStore = getBlobStore(); if (blobStore != null) { mkBuilder.setBlobStore(blobStore); } DocumentMK dmk = mkBuilder.open(); return newOak(dmk.getNodeStore()); } @Override public Oak[] setUpCluster(int n) throws Exception { Oak[] cluster = new Oak[n]; kernels = new DocumentMK[cluster.length]; for (int i = 0; i < cluster.length; i++) { BlobStore blobStore = getBlobStore(); DataSource ds = RDBDataSourceFactory.forJdbcUrl(jdbcuri, jdbcuser, jdbcpasswd); DocumentMK.Builder mkBuilder = new DocumentMK.Builder() .setRDBConnection(ds, getOptions(dropDBAfterTest, tablePrefix)) .memoryCacheSize(cacheSize) // FIXME: OAK-3389 .setLeaseCheck(false).setClusterId(i + 1).setLogging(false); if (blobStore != null) { mkBuilder.setBlobStore(blobStore); } kernels[i] = mkBuilder.open(); cluster[i] = newOak(kernels[i].getNodeStore()); } return cluster; } @Override public void tearDownCluster() { String dropped = ""; for (DocumentMK kernel : kernels) { kernel.dispose(); if (kernel.getDocumentStore() instanceof RDBDocumentStore) { dropped += ((RDBDocumentStore) kernel.getDocumentStore()).getDroppedTables(); } } if (dropDBAfterTest) { if (blobStoreFixture != null) { blobStoreFixture.tearDown(); } if (dropped.isEmpty()) { throw new RuntimeException("dropdb was set, but tables have not been dropped"); } } } }; } public static OakFixture getTar(final String name, final File base, final int maxFileSizeMB, final int cacheSizeMB, final boolean memoryMapping, final boolean useBlobStore) { return new SegmentFixture(name, base, maxFileSizeMB, cacheSizeMB, memoryMapping, useBlobStore); } public static OakFixture getSegmentTar(final String name, final File base, final int maxFileSizeMB, final int cacheSizeMB, final boolean memoryMapping, final boolean useBlobStore) { return new SegmentTarFixture(name, base, maxFileSizeMB, cacheSizeMB, memoryMapping, useBlobStore); } public static class SegmentFixture extends OakFixture { private FileStore[] stores; private BlobStoreFixture[] blobStoreFixtures = new BlobStoreFixture[0]; private final File base; private final int maxFileSizeMB; private final int cacheSizeMB; private final boolean memoryMapping; private final boolean useBlobStore; public SegmentFixture(String name, File base, int maxFileSizeMB, int cacheSizeMB, boolean memoryMapping, boolean useBlobStore) { super(name); this.base = base; this.maxFileSizeMB = maxFileSizeMB; this.cacheSizeMB = cacheSizeMB; this.memoryMapping = memoryMapping; this.useBlobStore = useBlobStore; } @Override public Oak getOak(int clusterId) throws Exception { FileStore fs = FileStore.builder(base).withMaxFileSize(maxFileSizeMB).withCacheSize(cacheSizeMB) .withMemoryMapping(memoryMapping).build(); return newOak(SegmentNodeStore.builder(fs).build()); } @Override public Oak[] setUpCluster(int n) throws Exception { Oak[] cluster = new Oak[n]; stores = new FileStore[cluster.length]; if (useBlobStore) { blobStoreFixtures = new BlobStoreFixture[cluster.length]; } for (int i = 0; i < cluster.length; i++) { BlobStore blobStore = null; if (useBlobStore) { blobStoreFixtures[i] = BlobStoreFixture.create(base, true); blobStore = blobStoreFixtures[i].setUp(); } FileStore.Builder builder = FileStore.builder(new File(base, unique)); if (blobStore != null) { builder.withBlobStore(blobStore); } stores[i] = builder.withRoot(EmptyNodeState.EMPTY_NODE).withMaxFileSize(maxFileSizeMB) .withCacheSize(cacheSizeMB).withMemoryMapping(memoryMapping).build(); cluster[i] = newOak(SegmentNodeStore.builder(stores[i]).build()); } return cluster; } @Override public void tearDownCluster() { for (SegmentStore store : stores) { store.close(); } for (BlobStoreFixture blobStore : blobStoreFixtures) { blobStore.tearDown(); } FileUtils.deleteQuietly(new File(base, unique)); } public BlobStoreFixture[] getBlobStoreFixtures() { return blobStoreFixtures; } public FileStore[] getStores() { return stores; } } static Oak newOak(NodeStore nodeStore) { return new Oak(nodeStore).with(ManagementFactory.getPlatformMBeanServer()); } }