Java tutorial
/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, * software distributed under the License is distributed on an * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY * KIND, either express or implied. See the License for the * specific language governing permissions and limitations * under the License. */ package org.apache.asterix.aql.translator; import java.io.BufferedReader; import java.io.File; import java.io.FileInputStream; import java.io.InputStream; import java.io.InputStreamReader; import java.io.PrintWriter; import java.rmi.RemoteException; import java.util.ArrayList; import java.util.Date; import java.util.HashMap; import java.util.HashSet; import java.util.Iterator; import java.util.List; import java.util.Map; import java.util.Map.Entry; import java.util.Properties; import java.util.Random; import java.util.Set; import java.util.logging.Level; import java.util.logging.Logger; import org.apache.asterix.api.common.APIFramework; import org.apache.asterix.api.common.Job; import org.apache.asterix.api.common.SessionConfig; import org.apache.asterix.api.common.SessionConfig.OutputFormat; import org.apache.asterix.aql.base.Statement; import org.apache.asterix.aql.expression.CompactStatement; import org.apache.asterix.aql.expression.ConnectFeedStatement; import org.apache.asterix.aql.expression.CreateDataverseStatement; import org.apache.asterix.aql.expression.CreateFeedPolicyStatement; import org.apache.asterix.aql.expression.CreateFeedStatement; import org.apache.asterix.aql.expression.CreateFunctionStatement; import org.apache.asterix.aql.expression.CreateIndexStatement; import org.apache.asterix.aql.expression.CreatePrimaryFeedStatement; import org.apache.asterix.aql.expression.CreateSecondaryFeedStatement; import org.apache.asterix.aql.expression.DatasetDecl; import org.apache.asterix.aql.expression.DataverseDecl; import org.apache.asterix.aql.expression.DataverseDropStatement; import org.apache.asterix.aql.expression.DeleteStatement; import org.apache.asterix.aql.expression.DisconnectFeedStatement; import org.apache.asterix.aql.expression.DropStatement; import org.apache.asterix.aql.expression.ExternalDetailsDecl; import org.apache.asterix.aql.expression.FeedDropStatement; import org.apache.asterix.aql.expression.FeedPolicyDropStatement; import org.apache.asterix.aql.expression.FunctionDecl; import org.apache.asterix.aql.expression.FunctionDropStatement; import org.apache.asterix.aql.expression.IDatasetDetailsDecl; import org.apache.asterix.aql.expression.Identifier; import org.apache.asterix.aql.expression.IndexDropStatement; import org.apache.asterix.aql.expression.InsertStatement; import org.apache.asterix.aql.expression.InternalDetailsDecl; import org.apache.asterix.aql.expression.LoadStatement; import org.apache.asterix.aql.expression.NodeGroupDropStatement; import org.apache.asterix.aql.expression.NodegroupDecl; import org.apache.asterix.aql.expression.Query; import org.apache.asterix.aql.expression.RefreshExternalDatasetStatement; import org.apache.asterix.aql.expression.RunStatement; import org.apache.asterix.aql.expression.SetStatement; import org.apache.asterix.aql.expression.SubscribeFeedStatement; import org.apache.asterix.aql.expression.TypeDecl; import org.apache.asterix.aql.expression.TypeDropStatement; import org.apache.asterix.aql.expression.TypeExpression; import org.apache.asterix.aql.expression.WriteStatement; import org.apache.asterix.aql.util.FunctionUtils; import org.apache.asterix.common.config.AsterixCompilerProperties; import org.apache.asterix.common.config.DatasetConfig.DatasetType; import org.apache.asterix.common.config.DatasetConfig.ExternalDatasetTransactionState; import org.apache.asterix.common.config.DatasetConfig.ExternalFilePendingOp; import org.apache.asterix.common.config.DatasetConfig.IndexType; import org.apache.asterix.common.config.GlobalConfig; import org.apache.asterix.common.exceptions.ACIDException; import org.apache.asterix.common.exceptions.AsterixException; import org.apache.asterix.common.feeds.FeedActivity.FeedActivityDetails; import org.apache.asterix.common.feeds.FeedConnectionId; import org.apache.asterix.common.feeds.FeedConnectionRequest; import org.apache.asterix.common.feeds.FeedId; import org.apache.asterix.common.feeds.FeedJointKey; import org.apache.asterix.common.feeds.FeedPolicyAccessor; import org.apache.asterix.common.feeds.api.IFeedJoint; import org.apache.asterix.common.feeds.api.IFeedJoint.FeedJointType; import org.apache.asterix.common.feeds.api.IFeedLifecycleEventSubscriber; import org.apache.asterix.common.feeds.api.IFeedLifecycleEventSubscriber.FeedLifecycleEvent; import org.apache.asterix.common.feeds.api.IFeedLifecycleListener.ConnectionLocation; import org.apache.asterix.common.functions.FunctionSignature; import org.apache.asterix.feeds.CentralFeedManager; import org.apache.asterix.feeds.FeedJoint; import org.apache.asterix.feeds.FeedLifecycleListener; import org.apache.asterix.file.DatasetOperations; import org.apache.asterix.file.DataverseOperations; import org.apache.asterix.file.ExternalIndexingOperations; import org.apache.asterix.file.FeedOperations; import org.apache.asterix.file.IndexOperations; import org.apache.asterix.formats.nontagged.AqlTypeTraitProvider; import org.apache.asterix.metadata.IDatasetDetails; import org.apache.asterix.metadata.MetadataException; import org.apache.asterix.metadata.MetadataManager; import org.apache.asterix.metadata.MetadataTransactionContext; import org.apache.asterix.metadata.api.IMetadataEntity; import org.apache.asterix.metadata.bootstrap.MetadataConstants; import org.apache.asterix.metadata.dataset.hints.DatasetHints; import org.apache.asterix.metadata.dataset.hints.DatasetHints.DatasetNodegroupCardinalityHint; import org.apache.asterix.metadata.declared.AqlMetadataProvider; import org.apache.asterix.metadata.entities.CompactionPolicy; import org.apache.asterix.metadata.entities.Dataset; import org.apache.asterix.metadata.entities.Datatype; import org.apache.asterix.metadata.entities.Dataverse; import org.apache.asterix.metadata.entities.ExternalDatasetDetails; import org.apache.asterix.metadata.entities.ExternalFile; import org.apache.asterix.metadata.entities.Feed; import org.apache.asterix.metadata.entities.Feed.FeedType; import org.apache.asterix.metadata.entities.FeedPolicy; import org.apache.asterix.metadata.entities.Function; import org.apache.asterix.metadata.entities.Index; import org.apache.asterix.metadata.entities.InternalDatasetDetails; import org.apache.asterix.metadata.entities.NodeGroup; import org.apache.asterix.metadata.entities.PrimaryFeed; import org.apache.asterix.metadata.entities.SecondaryFeed; import org.apache.asterix.metadata.feeds.FeedLifecycleEventSubscriber; import org.apache.asterix.metadata.feeds.FeedUtil; import org.apache.asterix.metadata.feeds.IFeedAdapterFactory; import org.apache.asterix.metadata.utils.DatasetUtils; import org.apache.asterix.metadata.utils.ExternalDatasetsRegistry; import org.apache.asterix.metadata.utils.MetadataLockManager; import org.apache.asterix.om.types.ARecordType; import org.apache.asterix.om.types.ATypeTag; import org.apache.asterix.om.types.IAType; import org.apache.asterix.om.types.TypeSignature; import org.apache.asterix.om.util.AsterixAppContextInfo; import org.apache.asterix.optimizer.rules.IntroduceSecondaryIndexInsertDeleteRule; import org.apache.asterix.result.ResultReader; import org.apache.asterix.result.ResultUtils; import org.apache.asterix.runtime.job.listener.JobEventListenerFactory; import org.apache.asterix.runtime.operators.std.FlushDatasetOperatorDescriptor; import org.apache.asterix.transaction.management.service.transaction.DatasetIdFactory; import org.apache.asterix.transaction.management.service.transaction.JobIdFactory; import org.apache.asterix.translator.AbstractAqlTranslator; import org.apache.asterix.translator.CompiledStatements.CompiledConnectFeedStatement; import org.apache.asterix.translator.CompiledStatements.CompiledCreateIndexStatement; import org.apache.asterix.translator.CompiledStatements.CompiledDatasetDropStatement; import org.apache.asterix.translator.CompiledStatements.CompiledDeleteStatement; import org.apache.asterix.translator.CompiledStatements.CompiledIndexCompactStatement; import org.apache.asterix.translator.CompiledStatements.CompiledIndexDropStatement; import org.apache.asterix.translator.CompiledStatements.CompiledInsertStatement; import org.apache.asterix.translator.CompiledStatements.CompiledLoadFromFileStatement; import org.apache.asterix.translator.CompiledStatements.CompiledSubscribeFeedStatement; import org.apache.asterix.translator.CompiledStatements.ICompiledDmlStatement; import org.apache.asterix.translator.TypeTranslator; import org.apache.commons.lang3.StringUtils; import org.apache.hyracks.algebricks.common.constraints.AlgebricksPartitionConstraint; import org.apache.hyracks.algebricks.common.constraints.AlgebricksPartitionConstraintHelper; import org.apache.hyracks.algebricks.common.exceptions.AlgebricksException; import org.apache.hyracks.algebricks.common.utils.Pair; import org.apache.hyracks.algebricks.common.utils.Triple; import org.apache.hyracks.algebricks.core.algebra.expressions.AbstractFunctionCallExpression.FunctionKind; import org.apache.hyracks.algebricks.data.IAWriterFactory; import org.apache.hyracks.algebricks.data.IResultSerializerFactoryProvider; import org.apache.hyracks.algebricks.runtime.base.IPushRuntimeFactory; import org.apache.hyracks.algebricks.runtime.operators.meta.AlgebricksMetaOperatorDescriptor; import org.apache.hyracks.algebricks.runtime.operators.std.EmptyTupleSourceRuntimeFactory; import org.apache.hyracks.algebricks.runtime.serializer.ResultSerializerFactoryProvider; import org.apache.hyracks.algebricks.runtime.writers.PrinterBasedWriterFactory; import org.apache.hyracks.api.client.IHyracksClientConnection; import org.apache.hyracks.api.dataflow.value.ISerializerDeserializer; import org.apache.hyracks.api.dataflow.value.ITypeTraits; import org.apache.hyracks.api.dataflow.value.RecordDescriptor; import org.apache.hyracks.api.dataset.IHyracksDataset; import org.apache.hyracks.api.dataset.ResultSetId; import org.apache.hyracks.api.io.FileReference; import org.apache.hyracks.api.job.JobId; import org.apache.hyracks.api.job.JobSpecification; import org.apache.hyracks.dataflow.std.connectors.OneToOneConnectorDescriptor; import org.apache.hyracks.dataflow.std.file.FileSplit; import org.apache.hyracks.dataflow.std.file.IFileSplitProvider; import org.apache.hyracks.storage.am.lsm.common.api.ILSMMergePolicyFactory; import org.json.JSONArray; import org.json.JSONException; import org.json.JSONObject; import com.google.common.collect.Lists; /* * Provides functionality for executing a batch of AQL statements (queries included) * sequentially. */ public class AqlTranslator extends AbstractAqlTranslator { private static Logger LOGGER = Logger.getLogger(AqlTranslator.class.getName()); private enum ProgressState { NO_PROGRESS, ADDED_PENDINGOP_RECORD_TO_METADATA } public static enum ResultDelivery { SYNC, ASYNC, ASYNC_DEFERRED } public static final boolean IS_DEBUG_MODE = false;//true private final List<Statement> aqlStatements; private final SessionConfig sessionConfig; private Dataverse activeDefaultDataverse; private final List<FunctionDecl> declaredFunctions; public AqlTranslator(List<Statement> aqlStatements, SessionConfig conf) throws MetadataException, AsterixException { this.aqlStatements = aqlStatements; this.sessionConfig = conf; declaredFunctions = getDeclaredFunctions(aqlStatements); } private List<FunctionDecl> getDeclaredFunctions(List<Statement> statements) { List<FunctionDecl> functionDecls = new ArrayList<FunctionDecl>(); for (Statement st : statements) { if (st.getKind().equals(Statement.Kind.FUNCTION_DECL)) { functionDecls.add((FunctionDecl) st); } } return functionDecls; } /** * Compiles and submits for execution a list of AQL statements. * * @param hcc * A Hyracks client connection that is used to submit a jobspec to Hyracks. * @param hdc * A Hyracks dataset client object that is used to read the results. * @param resultDelivery * True if the results should be read asynchronously or false if we should wait for results to be read. * @return A List<QueryResult> containing a QueryResult instance corresponding to each submitted query. * @throws Exception */ public void compileAndExecute(IHyracksClientConnection hcc, IHyracksDataset hdc, ResultDelivery resultDelivery) throws Exception { int resultSetIdCounter = 0; FileSplit outputFile = null; IAWriterFactory writerFactory = PrinterBasedWriterFactory.INSTANCE; IResultSerializerFactoryProvider resultSerializerFactoryProvider = ResultSerializerFactoryProvider.INSTANCE; Map<String, String> config = new HashMap<String, String>(); for (Statement stmt : aqlStatements) { if (sessionConfig.is(SessionConfig.FORMAT_HTML)) { sessionConfig.out().println(APIFramework.HTML_STATEMENT_SEPARATOR); } validateOperation(activeDefaultDataverse, stmt); AqlMetadataProvider metadataProvider = new AqlMetadataProvider(activeDefaultDataverse, CentralFeedManager.getInstance()); metadataProvider.setWriterFactory(writerFactory); metadataProvider.setResultSerializerFactoryProvider(resultSerializerFactoryProvider); metadataProvider.setOutputFile(outputFile); metadataProvider.setConfig(config); switch (stmt.getKind()) { case SET: { handleSetStatement(metadataProvider, stmt, config); break; } case DATAVERSE_DECL: { activeDefaultDataverse = handleUseDataverseStatement(metadataProvider, stmt); break; } case CREATE_DATAVERSE: { handleCreateDataverseStatement(metadataProvider, stmt); break; } case DATASET_DECL: { handleCreateDatasetStatement(metadataProvider, stmt, hcc); break; } case CREATE_INDEX: { handleCreateIndexStatement(metadataProvider, stmt, hcc); break; } case TYPE_DECL: { handleCreateTypeStatement(metadataProvider, stmt); break; } case NODEGROUP_DECL: { handleCreateNodeGroupStatement(metadataProvider, stmt); break; } case DATAVERSE_DROP: { handleDataverseDropStatement(metadataProvider, stmt, hcc); break; } case DATASET_DROP: { handleDatasetDropStatement(metadataProvider, stmt, hcc); break; } case INDEX_DROP: { handleIndexDropStatement(metadataProvider, stmt, hcc); break; } case TYPE_DROP: { handleTypeDropStatement(metadataProvider, stmt); break; } case NODEGROUP_DROP: { handleNodegroupDropStatement(metadataProvider, stmt); break; } case CREATE_FUNCTION: { handleCreateFunctionStatement(metadataProvider, stmt); break; } case FUNCTION_DROP: { handleFunctionDropStatement(metadataProvider, stmt); break; } case LOAD: { handleLoadStatement(metadataProvider, stmt, hcc); break; } case INSERT: { handleInsertStatement(metadataProvider, stmt, hcc); break; } case DELETE: { handleDeleteStatement(metadataProvider, stmt, hcc); break; } case CREATE_PRIMARY_FEED: case CREATE_SECONDARY_FEED: { handleCreateFeedStatement(metadataProvider, stmt, hcc); break; } case DROP_FEED: { handleDropFeedStatement(metadataProvider, stmt, hcc); break; } case DROP_FEED_POLICY: { handleDropFeedPolicyStatement(metadataProvider, stmt, hcc); break; } case CONNECT_FEED: { handleConnectFeedStatement(metadataProvider, stmt, hcc); break; } case DISCONNECT_FEED: { handleDisconnectFeedStatement(metadataProvider, stmt, hcc); break; } case SUBSCRIBE_FEED: { handleSubscribeFeedStatement(metadataProvider, stmt, hcc); break; } case CREATE_FEED_POLICY: { handleCreateFeedPolicyStatement(metadataProvider, stmt, hcc); break; } case QUERY: { metadataProvider.setResultSetId(new ResultSetId(resultSetIdCounter++)); metadataProvider.setResultAsyncMode( resultDelivery == ResultDelivery.ASYNC || resultDelivery == ResultDelivery.ASYNC_DEFERRED); handleQuery(metadataProvider, (Query) stmt, hcc, hdc, resultDelivery); break; } case COMPACT: { handleCompactStatement(metadataProvider, stmt, hcc); break; } case EXTERNAL_DATASET_REFRESH: { handleExternalDatasetRefreshStatement(metadataProvider, stmt, hcc); break; } case WRITE: { Pair<IAWriterFactory, FileSplit> result = handleWriteStatement(metadataProvider, stmt); if (result.first != null) { writerFactory = result.first; } outputFile = result.second; break; } case RUN: { handleRunStatement(metadataProvider, stmt, hcc); break; } } } } private void handleSetStatement(AqlMetadataProvider metadataProvider, Statement stmt, Map<String, String> config) throws RemoteException, ACIDException { SetStatement ss = (SetStatement) stmt; String pname = ss.getPropName(); String pvalue = ss.getPropValue(); config.put(pname, pvalue); } private Pair<IAWriterFactory, FileSplit> handleWriteStatement(AqlMetadataProvider metadataProvider, Statement stmt) throws InstantiationException, IllegalAccessException, ClassNotFoundException { WriteStatement ws = (WriteStatement) stmt; File f = new File(ws.getFileName()); FileSplit outputFile = new FileSplit(ws.getNcName().getValue(), new FileReference(f)); IAWriterFactory writerFactory = null; if (ws.getWriterClassName() != null) { writerFactory = (IAWriterFactory) Class.forName(ws.getWriterClassName()).newInstance(); } return new Pair<IAWriterFactory, FileSplit>(writerFactory, outputFile); } private Dataverse handleUseDataverseStatement(AqlMetadataProvider metadataProvider, Statement stmt) throws Exception { DataverseDecl dvd = (DataverseDecl) stmt; String dvName = dvd.getDataverseName().getValue(); MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); MetadataLockManager.INSTANCE.acquireDataverseReadLock(dvName); try { Dataverse dv = MetadataManager.INSTANCE.getDataverse(metadataProvider.getMetadataTxnContext(), dvName); if (dv == null) { throw new MetadataException("Unknown dataverse " + dvName); } MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); return dv; } catch (Exception e) { abort(e, e, mdTxnCtx); throw new MetadataException(e); } finally { MetadataLockManager.INSTANCE.releaseDataverseReadLock(dvName); } } private void handleCreateDataverseStatement(AqlMetadataProvider metadataProvider, Statement stmt) throws Exception { CreateDataverseStatement stmtCreateDataverse = (CreateDataverseStatement) stmt; String dvName = stmtCreateDataverse.getDataverseName().getValue(); MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); MetadataLockManager.INSTANCE.acquireDataverseReadLock(dvName); try { Dataverse dv = MetadataManager.INSTANCE.getDataverse(metadataProvider.getMetadataTxnContext(), dvName); if (dv != null) { if (stmtCreateDataverse.getIfNotExists()) { MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); return; } else { throw new AlgebricksException("A dataverse with this name " + dvName + " already exists."); } } MetadataManager.INSTANCE.addDataverse(metadataProvider.getMetadataTxnContext(), new Dataverse(dvName, stmtCreateDataverse.getFormat(), IMetadataEntity.PENDING_NO_OP)); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e) { abort(e, e, mdTxnCtx); throw e; } finally { MetadataLockManager.INSTANCE.releaseDataverseReadLock(dvName); } } private void validateCompactionPolicy(String compactionPolicy, Map<String, String> compactionPolicyProperties, MetadataTransactionContext mdTxnCtx, boolean isExternalDataset) throws AsterixException, Exception { CompactionPolicy compactionPolicyEntity = MetadataManager.INSTANCE.getCompactionPolicy(mdTxnCtx, MetadataConstants.METADATA_DATAVERSE_NAME, compactionPolicy); if (compactionPolicyEntity == null) { throw new AsterixException("Unknown compaction policy: " + compactionPolicy); } String compactionPolicyFactoryClassName = compactionPolicyEntity.getClassName(); ILSMMergePolicyFactory mergePolicyFactory = (ILSMMergePolicyFactory) Class .forName(compactionPolicyFactoryClassName).newInstance(); if (isExternalDataset && mergePolicyFactory.getName().compareTo("correlated-prefix") == 0) { throw new AsterixException("The correlated-prefix merge policy cannot be used with external dataset."); } if (compactionPolicyProperties == null) { if (mergePolicyFactory.getName().compareTo("no-merge") != 0) { throw new AsterixException("Compaction policy properties are missing."); } } else { for (Map.Entry<String, String> entry : compactionPolicyProperties.entrySet()) { if (!mergePolicyFactory.getPropertiesNames().contains(entry.getKey())) { throw new AsterixException("Invalid compaction policy property: " + entry.getKey()); } } for (String p : mergePolicyFactory.getPropertiesNames()) { if (!compactionPolicyProperties.containsKey(p)) { throw new AsterixException("Missing compaction policy property: " + p); } } } } private void handleCreateDatasetStatement(AqlMetadataProvider metadataProvider, Statement stmt, IHyracksClientConnection hcc) throws AsterixException, Exception { ProgressState progress = ProgressState.NO_PROGRESS; DatasetDecl dd = (DatasetDecl) stmt; String dataverseName = getActiveDataverse(dd.getDataverse()); String datasetName = dd.getName().getValue(); DatasetType dsType = dd.getDatasetType(); String itemTypeName = dd.getItemTypeName().getValue(); Identifier ngNameId = dd.getNodegroupName(); String nodegroupName = getNodeGroupName(ngNameId, dd, dataverseName); String compactionPolicy = dd.getCompactionPolicy(); Map<String, String> compactionPolicyProperties = dd.getCompactionPolicyProperties(); boolean defaultCompactionPolicy = (compactionPolicy == null); boolean temp = dd.getDatasetDetailsDecl().isTemp(); MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); boolean bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); MetadataLockManager.INSTANCE.createDatasetBegin(dataverseName, dataverseName + "." + itemTypeName, nodegroupName, compactionPolicy, dataverseName + "." + datasetName, defaultCompactionPolicy); Dataset dataset = null; try { IDatasetDetails datasetDetails = null; Dataset ds = MetadataManager.INSTANCE.getDataset(metadataProvider.getMetadataTxnContext(), dataverseName, datasetName); if (ds != null) { if (dd.getIfNotExists()) { MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); return; } else { throw new AlgebricksException("A dataset with this name " + datasetName + " already exists."); } } Datatype dt = MetadataManager.INSTANCE.getDatatype(metadataProvider.getMetadataTxnContext(), dataverseName, itemTypeName); if (dt == null) { throw new AlgebricksException(": type " + itemTypeName + " could not be found."); } String ngName = ngNameId != null ? ngNameId.getValue() : configureNodegroupForDataset(dd, dataverseName, mdTxnCtx); if (compactionPolicy == null) { compactionPolicy = GlobalConfig.DEFAULT_COMPACTION_POLICY_NAME; compactionPolicyProperties = GlobalConfig.DEFAULT_COMPACTION_POLICY_PROPERTIES; } else { validateCompactionPolicy(compactionPolicy, compactionPolicyProperties, mdTxnCtx, false); } switch (dd.getDatasetType()) { case INTERNAL: { IAType itemType = dt.getDatatype(); if (itemType.getTypeTag() != ATypeTag.RECORD) { throw new AlgebricksException("Can only partition ARecord's."); } List<List<String>> partitioningExprs = ((InternalDetailsDecl) dd.getDatasetDetailsDecl()) .getPartitioningExprs(); boolean autogenerated = ((InternalDetailsDecl) dd.getDatasetDetailsDecl()).isAutogenerated(); ARecordType aRecordType = (ARecordType) itemType; List<IAType> partitioningTypes = aRecordType.validatePartitioningExpressions(partitioningExprs, autogenerated); List<String> filterField = ((InternalDetailsDecl) dd.getDatasetDetailsDecl()).getFilterField(); if (filterField != null) { aRecordType.validateFilterField(filterField); } if (compactionPolicy == null) { if (filterField != null) { // If the dataset has a filter and the user didn't specify a merge policy, then we will pick the // correlated-prefix as the default merge policy. compactionPolicy = GlobalConfig.DEFAULT_FILTERED_DATASET_COMPACTION_POLICY_NAME; compactionPolicyProperties = GlobalConfig.DEFAULT_COMPACTION_POLICY_PROPERTIES; } } datasetDetails = new InternalDatasetDetails(InternalDatasetDetails.FileStructure.BTREE, InternalDatasetDetails.PartitioningStrategy.HASH, partitioningExprs, partitioningExprs, partitioningTypes, autogenerated, filterField, temp); break; } case EXTERNAL: { String adapter = ((ExternalDetailsDecl) dd.getDatasetDetailsDecl()).getAdapter(); Map<String, String> properties = ((ExternalDetailsDecl) dd.getDatasetDetailsDecl()).getProperties(); datasetDetails = new ExternalDatasetDetails(adapter, properties, new Date(), ExternalDatasetTransactionState.COMMIT); break; } } //#. initialize DatasetIdFactory if it is not initialized. if (!DatasetIdFactory.isInitialized()) { DatasetIdFactory.initialize(MetadataManager.INSTANCE.getMostRecentDatasetId()); } //#. add a new dataset with PendingAddOp dataset = new Dataset(dataverseName, datasetName, itemTypeName, ngName, compactionPolicy, compactionPolicyProperties, datasetDetails, dd.getHints(), dsType, DatasetIdFactory.generateDatasetId(), IMetadataEntity.PENDING_ADD_OP); MetadataManager.INSTANCE.addDataset(metadataProvider.getMetadataTxnContext(), dataset); if (dd.getDatasetType() == DatasetType.INTERNAL) { Dataverse dataverse = MetadataManager.INSTANCE .getDataverse(metadataProvider.getMetadataTxnContext(), dataverseName); JobSpecification jobSpec = DatasetOperations.createDatasetJobSpec(dataverse, datasetName, metadataProvider); //#. make metadataTxn commit before calling runJob. MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; progress = ProgressState.ADDED_PENDINGOP_RECORD_TO_METADATA; //#. runJob runJob(hcc, jobSpec, true); //#. begin new metadataTxn mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); } //#. add a new dataset with PendingNoOp after deleting the dataset with PendingAddOp MetadataManager.INSTANCE.dropDataset(metadataProvider.getMetadataTxnContext(), dataverseName, datasetName); dataset.setPendingOp(IMetadataEntity.PENDING_NO_OP); MetadataManager.INSTANCE.addDataset(metadataProvider.getMetadataTxnContext(), dataset); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e) { if (bActiveTxn) { abort(e, e, mdTxnCtx); } if (progress == ProgressState.ADDED_PENDINGOP_RECORD_TO_METADATA) { //#. execute compensation operations // remove the index in NC // [Notice] // As long as we updated(and committed) metadata, we should remove any effect of the job // because an exception occurs during runJob. mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); CompiledDatasetDropStatement cds = new CompiledDatasetDropStatement(dataverseName, datasetName); try { JobSpecification jobSpec = DatasetOperations.createDropDatasetJobSpec(cds, metadataProvider); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; runJob(hcc, jobSpec, true); } catch (Exception e2) { e.addSuppressed(e2); if (bActiveTxn) { abort(e, e2, mdTxnCtx); } } // remove the record from the metadata. mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); try { MetadataManager.INSTANCE.dropDataset(metadataProvider.getMetadataTxnContext(), dataverseName, datasetName); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e2) { e.addSuppressed(e2); abort(e, e2, mdTxnCtx); throw new IllegalStateException("System is inconsistent state: pending dataset(" + dataverseName + "." + datasetName + ") couldn't be removed from the metadata", e); } } throw e; } finally { MetadataLockManager.INSTANCE.createDatasetEnd(dataverseName, dataverseName + "." + itemTypeName, nodegroupName, compactionPolicy, dataverseName + "." + datasetName, defaultCompactionPolicy); } } private void validateIfResourceIsActiveInFeed(String dataverseName, String datasetName) throws AsterixException { List<FeedConnectionId> activeFeedConnections = FeedLifecycleListener.INSTANCE .getActiveFeedConnections(null); boolean resourceInUse = false; StringBuilder builder = new StringBuilder(); if (activeFeedConnections != null && !activeFeedConnections.isEmpty()) { for (FeedConnectionId connId : activeFeedConnections) { if (connId.getDatasetName().equals(datasetName)) { resourceInUse = true; builder.append(connId + "\n"); } } } if (resourceInUse) { throw new AsterixException( "Dataset " + datasetName + " is currently being " + "fed into by the following feed(s).\n" + builder.toString() + "\n" + "Operation not supported"); } } private String getNodeGroupName(Identifier ngNameId, DatasetDecl dd, String dataverse) { if (ngNameId != null) { return ngNameId.getValue(); } String hintValue = dd.getHints().get(DatasetNodegroupCardinalityHint.NAME); if (hintValue == null) { return MetadataConstants.METADATA_DEFAULT_NODEGROUP_NAME; } else { return (dataverse + ":" + dd.getName().getValue()); } } private String configureNodegroupForDataset(DatasetDecl dd, String dataverse, MetadataTransactionContext mdTxnCtx) throws AsterixException { int nodegroupCardinality = -1; String nodegroupName; String hintValue = dd.getHints().get(DatasetNodegroupCardinalityHint.NAME); if (hintValue == null) { nodegroupName = MetadataConstants.METADATA_DEFAULT_NODEGROUP_NAME; return nodegroupName; } else { int numChosen = 0; boolean valid = DatasetHints.validate(DatasetNodegroupCardinalityHint.NAME, dd.getHints().get(DatasetNodegroupCardinalityHint.NAME)).first; if (!valid) { throw new AsterixException("Incorrect use of hint:" + DatasetNodegroupCardinalityHint.NAME); } else { nodegroupCardinality = Integer.parseInt(dd.getHints().get(DatasetNodegroupCardinalityHint.NAME)); } Set<String> nodeNames = AsterixAppContextInfo.getInstance().getMetadataProperties().getNodeNames(); Set<String> nodeNamesClone = new HashSet<String>(); for (String node : nodeNames) { nodeNamesClone.add(node); } String metadataNodeName = AsterixAppContextInfo.getInstance().getMetadataProperties() .getMetadataNodeName(); List<String> selectedNodes = new ArrayList<String>(); selectedNodes.add(metadataNodeName); numChosen++; nodeNamesClone.remove(metadataNodeName); if (numChosen < nodegroupCardinality) { Random random = new Random(); String[] nodes = nodeNamesClone.toArray(new String[] {}); int[] b = new int[nodeNamesClone.size()]; for (int i = 0; i < b.length; i++) { b[i] = i; } for (int i = 0; i < nodegroupCardinality - numChosen; i++) { int selected = i + random.nextInt(nodeNamesClone.size() - i); int selNodeIndex = b[selected]; selectedNodes.add(nodes[selNodeIndex]); int temp = b[0]; b[0] = b[selected]; b[selected] = temp; } } nodegroupName = dataverse + ":" + dd.getName().getValue(); MetadataManager.INSTANCE.addNodegroup(mdTxnCtx, new NodeGroup(nodegroupName, selectedNodes)); return nodegroupName; } } @SuppressWarnings("unchecked") private void handleCreateIndexStatement(AqlMetadataProvider metadataProvider, Statement stmt, IHyracksClientConnection hcc) throws Exception { ProgressState progress = ProgressState.NO_PROGRESS; CreateIndexStatement stmtCreateIndex = (CreateIndexStatement) stmt; String dataverseName = getActiveDataverse(stmtCreateIndex.getDataverseName()); String datasetName = stmtCreateIndex.getDatasetName().getValue(); MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); boolean bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); MetadataLockManager.INSTANCE.createIndexBegin(dataverseName, dataverseName + "." + datasetName); String indexName = null; JobSpecification spec = null; Dataset ds = null; // For external datasets ArrayList<ExternalFile> externalFilesSnapshot = null; boolean firstExternalDatasetIndex = false; boolean filesIndexReplicated = false; Index filesIndex = null; boolean datasetLocked = false; try { ds = MetadataManager.INSTANCE.getDataset(metadataProvider.getMetadataTxnContext(), dataverseName, datasetName); if (ds == null) { throw new AlgebricksException( "There is no dataset with this name " + datasetName + " in dataverse " + dataverseName); } indexName = stmtCreateIndex.getIndexName().getValue(); Index idx = MetadataManager.INSTANCE.getIndex(metadataProvider.getMetadataTxnContext(), dataverseName, datasetName, indexName); String itemTypeName = ds.getItemTypeName(); Datatype dt = MetadataManager.INSTANCE.getDatatype(metadataProvider.getMetadataTxnContext(), dataverseName, itemTypeName); IAType itemType = dt.getDatatype(); ARecordType aRecordType = (ARecordType) itemType; List<List<String>> indexFields = new ArrayList<List<String>>(); List<IAType> indexFieldTypes = new ArrayList<IAType>(); for (Pair<List<String>, TypeExpression> fieldExpr : stmtCreateIndex.getFieldExprs()) { IAType fieldType = null; boolean isOpen = aRecordType.isOpen(); ARecordType subType = aRecordType; int i = 0; if (fieldExpr.first.size() > 1 && !isOpen) { for (; i < fieldExpr.first.size() - 1;) { subType = (ARecordType) subType.getFieldType(fieldExpr.first.get(i)); i++; if (subType.isOpen()) { isOpen = true; break; } ; } } if (fieldExpr.second == null) { fieldType = subType.getSubFieldType(fieldExpr.first.subList(i, fieldExpr.first.size())); } else { if (!stmtCreateIndex.isEnforced()) throw new AlgebricksException("Cannot create typed index on \"" + fieldExpr.first + "\" field without enforcing it's type"); if (!isOpen) throw new AlgebricksException("Typed index on \"" + fieldExpr.first + "\" field could be created only for open datatype"); Map<TypeSignature, IAType> typeMap = TypeTranslator.computeTypes(mdTxnCtx, fieldExpr.second, indexName, dataverseName); TypeSignature typeSignature = new TypeSignature(dataverseName, indexName); fieldType = typeMap.get(typeSignature); } if (fieldType == null) throw new AlgebricksException("Unknown type " + fieldExpr.second); indexFields.add(fieldExpr.first); indexFieldTypes.add(fieldType); } aRecordType.validateKeyFields(indexFields, indexFieldTypes, stmtCreateIndex.getIndexType()); if (idx != null) { if (stmtCreateIndex.getIfNotExists()) { MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); return; } else { throw new AlgebricksException("An index with this name " + indexName + " already exists."); } } // Checks whether a user is trying to create an inverted secondary index on a dataset with a variable-length primary key. // Currently, we do not support this. Therefore, as a temporary solution, we print an error message and stop. if (stmtCreateIndex.getIndexType() == IndexType.SINGLE_PARTITION_WORD_INVIX || stmtCreateIndex.getIndexType() == IndexType.SINGLE_PARTITION_NGRAM_INVIX || stmtCreateIndex.getIndexType() == IndexType.LENGTH_PARTITIONED_WORD_INVIX || stmtCreateIndex.getIndexType() == IndexType.LENGTH_PARTITIONED_NGRAM_INVIX) { List<List<String>> partitioningKeys = DatasetUtils.getPartitioningKeys(ds); for (List<String> partitioningKey : partitioningKeys) { IAType keyType = aRecordType.getSubFieldType(partitioningKey); ITypeTraits typeTrait = AqlTypeTraitProvider.INSTANCE.getTypeTrait(keyType); // If it is not a fixed length if (typeTrait.getFixedLength() < 0) { throw new AlgebricksException("The keyword or ngram index -" + indexName + " cannot be created on the dataset -" + datasetName + " due to its variable-length primary key field - " + partitioningKey); } } } if (ds.getDatasetType() == DatasetType.INTERNAL) { validateIfResourceIsActiveInFeed(dataverseName, datasetName); } else { // External dataset // Check if the dataset is indexible if (!ExternalIndexingOperations.isIndexible((ExternalDatasetDetails) ds.getDatasetDetails())) { throw new AlgebricksException( "dataset using " + ((ExternalDatasetDetails) ds.getDatasetDetails()).getAdapter() + " Adapter can't be indexed"); } // check if the name of the index is valid if (!ExternalIndexingOperations.isValidIndexName(datasetName, indexName)) { throw new AlgebricksException("external dataset index name is invalid"); } // Check if the files index exist filesIndex = MetadataManager.INSTANCE.getIndex(metadataProvider.getMetadataTxnContext(), dataverseName, datasetName, ExternalIndexingOperations.getFilesIndexName(datasetName)); firstExternalDatasetIndex = (filesIndex == null); // lock external dataset ExternalDatasetsRegistry.INSTANCE.buildIndexBegin(ds, firstExternalDatasetIndex); datasetLocked = true; if (firstExternalDatasetIndex) { // verify that no one has created an index before we acquire the lock filesIndex = MetadataManager.INSTANCE.getIndex(metadataProvider.getMetadataTxnContext(), dataverseName, datasetName, ExternalIndexingOperations.getFilesIndexName(datasetName)); if (filesIndex != null) { ExternalDatasetsRegistry.INSTANCE.buildIndexEnd(ds, firstExternalDatasetIndex); firstExternalDatasetIndex = false; ExternalDatasetsRegistry.INSTANCE.buildIndexBegin(ds, firstExternalDatasetIndex); } } if (firstExternalDatasetIndex) { // Get snapshot from External File System externalFilesSnapshot = ExternalIndexingOperations.getSnapshotFromExternalFileSystem(ds); // Add an entry for the files index filesIndex = new Index(dataverseName, datasetName, ExternalIndexingOperations.getFilesIndexName(datasetName), IndexType.BTREE, ExternalIndexingOperations.FILE_INDEX_FIELD_NAMES, ExternalIndexingOperations.FILE_INDEX_FIELD_TYPES, false, false, IMetadataEntity.PENDING_ADD_OP); MetadataManager.INSTANCE.addIndex(metadataProvider.getMetadataTxnContext(), filesIndex); // Add files to the external files index for (ExternalFile file : externalFilesSnapshot) { MetadataManager.INSTANCE.addExternalFile(mdTxnCtx, file); } // This is the first index for the external dataset, replicate the files index spec = ExternalIndexingOperations.buildFilesIndexReplicationJobSpec(ds, externalFilesSnapshot, metadataProvider, true); if (spec == null) { throw new AsterixException( "Failed to create job spec for replicating Files Index For external dataset"); } filesIndexReplicated = true; runJob(hcc, spec, true); } } //check whether there exists another enforced index on the same field if (stmtCreateIndex.isEnforced()) { List<Index> indexes = MetadataManager.INSTANCE .getDatasetIndexes(metadataProvider.getMetadataTxnContext(), dataverseName, datasetName); for (Index index : indexes) { if (index.getKeyFieldNames().equals(indexFields) && !index.getKeyFieldTypes().equals(indexFieldTypes) && index.isEnforcingKeyFileds()) throw new AsterixException( "Cannot create index " + indexName + " , enforced index " + index.getIndexName() + " on field \"" + StringUtils.join(indexFields, ',') + "\" already exist"); } } //#. add a new index with PendingAddOp Index index = new Index(dataverseName, datasetName, indexName, stmtCreateIndex.getIndexType(), indexFields, indexFieldTypes, stmtCreateIndex.getGramLength(), stmtCreateIndex.isEnforced(), false, IMetadataEntity.PENDING_ADD_OP); MetadataManager.INSTANCE.addIndex(metadataProvider.getMetadataTxnContext(), index); ARecordType enforcedType = null; if (stmtCreateIndex.isEnforced()) { enforcedType = IntroduceSecondaryIndexInsertDeleteRule.createEnforcedType(aRecordType, Lists.newArrayList(index)); } //#. prepare to create the index artifact in NC. CompiledCreateIndexStatement cis = new CompiledCreateIndexStatement(index.getIndexName(), dataverseName, index.getDatasetName(), index.getKeyFieldNames(), index.getKeyFieldTypes(), index.isEnforcingKeyFileds(), index.getGramLength(), index.getIndexType()); spec = IndexOperations.buildSecondaryIndexCreationJobSpec(cis, aRecordType, enforcedType, metadataProvider); if (spec == null) { throw new AsterixException("Failed to create job spec for creating index '" + stmtCreateIndex.getDatasetName() + "." + stmtCreateIndex.getIndexName() + "'"); } MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; progress = ProgressState.ADDED_PENDINGOP_RECORD_TO_METADATA; //#. create the index artifact in NC. runJob(hcc, spec, true); mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); //#. load data into the index in NC. cis = new CompiledCreateIndexStatement(index.getIndexName(), dataverseName, index.getDatasetName(), index.getKeyFieldNames(), index.getKeyFieldTypes(), index.isEnforcingKeyFileds(), index.getGramLength(), index.getIndexType()); spec = IndexOperations.buildSecondaryIndexLoadingJobSpec(cis, aRecordType, enforcedType, metadataProvider); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; runJob(hcc, spec, true); //#. begin new metadataTxn mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); //#. add another new index with PendingNoOp after deleting the index with PendingAddOp MetadataManager.INSTANCE.dropIndex(metadataProvider.getMetadataTxnContext(), dataverseName, datasetName, indexName); index.setPendingOp(IMetadataEntity.PENDING_NO_OP); MetadataManager.INSTANCE.addIndex(metadataProvider.getMetadataTxnContext(), index); // add another new files index with PendingNoOp after deleting the index with PendingAddOp if (firstExternalDatasetIndex) { MetadataManager.INSTANCE.dropIndex(metadataProvider.getMetadataTxnContext(), dataverseName, datasetName, filesIndex.getIndexName()); filesIndex.setPendingOp(IMetadataEntity.PENDING_NO_OP); MetadataManager.INSTANCE.addIndex(metadataProvider.getMetadataTxnContext(), filesIndex); // update transaction timestamp ((ExternalDatasetDetails) ds.getDatasetDetails()).setRefreshTimestamp(new Date()); MetadataManager.INSTANCE.updateDataset(mdTxnCtx, ds); } MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e) { if (bActiveTxn) { abort(e, e, mdTxnCtx); } // If files index was replicated for external dataset, it should be cleaned up on NC side if (filesIndexReplicated) { mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); bActiveTxn = true; CompiledIndexDropStatement cds = new CompiledIndexDropStatement(dataverseName, datasetName, ExternalIndexingOperations.getFilesIndexName(datasetName)); try { JobSpecification jobSpec = ExternalIndexingOperations.buildDropFilesIndexJobSpec(cds, metadataProvider, ds); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; runJob(hcc, jobSpec, true); } catch (Exception e2) { e.addSuppressed(e2); if (bActiveTxn) { abort(e, e2, mdTxnCtx); } } } if (progress == ProgressState.ADDED_PENDINGOP_RECORD_TO_METADATA) { //#. execute compensation operations // remove the index in NC mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); CompiledIndexDropStatement cds = new CompiledIndexDropStatement(dataverseName, datasetName, indexName); try { JobSpecification jobSpec = IndexOperations.buildDropSecondaryIndexJobSpec(cds, metadataProvider, ds); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; runJob(hcc, jobSpec, true); } catch (Exception e2) { e.addSuppressed(e2); if (bActiveTxn) { abort(e, e2, mdTxnCtx); } } if (firstExternalDatasetIndex) { mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); try { // Drop External Files from metadata MetadataManager.INSTANCE.dropDatasetExternalFiles(mdTxnCtx, ds); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e2) { e.addSuppressed(e2); abort(e, e2, mdTxnCtx); throw new IllegalStateException("System is inconsistent state: pending files for(" + dataverseName + "." + datasetName + ") couldn't be removed from the metadata", e); } mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); try { // Drop the files index from metadata MetadataManager.INSTANCE.dropIndex(metadataProvider.getMetadataTxnContext(), dataverseName, datasetName, ExternalIndexingOperations.getFilesIndexName(datasetName)); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e2) { e.addSuppressed(e2); abort(e, e2, mdTxnCtx); throw new IllegalStateException( "System is inconsistent state: pending index(" + dataverseName + "." + datasetName + "." + ExternalIndexingOperations.getFilesIndexName(datasetName) + ") couldn't be removed from the metadata", e); } } // remove the record from the metadata. mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); try { MetadataManager.INSTANCE.dropIndex(metadataProvider.getMetadataTxnContext(), dataverseName, datasetName, indexName); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e2) { e.addSuppressed(e2); abort(e, e2, mdTxnCtx); throw new IllegalStateException( "System is in inconsistent state: pending index(" + dataverseName + "." + datasetName + "." + indexName + ") couldn't be removed from the metadata", e); } } throw e; } finally { MetadataLockManager.INSTANCE.createIndexEnd(dataverseName, dataverseName + "." + datasetName); if (datasetLocked) { ExternalDatasetsRegistry.INSTANCE.buildIndexEnd(ds, firstExternalDatasetIndex); } } } private void handleCreateTypeStatement(AqlMetadataProvider metadataProvider, Statement stmt) throws Exception { TypeDecl stmtCreateType = (TypeDecl) stmt; String dataverseName = getActiveDataverse(stmtCreateType.getDataverseName()); String typeName = stmtCreateType.getIdent().getValue(); MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); MetadataLockManager.INSTANCE.createTypeBegin(dataverseName, dataverseName + "." + typeName); try { Dataverse dv = MetadataManager.INSTANCE.getDataverse(mdTxnCtx, dataverseName); if (dv == null) { throw new AlgebricksException("Unknown dataverse " + dataverseName); } Datatype dt = MetadataManager.INSTANCE.getDatatype(mdTxnCtx, dataverseName, typeName); if (dt != null) { if (!stmtCreateType.getIfNotExists()) { throw new AlgebricksException("A datatype with this name " + typeName + " already exists."); } } else { if (builtinTypeMap.get(typeName) != null) { throw new AlgebricksException("Cannot redefine builtin type " + typeName + "."); } else { Map<TypeSignature, IAType> typeMap = TypeTranslator.computeTypes(mdTxnCtx, stmtCreateType.getTypeDef(), stmtCreateType.getIdent().getValue(), dataverseName); TypeSignature typeSignature = new TypeSignature(dataverseName, typeName); IAType type = typeMap.get(typeSignature); MetadataManager.INSTANCE.addDatatype(mdTxnCtx, new Datatype(dataverseName, typeName, type, false)); } } MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e) { abort(e, e, mdTxnCtx); throw e; } finally { MetadataLockManager.INSTANCE.createTypeEnd(dataverseName, dataverseName + "." + typeName); } } private void handleDataverseDropStatement(AqlMetadataProvider metadataProvider, Statement stmt, IHyracksClientConnection hcc) throws Exception { DataverseDropStatement stmtDelete = (DataverseDropStatement) stmt; String dataverseName = stmtDelete.getDataverseName().getValue(); ProgressState progress = ProgressState.NO_PROGRESS; MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); boolean bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); MetadataLockManager.INSTANCE.acquireDataverseWriteLock(dataverseName); List<JobSpecification> jobsToExecute = new ArrayList<JobSpecification>(); try { Dataverse dv = MetadataManager.INSTANCE.getDataverse(mdTxnCtx, dataverseName); if (dv == null) { if (stmtDelete.getIfExists()) { MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); return; } else { throw new AlgebricksException("There is no dataverse with this name " + dataverseName + "."); } } //# disconnect all feeds from any datasets in the dataverse. List<FeedConnectionId> activeFeedConnections = FeedLifecycleListener.INSTANCE .getActiveFeedConnections(null); DisconnectFeedStatement disStmt = null; Identifier dvId = new Identifier(dataverseName); for (FeedConnectionId connection : activeFeedConnections) { FeedId feedId = connection.getFeedId(); if (feedId.getDataverse().equals(dataverseName)) { disStmt = new DisconnectFeedStatement(dvId, new Identifier(feedId.getFeedName()), new Identifier(connection.getDatasetName())); try { handleDisconnectFeedStatement(metadataProvider, disStmt, hcc); if (LOGGER.isLoggable(Level.INFO)) { LOGGER.info("Disconnected feed " + feedId.getFeedName() + " from dataset " + connection.getDatasetName()); } } catch (Exception exception) { if (LOGGER.isLoggable(Level.WARNING)) { LOGGER.warning("Unable to disconnect feed " + feedId.getFeedName() + " from dataset " + connection.getDatasetName() + ". Encountered exception " + exception); } } } } //#. prepare jobs which will drop corresponding datasets with indexes. List<Dataset> datasets = MetadataManager.INSTANCE.getDataverseDatasets(mdTxnCtx, dataverseName); for (int j = 0; j < datasets.size(); j++) { String datasetName = datasets.get(j).getDatasetName(); DatasetType dsType = datasets.get(j).getDatasetType(); if (dsType == DatasetType.INTERNAL) { List<Index> indexes = MetadataManager.INSTANCE.getDatasetIndexes(mdTxnCtx, dataverseName, datasetName); for (int k = 0; k < indexes.size(); k++) { if (indexes.get(k).isSecondaryIndex()) { CompiledIndexDropStatement cds = new CompiledIndexDropStatement(dataverseName, datasetName, indexes.get(k).getIndexName()); jobsToExecute.add(IndexOperations.buildDropSecondaryIndexJobSpec(cds, metadataProvider, datasets.get(j))); } } CompiledDatasetDropStatement cds = new CompiledDatasetDropStatement(dataverseName, datasetName); jobsToExecute.add(DatasetOperations.createDropDatasetJobSpec(cds, metadataProvider)); } else { // External dataset List<Index> indexes = MetadataManager.INSTANCE.getDatasetIndexes(mdTxnCtx, dataverseName, datasetName); for (int k = 0; k < indexes.size(); k++) { if (ExternalIndexingOperations.isFileIndex(indexes.get(k))) { CompiledIndexDropStatement cds = new CompiledIndexDropStatement(dataverseName, datasetName, indexes.get(k).getIndexName()); jobsToExecute.add(ExternalIndexingOperations.buildDropFilesIndexJobSpec(cds, metadataProvider, datasets.get(j))); } else { CompiledIndexDropStatement cds = new CompiledIndexDropStatement(dataverseName, datasetName, indexes.get(k).getIndexName()); jobsToExecute.add(IndexOperations.buildDropSecondaryIndexJobSpec(cds, metadataProvider, datasets.get(j))); } } ExternalDatasetsRegistry.INSTANCE.removeDatasetInfo(datasets.get(j)); } } jobsToExecute.add(DataverseOperations.createDropDataverseJobSpec(dv, metadataProvider)); //#. mark PendingDropOp on the dataverse record by // first, deleting the dataverse record from the DATAVERSE_DATASET // second, inserting the dataverse record with the PendingDropOp value into the DATAVERSE_DATASET MetadataManager.INSTANCE.dropDataverse(mdTxnCtx, dataverseName); MetadataManager.INSTANCE.addDataverse(mdTxnCtx, new Dataverse(dataverseName, dv.getDataFormat(), IMetadataEntity.PENDING_DROP_OP)); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; progress = ProgressState.ADDED_PENDINGOP_RECORD_TO_METADATA; for (JobSpecification jobSpec : jobsToExecute) { runJob(hcc, jobSpec, true); } mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); //#. finally, delete the dataverse. MetadataManager.INSTANCE.dropDataverse(mdTxnCtx, dataverseName); if (activeDefaultDataverse != null && activeDefaultDataverse.getDataverseName() == dataverseName) { activeDefaultDataverse = null; } MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e) { if (bActiveTxn) { abort(e, e, mdTxnCtx); } if (progress == ProgressState.ADDED_PENDINGOP_RECORD_TO_METADATA) { if (activeDefaultDataverse != null && activeDefaultDataverse.getDataverseName() == dataverseName) { activeDefaultDataverse = null; } //#. execute compensation operations // remove the all indexes in NC try { for (JobSpecification jobSpec : jobsToExecute) { runJob(hcc, jobSpec, true); } } catch (Exception e2) { //do no throw exception since still the metadata needs to be compensated. e.addSuppressed(e2); } // remove the record from the metadata. mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); try { MetadataManager.INSTANCE.dropDataverse(mdTxnCtx, dataverseName); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e2) { e.addSuppressed(e2); abort(e, e2, mdTxnCtx); throw new IllegalStateException("System is inconsistent state: pending dataverse(" + dataverseName + ") couldn't be removed from the metadata", e); } } throw e; } finally { MetadataLockManager.INSTANCE.releaseDataverseWriteLock(dataverseName); } } private void handleDatasetDropStatement(AqlMetadataProvider metadataProvider, Statement stmt, IHyracksClientConnection hcc) throws Exception { DropStatement stmtDelete = (DropStatement) stmt; String dataverseName = getActiveDataverse(stmtDelete.getDataverseName()); String datasetName = stmtDelete.getDatasetName().getValue(); ProgressState progress = ProgressState.NO_PROGRESS; MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); boolean bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); MetadataLockManager.INSTANCE.dropDatasetBegin(dataverseName, dataverseName + "." + datasetName); List<JobSpecification> jobsToExecute = new ArrayList<JobSpecification>(); try { Dataset ds = MetadataManager.INSTANCE.getDataset(mdTxnCtx, dataverseName, datasetName); if (ds == null) { if (stmtDelete.getIfExists()) { MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); return; } else { throw new AlgebricksException("There is no dataset with this name " + datasetName + " in dataverse " + dataverseName + "."); } } Map<FeedConnectionId, Pair<JobSpecification, Boolean>> disconnectJobList = new HashMap<FeedConnectionId, Pair<JobSpecification, Boolean>>(); if (ds.getDatasetType() == DatasetType.INTERNAL) { // prepare job spec(s) that would disconnect any active feeds involving the dataset. List<FeedConnectionId> feedConnections = FeedLifecycleListener.INSTANCE .getActiveFeedConnections(null); if (feedConnections != null && !feedConnections.isEmpty()) { for (FeedConnectionId connection : feedConnections) { Pair<JobSpecification, Boolean> p = FeedOperations .buildDisconnectFeedJobSpec(metadataProvider, connection); disconnectJobList.put(connection, p); if (LOGGER.isLoggable(Level.INFO)) { LOGGER.info("Disconnecting feed " + connection.getFeedId().getFeedName() + " from dataset " + datasetName + " as dataset is being dropped"); } } } //#. prepare jobs to drop the datatset and the indexes in NC List<Index> indexes = MetadataManager.INSTANCE.getDatasetIndexes(mdTxnCtx, dataverseName, datasetName); for (int j = 0; j < indexes.size(); j++) { if (indexes.get(j).isSecondaryIndex()) { CompiledIndexDropStatement cds = new CompiledIndexDropStatement(dataverseName, datasetName, indexes.get(j).getIndexName()); jobsToExecute .add(IndexOperations.buildDropSecondaryIndexJobSpec(cds, metadataProvider, ds)); } } CompiledDatasetDropStatement cds = new CompiledDatasetDropStatement(dataverseName, datasetName); jobsToExecute.add(DatasetOperations.createDropDatasetJobSpec(cds, metadataProvider)); //#. mark the existing dataset as PendingDropOp MetadataManager.INSTANCE.dropDataset(mdTxnCtx, dataverseName, datasetName); MetadataManager.INSTANCE.addDataset(mdTxnCtx, new Dataset(dataverseName, datasetName, ds.getItemTypeName(), ds.getNodeGroupName(), ds.getCompactionPolicy(), ds.getCompactionPolicyProperties(), ds.getDatasetDetails(), ds.getHints(), ds.getDatasetType(), ds.getDatasetId(), IMetadataEntity.PENDING_DROP_OP)); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; progress = ProgressState.ADDED_PENDINGOP_RECORD_TO_METADATA; //# disconnect the feeds for (Pair<JobSpecification, Boolean> p : disconnectJobList.values()) { runJob(hcc, p.first, true); } //#. run the jobs for (JobSpecification jobSpec : jobsToExecute) { runJob(hcc, jobSpec, true); } mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); } else { // External dataset ExternalDatasetsRegistry.INSTANCE.removeDatasetInfo(ds); //#. prepare jobs to drop the datatset and the indexes in NC List<Index> indexes = MetadataManager.INSTANCE.getDatasetIndexes(mdTxnCtx, dataverseName, datasetName); for (int j = 0; j < indexes.size(); j++) { if (ExternalIndexingOperations.isFileIndex(indexes.get(j))) { CompiledIndexDropStatement cds = new CompiledIndexDropStatement(dataverseName, datasetName, indexes.get(j).getIndexName()); jobsToExecute .add(IndexOperations.buildDropSecondaryIndexJobSpec(cds, metadataProvider, ds)); } else { CompiledIndexDropStatement cds = new CompiledIndexDropStatement(dataverseName, datasetName, indexes.get(j).getIndexName()); jobsToExecute.add( ExternalIndexingOperations.buildDropFilesIndexJobSpec(cds, metadataProvider, ds)); } } //#. mark the existing dataset as PendingDropOp MetadataManager.INSTANCE.dropDataset(mdTxnCtx, dataverseName, datasetName); MetadataManager.INSTANCE.addDataset(mdTxnCtx, new Dataset(dataverseName, datasetName, ds.getItemTypeName(), ds.getNodeGroupName(), ds.getCompactionPolicy(), ds.getCompactionPolicyProperties(), ds.getDatasetDetails(), ds.getHints(), ds.getDatasetType(), ds.getDatasetId(), IMetadataEntity.PENDING_DROP_OP)); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; progress = ProgressState.ADDED_PENDINGOP_RECORD_TO_METADATA; //#. run the jobs for (JobSpecification jobSpec : jobsToExecute) { runJob(hcc, jobSpec, true); } if (indexes.size() > 0) { ExternalDatasetsRegistry.INSTANCE.removeDatasetInfo(ds); } mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); } //#. finally, delete the dataset. MetadataManager.INSTANCE.dropDataset(mdTxnCtx, dataverseName, datasetName); // Drop the associated nodegroup String nodegroup = ds.getNodeGroupName(); if (!nodegroup.equalsIgnoreCase(MetadataConstants.METADATA_DEFAULT_NODEGROUP_NAME)) { MetadataManager.INSTANCE.dropNodegroup(mdTxnCtx, dataverseName + ":" + datasetName); } MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e) { if (bActiveTxn) { abort(e, e, mdTxnCtx); } if (progress == ProgressState.ADDED_PENDINGOP_RECORD_TO_METADATA) { //#. execute compensation operations // remove the all indexes in NC try { for (JobSpecification jobSpec : jobsToExecute) { runJob(hcc, jobSpec, true); } } catch (Exception e2) { //do no throw exception since still the metadata needs to be compensated. e.addSuppressed(e2); } // remove the record from the metadata. mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); try { MetadataManager.INSTANCE.dropDataset(metadataProvider.getMetadataTxnContext(), dataverseName, datasetName); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e2) { e.addSuppressed(e2); abort(e, e2, mdTxnCtx); throw new IllegalStateException("System is inconsistent state: pending dataset(" + dataverseName + "." + datasetName + ") couldn't be removed from the metadata", e); } } throw e; } finally { MetadataLockManager.INSTANCE.dropDatasetEnd(dataverseName, dataverseName + "." + datasetName); } } private void handleIndexDropStatement(AqlMetadataProvider metadataProvider, Statement stmt, IHyracksClientConnection hcc) throws Exception { IndexDropStatement stmtIndexDrop = (IndexDropStatement) stmt; String datasetName = stmtIndexDrop.getDatasetName().getValue(); String dataverseName = getActiveDataverse(stmtIndexDrop.getDataverseName()); ProgressState progress = ProgressState.NO_PROGRESS; MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); boolean bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); MetadataLockManager.INSTANCE.dropIndexBegin(dataverseName, dataverseName + "." + datasetName); String indexName = null; // For external index boolean dropFilesIndex = false; List<JobSpecification> jobsToExecute = new ArrayList<JobSpecification>(); try { Dataset ds = MetadataManager.INSTANCE.getDataset(mdTxnCtx, dataverseName, datasetName); if (ds == null) { throw new AlgebricksException( "There is no dataset with this name " + datasetName + " in dataverse " + dataverseName); } List<FeedConnectionId> feedConnections = FeedLifecycleListener.INSTANCE.getActiveFeedConnections(null); boolean resourceInUse = false; if (feedConnections != null && !feedConnections.isEmpty()) { StringBuilder builder = new StringBuilder(); for (FeedConnectionId connection : feedConnections) { if (connection.getDatasetName().equals(datasetName)) { resourceInUse = true; builder.append(connection + "\n"); } } if (resourceInUse) { throw new AsterixException( "Dataset" + datasetName + " is currently being fed into by the following feeds " + "." + builder.toString() + "\nOperation not supported."); } } if (ds.getDatasetType() == DatasetType.INTERNAL) { indexName = stmtIndexDrop.getIndexName().getValue(); Index index = MetadataManager.INSTANCE.getIndex(mdTxnCtx, dataverseName, datasetName, indexName); if (index == null) { if (stmtIndexDrop.getIfExists()) { MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); return; } else { throw new AlgebricksException("There is no index with this name " + indexName + "."); } } //#. prepare a job to drop the index in NC. CompiledIndexDropStatement cds = new CompiledIndexDropStatement(dataverseName, datasetName, indexName); jobsToExecute.add(IndexOperations.buildDropSecondaryIndexJobSpec(cds, metadataProvider, ds)); //#. mark PendingDropOp on the existing index MetadataManager.INSTANCE.dropIndex(mdTxnCtx, dataverseName, datasetName, indexName); MetadataManager.INSTANCE.addIndex(mdTxnCtx, new Index(dataverseName, datasetName, indexName, index.getIndexType(), index.getKeyFieldNames(), index.getKeyFieldTypes(), index.isEnforcingKeyFileds(), index.isPrimaryIndex(), IMetadataEntity.PENDING_DROP_OP)); //#. commit the existing transaction before calling runJob. MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; progress = ProgressState.ADDED_PENDINGOP_RECORD_TO_METADATA; for (JobSpecification jobSpec : jobsToExecute) { runJob(hcc, jobSpec, true); } //#. begin a new transaction mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); //#. finally, delete the existing index MetadataManager.INSTANCE.dropIndex(mdTxnCtx, dataverseName, datasetName, indexName); } else { // External dataset indexName = stmtIndexDrop.getIndexName().getValue(); Index index = MetadataManager.INSTANCE.getIndex(mdTxnCtx, dataverseName, datasetName, indexName); if (index == null) { if (stmtIndexDrop.getIfExists()) { MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); return; } else { throw new AlgebricksException("There is no index with this name " + indexName + "."); } } else if (ExternalIndexingOperations.isFileIndex(index)) { throw new AlgebricksException("Dropping a dataset's files index is not allowed."); } //#. prepare a job to drop the index in NC. CompiledIndexDropStatement cds = new CompiledIndexDropStatement(dataverseName, datasetName, indexName); jobsToExecute.add(IndexOperations.buildDropSecondaryIndexJobSpec(cds, metadataProvider, ds)); List<Index> datasetIndexes = MetadataManager.INSTANCE.getDatasetIndexes(mdTxnCtx, dataverseName, datasetName); if (datasetIndexes.size() == 2) { dropFilesIndex = true; // only one index + the files index, we need to delete both of the indexes for (Index externalIndex : datasetIndexes) { if (ExternalIndexingOperations.isFileIndex(externalIndex)) { cds = new CompiledIndexDropStatement(dataverseName, datasetName, externalIndex.getIndexName()); jobsToExecute.add(ExternalIndexingOperations.buildDropFilesIndexJobSpec(cds, metadataProvider, ds)); //#. mark PendingDropOp on the existing files index MetadataManager.INSTANCE.dropIndex(mdTxnCtx, dataverseName, datasetName, externalIndex.getIndexName()); MetadataManager.INSTANCE.addIndex(mdTxnCtx, new Index(dataverseName, datasetName, externalIndex.getIndexName(), externalIndex.getIndexType(), externalIndex.getKeyFieldNames(), index.getKeyFieldTypes(), index.isEnforcingKeyFileds(), externalIndex.isPrimaryIndex(), IMetadataEntity.PENDING_DROP_OP)); } } } //#. mark PendingDropOp on the existing index MetadataManager.INSTANCE.dropIndex(mdTxnCtx, dataverseName, datasetName, indexName); MetadataManager.INSTANCE.addIndex(mdTxnCtx, new Index(dataverseName, datasetName, indexName, index.getIndexType(), index.getKeyFieldNames(), index.getKeyFieldTypes(), index.isEnforcingKeyFileds(), index.isPrimaryIndex(), IMetadataEntity.PENDING_DROP_OP)); //#. commit the existing transaction before calling runJob. MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; progress = ProgressState.ADDED_PENDINGOP_RECORD_TO_METADATA; for (JobSpecification jobSpec : jobsToExecute) { runJob(hcc, jobSpec, true); } //#. begin a new transaction mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); //#. finally, delete the existing index MetadataManager.INSTANCE.dropIndex(mdTxnCtx, dataverseName, datasetName, indexName); if (dropFilesIndex) { // delete the files index too MetadataManager.INSTANCE.dropIndex(mdTxnCtx, dataverseName, datasetName, ExternalIndexingOperations.getFilesIndexName(datasetName)); MetadataManager.INSTANCE.dropDatasetExternalFiles(mdTxnCtx, ds); ExternalDatasetsRegistry.INSTANCE.removeDatasetInfo(ds); } } MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e) { if (bActiveTxn) { abort(e, e, mdTxnCtx); } if (progress == ProgressState.ADDED_PENDINGOP_RECORD_TO_METADATA) { //#. execute compensation operations // remove the all indexes in NC try { for (JobSpecification jobSpec : jobsToExecute) { runJob(hcc, jobSpec, true); } } catch (Exception e2) { //do no throw exception since still the metadata needs to be compensated. e.addSuppressed(e2); } // remove the record from the metadata. mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); try { MetadataManager.INSTANCE.dropIndex(metadataProvider.getMetadataTxnContext(), dataverseName, datasetName, indexName); if (dropFilesIndex) { MetadataManager.INSTANCE.dropIndex(metadataProvider.getMetadataTxnContext(), dataverseName, datasetName, ExternalIndexingOperations.getFilesIndexName(datasetName)); } MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e2) { e.addSuppressed(e2); abort(e, e2, mdTxnCtx); throw new IllegalStateException("System is inconsistent state: pending index(" + dataverseName + "." + datasetName + "." + indexName + ") couldn't be removed from the metadata", e); } } throw e; } finally { MetadataLockManager.INSTANCE.dropIndexEnd(dataverseName, dataverseName + "." + datasetName); } } private void handleTypeDropStatement(AqlMetadataProvider metadataProvider, Statement stmt) throws Exception { TypeDropStatement stmtTypeDrop = (TypeDropStatement) stmt; String dataverseName = getActiveDataverse(stmtTypeDrop.getDataverseName()); String typeName = stmtTypeDrop.getTypeName().getValue(); MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); MetadataLockManager.INSTANCE.dropTypeBegin(dataverseName, dataverseName + "." + typeName); try { Datatype dt = MetadataManager.INSTANCE.getDatatype(mdTxnCtx, dataverseName, typeName); if (dt == null) { if (!stmtTypeDrop.getIfExists()) throw new AlgebricksException("There is no datatype with this name " + typeName + "."); } else { MetadataManager.INSTANCE.dropDatatype(mdTxnCtx, dataverseName, typeName); } MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e) { abort(e, e, mdTxnCtx); throw e; } finally { MetadataLockManager.INSTANCE.dropTypeEnd(dataverseName, dataverseName + "." + typeName); } } private void handleNodegroupDropStatement(AqlMetadataProvider metadataProvider, Statement stmt) throws Exception { NodeGroupDropStatement stmtDelete = (NodeGroupDropStatement) stmt; String nodegroupName = stmtDelete.getNodeGroupName().getValue(); MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); MetadataLockManager.INSTANCE.acquireNodeGroupWriteLock(nodegroupName); try { NodeGroup ng = MetadataManager.INSTANCE.getNodegroup(mdTxnCtx, nodegroupName); if (ng == null) { if (!stmtDelete.getIfExists()) throw new AlgebricksException("There is no nodegroup with this name " + nodegroupName + "."); } else { MetadataManager.INSTANCE.dropNodegroup(mdTxnCtx, nodegroupName); } MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e) { abort(e, e, mdTxnCtx); throw e; } finally { MetadataLockManager.INSTANCE.releaseNodeGroupWriteLock(nodegroupName); } } private void handleCreateFunctionStatement(AqlMetadataProvider metadataProvider, Statement stmt) throws Exception { CreateFunctionStatement cfs = (CreateFunctionStatement) stmt; String dataverse = getActiveDataverseName(cfs.getSignature().getNamespace()); String functionName = cfs.getaAterixFunction().getName(); MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); MetadataLockManager.INSTANCE.functionStatementBegin(dataverse, dataverse + "." + functionName); try { Dataverse dv = MetadataManager.INSTANCE.getDataverse(mdTxnCtx, dataverse); if (dv == null) { throw new AlgebricksException("There is no dataverse with this name " + dataverse + "."); } Function function = new Function(dataverse, functionName, cfs.getaAterixFunction().getArity(), cfs.getParamList(), Function.RETURNTYPE_VOID, cfs.getFunctionBody(), Function.LANGUAGE_AQL, FunctionKind.SCALAR.toString()); MetadataManager.INSTANCE.addFunction(mdTxnCtx, function); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e) { abort(e, e, mdTxnCtx); throw e; } finally { MetadataLockManager.INSTANCE.functionStatementEnd(dataverse, dataverse + "." + functionName); } } private void handleFunctionDropStatement(AqlMetadataProvider metadataProvider, Statement stmt) throws Exception { FunctionDropStatement stmtDropFunction = (FunctionDropStatement) stmt; FunctionSignature signature = stmtDropFunction.getFunctionSignature(); MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); MetadataLockManager.INSTANCE.functionStatementBegin(signature.getNamespace(), signature.getNamespace() + "." + signature.getName()); try { Function function = MetadataManager.INSTANCE.getFunction(mdTxnCtx, signature); if (function == null) { if (!stmtDropFunction.getIfExists()) throw new AlgebricksException("Unknonw function " + signature); } else { MetadataManager.INSTANCE.dropFunction(mdTxnCtx, signature); } MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e) { abort(e, e, mdTxnCtx); throw e; } finally { MetadataLockManager.INSTANCE.functionStatementEnd(signature.getNamespace(), signature.getNamespace() + "." + signature.getName()); } } private void handleLoadStatement(AqlMetadataProvider metadataProvider, Statement stmt, IHyracksClientConnection hcc) throws Exception { LoadStatement loadStmt = (LoadStatement) stmt; String dataverseName = getActiveDataverse(loadStmt.getDataverseName()); String datasetName = loadStmt.getDatasetName().getValue(); MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); boolean bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); MetadataLockManager.INSTANCE.modifyDatasetBegin(dataverseName, dataverseName + "." + datasetName); try { CompiledLoadFromFileStatement cls = new CompiledLoadFromFileStatement(dataverseName, loadStmt.getDatasetName().getValue(), loadStmt.getAdapter(), loadStmt.getProperties(), loadStmt.dataIsAlreadySorted()); JobSpecification spec = APIFramework.compileQuery(null, metadataProvider, null, 0, null, sessionConfig, cls); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; if (spec != null) { runJob(hcc, spec, true); } } catch (Exception e) { if (bActiveTxn) { abort(e, e, mdTxnCtx); } throw e; } finally { MetadataLockManager.INSTANCE.modifyDatasetEnd(dataverseName, dataverseName + "." + datasetName); } } private void handleInsertStatement(AqlMetadataProvider metadataProvider, Statement stmt, IHyracksClientConnection hcc) throws Exception { InsertStatement stmtInsert = (InsertStatement) stmt; String dataverseName = getActiveDataverse(stmtInsert.getDataverseName()); Query query = stmtInsert.getQuery(); MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); boolean bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); MetadataLockManager.INSTANCE.insertDeleteBegin(dataverseName, dataverseName + "." + stmtInsert.getDatasetName(), query.getDataverses(), query.getDatasets()); try { metadataProvider.setWriteTransaction(true); CompiledInsertStatement clfrqs = new CompiledInsertStatement(dataverseName, stmtInsert.getDatasetName().getValue(), query, stmtInsert.getVarCounter()); JobSpecification compiled = rewriteCompileQuery(metadataProvider, query, clfrqs); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; if (compiled != null) { runJob(hcc, compiled, true); } } catch (Exception e) { if (bActiveTxn) { abort(e, e, mdTxnCtx); } throw e; } finally { MetadataLockManager.INSTANCE.insertDeleteEnd(dataverseName, dataverseName + "." + stmtInsert.getDatasetName(), query.getDataverses(), query.getDatasets()); } } private void handleDeleteStatement(AqlMetadataProvider metadataProvider, Statement stmt, IHyracksClientConnection hcc) throws Exception { DeleteStatement stmtDelete = (DeleteStatement) stmt; String dataverseName = getActiveDataverse(stmtDelete.getDataverseName()); MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); boolean bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); MetadataLockManager.INSTANCE.insertDeleteBegin(dataverseName, dataverseName + "." + stmtDelete.getDatasetName(), stmtDelete.getDataverses(), stmtDelete.getDatasets()); try { metadataProvider.setWriteTransaction(true); CompiledDeleteStatement clfrqs = new CompiledDeleteStatement(stmtDelete.getVariableExpr(), dataverseName, stmtDelete.getDatasetName().getValue(), stmtDelete.getCondition(), stmtDelete.getVarCounter(), metadataProvider); JobSpecification compiled = rewriteCompileQuery(metadataProvider, clfrqs.getQuery(), clfrqs); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; if (compiled != null) { runJob(hcc, compiled, true); } } catch (Exception e) { if (bActiveTxn) { abort(e, e, mdTxnCtx); } throw e; } finally { MetadataLockManager.INSTANCE.insertDeleteEnd(dataverseName, dataverseName + "." + stmtDelete.getDatasetName(), stmtDelete.getDataverses(), stmtDelete.getDatasets()); } } private JobSpecification rewriteCompileQuery(AqlMetadataProvider metadataProvider, Query query, ICompiledDmlStatement stmt) throws AsterixException, RemoteException, AlgebricksException, JSONException, ACIDException { // Query Rewriting (happens under the same ongoing metadata transaction) Pair<Query, Integer> reWrittenQuery = APIFramework.reWriteQuery(declaredFunctions, metadataProvider, query, sessionConfig); // Query Compilation (happens under the same ongoing metadata transaction) JobSpecification spec = APIFramework.compileQuery(declaredFunctions, metadataProvider, reWrittenQuery.first, reWrittenQuery.second, stmt == null ? null : stmt.getDatasetName(), sessionConfig, stmt); return spec; } private void handleCreateFeedStatement(AqlMetadataProvider metadataProvider, Statement stmt, IHyracksClientConnection hcc) throws Exception { CreateFeedStatement cfs = (CreateFeedStatement) stmt; String dataverseName = getActiveDataverse(cfs.getDataverseName()); String feedName = cfs.getFeedName().getValue(); MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); MetadataLockManager.INSTANCE.createFeedBegin(dataverseName, dataverseName + "." + feedName); Feed feed = null; try { feed = MetadataManager.INSTANCE.getFeed(metadataProvider.getMetadataTxnContext(), dataverseName, feedName); if (feed != null) { if (cfs.getIfNotExists()) { MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); return; } else { throw new AlgebricksException("A feed with this name " + feedName + " already exists."); } } switch (stmt.getKind()) { case CREATE_PRIMARY_FEED: CreatePrimaryFeedStatement cpfs = (CreatePrimaryFeedStatement) stmt; String adaptorName = cpfs.getAdaptorName(); feed = new PrimaryFeed(dataverseName, feedName, adaptorName, cpfs.getAdaptorConfiguration(), cfs.getAppliedFunction()); break; case CREATE_SECONDARY_FEED: CreateSecondaryFeedStatement csfs = (CreateSecondaryFeedStatement) stmt; feed = new SecondaryFeed(dataverseName, feedName, csfs.getSourceFeedName(), csfs.getAppliedFunction()); break; default: throw new IllegalStateException(); } MetadataManager.INSTANCE.addFeed(metadataProvider.getMetadataTxnContext(), feed); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e) { abort(e, e, mdTxnCtx); throw e; } finally { MetadataLockManager.INSTANCE.createFeedEnd(dataverseName, dataverseName + "." + feedName); } } private void handleCreateFeedPolicyStatement(AqlMetadataProvider metadataProvider, Statement stmt, IHyracksClientConnection hcc) throws Exception { MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); String dataverse; String policy; FeedPolicy newPolicy = null; CreateFeedPolicyStatement cfps = (CreateFeedPolicyStatement) stmt; dataverse = getActiveDataverse(null); policy = cfps.getPolicyName(); MetadataLockManager.INSTANCE.createFeedPolicyBegin(dataverse, dataverse + "." + policy); try { FeedPolicy feedPolicy = MetadataManager.INSTANCE.getFeedPolicy(metadataProvider.getMetadataTxnContext(), dataverse, policy); if (feedPolicy != null) { if (cfps.getIfNotExists()) { MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); return; } else { throw new AlgebricksException("A policy with this name " + policy + " already exists."); } } boolean extendingExisting = cfps.getSourcePolicyName() != null; String description = cfps.getDescription() == null ? "" : cfps.getDescription(); if (extendingExisting) { FeedPolicy sourceFeedPolicy = MetadataManager.INSTANCE.getFeedPolicy( metadataProvider.getMetadataTxnContext(), dataverse, cfps.getSourcePolicyName()); if (sourceFeedPolicy == null) { sourceFeedPolicy = MetadataManager.INSTANCE.getFeedPolicy( metadataProvider.getMetadataTxnContext(), MetadataConstants.METADATA_DATAVERSE_NAME, cfps.getSourcePolicyName()); if (sourceFeedPolicy == null) { throw new AlgebricksException("Unknown policy " + cfps.getSourcePolicyName()); } } Map<String, String> policyProperties = sourceFeedPolicy.getProperties(); policyProperties.putAll(cfps.getProperties()); newPolicy = new FeedPolicy(dataverse, policy, description, policyProperties); } else { Properties prop = new Properties(); try { InputStream stream = new FileInputStream(cfps.getSourcePolicyFile()); prop.load(stream); } catch (Exception e) { throw new AlgebricksException("Unable to read policy file" + cfps.getSourcePolicyFile()); } Map<String, String> policyProperties = new HashMap<String, String>(); for (Entry<Object, Object> entry : prop.entrySet()) { policyProperties.put((String) entry.getKey(), (String) entry.getValue()); } newPolicy = new FeedPolicy(dataverse, policy, description, policyProperties); } MetadataManager.INSTANCE.addFeedPolicy(mdTxnCtx, newPolicy); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e) { abort(e, e, mdTxnCtx); throw e; } finally { MetadataLockManager.INSTANCE.createFeedPolicyEnd(dataverse, dataverse + "." + policy); } } private void handleDropFeedStatement(AqlMetadataProvider metadataProvider, Statement stmt, IHyracksClientConnection hcc) throws Exception { FeedDropStatement stmtFeedDrop = (FeedDropStatement) stmt; String dataverseName = getActiveDataverse(stmtFeedDrop.getDataverseName()); String feedName = stmtFeedDrop.getFeedName().getValue(); MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); MetadataLockManager.INSTANCE.dropFeedBegin(dataverseName, dataverseName + "." + feedName); try { Feed feed = MetadataManager.INSTANCE.getFeed(mdTxnCtx, dataverseName, feedName); if (feed == null) { if (!stmtFeedDrop.getIfExists()) { throw new AlgebricksException("There is no feed with this name " + feedName + "."); } } FeedId feedId = new FeedId(dataverseName, feedName); List<FeedConnectionId> activeConnections = FeedLifecycleListener.INSTANCE .getActiveFeedConnections(feedId); if (activeConnections != null && !activeConnections.isEmpty()) { StringBuilder builder = new StringBuilder(); for (FeedConnectionId connectionId : activeConnections) { builder.append(connectionId.getDatasetName() + "\n"); } throw new AlgebricksException("Feed " + feedId + " is currently active and connected to the following dataset(s) \n" + builder.toString()); } else { MetadataManager.INSTANCE.dropFeed(mdTxnCtx, dataverseName, feedName); } if (LOGGER.isLoggable(Level.INFO)) { LOGGER.info("Removed feed " + feedId); } MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e) { abort(e, e, mdTxnCtx); throw e; } finally { MetadataLockManager.INSTANCE.dropFeedEnd(dataverseName, dataverseName + "." + feedName); } } private void handleDropFeedPolicyStatement(AqlMetadataProvider metadataProvider, Statement stmt, IHyracksClientConnection hcc) throws Exception { MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); FeedPolicyDropStatement stmtFeedPolicyDrop = (FeedPolicyDropStatement) stmt; String dataverseName = getActiveDataverse(stmtFeedPolicyDrop.getDataverseName()); String policyName = stmtFeedPolicyDrop.getPolicyName().getValue(); MetadataLockManager.INSTANCE.dropFeedPolicyBegin(dataverseName, dataverseName + "." + policyName); try { FeedPolicy feedPolicy = MetadataManager.INSTANCE.getFeedPolicy(mdTxnCtx, dataverseName, policyName); if (feedPolicy == null) { if (!stmtFeedPolicyDrop.getIfExists()) { throw new AlgebricksException( "Unknown policy " + policyName + " in dataverse " + dataverseName); } } MetadataManager.INSTANCE.dropFeedPolicy(mdTxnCtx, dataverseName, policyName); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e) { abort(e, e, mdTxnCtx); throw e; } finally { MetadataLockManager.INSTANCE.dropFeedPolicyEnd(dataverseName, dataverseName + "." + policyName); } } private void handleConnectFeedStatement(AqlMetadataProvider metadataProvider, Statement stmt, IHyracksClientConnection hcc) throws Exception { ConnectFeedStatement cfs = (ConnectFeedStatement) stmt; String dataverseName = getActiveDataverse(cfs.getDataverseName()); String feedName = cfs.getFeedName(); String datasetName = cfs.getDatasetName().getValue(); boolean bActiveTxn = true; MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); boolean readLatchAcquired = true; boolean subscriberRegistered = false; IFeedLifecycleEventSubscriber eventSubscriber = new FeedLifecycleEventSubscriber(); FeedConnectionId feedConnId = null; MetadataLockManager.INSTANCE.connectFeedBegin(dataverseName, dataverseName + "." + datasetName, dataverseName + "." + feedName); try { metadataProvider.setWriteTransaction(true); CompiledConnectFeedStatement cbfs = new CompiledConnectFeedStatement(dataverseName, cfs.getFeedName(), cfs.getDatasetName().getValue(), cfs.getPolicy(), cfs.getQuery(), cfs.getVarCounter()); FeedUtil.validateIfDatasetExists(dataverseName, cfs.getDatasetName().getValue(), metadataProvider.getMetadataTxnContext()); Feed feed = FeedUtil.validateIfFeedExists(dataverseName, cfs.getFeedName(), metadataProvider.getMetadataTxnContext()); feedConnId = new FeedConnectionId(dataverseName, cfs.getFeedName(), cfs.getDatasetName().getValue()); if (FeedLifecycleListener.INSTANCE.isFeedConnectionActive(feedConnId)) { throw new AsterixException("Feed " + cfs.getFeedName() + " is already connected to dataset " + cfs.getDatasetName().getValue()); } FeedPolicy feedPolicy = FeedUtil.validateIfPolicyExists(dataverseName, cbfs.getPolicyName(), mdTxnCtx); // All Metadata checks have passed. Feed connect request is valid. // FeedPolicyAccessor policyAccessor = new FeedPolicyAccessor(feedPolicy.getProperties()); Triple<FeedConnectionRequest, Boolean, List<IFeedJoint>> triple = getFeedConnectionRequest( dataverseName, feed, cbfs.getDatasetName(), feedPolicy, mdTxnCtx); FeedConnectionRequest connectionRequest = triple.first; boolean createFeedIntakeJob = triple.second; FeedLifecycleListener.INSTANCE.registerFeedEventSubscriber(feedConnId, eventSubscriber); subscriberRegistered = true; if (createFeedIntakeJob) { FeedId feedId = connectionRequest.getFeedJointKey().getFeedId(); PrimaryFeed primaryFeed = (PrimaryFeed) MetadataManager.INSTANCE.getFeed(mdTxnCtx, feedId.getDataverse(), feedId.getFeedName()); Pair<JobSpecification, IFeedAdapterFactory> pair = FeedOperations .buildFeedIntakeJobSpec(primaryFeed, metadataProvider, policyAccessor); // adapter configuration are valid at this stage // register the feed joints (these are auto-de-registered) for (IFeedJoint fj : triple.third) { FeedLifecycleListener.INSTANCE.registerFeedJoint(fj); } runJob(hcc, pair.first, false); IFeedAdapterFactory adapterFactory = pair.second; if (adapterFactory.isRecordTrackingEnabled()) { FeedLifecycleListener.INSTANCE.registerFeedIntakeProgressTracker(feedConnId, adapterFactory.createIntakeProgressTracker()); } eventSubscriber.assertEvent(FeedLifecycleEvent.FEED_INTAKE_STARTED); } else { for (IFeedJoint fj : triple.third) { FeedLifecycleListener.INSTANCE.registerFeedJoint(fj); } } MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; readLatchAcquired = false; eventSubscriber.assertEvent(FeedLifecycleEvent.FEED_COLLECT_STARTED); if (Boolean.valueOf(metadataProvider.getConfig().get(ConnectFeedStatement.WAIT_FOR_COMPLETION))) { eventSubscriber.assertEvent(FeedLifecycleEvent.FEED_ENDED); // blocking call } String waitForCompletionParam = metadataProvider.getConfig() .get(ConnectFeedStatement.WAIT_FOR_COMPLETION); boolean waitForCompletion = waitForCompletionParam == null ? false : Boolean.valueOf(waitForCompletionParam); if (waitForCompletion) { MetadataLockManager.INSTANCE.connectFeedEnd(dataverseName, dataverseName + "." + datasetName, dataverseName + "." + feedName); readLatchAcquired = false; } } catch (Exception e) { if (bActiveTxn) { abort(e, e, mdTxnCtx); } throw e; } finally { if (readLatchAcquired) { MetadataLockManager.INSTANCE.connectFeedEnd(dataverseName, dataverseName + "." + datasetName, dataverseName + "." + feedName); } if (subscriberRegistered) { FeedLifecycleListener.INSTANCE.deregisterFeedEventSubscriber(feedConnId, eventSubscriber); } } } /** * Generates a subscription request corresponding to a connect feed request. In addition, provides a boolean * flag indicating if feed intake job needs to be started (source primary feed not found to be active). * * @param dataverse * @param feed * @param dataset * @param feedPolicy * @param mdTxnCtx * @return * @throws MetadataException */ private Triple<FeedConnectionRequest, Boolean, List<IFeedJoint>> getFeedConnectionRequest(String dataverse, Feed feed, String dataset, FeedPolicy feedPolicy, MetadataTransactionContext mdTxnCtx) throws MetadataException { IFeedJoint sourceFeedJoint = null; FeedConnectionRequest request = null; List<String> functionsToApply = new ArrayList<String>(); boolean needIntakeJob = false; List<IFeedJoint> jointsToRegister = new ArrayList<IFeedJoint>(); FeedConnectionId connectionId = new FeedConnectionId(feed.getFeedId(), dataset); ConnectionLocation connectionLocation = null; FeedJointKey feedJointKey = getFeedJointKey(feed, mdTxnCtx); boolean isFeedJointAvailable = FeedLifecycleListener.INSTANCE.isFeedJointAvailable(feedJointKey); if (!isFeedJointAvailable) { sourceFeedJoint = FeedLifecycleListener.INSTANCE.getAvailableFeedJoint(feedJointKey); if (sourceFeedJoint == null) { // the feed is currently not being ingested, i.e., it is unavailable. connectionLocation = ConnectionLocation.SOURCE_FEED_INTAKE_STAGE; FeedId sourceFeedId = feedJointKey.getFeedId(); // the root/primary feedId Feed primaryFeed = MetadataManager.INSTANCE.getFeed(mdTxnCtx, dataverse, sourceFeedId.getFeedName()); FeedJointKey intakeFeedJointKey = new FeedJointKey(sourceFeedId, new ArrayList<String>()); sourceFeedJoint = new FeedJoint(intakeFeedJointKey, primaryFeed.getFeedId(), connectionLocation, FeedJointType.INTAKE, connectionId); jointsToRegister.add(sourceFeedJoint); needIntakeJob = true; } else { connectionLocation = sourceFeedJoint.getConnectionLocation(); } String[] functions = feedJointKey.getStringRep() .substring(sourceFeedJoint.getFeedJointKey().getStringRep().length()).trim().split(":"); for (String f : functions) { if (f.trim().length() > 0) { functionsToApply.add(f); } } // register the compute feed point that represents the final output from the collection of // functions that will be applied. if (!functionsToApply.isEmpty()) { FeedJointKey computeFeedJointKey = new FeedJointKey(feed.getFeedId(), functionsToApply); IFeedJoint computeFeedJoint = new FeedJoint(computeFeedJointKey, feed.getFeedId(), ConnectionLocation.SOURCE_FEED_COMPUTE_STAGE, FeedJointType.COMPUTE, connectionId); jointsToRegister.add(computeFeedJoint); } } else { sourceFeedJoint = FeedLifecycleListener.INSTANCE.getFeedJoint(feedJointKey); connectionLocation = sourceFeedJoint.getConnectionLocation(); if (LOGGER.isLoggable(Level.INFO)) { LOGGER.info( "Feed joint " + sourceFeedJoint + " is available! need not apply any further computation"); } } request = new FeedConnectionRequest(sourceFeedJoint.getFeedJointKey(), connectionLocation, functionsToApply, dataset, feedPolicy.getPolicyName(), feedPolicy.getProperties(), feed.getFeedId()); sourceFeedJoint.addConnectionRequest(request); return new Triple<FeedConnectionRequest, Boolean, List<IFeedJoint>>(request, needIntakeJob, jointsToRegister); } /* * Gets the feed joint corresponding to the feed definition. Tuples constituting the feed are * available at this feed joint. */ private FeedJointKey getFeedJointKey(Feed feed, MetadataTransactionContext ctx) throws MetadataException { Feed sourceFeed = feed; List<String> appliedFunctions = new ArrayList<String>(); while (sourceFeed.getFeedType().equals(FeedType.SECONDARY)) { if (sourceFeed.getAppliedFunction() != null) { appliedFunctions.add(0, sourceFeed.getAppliedFunction().getName()); } Feed parentFeed = MetadataManager.INSTANCE.getFeed(ctx, feed.getDataverseName(), ((SecondaryFeed) sourceFeed).getSourceFeedName()); sourceFeed = parentFeed; } if (sourceFeed.getAppliedFunction() != null) { appliedFunctions.add(0, sourceFeed.getAppliedFunction().getName()); } return new FeedJointKey(sourceFeed.getFeedId(), appliedFunctions); } private void handleDisconnectFeedStatement(AqlMetadataProvider metadataProvider, Statement stmt, IHyracksClientConnection hcc) throws Exception { DisconnectFeedStatement cfs = (DisconnectFeedStatement) stmt; String dataverseName = getActiveDataverse(cfs.getDataverseName()); String datasetName = cfs.getDatasetName().getValue(); MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); boolean bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); FeedUtil.validateIfDatasetExists(dataverseName, cfs.getDatasetName().getValue(), mdTxnCtx); Feed feed = FeedUtil.validateIfFeedExists(dataverseName, cfs.getFeedName().getValue(), mdTxnCtx); FeedConnectionId connectionId = new FeedConnectionId(feed.getFeedId(), cfs.getDatasetName().getValue()); boolean isFeedConnectionActive = FeedLifecycleListener.INSTANCE.isFeedConnectionActive(connectionId); if (!isFeedConnectionActive) { throw new AsterixException("Feed " + feed.getFeedId().getFeedName() + " is currently not connected to " + cfs.getDatasetName().getValue() + ". Invalid operation!"); } MetadataLockManager.INSTANCE.disconnectFeedBegin(dataverseName, dataverseName + "." + datasetName, dataverseName + "." + cfs.getFeedName()); try { Dataset dataset = MetadataManager.INSTANCE.getDataset(metadataProvider.getMetadataTxnContext(), dataverseName, cfs.getDatasetName().getValue()); if (dataset == null) { throw new AsterixException( "Unknown dataset :" + cfs.getDatasetName().getValue() + " in dataverse " + dataverseName); } Pair<JobSpecification, Boolean> specDisconnectType = FeedOperations .buildDisconnectFeedJobSpec(metadataProvider, connectionId); JobSpecification jobSpec = specDisconnectType.first; MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; runJob(hcc, jobSpec, true); if (!specDisconnectType.second) { CentralFeedManager.getInstance().getFeedLoadManager().removeFeedActivity(connectionId); FeedLifecycleListener.INSTANCE.reportPartialDisconnection(connectionId); } } catch (Exception e) { if (bActiveTxn) { abort(e, e, mdTxnCtx); } throw e; } finally { MetadataLockManager.INSTANCE.disconnectFeedEnd(dataverseName, dataverseName + "." + datasetName, dataverseName + "." + cfs.getFeedName()); } } private void handleSubscribeFeedStatement(AqlMetadataProvider metadataProvider, Statement stmt, IHyracksClientConnection hcc) throws Exception { if (LOGGER.isLoggable(Level.INFO)) { LOGGER.info("Subscriber Feed Statement :" + stmt); } MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); boolean bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); metadataProvider.setWriteTransaction(true); SubscribeFeedStatement bfs = (SubscribeFeedStatement) stmt; bfs.initialize(metadataProvider.getMetadataTxnContext()); CompiledSubscribeFeedStatement csfs = new CompiledSubscribeFeedStatement(bfs.getSubscriptionRequest(), bfs.getQuery(), bfs.getVarCounter()); metadataProvider.getConfig().put(FunctionUtils.IMPORT_PRIVATE_FUNCTIONS, "" + Boolean.TRUE); metadataProvider.getConfig().put(FeedActivityDetails.FEED_POLICY_NAME, "" + bfs.getPolicy()); metadataProvider.getConfig().put(FeedActivityDetails.COLLECT_LOCATIONS, StringUtils.join(bfs.getLocations(), ',')); JobSpecification compiled = rewriteCompileQuery(metadataProvider, bfs.getQuery(), csfs); FeedConnectionId feedConnectionId = new FeedConnectionId(bfs.getSubscriptionRequest().getReceivingFeedId(), bfs.getSubscriptionRequest().getTargetDataset()); String dataverse = feedConnectionId.getFeedId().getDataverse(); String dataset = feedConnectionId.getDatasetName(); MetadataLockManager.INSTANCE.subscribeFeedBegin(dataverse, dataverse + "." + dataset, dataverse + "." + feedConnectionId.getFeedId().getFeedName()); try { JobSpecification alteredJobSpec = FeedUtil.alterJobSpecificationForFeed(compiled, feedConnectionId, bfs.getSubscriptionRequest().getPolicyParameters()); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; if (compiled != null) { runJob(hcc, alteredJobSpec, false); } } catch (Exception e) { e.printStackTrace(); if (bActiveTxn) { abort(e, e, mdTxnCtx); } throw e; } finally { MetadataLockManager.INSTANCE.subscribeFeedEnd(dataverse, dataverse + "." + dataset, dataverse + "." + feedConnectionId.getFeedId().getFeedName()); } } private void handleCompactStatement(AqlMetadataProvider metadataProvider, Statement stmt, IHyracksClientConnection hcc) throws Exception { CompactStatement compactStatement = (CompactStatement) stmt; String dataverseName = getActiveDataverse(compactStatement.getDataverseName()); String datasetName = compactStatement.getDatasetName().getValue(); MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); boolean bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); MetadataLockManager.INSTANCE.compactBegin(dataverseName, dataverseName + "." + datasetName); List<JobSpecification> jobsToExecute = new ArrayList<JobSpecification>(); try { Dataset ds = MetadataManager.INSTANCE.getDataset(mdTxnCtx, dataverseName, datasetName); if (ds == null) { throw new AlgebricksException("There is no dataset with this name " + datasetName + " in dataverse " + dataverseName + "."); } String itemTypeName = ds.getItemTypeName(); Datatype dt = MetadataManager.INSTANCE.getDatatype(metadataProvider.getMetadataTxnContext(), dataverseName, itemTypeName); // Prepare jobs to compact the datatset and its indexes List<Index> indexes = MetadataManager.INSTANCE.getDatasetIndexes(mdTxnCtx, dataverseName, datasetName); if (indexes.size() == 0) { throw new AlgebricksException( "Cannot compact the extrenal dataset " + datasetName + " because it has no indexes"); } Dataverse dataverse = MetadataManager.INSTANCE.getDataverse(metadataProvider.getMetadataTxnContext(), dataverseName); jobsToExecute.add(DatasetOperations.compactDatasetJobSpec(dataverse, datasetName, metadataProvider)); ARecordType aRecordType = (ARecordType) dt.getDatatype(); ARecordType enforcedType = IntroduceSecondaryIndexInsertDeleteRule.createEnforcedType(aRecordType, indexes); for (int j = 0; j < indexes.size(); j++) { if (ds.getDatasetType() == DatasetType.INTERNAL && indexes.get(j).isSecondaryIndex() || ds.getDatasetType() == DatasetType.EXTERNAL && !ExternalIndexingOperations.isFileIndex(indexes.get(j))) { CompiledIndexCompactStatement cics = new CompiledIndexCompactStatement(dataverseName, datasetName, indexes.get(j).getIndexName(), indexes.get(j).getKeyFieldNames(), indexes.get(j).getKeyFieldTypes(), indexes.get(j).isEnforcingKeyFileds(), indexes.get(j).getGramLength(), indexes.get(j).getIndexType()); jobsToExecute.add(IndexOperations.buildSecondaryIndexCompactJobSpec(cics, aRecordType, enforcedType, metadataProvider, ds)); } } MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; //#. run the jobs for (JobSpecification jobSpec : jobsToExecute) { runJob(hcc, jobSpec, true); } } catch (Exception e) { if (bActiveTxn) { abort(e, e, mdTxnCtx); } throw e; } finally { MetadataLockManager.INSTANCE.compactEnd(dataverseName, dataverseName + "." + datasetName); } } private void handleQuery(AqlMetadataProvider metadataProvider, Query query, IHyracksClientConnection hcc, IHyracksDataset hdc, ResultDelivery resultDelivery) throws Exception { MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); boolean bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); MetadataLockManager.INSTANCE.queryBegin(activeDefaultDataverse, query.getDataverses(), query.getDatasets()); JobSpecification compiled = null; try { compiled = rewriteCompileQuery(metadataProvider, query, null); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; if (sessionConfig.isExecuteQuery() && compiled != null) { GlobalConfig.ASTERIX_LOGGER.info(compiled.toJSON().toString(1)); JobId jobId = runJob(hcc, compiled, false); JSONObject response = new JSONObject(); switch (resultDelivery) { case ASYNC: JSONArray handle = new JSONArray(); handle.put(jobId.getId()); handle.put(metadataProvider.getResultSetId().getId()); response.put("handle", handle); sessionConfig.out().print(response); sessionConfig.out().flush(); hcc.waitForCompletion(jobId); break; case SYNC: ResultReader resultReader = new ResultReader(hcc, hdc); resultReader.open(jobId, metadataProvider.getResultSetId()); // In this case (the normal case), we don't use the // "response" JSONObject - just stream the results // to the "out" PrintWriter if (sessionConfig.fmt() == OutputFormat.CSV && sessionConfig.is(SessionConfig.FORMAT_CSV_HEADER)) { ResultUtils.displayCSVHeader(metadataProvider.findOutputRecordType(), sessionConfig); } ResultUtils.displayResults(resultReader, sessionConfig); hcc.waitForCompletion(jobId); break; case ASYNC_DEFERRED: handle = new JSONArray(); handle.put(jobId.getId()); handle.put(metadataProvider.getResultSetId().getId()); response.put("handle", handle); hcc.waitForCompletion(jobId); sessionConfig.out().print(response); sessionConfig.out().flush(); break; default: break; } } } catch (Exception e) { e.printStackTrace(); if (bActiveTxn) { abort(e, e, mdTxnCtx); } throw e; } finally { MetadataLockManager.INSTANCE.queryEnd(query.getDataverses(), query.getDatasets()); // release external datasets' locks acquired during compilation of the query ExternalDatasetsRegistry.INSTANCE.releaseAcquiredLocks(metadataProvider); } } private void handleCreateNodeGroupStatement(AqlMetadataProvider metadataProvider, Statement stmt) throws Exception { NodegroupDecl stmtCreateNodegroup = (NodegroupDecl) stmt; String ngName = stmtCreateNodegroup.getNodegroupName().getValue(); MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); MetadataLockManager.INSTANCE.acquireNodeGroupWriteLock(ngName); try { NodeGroup ng = MetadataManager.INSTANCE.getNodegroup(mdTxnCtx, ngName); if (ng != null) { if (!stmtCreateNodegroup.getIfNotExists()) throw new AlgebricksException("A nodegroup with this name " + ngName + " already exists."); } else { List<Identifier> ncIdentifiers = stmtCreateNodegroup.getNodeControllerNames(); List<String> ncNames = new ArrayList<String>(ncIdentifiers.size()); for (Identifier id : ncIdentifiers) { ncNames.add(id.getValue()); } MetadataManager.INSTANCE.addNodegroup(mdTxnCtx, new NodeGroup(ngName, ncNames)); } MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e) { abort(e, e, mdTxnCtx); throw e; } finally { MetadataLockManager.INSTANCE.releaseNodeGroupWriteLock(ngName); } } private void handleExternalDatasetRefreshStatement(AqlMetadataProvider metadataProvider, Statement stmt, IHyracksClientConnection hcc) throws Exception { RefreshExternalDatasetStatement stmtRefresh = (RefreshExternalDatasetStatement) stmt; String dataverseName = getActiveDataverse(stmtRefresh.getDataverseName()); String datasetName = stmtRefresh.getDatasetName().getValue(); ExternalDatasetTransactionState transactionState = ExternalDatasetTransactionState.COMMIT; MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); MetadataLockManager.INSTANCE.refreshDatasetBegin(dataverseName, dataverseName + "." + datasetName); boolean bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); JobSpecification spec = null; Dataset ds = null; List<ExternalFile> metadataFiles = null; List<ExternalFile> deletedFiles = null; List<ExternalFile> addedFiles = null; List<ExternalFile> appendedFiles = null; List<Index> indexes = null; Dataset transactionDataset = null; boolean lockAquired = false; boolean success = false; try { ds = MetadataManager.INSTANCE.getDataset(metadataProvider.getMetadataTxnContext(), dataverseName, datasetName); // Dataset exists ? if (ds == null) { throw new AlgebricksException( "There is no dataset with this name " + datasetName + " in dataverse " + dataverseName); } // Dataset external ? if (ds.getDatasetType() != DatasetType.EXTERNAL) { throw new AlgebricksException("dataset " + datasetName + " in dataverse " + dataverseName + " is not an external dataset"); } // Dataset has indexes ? indexes = MetadataManager.INSTANCE.getDatasetIndexes(mdTxnCtx, dataverseName, datasetName); if (indexes.size() == 0) { throw new AlgebricksException("External dataset " + datasetName + " in dataverse " + dataverseName + " doesn't have any index"); } // Record transaction time Date txnTime = new Date(); // refresh lock here ExternalDatasetsRegistry.INSTANCE.refreshBegin(ds); lockAquired = true; // Get internal files metadataFiles = MetadataManager.INSTANCE.getDatasetExternalFiles(mdTxnCtx, ds); deletedFiles = new ArrayList<ExternalFile>(); addedFiles = new ArrayList<ExternalFile>(); appendedFiles = new ArrayList<ExternalFile>(); // Compute delta // Now we compare snapshot with external file system if (ExternalIndexingOperations.isDatasetUptodate(ds, metadataFiles, addedFiles, deletedFiles, appendedFiles)) { ((ExternalDatasetDetails) ds.getDatasetDetails()).setRefreshTimestamp(txnTime); MetadataManager.INSTANCE.updateDataset(mdTxnCtx, ds); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); // latch will be released in the finally clause return; } // At this point, we know data has changed in the external file system, record transaction in metadata and start transactionDataset = ExternalIndexingOperations.createTransactionDataset(ds); /* * Remove old dataset record and replace it with a new one */ MetadataManager.INSTANCE.updateDataset(mdTxnCtx, transactionDataset); // Add delta files to the metadata for (ExternalFile file : addedFiles) { MetadataManager.INSTANCE.addExternalFile(mdTxnCtx, file); } for (ExternalFile file : appendedFiles) { MetadataManager.INSTANCE.addExternalFile(mdTxnCtx, file); } for (ExternalFile file : deletedFiles) { MetadataManager.INSTANCE.addExternalFile(mdTxnCtx, file); } // Create the files index update job spec = ExternalIndexingOperations.buildFilesIndexUpdateOp(ds, metadataFiles, deletedFiles, addedFiles, appendedFiles, metadataProvider); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; transactionState = ExternalDatasetTransactionState.BEGIN; //run the files update job runJob(hcc, spec, true); for (Index index : indexes) { if (!ExternalIndexingOperations.isFileIndex(index)) { spec = ExternalIndexingOperations.buildIndexUpdateOp(ds, index, metadataFiles, deletedFiles, addedFiles, appendedFiles, metadataProvider); //run the files update job runJob(hcc, spec, true); } } // all index updates has completed successfully, record transaction state spec = ExternalIndexingOperations.buildCommitJob(ds, indexes, metadataProvider); // Aquire write latch again -> start a transaction and record the decision to commit mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); bActiveTxn = true; ((ExternalDatasetDetails) transactionDataset.getDatasetDetails()) .setState(ExternalDatasetTransactionState.READY_TO_COMMIT); ((ExternalDatasetDetails) transactionDataset.getDatasetDetails()).setRefreshTimestamp(txnTime); MetadataManager.INSTANCE.updateDataset(mdTxnCtx, transactionDataset); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; transactionState = ExternalDatasetTransactionState.READY_TO_COMMIT; // We don't release the latch since this job is expected to be quick runJob(hcc, spec, true); // Start a new metadata transaction to record the final state of the transaction mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); bActiveTxn = true; for (ExternalFile file : metadataFiles) { if (file.getPendingOp() == ExternalFilePendingOp.PENDING_DROP_OP) { MetadataManager.INSTANCE.dropExternalFile(mdTxnCtx, file); } else if (file.getPendingOp() == ExternalFilePendingOp.PENDING_NO_OP) { Iterator<ExternalFile> iterator = appendedFiles.iterator(); while (iterator.hasNext()) { ExternalFile appendedFile = iterator.next(); if (file.getFileName().equals(appendedFile.getFileName())) { // delete existing file MetadataManager.INSTANCE.dropExternalFile(mdTxnCtx, file); // delete existing appended file MetadataManager.INSTANCE.dropExternalFile(mdTxnCtx, appendedFile); // add the original file with appended information appendedFile.setFileNumber(file.getFileNumber()); appendedFile.setPendingOp(ExternalFilePendingOp.PENDING_NO_OP); MetadataManager.INSTANCE.addExternalFile(mdTxnCtx, appendedFile); iterator.remove(); } } } } // remove the deleted files delta for (ExternalFile file : deletedFiles) { MetadataManager.INSTANCE.dropExternalFile(mdTxnCtx, file); } // insert new files for (ExternalFile file : addedFiles) { MetadataManager.INSTANCE.dropExternalFile(mdTxnCtx, file); file.setPendingOp(ExternalFilePendingOp.PENDING_NO_OP); MetadataManager.INSTANCE.addExternalFile(mdTxnCtx, file); } // mark the transaction as complete ((ExternalDatasetDetails) transactionDataset.getDatasetDetails()) .setState(ExternalDatasetTransactionState.COMMIT); MetadataManager.INSTANCE.updateDataset(mdTxnCtx, transactionDataset); // commit metadata transaction MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); success = true; } catch (Exception e) { if (bActiveTxn) { abort(e, e, mdTxnCtx); } if (transactionState == ExternalDatasetTransactionState.READY_TO_COMMIT) { throw new IllegalStateException("System is inconsistent state: commit of (" + dataverseName + "." + datasetName + ") refresh couldn't carry out the commit phase", e); } if (transactionState == ExternalDatasetTransactionState.COMMIT) { // Nothing to do , everything should be clean throw e; } if (transactionState == ExternalDatasetTransactionState.BEGIN) { // transaction failed, need to do the following // clean NCs removing transaction components mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); bActiveTxn = true; metadataProvider.setMetadataTxnContext(mdTxnCtx); spec = ExternalIndexingOperations.buildAbortOp(ds, indexes, metadataProvider); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; try { runJob(hcc, spec, true); } catch (Exception e2) { // This should never happen -- fix throw illegal e.addSuppressed(e2); throw new IllegalStateException("System is in inconsistent state. Failed to abort refresh", e); } // remove the delta of files // return the state of the dataset to committed try { mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); for (ExternalFile file : deletedFiles) { MetadataManager.INSTANCE.dropExternalFile(mdTxnCtx, file); } for (ExternalFile file : addedFiles) { MetadataManager.INSTANCE.dropExternalFile(mdTxnCtx, file); } for (ExternalFile file : appendedFiles) { MetadataManager.INSTANCE.dropExternalFile(mdTxnCtx, file); } MetadataManager.INSTANCE.updateDataset(mdTxnCtx, ds); // commit metadata transaction MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); } catch (Exception e2) { abort(e, e2, mdTxnCtx); e.addSuppressed(e2); throw new IllegalStateException("System is in inconsistent state. Failed to drop delta files", e); } } } finally { if (lockAquired) { ExternalDatasetsRegistry.INSTANCE.refreshEnd(ds, success); } MetadataLockManager.INSTANCE.refreshDatasetEnd(dataverseName, dataverseName + "." + datasetName); } } private void handleRunStatement(AqlMetadataProvider metadataProvider, Statement stmt, IHyracksClientConnection hcc) throws AsterixException, Exception { RunStatement runStmt = (RunStatement) stmt; switch (runStmt.getSystem()) { case "pregel": case "pregelix": handlePregelixStatement(metadataProvider, runStmt, hcc); break; default: throw new AlgebricksException( "The system \"" + runStmt.getSystem() + "\" specified in your run statement is not supported."); } } private void handlePregelixStatement(AqlMetadataProvider metadataProvider, Statement stmt, IHyracksClientConnection hcc) throws AsterixException, Exception { RunStatement pregelixStmt = (RunStatement) stmt; boolean bActiveTxn = true; String dataverseNameFrom = getActiveDataverse(pregelixStmt.getDataverseNameFrom()); String dataverseNameTo = getActiveDataverse(pregelixStmt.getDataverseNameTo()); String datasetNameFrom = pregelixStmt.getDatasetNameFrom().getValue(); String datasetNameTo = pregelixStmt.getDatasetNameTo().getValue(); if (dataverseNameFrom != dataverseNameTo) { throw new AlgebricksException("Pregelix statements across different dataverses are not supported."); } MetadataTransactionContext mdTxnCtx = MetadataManager.INSTANCE.beginTransaction(); metadataProvider.setMetadataTxnContext(mdTxnCtx); MetadataLockManager.INSTANCE.pregelixBegin(dataverseNameFrom, datasetNameFrom, datasetNameTo); try { // construct input paths Index fromIndex = null; List<Index> indexes = MetadataManager.INSTANCE.getDatasetIndexes(mdTxnCtx, dataverseNameFrom, pregelixStmt.getDatasetNameFrom().getValue()); for (Index ind : indexes) { if (ind.isPrimaryIndex()) fromIndex = ind; } if (fromIndex == null) { throw new AlgebricksException("Tried to access non-existing dataset: " + datasetNameFrom); } Dataset datasetFrom = MetadataManager.INSTANCE.getDataset(mdTxnCtx, dataverseNameFrom, datasetNameFrom); IFileSplitProvider fromSplits = metadataProvider.splitProviderAndPartitionConstraintsForDataset( dataverseNameFrom, datasetNameFrom, fromIndex.getIndexName(), datasetFrom.getDatasetDetails().isTemp()).first; StringBuilder fromSplitsPaths = new StringBuilder(); for (FileSplit f : fromSplits.getFileSplits()) { fromSplitsPaths .append("asterix://" + f.getNodeName() + f.getLocalFile().getFile().getAbsolutePath()); fromSplitsPaths.append(","); } fromSplitsPaths.setLength(fromSplitsPaths.length() - 1); // Construct output paths Index toIndex = null; indexes = MetadataManager.INSTANCE.getDatasetIndexes(mdTxnCtx, dataverseNameTo, pregelixStmt.getDatasetNameTo().getValue()); for (Index ind : indexes) { if (ind.isPrimaryIndex()) toIndex = ind; } if (toIndex == null) { throw new AlgebricksException("Tried to access non-existing dataset: " + datasetNameTo); } Dataset datasetTo = MetadataManager.INSTANCE.getDataset(mdTxnCtx, dataverseNameTo, datasetNameTo); IFileSplitProvider toSplits = metadataProvider.splitProviderAndPartitionConstraintsForDataset( dataverseNameTo, datasetNameTo, toIndex.getIndexName(), datasetTo.getDatasetDetails().isTemp()).first; StringBuilder toSplitsPaths = new StringBuilder(); for (FileSplit f : toSplits.getFileSplits()) { toSplitsPaths.append("asterix://" + f.getNodeName() + f.getLocalFile().getFile().getAbsolutePath()); toSplitsPaths.append(","); } toSplitsPaths.setLength(toSplitsPaths.length() - 1); try { Dataset toDataset = MetadataManager.INSTANCE.getDataset(mdTxnCtx, dataverseNameTo, datasetNameTo); DropStatement dropStmt = new DropStatement(new Identifier(dataverseNameTo), pregelixStmt.getDatasetNameTo(), true); this.handleDatasetDropStatement(metadataProvider, dropStmt, hcc); IDatasetDetailsDecl idd = new InternalDetailsDecl(toIndex.getKeyFieldNames(), false, null, toDataset.getDatasetDetails().isTemp()); DatasetDecl createToDataset = new DatasetDecl(new Identifier(dataverseNameTo), pregelixStmt.getDatasetNameTo(), new Identifier(toDataset.getItemTypeName()), new Identifier(toDataset.getNodeGroupName()), toDataset.getCompactionPolicy(), toDataset.getCompactionPolicyProperties(), toDataset.getHints(), toDataset.getDatasetType(), idd, false); this.handleCreateDatasetStatement(metadataProvider, createToDataset, hcc); } catch (Exception e) { e.printStackTrace(); throw new AlgebricksException("Error cleaning the result dataset. This should not happen."); } // Flush source dataset flushDataset(hcc, metadataProvider, mdTxnCtx, dataverseNameFrom, datasetNameFrom, fromIndex.getIndexName()); // call Pregelix String pregelix_home = System.getenv("PREGELIX_HOME"); if (pregelix_home == null) { throw new AlgebricksException("PREGELIX_HOME is not defined!"); } // construct command ArrayList<String> cmd = new ArrayList<String>(); cmd.add("bin/pregelix"); cmd.add(pregelixStmt.getParameters().get(0)); // jar cmd.add(pregelixStmt.getParameters().get(1)); // class for (String s : pregelixStmt.getParameters().get(2).split(" ")) { cmd.add(s); } cmd.add("-inputpaths"); cmd.add(fromSplitsPaths.toString()); cmd.add("-outputpath"); cmd.add(toSplitsPaths.toString()); StringBuilder command = new StringBuilder(); for (String s : cmd) { command.append(s); command.append(" "); } LOGGER.info("Running Pregelix Command: " + command.toString()); ProcessBuilder pb = new ProcessBuilder(cmd); pb.directory(new File(pregelix_home)); pb.redirectErrorStream(true); MetadataManager.INSTANCE.commitTransaction(mdTxnCtx); bActiveTxn = false; Process pr = pb.start(); int resultState = 0; BufferedReader in = new BufferedReader(new InputStreamReader(pr.getInputStream())); String line; while ((line = in.readLine()) != null) { System.out.println(line); if (line.contains("job finished")) { resultState = 1; } if (line.contains("Exception") || line.contains("Error")) { if (line.contains("Connection refused")) { throw new AlgebricksException( "The connection to your Pregelix cluster was refused. Is it running? Is the port in the query correct?"); } if (line.contains("Could not find or load main class")) { throw new AlgebricksException( "The main class of your Pregelix query was not found. Is the path to your .jar file correct?"); } if (line.contains("ClassNotFoundException")) { throw new AlgebricksException( "The vertex class of your Pregelix query was not found. Does it exist? Is the spelling correct?"); } if (line.contains("HyracksException")) { throw new AlgebricksException( "Something went wrong executing your Pregelix Job (HyracksException). Check the configuration of STORAGE_BUFFERCACHE_PAGESIZE and STORAGE_MEMORYCOMPONENT_PAGESIZE." + "It must match the one of Asterix. You can use managix describe -admin to find out the right configuration. " + "Check also if your datatypes in Pregelix and Asterix are matching."); } throw new AlgebricksException( "Something went wrong executing your Pregelix Job. Perhaps the Pregelix cluster needs to be restartet. " + "Check the following things: Are the datatypes of Asterix and Pregelix matching? " + "Is the server configuration correct (node names, buffer sizes, framesize)? Check the logfiles for more details."); } } pr.waitFor(); in.close(); if (resultState != 1) { throw new AlgebricksException( "Something went wrong executing your Pregelix Job. Perhaps the Pregelix cluster needs to be restartet. " + "Check the following things: Are the datatypes of Asterix and Pregelix matching? " + "Is the server configuration correct (node names, buffer sizes, framesize)? Check the logfiles for more details."); } } catch (Exception e) { if (bActiveTxn) { abort(e, e, mdTxnCtx); } throw e; } finally { MetadataLockManager.INSTANCE.pregelixEnd(dataverseNameFrom, datasetNameFrom, datasetNameTo); } } private void flushDataset(IHyracksClientConnection hcc, AqlMetadataProvider metadataProvider, MetadataTransactionContext mdTxnCtx, String dataverseName, String datasetName, String indexName) throws Exception { AsterixCompilerProperties compilerProperties = AsterixAppContextInfo.getInstance().getCompilerProperties(); int frameSize = compilerProperties.getFrameSize(); JobSpecification spec = new JobSpecification(frameSize); RecordDescriptor[] rDescs = new RecordDescriptor[] { new RecordDescriptor(new ISerializerDeserializer[] {}) }; AlgebricksMetaOperatorDescriptor emptySource = new AlgebricksMetaOperatorDescriptor(spec, 0, 1, new IPushRuntimeFactory[] { new EmptyTupleSourceRuntimeFactory() }, rDescs); org.apache.asterix.common.transactions.JobId jobId = JobIdFactory.generateJobId(); Dataset dataset = MetadataManager.INSTANCE.getDataset(mdTxnCtx, dataverseName, datasetName); FlushDatasetOperatorDescriptor flushOperator = new FlushDatasetOperatorDescriptor(spec, jobId, dataset.getDatasetId()); spec.connect(new OneToOneConnectorDescriptor(spec), emptySource, 0, flushOperator, 0); Pair<IFileSplitProvider, AlgebricksPartitionConstraint> primarySplitsAndConstraint = metadataProvider .splitProviderAndPartitionConstraintsForDataset(dataverseName, datasetName, indexName, dataset.getDatasetDetails().isTemp()); AlgebricksPartitionConstraint primaryPartitionConstraint = primarySplitsAndConstraint.second; AlgebricksPartitionConstraintHelper.setPartitionConstraintInJobSpec(spec, emptySource, primaryPartitionConstraint); JobEventListenerFactory jobEventListenerFactory = new JobEventListenerFactory(jobId, true); spec.setJobletEventListenerFactory(jobEventListenerFactory); runJob(hcc, spec, true); } private JobId runJob(IHyracksClientConnection hcc, JobSpecification spec, boolean waitForCompletion) throws Exception { JobId[] jobIds = executeJobArray(hcc, new Job[] { new Job(spec) }, sessionConfig.out(), waitForCompletion); return jobIds[0]; } public JobId[] executeJobArray(IHyracksClientConnection hcc, Job[] jobs, PrintWriter out, boolean waitForCompletion) throws Exception { JobId[] startedJobIds = new JobId[jobs.length]; for (int i = 0; i < jobs.length; i++) { JobSpecification spec = jobs[i].getJobSpec(); spec.setMaxReattempts(0); JobId jobId = hcc.startJob(spec); startedJobIds[i] = jobId; if (waitForCompletion) { hcc.waitForCompletion(jobId); } } return startedJobIds; } private String getActiveDataverseName(String dataverse) throws AlgebricksException { if (dataverse != null) { return dataverse; } if (activeDefaultDataverse != null) { return activeDefaultDataverse.getDataverseName(); } throw new AlgebricksException("dataverse not specified"); } private String getActiveDataverse(Identifier dataverse) throws AlgebricksException { return getActiveDataverseName(dataverse != null ? dataverse.getValue() : null); } private void abort(Exception rootE, Exception parentE, MetadataTransactionContext mdTxnCtx) { try { if (IS_DEBUG_MODE) { rootE.printStackTrace(); } MetadataManager.INSTANCE.abortTransaction(mdTxnCtx); } catch (Exception e2) { parentE.addSuppressed(e2); throw new IllegalStateException(rootE); } } }