Java tutorial
/* * Licensed to the Apache Software Foundation (ASF) under one * or more contributor license agreements. See the NOTICE file * distributed with this work for additional information * regarding copyright ownership. The ASF licenses this file * to you under the Apache License, Version 2.0 (the * "License"); you may not use this file except in compliance * with the License. You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package org.apache.phoenix.util; import static org.apache.phoenix.util.SchemaUtil.getVarChars; import java.io.IOException; import java.sql.SQLException; import java.util.ArrayList; import java.util.Collection; import java.util.Collections; import java.util.List; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.Cell; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.HRegionLocation; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.KeyValue; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.client.HConnection; import org.apache.hadoop.hbase.client.HConnectionManager; import org.apache.hadoop.hbase.client.Mutation; import org.apache.hadoop.hbase.client.Put; import org.apache.hadoop.hbase.client.Scan; import org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment; import org.apache.hadoop.hbase.io.ImmutableBytesWritable; import org.apache.hadoop.hbase.protobuf.ProtobufUtil; import org.apache.hadoop.hbase.protobuf.RequestConverter; import org.apache.hadoop.hbase.protobuf.generated.AdminProtos.AdminService; import org.apache.hadoop.hbase.protobuf.generated.AdminProtos.GetRegionInfoRequest; import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.ipc.RemoteException; import org.apache.phoenix.coprocessor.MetaDataProtocol; import org.apache.phoenix.hbase.index.util.ImmutableBytesPtr; import org.apache.phoenix.hbase.index.util.IndexManagementUtil; import org.apache.phoenix.hbase.index.util.KeyValueBuilder; import org.apache.phoenix.hbase.index.util.VersionUtil; import org.apache.phoenix.jdbc.PhoenixConnection; import org.apache.phoenix.jdbc.PhoenixDatabaseMetaData; import org.apache.phoenix.query.QueryConstants; import org.apache.phoenix.schema.PColumn; import org.apache.phoenix.schema.PName; import org.apache.phoenix.schema.PNameFactory; import org.apache.phoenix.schema.PTable; import org.apache.phoenix.schema.PTable.LinkType; import org.apache.phoenix.schema.PTableType; import org.apache.phoenix.schema.SequenceKey; import org.apache.phoenix.schema.SortOrder; import org.apache.phoenix.schema.TableNotFoundException; import org.apache.phoenix.schema.TableProperty; import org.apache.phoenix.schema.types.PBoolean; import org.apache.phoenix.schema.types.PDataType; import org.apache.phoenix.schema.types.PLong; import org.apache.phoenix.schema.types.PSmallint; import org.apache.phoenix.schema.types.PUnsignedTinyint; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import com.google.protobuf.ServiceException; public class MetaDataUtil { private static final Logger logger = LoggerFactory.getLogger(MetaDataUtil.class); public static final String VIEW_INDEX_TABLE_PREFIX = "_IDX_"; public static final String LOCAL_INDEX_TABLE_PREFIX = "_LOCAL_IDX_"; public static final String VIEW_INDEX_SEQUENCE_PREFIX = "_SEQ_"; public static final String VIEW_INDEX_SEQUENCE_NAME_PREFIX = "_ID_"; public static final byte[] VIEW_INDEX_SEQUENCE_PREFIX_BYTES = Bytes.toBytes(VIEW_INDEX_SEQUENCE_PREFIX); private static final String VIEW_INDEX_ID_COLUMN_NAME = "_INDEX_ID"; public static final String PARENT_TABLE_KEY = "PARENT_TABLE"; public static final byte[] PARENT_TABLE_KEY_BYTES = Bytes.toBytes("PARENT_TABLE"); public static boolean areClientAndServerCompatible(long serverHBaseAndPhoenixVersion) { // As of 3.0, we allow a client and server to differ for the minor version. // Care has to be taken to upgrade the server before the client, as otherwise // the client may call expressions that don't yet exist on the server. // Differing by the patch version has always been allowed. // Only differing by the major version is not allowed. return areClientAndServerCompatible(MetaDataUtil.decodePhoenixVersion(serverHBaseAndPhoenixVersion), MetaDataProtocol.PHOENIX_MAJOR_VERSION, MetaDataProtocol.PHOENIX_MINOR_VERSION); } // Default scope for testing static boolean areClientAndServerCompatible(int serverVersion, int clientMajorVersion, int clientMinorVersion) { // A server and client with the same major and minor version number must be compatible. // So it's important that we roll the PHOENIX_MAJOR_VERSION or PHOENIX_MINOR_VERSION // when we make an incompatible change. return VersionUtil.encodeMinPatchVersion(clientMajorVersion, clientMinorVersion) <= serverVersion && // Minor major and minor cannot be ahead of server VersionUtil.encodeMaxMinorVersion(clientMajorVersion) >= serverVersion; // Major version must at least be up to server version } // Given the encoded integer representing the phoenix version in the encoded version value. // The second byte in int would be the major version, 3rd byte minor version, and 4th byte // patch version. public static int decodePhoenixVersion(long version) { return (int) ((version << Byte.SIZE * 4) >>> Byte.SIZE * 5); } // TODO: generalize this to use two bytes to return a SQL error code instead public static long encodeHasIndexWALCodec(long version, boolean isValid) { if (!isValid) { return version | 1; } return version; } public static boolean decodeHasIndexWALCodec(long version) { return (version & 0xF) == 0; } // Given the encoded integer representing the client hbase version in the encoded version value. // The second byte in int would be the major version, 3rd byte minor version, and 4th byte // patch version. public static int decodeHBaseVersion(long version) { return (int) (version >>> Byte.SIZE * 5); } public static String decodeHBaseVersionAsString(int version) { int major = (version >>> Byte.SIZE * 2) & 0xFF; int minor = (version >>> Byte.SIZE * 1) & 0xFF; int patch = version & 0xFF; return major + "." + minor + "." + patch; } // Given the encoded integer representing the phoenix version in the encoded version value. // The second byte in int would be the major version, 3rd byte minor version, and 4th byte // patch version. public static boolean decodeTableNamespaceMappingEnabled(long version) { return ((int) ((version << Byte.SIZE * 3) >>> Byte.SIZE * 7) & 0x1) != 0; } // The first 3 bytes of the long is used to encoding the HBase version as major.minor.patch. // The next 4 bytes of the value is used to encode the Phoenix version as major.minor.patch. /** * Encode HBase and Phoenix version along with some server-side config information such as whether WAL codec is * installed (necessary for non transactional, mutable secondar indexing), and whether systemNamespace mapping is enabled. * * @param env * RegionCoprocessorEnvironment to access HBase version and Configuration. * @return long value sent back during initialization of a cluster connection. */ public static long encodeVersion(String hbaseVersionStr, Configuration config) { long hbaseVersion = VersionUtil.encodeVersion(hbaseVersionStr); long isTableNamespaceMappingEnabled = SchemaUtil.isNamespaceMappingEnabled(PTableType.TABLE, new ReadOnlyProps(config.iterator())) ? 1 : 0; long phoenixVersion = VersionUtil.encodeVersion(MetaDataProtocol.PHOENIX_MAJOR_VERSION, MetaDataProtocol.PHOENIX_MINOR_VERSION, MetaDataProtocol.PHOENIX_PATCH_NUMBER); long walCodec = IndexManagementUtil.isWALEditCodecSet(config) ? 0 : 1; long version = // Encode HBase major, minor, patch version (hbaseVersion << (Byte.SIZE * 5)) // Encode if systemMappingEnabled are enabled on the server side | (isTableNamespaceMappingEnabled << (Byte.SIZE * 4)) // Encode Phoenix major, minor, patch version | (phoenixVersion << (Byte.SIZE * 1)) // Encode whether or not non transactional, mutable secondary indexing was configured properly. | walCodec; return version; } public static void getTenantIdAndSchemaAndTableName(List<Mutation> tableMetadata, byte[][] rowKeyMetaData) { Mutation m = getTableHeaderRow(tableMetadata); getVarChars(m.getRow(), 3, rowKeyMetaData); } public static void getTenantIdAndFunctionName(List<Mutation> functionMetadata, byte[][] rowKeyMetaData) { Mutation m = getTableHeaderRow(functionMetadata); getVarChars(m.getRow(), 2, rowKeyMetaData); } public static byte[] getParentTableName(List<Mutation> tableMetadata) { if (tableMetadata.size() == 1) { return null; } byte[][] rowKeyMetaData = new byte[3][]; getTenantIdAndSchemaAndTableName(tableMetadata, rowKeyMetaData); byte[] schemaName = rowKeyMetaData[PhoenixDatabaseMetaData.SCHEMA_NAME_INDEX]; byte[] tableName = rowKeyMetaData[PhoenixDatabaseMetaData.TABLE_NAME_INDEX]; Mutation m = getParentTableHeaderRow(tableMetadata); getVarChars(m.getRow(), 3, rowKeyMetaData); if (Bytes.compareTo(schemaName, rowKeyMetaData[PhoenixDatabaseMetaData.SCHEMA_NAME_INDEX]) == 0 && Bytes.compareTo(tableName, rowKeyMetaData[PhoenixDatabaseMetaData.TABLE_NAME_INDEX]) == 0) { return null; } return rowKeyMetaData[PhoenixDatabaseMetaData.TABLE_NAME_INDEX]; } public static long getSequenceNumber(Mutation tableMutation) { List<Cell> kvs = tableMutation.getFamilyCellMap().get(PhoenixDatabaseMetaData.TABLE_FAMILY_BYTES); if (kvs != null) { for (Cell kv : kvs) { // list is not ordered, so search. TODO: we could potentially assume the position if (Bytes.compareTo(kv.getQualifierArray(), kv.getQualifierOffset(), kv.getQualifierLength(), PhoenixDatabaseMetaData.TABLE_SEQ_NUM_BYTES, 0, PhoenixDatabaseMetaData.TABLE_SEQ_NUM_BYTES.length) == 0) { return PLong.INSTANCE.getCodec().decodeLong(kv.getValueArray(), kv.getValueOffset(), SortOrder.getDefault()); } } } throw new IllegalStateException(); } public static long getSequenceNumber(List<Mutation> tableMetaData) { return getSequenceNumber(getPutOnlyTableHeaderRow(tableMetaData)); } public static PTableType getTableType(List<Mutation> tableMetaData, KeyValueBuilder builder, ImmutableBytesWritable value) { if (getMutationValue(getPutOnlyTableHeaderRow(tableMetaData), PhoenixDatabaseMetaData.TABLE_TYPE_BYTES, builder, value)) { return PTableType.fromSerializedValue(value.get()[value.getOffset()]); } return null; } public static long getParentSequenceNumber(List<Mutation> tableMetaData) { return getSequenceNumber(getParentTableHeaderRow(tableMetaData)); } public static Mutation getTableHeaderRow(List<Mutation> tableMetaData) { return tableMetaData.get(0); } /** * Get the mutation who's qualifier matches the passed key * <p> * We need to pass in an {@link ImmutableBytesPtr} to pass the result back to make life easier * when dealing with a regular {@link KeyValue} vs. a custom KeyValue as the latter may not * support things like {@link KeyValue#getBuffer()} * @param headerRow mutation to check * @param key to check * @param builder that created the {@link KeyValue KeyValues} in the {@link Mutation} * @param ptr to point to the KeyValue's value if found * @return true if the KeyValue was found and false otherwise */ public static boolean getMutationValue(Mutation headerRow, byte[] key, KeyValueBuilder builder, ImmutableBytesWritable ptr) { List<Cell> kvs = headerRow.getFamilyCellMap().get(PhoenixDatabaseMetaData.TABLE_FAMILY_BYTES); if (kvs != null) { for (Cell cell : kvs) { KeyValue kv = org.apache.hadoop.hbase.KeyValueUtil.ensureKeyValue(cell); if (builder.compareQualifier(kv, key, 0, key.length) == 0) { builder.getValueAsPtr(kv, ptr); return true; } } } return false; } /** * Returns the first Put element in <code>tableMetaData</code>. There could be leading Delete elements before the * table header row */ public static Put getPutOnlyTableHeaderRow(List<Mutation> tableMetaData) { for (Mutation m : tableMetaData) { if (m instanceof Put) { return (Put) m; } } throw new IllegalStateException("No table header row found in table metadata"); } public static Put getPutOnlyAutoPartitionColumn(PTable parentTable, List<Mutation> tableMetaData) { int autoPartitionPutIndex = parentTable.isMultiTenant() ? 2 : 1; int i = 0; for (Mutation m : tableMetaData) { if (m instanceof Put && i++ == autoPartitionPutIndex) { return (Put) m; } } throw new IllegalStateException("No auto partition column row found in table metadata"); } public static Mutation getParentTableHeaderRow(List<Mutation> tableMetaData) { return tableMetaData.get(tableMetaData.size() - 1); } public static long getClientTimeStamp(List<Mutation> tableMetadata) { Mutation m = tableMetadata.get(0); return getClientTimeStamp(m); } public static long getClientTimeStamp(Mutation m) { Collection<List<Cell>> kvs = m.getFamilyCellMap().values(); // Empty if Mutation is a Delete // TODO: confirm that Delete timestamp is reset like Put return kvs.isEmpty() ? m.getTimeStamp() : kvs.iterator().next().get(0).getTimestamp(); } public static byte[] getParentLinkKey(String tenantId, String schemaName, String tableName, String indexName) { return ByteUtil.concat(tenantId == null ? ByteUtil.EMPTY_BYTE_ARRAY : Bytes.toBytes(tenantId), QueryConstants.SEPARATOR_BYTE_ARRAY, schemaName == null ? ByteUtil.EMPTY_BYTE_ARRAY : Bytes.toBytes(schemaName), QueryConstants.SEPARATOR_BYTE_ARRAY, Bytes.toBytes(tableName), QueryConstants.SEPARATOR_BYTE_ARRAY, QueryConstants.SEPARATOR_BYTE_ARRAY, Bytes.toBytes(indexName)); } public static byte[] getParentLinkKey(byte[] tenantId, byte[] schemaName, byte[] tableName, byte[] indexName) { return ByteUtil.concat(tenantId == null ? ByteUtil.EMPTY_BYTE_ARRAY : tenantId, QueryConstants.SEPARATOR_BYTE_ARRAY, schemaName == null ? ByteUtil.EMPTY_BYTE_ARRAY : schemaName, QueryConstants.SEPARATOR_BYTE_ARRAY, tableName, QueryConstants.SEPARATOR_BYTE_ARRAY, QueryConstants.SEPARATOR_BYTE_ARRAY, indexName); } public static byte[] getChildLinkKey(PName parentTenantId, PName parentSchemaName, PName parentTableName, PName viewTenantId, PName viewName) { return ByteUtil.concat(parentTenantId == null ? ByteUtil.EMPTY_BYTE_ARRAY : parentTenantId.getBytes(), QueryConstants.SEPARATOR_BYTE_ARRAY, parentSchemaName == null ? ByteUtil.EMPTY_BYTE_ARRAY : parentSchemaName.getBytes(), QueryConstants.SEPARATOR_BYTE_ARRAY, parentTableName.getBytes(), QueryConstants.SEPARATOR_BYTE_ARRAY, viewTenantId == null ? ByteUtil.EMPTY_BYTE_ARRAY : viewTenantId.getBytes(), QueryConstants.SEPARATOR_BYTE_ARRAY, viewName.getBytes()); } public static Cell getCell(List<Cell> cells, byte[] cq) { for (Cell cell : cells) { if (Bytes.compareTo(cell.getQualifierArray(), cell.getQualifierOffset(), cell.getQualifierLength(), cq, 0, cq.length) == 0) { return cell; } } return null; } public static boolean isMultiTenant(Mutation m, KeyValueBuilder builder, ImmutableBytesWritable ptr) { if (getMutationValue(m, PhoenixDatabaseMetaData.MULTI_TENANT_BYTES, builder, ptr)) { return Boolean.TRUE.equals(PBoolean.INSTANCE.toObject(ptr)); } return false; } public static boolean isTransactional(Mutation m, KeyValueBuilder builder, ImmutableBytesWritable ptr) { if (getMutationValue(m, PhoenixDatabaseMetaData.TRANSACTIONAL_BYTES, builder, ptr)) { return Boolean.TRUE.equals(PBoolean.INSTANCE.toObject(ptr)); } return false; } public static boolean isSalted(Mutation m, KeyValueBuilder builder, ImmutableBytesWritable ptr) { return MetaDataUtil.getMutationValue(m, PhoenixDatabaseMetaData.SALT_BUCKETS_BYTES, builder, ptr); } public static byte[] getViewIndexPhysicalName(byte[] physicalTableName) { return getIndexPhysicalName(physicalTableName, VIEW_INDEX_TABLE_PREFIX); } public static String getViewIndexTableName(String tableName) { return VIEW_INDEX_TABLE_PREFIX + tableName; } public static String getViewIndexSchemaName(String schemaName) { return schemaName; } public static String getViewIndexName(String schemaName, String tableName) { return SchemaUtil.getTableName(getViewIndexSchemaName(schemaName), getViewIndexTableName(tableName)); } public static byte[] getIndexPhysicalName(byte[] physicalTableName, String indexPrefix) { return getIndexPhysicalName(Bytes.toString(physicalTableName), indexPrefix).getBytes(); } public static String getIndexPhysicalName(String physicalTableName, String indexPrefix) { if (physicalTableName.contains(QueryConstants.NAMESPACE_SEPARATOR)) { String schemaName = SchemaUtil.getSchemaNameFromFullName(physicalTableName, QueryConstants.NAMESPACE_SEPARATOR); String tableName = SchemaUtil.getTableNameFromFullName(physicalTableName, QueryConstants.NAMESPACE_SEPARATOR); return (schemaName + QueryConstants.NAMESPACE_SEPARATOR + indexPrefix + tableName); } return indexPrefix + physicalTableName; } public static byte[] getLocalIndexPhysicalName(byte[] physicalTableName) { return getIndexPhysicalName(physicalTableName, LOCAL_INDEX_TABLE_PREFIX); } public static String getLocalIndexTableName(String tableName) { return LOCAL_INDEX_TABLE_PREFIX + tableName; } public static String getLocalIndexSchemaName(String schemaName) { return schemaName; } public static String getLocalIndexUserTableName(String localIndexTableName) { if (localIndexTableName.contains(QueryConstants.NAMESPACE_SEPARATOR)) { String schemaName = SchemaUtil.getSchemaNameFromFullName(localIndexTableName, QueryConstants.NAMESPACE_SEPARATOR); String tableName = SchemaUtil.getTableNameFromFullName(localIndexTableName, QueryConstants.NAMESPACE_SEPARATOR); String userTableName = tableName.substring(LOCAL_INDEX_TABLE_PREFIX.length()); return (schemaName + QueryConstants.NAMESPACE_SEPARATOR + userTableName); } else { String schemaName = SchemaUtil.getSchemaNameFromFullName(localIndexTableName); if (!schemaName.isEmpty()) schemaName = schemaName.substring(LOCAL_INDEX_TABLE_PREFIX.length()); String tableName = localIndexTableName.substring( (schemaName.isEmpty() ? 0 : (schemaName.length() + QueryConstants.NAME_SEPARATOR.length())) + LOCAL_INDEX_TABLE_PREFIX.length()); return SchemaUtil.getTableName(schemaName, tableName); } } public static String getViewIndexUserTableName(String viewIndexTableName) { if (viewIndexTableName.contains(QueryConstants.NAMESPACE_SEPARATOR)) { String schemaName = SchemaUtil.getSchemaNameFromFullName(viewIndexTableName, QueryConstants.NAMESPACE_SEPARATOR); String tableName = SchemaUtil.getTableNameFromFullName(viewIndexTableName, QueryConstants.NAMESPACE_SEPARATOR); String userTableName = tableName.substring(VIEW_INDEX_TABLE_PREFIX.length()); return (schemaName + QueryConstants.NAMESPACE_SEPARATOR + userTableName); } else { String schemaName = SchemaUtil.getSchemaNameFromFullName(viewIndexTableName); if (!schemaName.isEmpty()) schemaName = schemaName.substring(VIEW_INDEX_TABLE_PREFIX.length()); String tableName = viewIndexTableName.substring( (schemaName.isEmpty() ? 0 : (schemaName.length() + QueryConstants.NAME_SEPARATOR.length())) + VIEW_INDEX_TABLE_PREFIX.length()); return SchemaUtil.getTableName(schemaName, tableName); } } public static String getViewIndexSequenceSchemaName(PName physicalName, boolean isNamespaceMapped) { if (!isNamespaceMapped) { return VIEW_INDEX_SEQUENCE_PREFIX + physicalName.getString(); } return SchemaUtil.getSchemaNameFromFullName(physicalName.toString()); } public static String getViewIndexSequenceName(PName physicalName, PName tenantId, boolean isNamespaceMapped) { if (!isNamespaceMapped) { return VIEW_INDEX_SEQUENCE_NAME_PREFIX + (tenantId == null ? "" : tenantId); } return SchemaUtil.getTableNameFromFullName(physicalName.toString()) + VIEW_INDEX_SEQUENCE_NAME_PREFIX; } public static SequenceKey getViewIndexSequenceKey(String tenantId, PName physicalName, int nSaltBuckets, boolean isNamespaceMapped) { // Create global sequence of the form: <prefixed base table name><tenant id> // rather than tenant-specific sequence, as it makes it much easier // to cleanup when the physical table is dropped, as we can delete // all global sequences leading with <prefix> + physical name. String schemaName = getViewIndexSequenceSchemaName(physicalName, isNamespaceMapped); String tableName = getViewIndexSequenceName(physicalName, PNameFactory.newName(tenantId), isNamespaceMapped); return new SequenceKey(isNamespaceMapped ? tenantId : null, schemaName, tableName, nSaltBuckets); } public static PDataType getViewIndexIdDataType() { return PSmallint.INSTANCE; } public static String getViewIndexIdColumnName() { return VIEW_INDEX_ID_COLUMN_NAME; } public static boolean hasViewIndexTable(PhoenixConnection connection, PName physicalName) throws SQLException { return hasViewIndexTable(connection, physicalName.getBytes()); } public static boolean hasViewIndexTable(PhoenixConnection connection, byte[] physicalTableName) throws SQLException { byte[] physicalIndexName = MetaDataUtil.getViewIndexPhysicalName(physicalTableName); try { HTableDescriptor desc = connection.getQueryServices().getTableDescriptor(physicalIndexName); return desc != null && Boolean.TRUE .equals(PBoolean.INSTANCE.toObject(desc.getValue(IS_VIEW_INDEX_TABLE_PROP_BYTES))); } catch (TableNotFoundException e) { return false; } } public static boolean hasLocalIndexTable(PhoenixConnection connection, PName physicalName) throws SQLException { return hasLocalIndexTable(connection, physicalName.getBytes()); } public static boolean hasLocalIndexTable(PhoenixConnection connection, byte[] physicalTableName) throws SQLException { try { HTableDescriptor desc = connection.getQueryServices().getTableDescriptor(physicalTableName); if (desc == null) return false; return hasLocalIndexColumnFamily(desc); } catch (TableNotFoundException e) { return false; } } public static boolean hasLocalIndexColumnFamily(HTableDescriptor desc) { for (HColumnDescriptor cf : desc.getColumnFamilies()) { if (cf.getNameAsString().startsWith(QueryConstants.LOCAL_INDEX_COLUMN_FAMILY_PREFIX)) { return true; } } return false; } public static List<byte[]> getNonLocalIndexColumnFamilies(HTableDescriptor desc) { List<byte[]> families = new ArrayList<byte[]>(desc.getColumnFamilies().length); for (HColumnDescriptor cf : desc.getColumnFamilies()) { if (!cf.getNameAsString().startsWith(QueryConstants.LOCAL_INDEX_COLUMN_FAMILY_PREFIX)) { families.add(cf.getName()); } } return families; } public static List<byte[]> getLocalIndexColumnFamilies(PhoenixConnection conn, byte[] physicalTableName) throws SQLException { HTableDescriptor desc = conn.getQueryServices().getTableDescriptor(physicalTableName); if (desc == null) return Collections.emptyList(); List<byte[]> families = new ArrayList<byte[]>(desc.getColumnFamilies().length / 2); for (HColumnDescriptor cf : desc.getColumnFamilies()) { if (cf.getNameAsString().startsWith(QueryConstants.LOCAL_INDEX_COLUMN_FAMILY_PREFIX)) { families.add(cf.getName()); } } return families; } public static void deleteViewIndexSequences(PhoenixConnection connection, PName name, boolean isNamespaceMapped) throws SQLException { String schemaName = getViewIndexSequenceSchemaName(name, isNamespaceMapped); String sequenceName = getViewIndexSequenceName(name, null, isNamespaceMapped); connection.createStatement() .executeUpdate("DELETE FROM " + PhoenixDatabaseMetaData.SYSTEM_SEQUENCE + " WHERE " + PhoenixDatabaseMetaData.SEQUENCE_SCHEMA + (schemaName.length() > 0 ? "='" + schemaName + "'" : " IS NULL") + (isNamespaceMapped ? " AND " + PhoenixDatabaseMetaData.SEQUENCE_NAME + " = '" + sequenceName + "'" : "")); } /** * This function checks if all regions of a table is online * @param table * @return true when all regions of a table are online * @throws IOException * @throws */ public static boolean tableRegionsOnline(Configuration conf, PTable table) { HConnection hcon = null; try { hcon = HConnectionManager.getConnection(conf); List<HRegionLocation> locations = hcon .locateRegions(org.apache.hadoop.hbase.TableName.valueOf(table.getPhysicalName().getBytes())); for (HRegionLocation loc : locations) { try { ServerName sn = loc.getServerName(); if (sn == null) continue; AdminService.BlockingInterface admin = hcon.getAdmin(sn); GetRegionInfoRequest request = RequestConverter .buildGetRegionInfoRequest(loc.getRegionInfo().getRegionName()); admin.getRegionInfo(null, request); } catch (ServiceException e) { IOException ie = ProtobufUtil.getRemoteException(e); logger.debug("Region " + loc.getRegionInfo().getEncodedName() + " isn't online due to:" + ie); return false; } catch (RemoteException e) { logger.debug("Cannot get region " + loc.getRegionInfo().getEncodedName() + " info due to error:" + e); return false; } } } catch (IOException ex) { logger.warn("tableRegionsOnline failed due to:" + ex); return false; } finally { if (hcon != null) { try { hcon.close(); } catch (IOException ignored) { } } } return true; } public static final String IS_VIEW_INDEX_TABLE_PROP_NAME = "IS_VIEW_INDEX_TABLE"; public static final byte[] IS_VIEW_INDEX_TABLE_PROP_BYTES = Bytes.toBytes(IS_VIEW_INDEX_TABLE_PROP_NAME); public static final String IS_LOCAL_INDEX_TABLE_PROP_NAME = "IS_LOCAL_INDEX_TABLE"; public static final byte[] IS_LOCAL_INDEX_TABLE_PROP_BYTES = Bytes.toBytes(IS_LOCAL_INDEX_TABLE_PROP_NAME); public static final String DATA_TABLE_NAME_PROP_NAME = "DATA_TABLE_NAME"; public static final byte[] DATA_TABLE_NAME_PROP_BYTES = Bytes.toBytes(DATA_TABLE_NAME_PROP_NAME); public static Scan newTableRowsScan(byte[] key, long startTimeStamp, long stopTimeStamp) { return newTableRowsScan(key, null, startTimeStamp, stopTimeStamp); } public static Scan newTableRowsScan(byte[] startKey, byte[] stopKey, long startTimeStamp, long stopTimeStamp) { Scan scan = new Scan(); ScanUtil.setTimeRange(scan, startTimeStamp, stopTimeStamp); scan.setStartRow(startKey); if (stopKey == null) { stopKey = ByteUtil.concat(startKey, QueryConstants.SEPARATOR_BYTE_ARRAY); ByteUtil.nextKey(stopKey, stopKey.length); } scan.setStopRow(stopKey); return scan; } public static LinkType getLinkType(Mutation tableMutation) { List<Cell> kvs = tableMutation.getFamilyCellMap().get(PhoenixDatabaseMetaData.TABLE_FAMILY_BYTES); if (kvs != null) { for (Cell kv : kvs) { if (Bytes.compareTo(kv.getQualifierArray(), kv.getQualifierOffset(), kv.getQualifierLength(), PhoenixDatabaseMetaData.LINK_TYPE_BYTES, 0, PhoenixDatabaseMetaData.LINK_TYPE_BYTES.length) == 0) { return LinkType.fromSerializedValue(PUnsignedTinyint.INSTANCE.getCodec() .decodeByte(kv.getValueArray(), kv.getValueOffset(), SortOrder.getDefault())); } } } return null; } public static boolean isLocalIndex(String physicalName) { if (physicalName.contains(LOCAL_INDEX_TABLE_PREFIX)) { return true; } return false; } public static boolean isViewIndex(String physicalName) { if (physicalName.contains(QueryConstants.NAMESPACE_SEPARATOR)) { return SchemaUtil.getTableNameFromFullName(physicalName).startsWith(VIEW_INDEX_TABLE_PREFIX); } else { return physicalName.startsWith(VIEW_INDEX_TABLE_PREFIX); } } public static String getAutoPartitionColumnName(PTable parentTable) { List<PColumn> parentTableColumns = parentTable.getPKColumns(); PColumn column = parentTableColumns.get(getAutoPartitionColIndex(parentTable)); return column.getName().getString(); } // this method should only be called on the parent table (since it has the _SALT column) public static int getAutoPartitionColIndex(PTable parentTable) { boolean isMultiTenant = parentTable.isMultiTenant(); boolean isSalted = parentTable.getBucketNum() != null; return (isMultiTenant && isSalted) ? 2 : (isMultiTenant || isSalted) ? 1 : 0; } public static String getJdbcUrl(RegionCoprocessorEnvironment env) { String zkQuorum = env.getConfiguration().get(HConstants.ZOOKEEPER_QUORUM); String zkClientPort = env.getConfiguration().get(HConstants.ZOOKEEPER_CLIENT_PORT, Integer.toString(HConstants.DEFAULT_ZOOKEPER_CLIENT_PORT)); String zkParentNode = env.getConfiguration().get(HConstants.ZOOKEEPER_ZNODE_PARENT, HConstants.DEFAULT_ZOOKEEPER_ZNODE_PARENT); return PhoenixRuntime.JDBC_PROTOCOL + PhoenixRuntime.JDBC_PROTOCOL_SEPARATOR + zkQuorum + PhoenixRuntime.JDBC_PROTOCOL_SEPARATOR + zkClientPort + PhoenixRuntime.JDBC_PROTOCOL_SEPARATOR + zkParentNode; } public static boolean isHColumnProperty(String propName) { return HColumnDescriptor.getDefaultValues().containsKey(propName); } public static boolean isHTableProperty(String propName) { return !isHColumnProperty(propName) && !TableProperty.isPhoenixTableProperty(propName); } public static boolean isLocalIndexFamily(ImmutableBytesPtr cfPtr) { return cfPtr.getLength() >= QueryConstants.LOCAL_INDEX_COLUMN_FAMILY_PREFIX_BYTES.length && Bytes.compareTo(cfPtr.get(), cfPtr.getOffset(), QueryConstants.LOCAL_INDEX_COLUMN_FAMILY_PREFIX_BYTES.length, QueryConstants.LOCAL_INDEX_COLUMN_FAMILY_PREFIX_BYTES, 0, QueryConstants.LOCAL_INDEX_COLUMN_FAMILY_PREFIX_BYTES.length) == 0; } public static boolean isLocalIndexFamily(byte[] cf) { return Bytes.startsWith(cf, QueryConstants.LOCAL_INDEX_COLUMN_FAMILY_PREFIX_BYTES); } public static final byte[] getPhysicalTableRowForView(PTable view) { byte[] physicalTableSchemaName = Bytes .toBytes(SchemaUtil.getSchemaNameFromFullName(view.getPhysicalName().getString())); byte[] physicalTableName = Bytes .toBytes(SchemaUtil.getTableNameFromFullName(view.getPhysicalName().getString())); return SchemaUtil.getTableKey(ByteUtil.EMPTY_BYTE_ARRAY, physicalTableSchemaName, physicalTableName); } }