Modifier and Type | Field and Description |
---|---|
static TableName |
HConstants.ENSEMBLE_TABLE_NAME
The name of the ensemble table
|
static TableName |
TableName.META_TABLE_NAME
The hbase:meta table's name.
|
static TableName |
TableName.NAMESPACE_TABLE_NAME
Deprecated.
since 3.0.0 and will be removed in 4.0.0. We have folded the data in namespace
table into meta table, so do not use it any more.
|
static TableName |
TableName.OLD_META_TABLE_NAME
TableName for old .META.
|
static TableName |
TableName.OLD_ROOT_TABLE_NAME
TableName for old -ROOT- table.
|
Modifier and Type | Field and Description |
---|---|
private static Set<TableName> |
TableName.tableCache
|
private Map<TableName,RegionStatesCount> |
ClusterMetricsBuilder.tableRegionStatesCount |
private Map<TableName,RegionStatesCount> |
ClusterMetricsBuilder.ClusterMetricsImpl.tableRegionStatesCount |
Modifier and Type | Method and Description |
---|---|
private static TableName |
TableName.createTableNameIfNecessary(ByteBuffer bns,
ByteBuffer qns)
Check that the object does not exist already.
|
private static TableName |
TableName.getADummyTableName(String qualifier)
It is used to create table names for old META, and ROOT table.
|
static TableName |
TableName.valueOf(byte[] fullName)
Construct a TableName
|
static TableName |
TableName.valueOf(byte[] namespace,
byte[] qualifier) |
static TableName |
TableName.valueOf(byte[] fullName,
int offset,
int length)
Construct a TableName
|
static TableName |
TableName.valueOf(ByteBuffer fullname)
Construct a TableName
|
static TableName |
TableName.valueOf(ByteBuffer namespace,
ByteBuffer qualifier) |
static TableName |
TableName.valueOf(String name)
Construct a TableName
|
static TableName |
TableName.valueOf(String namespaceAsString,
String qualifierAsString) |
Modifier and Type | Method and Description |
---|---|
Map<TableName,RegionStatesCount> |
ClusterMetricsBuilder.ClusterMetricsImpl.getTableRegionStatesCount() |
Map<TableName,RegionStatesCount> |
ClusterMetrics.getTableRegionStatesCount()
Provide region states count for given table.
|
static Map<TableName,TableState> |
MetaTableAccessor.getTableStates(Connection conn)
Fetch table states from META table
|
Modifier and Type | Method and Description |
---|---|
void |
HBaseTestingUtility.compact(TableName tableName,
boolean major)
Deprecated.
Compact all of a table's reagion in the mini hbase cluster
|
void |
MiniHBaseCluster.compact(TableName tableName,
boolean major)
Deprecated.
Call flushCache on all regions of the specified table.
|
int |
TableName.compareTo(TableName tableName) |
int |
HBaseTestingUtility.countRows(TableName tableName)
Deprecated.
Return the number of rows in the given table.
|
HRegion |
HBaseTestingUtility.createLocalHRegion(TableName tableName,
byte[] startKey,
byte[] stopKey,
org.apache.hadoop.conf.Configuration conf,
boolean isReadOnly,
Durability durability,
WAL wal,
byte[]... families)
Deprecated.
Return a region on which you must call
HBaseTestingUtility.closeRegionAndWAL(HRegion)
when done. |
HRegion |
HBaseTestingUtility.createLocalHRegionWithInMemoryFlags(TableName tableName,
byte[] startKey,
byte[] stopKey,
org.apache.hadoop.conf.Configuration conf,
boolean isReadOnly,
Durability durability,
WAL wal,
boolean[] compactedMemStore,
byte[]... families)
Deprecated.
|
TableDescriptorBuilder |
HBaseTestingUtility.createModifyableTableDescriptor(TableName name,
int minVersions,
int versions,
int ttl,
KeepDeletedCells keepDeleted)
Deprecated.
|
Table |
HBaseTestingUtility.createMultiRegionTable(TableName tableName,
byte[] family)
Deprecated.
Create a table with multiple regions.
|
Table |
HBaseTestingUtility.createMultiRegionTable(TableName tableName,
byte[][] families)
Deprecated.
Create a table with multiple regions.
|
Table |
HBaseTestingUtility.createMultiRegionTable(TableName tableName,
byte[][] families,
int numVersions)
Deprecated.
Create a table with multiple regions.
|
Table |
HBaseTestingUtility.createMultiRegionTable(TableName tableName,
byte[] family,
int numRegions)
Deprecated.
Create a table with multiple regions.
|
Table |
HBaseTestingUtility.createMultiRegionTable(TableName tableName,
int replicaCount,
byte[][] families)
Deprecated.
Create a table with multiple regions.
|
static int |
HBaseTestingUtility.createPreSplitLoadTestTable(org.apache.hadoop.conf.Configuration conf,
TableName tableName,
byte[][] columnFamilies,
Compression.Algorithm compression,
DataBlockEncoding dataBlockEncoding,
int numRegionsPerServer,
int regionReplication,
Durability durability)
Deprecated.
Creates a pre-split table for load testing.
|
static int |
HBaseTestingUtility.createPreSplitLoadTestTable(org.apache.hadoop.conf.Configuration conf,
TableName tableName,
byte[] columnFamily,
Compression.Algorithm compression,
DataBlockEncoding dataBlockEncoding)
Deprecated.
Creates a pre-split table for load testing.
|
static int |
HBaseTestingUtility.createPreSplitLoadTestTable(org.apache.hadoop.conf.Configuration conf,
TableName tableName,
byte[] columnFamily,
Compression.Algorithm compression,
DataBlockEncoding dataBlockEncoding,
int numRegionsPerServer,
int regionReplication,
Durability durability)
Deprecated.
Creates a pre-split table for load testing.
|
Table |
HBaseTestingUtility.createRandomTable(TableName tableName,
Collection<String> families,
int maxVersions,
int numColsPerRow,
int numFlushes,
int numRegions,
int numRowsPerFlush)
Deprecated.
Creates a random table with the given parameters
|
Table |
HBaseTestingUtility.createTable(TableName tableName,
byte[] family)
Deprecated.
Create a table.
|
Table |
HBaseTestingUtility.createTable(TableName tableName,
byte[][] families)
Deprecated.
Create a table.
|
Table |
HBaseTestingUtility.createTable(TableName tableName,
byte[][] families,
byte[][] splitKeys)
Deprecated.
Create a table.
|
Table |
HBaseTestingUtility.createTable(TableName tableName,
byte[][] families,
byte[][] splitKeys,
int replicaCount)
Deprecated.
Create a table.
|
Table |
HBaseTestingUtility.createTable(TableName tableName,
byte[][] families,
byte[][] splitKeys,
int replicaCount,
org.apache.hadoop.conf.Configuration c)
Deprecated.
Create a table.
|
Table |
HBaseTestingUtility.createTable(TableName tableName,
byte[][] families,
int numVersions)
Deprecated.
Create a table.
|
Table |
HBaseTestingUtility.createTable(TableName tableName,
byte[][] families,
int[] numVersions)
Deprecated.
Create a table.
|
Table |
HBaseTestingUtility.createTable(TableName tableName,
byte[][] families,
int numVersions,
byte[][] splitKeys)
Deprecated.
Create a table.
|
Table |
HBaseTestingUtility.createTable(TableName tableName,
byte[][] families,
int numVersions,
byte[] startKey,
byte[] endKey,
int numRegions)
Deprecated.
|
Table |
HBaseTestingUtility.createTable(TableName tableName,
byte[][] families,
int numVersions,
int blockSize)
Deprecated.
Create a table.
|
Table |
HBaseTestingUtility.createTable(TableName tableName,
byte[][] families,
int numVersions,
int blockSize,
String cpName)
Deprecated.
|
Table |
HBaseTestingUtility.createTable(TableName tableName,
byte[] family,
byte[][] splitRows)
Deprecated.
Create a table.
|
Table |
HBaseTestingUtility.createTable(TableName tableName,
byte[] family,
int numVersions)
Deprecated.
Create a table.
|
Table |
HBaseTestingUtility.createTable(TableName tableName,
String family)
Deprecated.
Create a table.
|
Table |
HBaseTestingUtility.createTable(TableName tableName,
String[] families)
Deprecated.
Create a table.
|
TableDescriptor |
HBaseTestingUtility.createTableDescriptor(TableName name)
Deprecated.
Create a table of name
name . |
TableDescriptor |
HBaseTestingUtility.createTableDescriptor(TableName tableName,
byte[] family)
Deprecated.
|
TableDescriptor |
HBaseTestingUtility.createTableDescriptor(TableName tableName,
byte[][] families,
int maxVersions)
Deprecated.
|
TableDescriptor |
HBaseTestingUtility.createTableDescriptor(TableName name,
int minVersions,
int versions,
int ttl,
KeepDeletedCells keepDeleted)
Deprecated.
|
void |
HBaseTestingUtility.deleteTable(TableName tableName)
Deprecated.
Drop an existing table
|
Table |
HBaseTestingUtility.deleteTableData(TableName tableName)
Deprecated.
Provide an existing table name to truncate.
|
void |
HBaseTestingUtility.deleteTableIfAny(TableName tableName)
Deprecated.
Drop an existing table
|
static void |
MetaTableAccessor.deleteTableState(Connection connection,
TableName table)
Remove state for table from meta
|
default boolean |
TableDescriptors.exists(TableName tableName)
Test whether a given table exists, i.e, has a table descriptor.
|
String |
HBaseTestingUtility.explainTableAvailability(TableName tableName)
Deprecated.
|
String |
HBaseTestingUtility.explainTableState(TableName table,
TableState.State state)
Deprecated.
|
TableState |
HBaseTestingUtility.findLastTableState(TableName table)
Deprecated.
|
List<HRegion> |
MiniHBaseCluster.findRegionsForTable(TableName tableName)
Deprecated.
|
void |
HBaseTestingUtility.flush(TableName tableName)
Deprecated.
Flushes all caches in the mini hbase cluster
|
void |
MiniHBaseCluster.flushcache(TableName tableName)
Deprecated.
Call flushCache on all regions of the specified table.
|
TableDescriptor |
TableDescriptors.get(TableName tableName)
Returns TableDescriptor for tablename
|
static CellComparator |
CellComparatorImpl.getCellComparator(TableName tableName)
Utility method that makes a guess at comparator to use based off passed tableName.
|
private static RegionInfo |
MetaTableAccessor.getClosestRegionInfo(Connection connection,
TableName tableName,
byte[] row)
Returns Get closest metatable region row to passed
row |
static CellComparator |
InnerStoreCellComparator.getInnerStoreCellComparator(TableName tableName)
Utility method that makes a guess at comparator to use based off passed tableName.
|
default long |
ClusterMetrics.getLastMajorCompactionTimestamp(TableName table) |
List<byte[]> |
HBaseTestingUtility.getMetaTableRows(TableName tableName)
Deprecated.
Returns all rows from the hbase:meta table for a given user table
|
int |
HBaseTestingUtility.getNumHFiles(TableName tableName,
byte[] family)
Deprecated.
|
int |
HBaseTestingUtility.getNumHFilesForRS(HRegionServer rs,
TableName tableName,
byte[] family)
Deprecated.
|
private List<RegionInfo> |
HBaseTestingUtility.getRegions(TableName tableName)
Deprecated.
Returns all regions of the specified table
|
List<HRegion> |
MiniHBaseCluster.getRegions(TableName tableName)
Deprecated.
|
HRegionServer |
HBaseTestingUtility.getRSForFirstRegionInTable(TableName tableName)
Deprecated.
Tool to get the reference to the region server object that holds the region of the specified
user table.
|
static Scan |
MetaTableAccessor.getScanForTableName(org.apache.hadoop.conf.Configuration conf,
TableName tableName)
This method creates a Scan object that will only scan catalog rows that belong to the specified
table.
|
ServerName |
MiniHBaseCluster.getServerHoldingRegion(TableName tn,
byte[] regionName)
Deprecated.
|
abstract ServerName |
HBaseCluster.getServerHoldingRegion(TableName tn,
byte[] regionName)
Deprecated.
Get the ServerName of region server serving the specified region
|
HRegion |
HBaseTestingUtility.getSplittableRegion(TableName tableName,
int maxAttempts)
Deprecated.
Retrieves a splittable region randomly from tableName
|
static CompletableFuture<List<HRegionLocation>> |
ClientMetaTableAccessor.getTableHRegionLocations(AsyncTable<AdvancedScanResultConsumer> metaTable,
TableName tableName)
Used to get all region locations for the specific table
|
static List<RegionInfo> |
MetaTableAccessor.getTableRegions(Connection connection,
TableName tableName)
Gets all of the regions of the specified table.
|
static List<RegionInfo> |
MetaTableAccessor.getTableRegions(Connection connection,
TableName tableName,
boolean excludeOfflinedSplitParents)
Gets all of the regions of the specified table.
|
private static CompletableFuture<List<Pair<RegionInfo,ServerName>>> |
ClientMetaTableAccessor.getTableRegionsAndLocations(AsyncTable<AdvancedScanResultConsumer> metaTable,
TableName tableName,
boolean excludeOfflinedSplitParents)
Used to get table regions' info and server.
|
static List<Pair<RegionInfo,ServerName>> |
MetaTableAccessor.getTableRegionsAndLocations(Connection connection,
TableName tableName)
Do not use this method to get meta table regions, use methods in MetaTableLocator instead.
|
static List<Pair<RegionInfo,ServerName>> |
MetaTableAccessor.getTableRegionsAndLocations(Connection connection,
TableName tableName,
boolean excludeOfflinedSplitParents)
Do not use this method to get meta table regions, use methods in MetaTableLocator instead.
|
static byte[] |
ClientMetaTableAccessor.getTableStartRowForMeta(TableName tableName,
ClientMetaTableAccessor.QueryType type)
Returns start row for scanning META according to query type
|
static CompletableFuture<Optional<TableState>> |
ClientMetaTableAccessor.getTableState(AsyncTable<?> metaTable,
TableName tableName) |
static TableState |
MetaTableAccessor.getTableState(Connection conn,
TableName tableName)
Fetch table state for given table from META table
|
static byte[] |
ClientMetaTableAccessor.getTableStopRowForMeta(TableName tableName,
ClientMetaTableAccessor.QueryType type)
Returns stop row for scanning META according to query type
|
static boolean |
TableName.isMetaTableName(TableName tn)
Returns True if
tn is the hbase:meta table name. |
<any> |
HBaseTestingUtility.predicateTableAvailable(TableName tableName)
Deprecated.
Returns a
Predicate for checking that table is enabled |
<any> |
HBaseTestingUtility.predicateTableDisabled(TableName tableName)
Deprecated.
Returns a
Predicate for checking that table is enabled |
<any> |
HBaseTestingUtility.predicateTableEnabled(TableName tableName)
Deprecated.
Returns a
Predicate for checking that table is enabled |
TableDescriptor |
TableDescriptors.remove(TableName tablename)
Returns Instance of table descriptor or null if none found.
|
private static CompletableFuture<Void> |
ClientMetaTableAccessor.scanMeta(AsyncTable<AdvancedScanResultConsumer> metaTable,
TableName tableName,
ClientMetaTableAccessor.QueryType type,
ClientMetaTableAccessor.Visitor visitor)
Performs a scan of META table for given table.
|
static void |
MetaTableAccessor.scanMeta(Connection connection,
ClientMetaTableAccessor.Visitor visitor,
TableName tableName,
byte[] row,
int rowLimit)
Performs a scan of META table for given table starting from given row.
|
private static void |
MetaTableAccessor.scanMeta(Connection connection,
TableName table,
ClientMetaTableAccessor.QueryType type,
int maxRows,
ClientMetaTableAccessor.Visitor visitor) |
static void |
MetaTableAccessor.scanMetaForTableRegions(Connection connection,
ClientMetaTableAccessor.Visitor visitor,
TableName tableName) |
static void |
HBaseTestingUtility.setReplicas(Admin admin,
TableName table,
int replicaCount)
Deprecated.
Set the number of Region replicas.
|
static CompletableFuture<Boolean> |
ClientMetaTableAccessor.tableExists(AsyncTable<?> metaTable,
TableName tableName) |
Table |
HBaseTestingUtility.truncateTable(TableName tableName)
Deprecated.
Truncate a table using the admin command.
|
Table |
HBaseTestingUtility.truncateTable(TableName tableName,
boolean preserveRegions)
Deprecated.
Truncate a table using the admin command.
|
static void |
MetaTableAccessor.updateTableState(Connection conn,
TableName tableName,
TableState.State actual)
Updates state in META Do not use.
|
void |
HBaseTestingUtility.waitTableAvailable(TableName table)
Deprecated.
Wait until all regions in a table have been assigned.
|
void |
HBaseTestingUtility.waitTableAvailable(TableName table,
long timeoutMillis)
Deprecated.
|
void |
HBaseTestingUtility.waitTableDisabled(TableName table,
long millisTimeout)
Deprecated.
|
void |
HBaseTestingUtility.waitTableEnabled(TableName table)
Deprecated.
Waits for a table to be 'enabled'.
|
void |
HBaseTestingUtility.waitTableEnabled(TableName table,
long timeoutMillis)
Deprecated.
|
void |
HBaseTestingUtility.waitUntilAllRegionsAssigned(TableName tableName)
Deprecated.
Wait until all regions for a table in hbase:meta have a non-empty info:server, up to a
configuable timeout value (default is 60 seconds) This means all regions have been deployed,
master has been informed and updated hbase:meta with the regions deployed server.
|
void |
HBaseTestingUtility.waitUntilAllRegionsAssigned(TableName tableName,
long timeout)
Deprecated.
Wait until all regions for a table in hbase:meta have a non-empty info:server, or until
timeout.
|
Modifier and Type | Method and Description |
---|---|
ClusterMetricsBuilder |
ClusterMetricsBuilder.setTableRegionStatesCount(Map<TableName,RegionStatesCount> tableRegionStatesCount) |
Constructor and Description |
---|
ConcurrentTableModificationException(TableName tableName) |
TableExistsException(TableName tableName) |
TableNotDisabledException(TableName tableName) |
TableNotEnabledException(TableName tableName) |
TableNotFoundException(TableName tableName) |
Constructor and Description |
---|
ClusterMetricsImpl(String hbaseVersion,
List<ServerName> deadServerNames,
List<ServerName> unknownServerNames,
Map<ServerName,ServerMetrics> liveServerMetrics,
ServerName masterName,
List<ServerName> backupMasterNames,
List<RegionState> regionsInTransition,
String clusterId,
List<String> masterCoprocessorNames,
Boolean balancerOn,
int masterInfoPort,
List<ServerName> serversName,
Map<TableName,RegionStatesCount> tableRegionStatesCount,
List<ServerTask> masterTasks,
List<ServerName> decommissionedServerNames) |
Modifier and Type | Field and Description |
---|---|
private TableName[] |
RestoreRequest.fromTables |
private TableName |
BackupTableInfo.table |
private TableName[] |
RestoreRequest.toTables |
Modifier and Type | Field and Description |
---|---|
private Map<TableName,BackupTableInfo> |
BackupInfo.backupTableInfoMap
Backup status map for all tables
|
private List<TableName> |
BackupHFileCleaner.fullyBackedUpTables |
private Map<TableName,Map<String,Long>> |
BackupInfo.incrTimestampMap
Previous Region server log timestamps for table set after distributed log roll key - table
name, value - map of RegionServer hostname -> last log rolled timestamp
|
private List<TableName> |
BackupRequest.tableList |
private Map<TableName,Map<String,Long>> |
BackupInfo.tableSetTimestampMap
New region server log timestamps for table set after distributed log roll key - table name,
value - map of RegionServer hostname -> last log rolled timestamp
|
Modifier and Type | Method and Description |
---|---|
TableName[] |
RestoreRequest.getFromTables() |
TableName |
BackupTableInfo.getTable() |
TableName |
BackupInfo.getTableBySnapshot(String snapshotName) |
TableName[] |
RestoreRequest.getToTables() |
Modifier and Type | Method and Description |
---|---|
Map<TableName,Map<String,Long>> |
BackupInfo.getIncrTimestampMap()
Get new region server log timestamps after distributed log roll
|
List<TableName> |
BackupRequest.getTableList() |
List<TableName> |
BackupInfo.getTableNames() |
Set<TableName> |
BackupInfo.getTables() |
Map<TableName,Map<String,Long>> |
BackupInfo.getTableSetTimestampMap() |
private static Map<TableName,Map<String,Long>> |
BackupInfo.getTableSetTimestampMap(Map<String,org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.BackupInfo.RSTimestampMap> map) |
private static Map<TableName,BackupTableInfo> |
BackupInfo.toMap(List<org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.BackupTableInfo> list) |
Modifier and Type | Method and Description |
---|---|
void |
BackupInfo.addTables(TableName[] tables) |
void |
BackupAdmin.addToBackupSet(String name,
TableName[] tables)
Add tables to backup set command
|
static void |
HBackupFileSystem.checkImageManifestExist(HashMap<TableName,BackupManifest> backupManifestMap,
TableName[] tableArray,
org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.Path backupRootPath,
String backupId)
Check whether the backup image path and there is manifest file in the path.
|
BackupTableInfo |
BackupInfo.getBackupTableInfo(TableName table) |
String |
BackupInfo.getSnapshotName(TableName table) |
static String |
HBackupFileSystem.getTableBackupDataDir(String backupRootDir,
String backupId,
TableName tableName) |
static String |
HBackupFileSystem.getTableBackupDir(String backupRootDir,
String backupId,
TableName tableName)
Given the backup root dir, backup id and the table name, return the backup image location,
which is also where the backup manifest file is.
|
String |
BackupInfo.getTableBackupDir(TableName tableName) |
static org.apache.hadoop.fs.Path |
HBackupFileSystem.getTableBackupPath(TableName tableName,
org.apache.hadoop.fs.Path backupRootPath,
String backupId)
Given the backup root dir, backup id and the table name, return the backup image location,
which is also where the backup manifest file is.
|
void |
BackupAdmin.removeFromBackupSet(String name,
TableName[] tables)
Remove tables from backup set
|
void |
RestoreJob.run(org.apache.hadoop.fs.Path[] dirPaths,
TableName[] fromTables,
org.apache.hadoop.fs.Path restoreRootDir,
TableName[] toTables,
boolean fullBackupRestore)
Run restore operation
|
void |
RestoreJob.run(org.apache.hadoop.fs.Path[] dirPaths,
TableName[] fromTables,
org.apache.hadoop.fs.Path restoreRootDir,
TableName[] toTables,
boolean fullBackupRestore)
Run restore operation
|
private RestoreRequest |
RestoreRequest.setFromTables(TableName[] fromTables) |
void |
BackupInfo.setSnapshotName(TableName table,
String snapshotName) |
private RestoreRequest |
RestoreRequest.setToTables(TableName[] toTables) |
RestoreRequest.Builder |
RestoreRequest.Builder.withFromTables(TableName[] fromTables) |
RestoreRequest.Builder |
RestoreRequest.Builder.withToTables(TableName[] toTables) |
Modifier and Type | Method and Description |
---|---|
static void |
HBackupFileSystem.checkImageManifestExist(HashMap<TableName,BackupManifest> backupManifestMap,
TableName[] tableArray,
org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.Path backupRootPath,
String backupId)
Check whether the backup image path and there is manifest file in the path.
|
private Set<String> |
BackupHFileCleaner.loadHFileRefs(List<TableName> tableList) |
void |
BackupInfo.setBackupTableInfoMap(Map<TableName,BackupTableInfo> backupTableInfoMap) |
void |
BackupInfo.setIncrTimestampMap(Map<TableName,Map<String,Long>> prevTableSetTimestampMap)
Set the new region server log timestamps after distributed log roll
|
private BackupRequest |
BackupRequest.setTableList(List<TableName> tableList) |
void |
BackupInfo.setTables(List<TableName> tables) |
void |
BackupInfo.setTableSetTimestampMap(Map<TableName,Map<String,Long>> tableSetTimestampMap) |
BackupRequest.Builder |
BackupRequest.Builder.withTableList(List<TableName> tables) |
Constructor and Description |
---|
BackupInfo(String backupId,
BackupType type,
TableName[] tables,
String targetRootDir) |
BackupTableInfo(TableName table,
String targetRootDir,
String backupId) |
Modifier and Type | Field and Description |
---|---|
private TableName |
BackupSystemTable.bulkLoadTableName
Backup System table name for bulk loaded files.
|
private TableName[] |
RestoreTablesClient.sTableArray |
private TableName |
BackupSystemTable.tableName
Backup system table (main) name
|
private TableName[] |
RestoreTablesClient.tTableArray |
Modifier and Type | Field and Description |
---|---|
private Map<TableName,Map<String,Long>> |
BackupManifest.BackupImage.incrTimeRanges |
private List<TableName> |
BackupManifest.BackupImage.tableList |
protected List<TableName> |
TableBackupClient.tableList |
Modifier and Type | Method and Description |
---|---|
private TableName |
BackupCommands.HistoryCommand.getTableName() |
static TableName |
BackupSystemTable.getTableName(org.apache.hadoop.conf.Configuration conf) |
static TableName |
BackupSystemTable.getTableNameForBulkLoadedData(org.apache.hadoop.conf.Configuration conf) |
private TableName[] |
BackupCommands.BackupSetCommand.toTableNames(String[] tables) |
Modifier and Type | Method and Description |
---|---|
List<TableName> |
BackupSystemTable.describeBackupSet(String name)
Get backup set description (list of tables)
|
private List<TableName> |
BackupAdminImpl.excludeNonExistingTables(List<TableName> tableList,
List<TableName> nonExistingTableList) |
Map<TableName,ArrayList<BackupInfo>> |
BackupSystemTable.getBackupHistoryForTableSet(Set<TableName> set,
String backupRoot) |
Set<TableName> |
BackupManager.getIncrementalBackupTableSet()
Return the current tables covered by incremental backup.
|
Set<TableName> |
BackupSystemTable.getIncrementalBackupTableSet(String backupRoot)
Return the current tables covered by incremental backup.
|
Map<TableName,Map<String,Long>> |
BackupManifest.BackupImage.getIncrTimeRanges() |
Map<TableName,Map<String,Long>> |
BackupManifest.getIncrTimestampMap() |
List<TableName> |
BackupManifest.getTableList()
Get the table set of this image.
|
List<TableName> |
BackupManifest.BackupImage.getTableNames() |
List<TableName> |
BackupSystemTable.getTablesForBackupType(BackupType type) |
private static Map<TableName,Map<String,Long>> |
BackupManifest.BackupImage.loadIncrementalTimestampMap(org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.BackupImage proto) |
Pair<Map<TableName,Map<String,Map<String,List<Pair<String,Boolean>>>>>,List<byte[]>> |
BackupManager.readBulkloadRows(List<TableName> tableList) |
Pair<Map<TableName,Map<String,Map<String,List<Pair<String,Boolean>>>>>,List<byte[]>> |
BackupSystemTable.readBulkloadRows(List<TableName> tableList) |
Map<TableName,Map<String,Long>> |
BackupManager.readLogTimestampMap()
Read the timestamp for each region server log after the last successful backup.
|
Map<TableName,Map<String,Long>> |
BackupSystemTable.readLogTimestampMap(String backupRoot)
Read the timestamp for each region server log after the last successful backup.
|
Modifier and Type | Method and Description |
---|---|
void |
BackupAdminImpl.addToBackupSet(String name,
TableName[] tables) |
private void |
RestoreTablesClient.checkTargetTables(TableName[] tTableArray,
boolean isOverwrite)
Validate target tables.
|
private void |
BackupAdminImpl.cleanupBackupDir(BackupInfo backupInfo,
TableName table,
org.apache.hadoop.conf.Configuration conf)
Clean up the data at target directory
|
(package private) static Put |
BackupSystemTable.createPutForBulkLoadedFile(TableName tn,
byte[] fam,
String p,
String backupId,
long ts,
int idx) |
(package private) static List<Put> |
BackupSystemTable.createPutForCommittedBulkload(TableName table,
byte[] region,
Map<byte[],List<org.apache.hadoop.fs.Path>> finalPaths) |
(package private) static List<Put> |
BackupSystemTable.createPutForPreparedBulkload(TableName table,
byte[] region,
byte[] family,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs) |
private Put |
BackupSystemTable.createPutForWriteRegionServerLogTimestamp(TableName table,
byte[] smap,
String backupRoot)
Creates Put to write RS last roll log timestamp map
|
(package private) static Scan |
BackupSystemTable.createScanForOrigBulkLoadedFiles(TableName table) |
private List<BackupInfo> |
BackupAdminImpl.getAffectedBackupSessions(BackupInfo backupInfo,
TableName tn,
BackupSystemTable table) |
ArrayList<BackupManifest.BackupImage> |
BackupManifest.getAllDependentListByTable(TableName table)
Get the full dependent image list in the whole dependency scope for a specific table of this
backup in time order from old to new.
|
ArrayList<BackupManifest.BackupImage> |
BackupManager.getAncestors(BackupInfo backupInfo,
TableName table)
Get the direct ancestors of this backup for one table involved.
|
List<BackupInfo> |
BackupSystemTable.getBackupHistoryForTable(TableName name)
Get history for a table
|
protected org.apache.hadoop.fs.Path |
IncrementalTableBackupClient.getBulkOutputDirForTable(TableName table) |
ArrayList<BackupManifest.BackupImage> |
BackupManifest.getDependentListByTable(TableName table)
Get the dependent image list for a specific table of this backup in time order from old to new
if want to restore to this backup image level.
|
protected static int |
IncrementalTableBackupClient.getIndex(TableName tbl,
List<TableName> sTableList) |
boolean |
BackupManifest.BackupImage.hasTable(TableName table) |
private boolean |
BackupAdminImpl.isLastBackupSession(BackupSystemTable table,
TableName tn,
long startTime) |
void |
BackupAdminImpl.removeFromBackupSet(String name,
TableName[] tables) |
private void |
BackupAdminImpl.removeTableFromBackupImage(BackupInfo info,
TableName tn,
BackupSystemTable sysTable) |
private void |
RestoreTablesClient.restore(HashMap<TableName,BackupManifest> backupManifestMap,
TableName[] sTableArray,
TableName[] tTableArray,
boolean isOverwrite)
Restore operation.
|
private void |
RestoreTablesClient.restore(HashMap<TableName,BackupManifest> backupManifestMap,
TableName[] sTableArray,
TableName[] tTableArray,
boolean isOverwrite)
Restore operation.
|
private void |
RestoreTablesClient.restoreImages(BackupManifest.BackupImage[] images,
TableName sTable,
TableName tTable,
boolean truncateIfExists)
Restore operation handle each backupImage in array.
|
protected void |
FullTableBackupClient.snapshotTable(Admin admin,
TableName tableName,
String snapshotName) |
protected boolean |
IncrementalTableBackupClient.tableExists(TableName table,
Connection conn) |
private String[] |
BackupAdminImpl.toStringArray(TableName[] list) |
private org.apache.hadoop.hbase.shaded.protobuf.generated.BackupProtos.TableServerTimestamp |
BackupSystemTable.toTableServerTimestampProto(TableName table,
Map<String,Long> map) |
private void |
BackupSystemTable.waitForSystemTable(Admin admin,
TableName tableName) |
void |
BackupSystemTable.writeFilesForBulkLoadPreCommit(TableName tabName,
byte[] region,
byte[] family,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs) |
void |
BackupSystemTable.writePathsPostBulkLoad(TableName tabName,
byte[] region,
Map<byte[],List<org.apache.hadoop.fs.Path>> finalPaths) |
Modifier and Type | Method and Description |
---|---|
void |
BackupManager.addIncrementalBackupTableSet(Set<TableName> tables)
Adds set of tables to overall incremental backup table set
|
void |
BackupSystemTable.addIncrementalBackupTableSet(Set<TableName> tables,
String backupRoot)
Add tables to global incremental backup set
|
BackupInfo |
BackupManager.createBackupInfo(String backupId,
BackupType type,
List<TableName> tableList,
String targetRootDir,
int workers,
long bandwidth)
Creates a backup info based on input backup request.
|
static List<Delete> |
BackupSystemTable.createDeleteForOrigBulkLoad(List<TableName> lst) |
private Put |
BackupSystemTable.createPutForIncrBackupTableSet(Set<TableName> tables,
String backupRoot)
Creates Put to store incremental backup table set
|
private Put |
BackupSystemTable.createPutForUpdateTablesForMerge(List<TableName> tables) |
private List<TableName> |
BackupAdminImpl.excludeNonExistingTables(List<TableName> tableList,
List<TableName> nonExistingTableList) |
private List<TableName> |
BackupAdminImpl.excludeNonExistingTables(List<TableName> tableList,
List<TableName> nonExistingTableList) |
private void |
BackupAdminImpl.finalizeDelete(Map<String,HashSet<TableName>> tablesMap,
BackupSystemTable table)
Updates incremental backup set for every backupRoot
|
Map<TableName,ArrayList<BackupInfo>> |
BackupSystemTable.getBackupHistoryForTableSet(Set<TableName> set,
String backupRoot) |
protected static int |
IncrementalTableBackupClient.getIndex(TableName tbl,
List<TableName> sTableList) |
protected Map<byte[],List<org.apache.hadoop.fs.Path>>[] |
IncrementalTableBackupClient.handleBulkLoad(List<TableName> sTableList) |
Map<byte[],List<org.apache.hadoop.fs.Path>>[] |
BackupSystemTable.readBulkLoadedFiles(String backupId,
List<TableName> sTableList) |
Pair<Map<TableName,Map<String,Map<String,List<Pair<String,Boolean>>>>>,List<byte[]>> |
BackupManager.readBulkloadRows(List<TableName> tableList) |
Pair<Map<TableName,Map<String,Map<String,List<Pair<String,Boolean>>>>>,List<byte[]>> |
BackupSystemTable.readBulkloadRows(List<TableName> tableList) |
private void |
RestoreTablesClient.restore(HashMap<TableName,BackupManifest> backupManifestMap,
TableName[] sTableArray,
TableName[] tTableArray,
boolean isOverwrite)
Restore operation.
|
private void |
BackupManifest.BackupImage.setIncrTimeRanges(Map<TableName,Map<String,Long>> incrTimeRanges) |
void |
BackupManifest.setIncrTimestampMap(Map<TableName,Map<String,Long>> incrTimestampMap)
Set the incremental timestamp map directly.
|
private void |
BackupManifest.BackupImage.setTableList(List<TableName> tableList) |
void |
BackupSystemTable.updateProcessedTablesForMerge(List<TableName> tables) |
(package private) BackupManifest.BackupImage.Builder |
BackupManifest.BackupImage.Builder.withTableList(List<TableName> tableList) |
void |
BackupSystemTable.writeBulkLoadedFiles(List<TableName> sTableList,
Map<byte[],List<org.apache.hadoop.fs.Path>>[] maps,
String backupId) |
void |
BackupManager.writeRegionServerLogTimestamp(Set<TableName> tables,
Map<String,Long> newTimestamps)
Write the current timestamps for each regionserver to backup system table after a successful
full or incremental backup.
|
void |
BackupSystemTable.writeRegionServerLogTimestamp(Set<TableName> tables,
Map<String,Long> newTimestamps,
String backupRoot)
Write the current timestamps for each regionserver to backup system table after a successful
full or incremental backup.
|
Constructor and Description |
---|
BackupManifest(BackupInfo backup,
TableName table)
Construct a table level manifest for a backup of the named table.
|
Constructor and Description |
---|
BackupImage(String backupId,
BackupType type,
String rootDir,
List<TableName> tableList,
long startTs,
long completeTs) |
Modifier and Type | Field and Description |
---|---|
private TableName |
MapReduceBackupCopyJob.SnapshotCopy.table |
Modifier and Type | Method and Description |
---|---|
TableName |
MapReduceBackupCopyJob.SnapshotCopy.getTable() |
protected TableName[] |
MapReduceBackupMergeJob.getTableNamesInBackupImages(String[] backupIds) |
Modifier and Type | Method and Description |
---|---|
protected List<TableName> |
MapReduceBackupMergeJob.toTableNameList(List<Pair<TableName,org.apache.hadoop.fs.Path>> processedTableList) |
Modifier and Type | Method and Description |
---|---|
protected org.apache.hadoop.fs.Path[] |
MapReduceBackupMergeJob.findInputDirectories(org.apache.hadoop.fs.FileSystem fs,
String backupRoot,
TableName tableName,
String[] backupIds) |
protected void |
MapReduceBackupMergeJob.moveData(org.apache.hadoop.fs.FileSystem fs,
String backupRoot,
org.apache.hadoop.fs.Path bulkOutputPath,
TableName tableName,
String mergedBackupId) |
void |
MapReduceRestoreJob.run(org.apache.hadoop.fs.Path[] dirPaths,
TableName[] tableNames,
org.apache.hadoop.fs.Path restoreRootDir,
TableName[] newTableNames,
boolean fullBackupRestore) |
void |
MapReduceRestoreJob.run(org.apache.hadoop.fs.Path[] dirPaths,
TableName[] tableNames,
org.apache.hadoop.fs.Path restoreRootDir,
TableName[] newTableNames,
boolean fullBackupRestore) |
Modifier and Type | Method and Description |
---|---|
protected List<org.apache.hadoop.fs.Path> |
MapReduceBackupMergeJob.toPathList(List<Pair<TableName,org.apache.hadoop.fs.Path>> processedTableList) |
protected List<TableName> |
MapReduceBackupMergeJob.toTableNameList(List<Pair<TableName,org.apache.hadoop.fs.Path>> processedTableList) |
Constructor and Description |
---|
SnapshotCopy(BackupInfo backupInfo,
TableName table) |
Modifier and Type | Field and Description |
---|---|
private HashMap<TableName,org.apache.hadoop.fs.Path> |
RestoreTool.snapshotMap |
private List<TableName> |
BackupSet.tables |
Modifier and Type | Method and Description |
---|---|
static TableName[] |
BackupUtils.parseTableNames(String tables) |
Modifier and Type | Method and Description |
---|---|
List<TableName> |
BackupSet.getTables() |
Modifier and Type | Method and Description |
---|---|
private void |
RestoreTool.checkAndCreateTable(Connection conn,
TableName targetTableName,
ArrayList<org.apache.hadoop.fs.Path> regionDirList,
TableDescriptor htd,
boolean truncateIfExists)
Prepare the table for bulkload, most codes copied from
createTable method in
BulkLoadHFilesTool . |
private void |
RestoreTool.createAndRestoreTable(Connection conn,
TableName tableName,
TableName newTableName,
org.apache.hadoop.fs.Path tableBackupPath,
boolean truncateIfExists,
String lastIncrBackupId) |
static RestoreRequest |
BackupUtils.createRestoreRequest(String backupRootDir,
String backupId,
boolean check,
TableName[] fromTables,
TableName[] toTables,
boolean isOverwrite)
Create restore request.
|
static RestoreRequest |
BackupUtils.createRestoreRequest(String backupRootDir,
String backupId,
boolean check,
TableName[] fromTables,
TableName[] toTables,
boolean isOverwrite)
Create restore request.
|
void |
RestoreTool.fullRestoreTable(Connection conn,
org.apache.hadoop.fs.Path tableBackupPath,
TableName tableName,
TableName newTableName,
boolean truncateIfExists,
String lastIncrBackupId) |
static String |
BackupUtils.getFileNameCompatibleString(TableName table) |
(package private) ArrayList<org.apache.hadoop.fs.Path> |
RestoreTool.getRegionList(TableName tableName)
Gets region list
|
(package private) org.apache.hadoop.fs.Path |
RestoreTool.getTableArchivePath(TableName tableName)
return value represent path for:
".../user/biadmin/backup1/default/t1_dn/backup_1396650096738/archive/data/default/t1_dn"
|
static String |
BackupUtils.getTableBackupDir(String backupRootDir,
String backupId,
TableName tableName)
Given the backup root dir, backup id and the table name, return the backup image location,
which is also where the backup manifest file is.
|
(package private) TableDescriptor |
RestoreTool.getTableDesc(TableName tableName)
Get table descriptor
|
private TableDescriptor |
RestoreTool.getTableDescriptor(org.apache.hadoop.fs.FileSystem fileSys,
TableName tableName,
String lastIncrBackupId) |
(package private) org.apache.hadoop.fs.Path |
RestoreTool.getTableInfoPath(TableName tableName)
Returns value represent path for:
""/$USER/SBACKUP_ROOT/backup_id/namespace/table/.hbase-snapshot/
snapshot_1396650097621_namespace_table" this path contains .snapshotinfo, .tabledesc (0.96 and
0.98) this path contains .snapshotinfo, .data.manifest (trunk)
|
(package private) org.apache.hadoop.fs.Path |
RestoreTool.getTableSnapshotPath(org.apache.hadoop.fs.Path backupRootPath,
TableName tableName,
String backupId)
Returns value represent path for path to backup table snapshot directory:
"/$USER/SBACKUP_ROOT/backup_id/namespace/table/.hbase-snapshot"
|
void |
RestoreTool.incrementalRestoreTable(Connection conn,
org.apache.hadoop.fs.Path tableBackupPath,
org.apache.hadoop.fs.Path[] logDirs,
TableName[] tableNames,
TableName[] newTableNames,
String incrBackupId)
During incremental backup operation.
|
void |
RestoreTool.incrementalRestoreTable(Connection conn,
org.apache.hadoop.fs.Path tableBackupPath,
org.apache.hadoop.fs.Path[] logDirs,
TableName[] tableNames,
TableName[] newTableNames,
String incrBackupId)
During incremental backup operation.
|
Modifier and Type | Method and Description |
---|---|
static Map<String,Long> |
BackupUtils.getRSLogTimestampMins(Map<TableName,Map<String,Long>> rsLogTimestampMap)
Loop through the RS log timestamp map for the tables, for each RS, find the min timestamp value
for the RS among the tables.
|
static boolean |
BackupUtils.validate(HashMap<TableName,BackupManifest> backupManifestMap,
org.apache.hadoop.conf.Configuration conf) |
Constructor and Description |
---|
BackupSet(String name,
List<TableName> tables) |
Modifier and Type | Field and Description |
---|---|
private TableName |
TableDescriptorBuilder.ModifyableTableDescriptor.name |
private TableName |
SnapshotDescription.table |
protected TableName |
AsyncTableBuilderBase.tableName |
protected TableName |
TableBuilderBase.tableName |
protected TableName |
RawAsyncHBaseAdmin.TableProcedureBiConsumer.tableName |
private TableName |
AsyncBatchRpcRetryingCaller.tableName |
private TableName |
AsyncSingleRequestRpcRetryingCaller.tableName |
private TableName |
AsyncRegionLocationCache.tableName |
private TableName |
CatalogReplicaLoadBalanceSimpleSelector.tableName |
private TableName |
TableState.tableName |
private TableName |
AsyncTableResultScanner.tableName |
private TableName |
AsyncClientScanner.tableName |
private TableName |
RegionCoprocessorRpcChannelImpl.tableName |
private TableName |
RawAsyncTableImpl.tableName |
private TableName |
AsyncTableRegionLocatorImpl.tableName |
private TableName |
BufferedMutatorParams.tableName |
private TableName |
RegionInfoBuilder.tableName |
private TableName |
MutableRegionInfo.tableName |
private TableName |
AsyncRpcRetryingCallerFactory.SingleRequestCallerBuilder.tableName |
private TableName |
AsyncRpcRetryingCallerFactory.BatchCallerBuilder.tableName |
Modifier and Type | Field and Description |
---|---|
private ConcurrentMap<TableName,AsyncNonMetaRegionLocator.TableCache> |
AsyncNonMetaRegionLocator.cache |
private ConcurrentMap<TableName,ConcurrentNavigableMap<byte[],CatalogReplicaLoadBalanceSimpleSelector.StaleLocationCacheEntry>> |
CatalogReplicaLoadBalanceSimpleSelector.staleCache |
private List<TableName> |
NormalizeTableFilterParams.tableNames |
private List<TableName> |
NormalizeTableFilterParams.Builder.tableNames |
Modifier and Type | Method and Description |
---|---|
private static TableName |
MutableRegionInfo.checkTableName(TableName tableName) |
TableName |
RegionLocator.getName()
Gets the fully qualified table name instance of this table.
|
TableName |
Table.getName()
Gets the fully qualified table name instance of this table.
|
TableName |
BufferedMutatorOverAsyncBufferedMutator.getName() |
TableName |
RegionLocatorOverAsyncTableRegionLocator.getName() |
TableName |
BufferedMutator.getName()
Gets the fully qualified table name instance of the table that this BufferedMutator writes to.
|
TableName |
AsyncTableRegionLocator.getName()
Gets the fully qualified table name instance of the table whose region we want to locate.
|
TableName |
AsyncTableImpl.getName() |
TableName |
AsyncTable.getName()
Gets the fully qualified table name instance of this table.
|
TableName |
RawAsyncTableImpl.getName() |
TableName |
AsyncTableRegionLocatorImpl.getName() |
TableName |
AsyncBufferedMutatorImpl.getName() |
TableName |
AsyncBufferedMutator.getName()
Gets the fully qualified table name instance of the table that this
AsyncBufferedMutator writes to. |
TableName |
TableOverAsyncTable.getName() |
TableName |
RegionInfo.getTable()
Returns current table name of the region
|
TableName |
MutableRegionInfo.getTable()
Get current table name of the region
|
static TableName |
RegionInfo.getTable(byte[] regionName)
Gets the table name from the specified region name.
|
TableName |
TableDescriptor.getTableName()
Get the name of the table
|
TableName |
TableDescriptorBuilder.ModifyableTableDescriptor.getTableName()
Get the name of the table
|
TableName |
TableState.getTableName()
Table name for state
|
TableName |
SnapshotDescription.getTableName() |
TableName |
BufferedMutatorParams.getTableName() |
TableName[] |
Admin.listTableNames()
List all of the names of userspace tables.
|
TableName[] |
AdminOverAsyncAdmin.listTableNames() |
default TableName[] |
Admin.listTableNames(Pattern pattern)
List all of the names of userspace tables.
|
TableName[] |
Admin.listTableNames(Pattern pattern,
boolean includeSysTables)
List all of the names of userspace tables.
|
TableName[] |
AdminOverAsyncAdmin.listTableNames(Pattern pattern,
boolean includeSysTables) |
TableName[] |
Admin.listTableNamesByNamespace(String name)
Get list of table names by namespace.
|
TableName[] |
AdminOverAsyncAdmin.listTableNamesByNamespace(String name) |
Modifier and Type | Method and Description |
---|---|
private CompletableFuture<TableName> |
RawAsyncHBaseAdmin.checkRegionsAndGetTableName(byte[][] encodedRegionNames) |
CompletableFuture<Pair<List<String>,List<TableName>>> |
AsyncAdmin.getConfiguredNamespacesAndTablesInRSGroup(String groupName)
Get the namespaces and tables which have this RegionServer group in descriptor.
|
CompletableFuture<Pair<List<String>,List<TableName>>> |
RawAsyncHBaseAdmin.getConfiguredNamespacesAndTablesInRSGroup(String groupName) |
Pair<List<String>,List<TableName>> |
Admin.getConfiguredNamespacesAndTablesInRSGroup(String groupName)
Get the namespaces and tables which have this RegionServer group in descriptor.
|
Pair<List<String>,List<TableName>> |
AdminOverAsyncAdmin.getConfiguredNamespacesAndTablesInRSGroup(String groupName) |
CompletableFuture<Pair<List<String>,List<TableName>>> |
AsyncHBaseAdmin.getConfiguredNamespacesAndTablesInRSGroup(String groupName) |
CompletableFuture<? extends Map<TableName,? extends SpaceQuotaSnapshotView>> |
AsyncAdmin.getRegionServerSpaceQuotaSnapshots(ServerName serverName)
Fetches the observed
SpaceQuotaSnapshotView s observed by a RegionServer. |
CompletableFuture<Map<TableName,SpaceQuotaSnapshot>> |
RawAsyncHBaseAdmin.getRegionServerSpaceQuotaSnapshots(ServerName serverName) |
Map<TableName,? extends SpaceQuotaSnapshotView> |
Admin.getRegionServerSpaceQuotaSnapshots(ServerName serverName)
Fetches the observed
SpaceQuotaSnapshotView s observed by a RegionServer. |
Map<TableName,? extends SpaceQuotaSnapshotView> |
AdminOverAsyncAdmin.getRegionServerSpaceQuotaSnapshots(ServerName serverName) |
CompletableFuture<Map<TableName,SpaceQuotaSnapshot>> |
AsyncHBaseAdmin.getRegionServerSpaceQuotaSnapshots(ServerName serverName) |
CompletableFuture<Map<TableName,Long>> |
AsyncAdmin.getSpaceQuotaTableSizes()
Fetches the table sizes on the filesystem as tracked by the HBase Master.
|
CompletableFuture<Map<TableName,Long>> |
RawAsyncHBaseAdmin.getSpaceQuotaTableSizes() |
Map<TableName,Long> |
Admin.getSpaceQuotaTableSizes()
Fetches the table sizes on the filesystem as tracked by the HBase Master.
|
Map<TableName,Long> |
AdminOverAsyncAdmin.getSpaceQuotaTableSizes() |
CompletableFuture<Map<TableName,Long>> |
AsyncHBaseAdmin.getSpaceQuotaTableSizes() |
protected Optional<TableName> |
AsyncSingleRequestRpcRetryingCaller.getTableName() |
protected Optional<TableName> |
AsyncRpcRetryingCaller.getTableName() |
List<TableName> |
NormalizeTableFilterParams.getTableNames() |
private CompletableFuture<List<TableName>> |
RawAsyncHBaseAdmin.getTableNames(org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.GetTableNamesRequest request) |
default CompletableFuture<List<TableName>> |
AsyncAdmin.listTableNames()
List all of the names of userspace tables.
|
CompletableFuture<List<TableName>> |
AsyncAdmin.listTableNames(boolean includeSysTables)
List all of the names of tables.
|
CompletableFuture<List<TableName>> |
RawAsyncHBaseAdmin.listTableNames(boolean includeSysTables) |
CompletableFuture<List<TableName>> |
AsyncHBaseAdmin.listTableNames(boolean includeSysTables) |
CompletableFuture<List<TableName>> |
AsyncAdmin.listTableNames(Pattern pattern,
boolean includeSysTables)
List all of the names of userspace tables.
|
CompletableFuture<List<TableName>> |
RawAsyncHBaseAdmin.listTableNames(Pattern pattern,
boolean includeSysTables) |
CompletableFuture<List<TableName>> |
AsyncHBaseAdmin.listTableNames(Pattern pattern,
boolean includeSysTables) |
CompletableFuture<List<TableName>> |
AsyncAdmin.listTableNamesByNamespace(String name)
Get list of table names by namespace.
|
CompletableFuture<List<TableName>> |
RawAsyncHBaseAdmin.listTableNamesByNamespace(String name) |
CompletableFuture<List<TableName>> |
AsyncHBaseAdmin.listTableNamesByNamespace(String name) |
CompletableFuture<List<TableName>> |
AsyncAdmin.listTableNamesByState(boolean isEnabled)
List all enabled or disabled table names
|
CompletableFuture<List<TableName>> |
RawAsyncHBaseAdmin.listTableNamesByState(boolean isEnabled) |
List<TableName> |
Admin.listTableNamesByState(boolean isEnabled)
List all enabled or disabled table names
|
List<TableName> |
AdminOverAsyncAdmin.listTableNamesByState(boolean isEnabled) |
CompletableFuture<List<TableName>> |
AsyncHBaseAdmin.listTableNamesByState(boolean isEnabled) |
CompletableFuture<List<TableName>> |
AsyncAdmin.listTablesInRSGroup(String groupName)
Get all tables in this RegionServer group.
|
CompletableFuture<List<TableName>> |
RawAsyncHBaseAdmin.listTablesInRSGroup(String groupName) |
List<TableName> |
Admin.listTablesInRSGroup(String groupName)
Get all tables in this RegionServer group.
|
List<TableName> |
AdminOverAsyncAdmin.listTablesInRSGroup(String groupName) |
CompletableFuture<List<TableName>> |
AsyncHBaseAdmin.listTablesInRSGroup(String groupName) |
Modifier and Type | Method and Description |
---|---|
CompletableFuture<Void> |
AsyncAdmin.addColumnFamily(TableName tableName,
ColumnFamilyDescriptor columnFamily)
Add a column family to an existing table.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.addColumnFamily(TableName tableName,
ColumnFamilyDescriptor columnFamily) |
default void |
Admin.addColumnFamily(TableName tableName,
ColumnFamilyDescriptor columnFamily)
Add a column family to an existing table.
|
CompletableFuture<Void> |
AsyncHBaseAdmin.addColumnFamily(TableName tableName,
ColumnFamilyDescriptor columnFamily) |
Future<Void> |
Admin.addColumnFamilyAsync(TableName tableName,
ColumnFamilyDescriptor columnFamily)
Add a column family to an existing table.
|
Future<Void> |
AdminOverAsyncAdmin.addColumnFamilyAsync(TableName tableName,
ColumnFamilyDescriptor columnFamily) |
CompletableFuture<Boolean> |
AsyncClusterConnection.bulkLoad(TableName tableName,
List<Pair<byte[],String>> familyPaths,
byte[] row,
boolean assignSeqNum,
org.apache.hadoop.security.token.Token<?> userToken,
String bulkToken,
boolean copyFiles,
List<String> clusterIds,
boolean replicate)
Securely bulk load a list of HFiles, passing additional list of clusters ids tracking clusters
where the given bulk load has already been processed (important for bulk loading replication).
|
CompletableFuture<Boolean> |
AsyncClusterConnectionImpl.bulkLoad(TableName tableName,
List<Pair<byte[],String>> familyPaths,
byte[] row,
boolean assignSeqNum,
org.apache.hadoop.security.token.Token<?> userToken,
String bulkToken,
boolean copyFiles,
List<String> clusterIds,
boolean replicate) |
(package private) static int |
ConnectionUtils.calcPriority(int priority,
TableName tableName)
Select the priority for the rpc call.
|
private CompletableFuture<Void> |
RawAsyncHBaseAdmin.checkAndSyncTableToPeerClusters(TableName tableName,
byte[][] splits)
Connect to peer and check the table descriptor on peer:
Create the same table on peer when not exist.
Throw an exception if the table already has replication enabled on any of the column
families.
Throw an exception if the table exists on peer cluster but descriptors are not same.
|
private static TableName |
MutableRegionInfo.checkTableName(TableName tableName) |
CompletableFuture<Void> |
AsyncClusterConnection.cleanupBulkLoad(TableName tableName,
String bulkToken)
Clean up after finishing bulk load, no matter success or not.
|
CompletableFuture<Void> |
AsyncClusterConnectionImpl.cleanupBulkLoad(TableName tableName,
String bulkToken) |
CompletableFuture<CacheEvictionStats> |
AsyncAdmin.clearBlockCache(TableName tableName)
Clear all the blocks corresponding to this table from BlockCache.
|
CompletableFuture<CacheEvictionStats> |
RawAsyncHBaseAdmin.clearBlockCache(TableName tableName) |
CacheEvictionStats |
Admin.clearBlockCache(TableName tableName)
Clear all the blocks corresponding to this table from BlockCache.
|
CacheEvictionStats |
AdminOverAsyncAdmin.clearBlockCache(TableName tableName) |
CompletableFuture<CacheEvictionStats> |
AsyncHBaseAdmin.clearBlockCache(TableName tableName) |
(package private) void |
AsyncRegionLocator.clearCache(TableName tableName) |
(package private) void |
AsyncNonMetaRegionLocator.clearCache(TableName tableName) |
default CompletableFuture<Void> |
AsyncAdmin.cloneSnapshot(String snapshotName,
TableName tableName)
Create a new table by cloning the snapshot content.
|
default void |
Admin.cloneSnapshot(String snapshotName,
TableName tableName)
Create a new table by cloning the snapshot content.
|
default CompletableFuture<Void> |
AsyncAdmin.cloneSnapshot(String snapshotName,
TableName tableName,
boolean restoreAcl)
Create a new table by cloning the snapshot content.
|
default void |
Admin.cloneSnapshot(String snapshotName,
TableName tableName,
boolean restoreAcl)
Create a new table by cloning the snapshot content.
|
CompletableFuture<Void> |
AsyncAdmin.cloneSnapshot(String snapshotName,
TableName tableName,
boolean restoreAcl,
String customSFT)
Create a new table by cloning the snapshot content.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.cloneSnapshot(String snapshotName,
TableName tableName,
boolean restoreAcl,
String customSFT) |
default void |
Admin.cloneSnapshot(String snapshotName,
TableName tableName,
boolean restoreAcl,
String customSFT)
Create a new table by cloning the snapshot content.
|
CompletableFuture<Void> |
AsyncHBaseAdmin.cloneSnapshot(String snapshotName,
TableName tableName,
boolean restoreAcl,
String customSFT) |
default Future<Void> |
Admin.cloneSnapshotAsync(String snapshotName,
TableName tableName)
Create a new table by cloning the snapshot content, but does not block and wait for it to be
completely cloned.
|
default Future<Void> |
Admin.cloneSnapshotAsync(String snapshotName,
TableName tableName,
boolean restoreAcl)
Create a new table by cloning the snapshot content.
|
Future<Void> |
Admin.cloneSnapshotAsync(String snapshotName,
TableName tableName,
boolean restoreAcl,
String customSFT)
Create a new table by cloning the snapshot content.
|
Future<Void> |
AdminOverAsyncAdmin.cloneSnapshotAsync(String snapshotName,
TableName tableName,
boolean restoreAcl,
String customSFT) |
CompletableFuture<Void> |
AsyncAdmin.cloneTableSchema(TableName tableName,
TableName newTableName,
boolean preserveSplits)
Create a new table by cloning the existent table schema.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.cloneTableSchema(TableName tableName,
TableName newTableName,
boolean preserveSplits) |
void |
Admin.cloneTableSchema(TableName tableName,
TableName newTableName,
boolean preserveSplits)
Create a new table by cloning the existent table schema.
|
void |
AdminOverAsyncAdmin.cloneTableSchema(TableName tableName,
TableName newTableName,
boolean preserveSplits) |
CompletableFuture<Void> |
AsyncHBaseAdmin.cloneTableSchema(TableName tableName,
TableName newTableName,
boolean preserveSplits) |
default CompletableFuture<Void> |
AsyncAdmin.compact(TableName tableName)
Compact a table.
|
void |
Admin.compact(TableName tableName)
Compact a table.
|
void |
AdminOverAsyncAdmin.compact(TableName tableName) |
default CompletableFuture<Void> |
AsyncAdmin.compact(TableName tableName,
byte[] columnFamily)
Compact a column family within a table.
|
void |
Admin.compact(TableName tableName,
byte[] columnFamily)
Compact a column family within a table.
|
void |
AdminOverAsyncAdmin.compact(TableName tableName,
byte[] columnFamily) |
private CompletableFuture<Void> |
RawAsyncHBaseAdmin.compact(TableName tableName,
byte[] columnFamily,
boolean major,
CompactType compactType)
Compact column family of a table, Asynchronous operation even if CompletableFuture.get()
|
CompletableFuture<Void> |
AsyncAdmin.compact(TableName tableName,
byte[] columnFamily,
CompactType compactType)
Compact a column family within a table.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.compact(TableName tableName,
byte[] columnFamily,
CompactType compactType) |
void |
Admin.compact(TableName tableName,
byte[] columnFamily,
CompactType compactType)
Compact a column family within a table.
|
void |
AdminOverAsyncAdmin.compact(TableName tableName,
byte[] columnFamily,
CompactType compactType) |
CompletableFuture<Void> |
AsyncHBaseAdmin.compact(TableName tableName,
byte[] columnFamily,
CompactType compactType) |
CompletableFuture<Void> |
AsyncAdmin.compact(TableName tableName,
CompactType compactType)
Compact a table.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.compact(TableName tableName,
CompactType compactType) |
void |
Admin.compact(TableName tableName,
CompactType compactType)
Compact a table.
|
void |
AdminOverAsyncAdmin.compact(TableName tableName,
CompactType compactType) |
CompletableFuture<Void> |
AsyncHBaseAdmin.compact(TableName tableName,
CompactType compactType) |
private CompletableFuture<Void> |
RawAsyncHBaseAdmin.compareTableWithPeerCluster(TableName tableName,
TableDescriptor tableDesc,
ReplicationPeerDescription peer,
AsyncAdmin peerAdmin) |
private void |
AsyncNonMetaRegionLocator.complete(TableName tableName,
AsyncNonMetaRegionLocator.LocateRequest req,
RegionLocations locs,
Throwable error) |
private static CompletableFuture<Boolean> |
RawAsyncHBaseAdmin.completeCheckTableState(CompletableFuture<Boolean> future,
TableState tableState,
Throwable error,
TableState.State targetState,
TableName tableName)
Utility for completing passed TableState
CompletableFuture future using
passed parameters. |
static TableState |
TableState.convert(TableName tableName,
org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableState tableState)
Covert from PB version of TableState
|
static TableDescriptor |
TableDescriptorBuilder.copy(TableName name,
TableDescriptor desc) |
static RegionInfo |
RegionInfo.createMobRegionInfo(TableName tableName)
Creates a RegionInfo object for MOB data.
|
static byte[] |
RegionInfo.createRegionName(TableName tableName,
byte[] startKey,
byte[] id,
boolean newFormat)
Make a region name of passed parameters.
|
static byte[] |
RegionInfo.createRegionName(TableName tableName,
byte[] startKey,
byte[] id,
int replicaId,
boolean newFormat)
Make a region name of passed parameters.
|
static byte[] |
RegionInfo.createRegionName(TableName tableName,
byte[] startKey,
long regionid,
boolean newFormat)
Make a region name of passed parameters.
|
static byte[] |
RegionInfo.createRegionName(TableName tableName,
byte[] startKey,
long regionid,
int replicaId,
boolean newFormat)
Make a region name of passed parameters.
|
static byte[] |
RegionInfo.createRegionName(TableName tableName,
byte[] startKey,
String id,
boolean newFormat)
Make a region name of passed parameters.
|
static CatalogReplicaLoadBalanceSelector |
CatalogReplicaLoadBalanceSelectorFactory.createSelector(String replicaSelectorClass,
TableName tableName,
AsyncConnectionImpl conn,
IntSupplier getReplicaCount)
Create a CatalogReplicaLoadBalanceReplicaSelector based on input config.
|
private CompletableFuture<Void> |
RawAsyncHBaseAdmin.createTable(TableName tableName,
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.CreateTableRequest request) |
CompletableFuture<Void> |
AsyncAdmin.deleteColumnFamily(TableName tableName,
byte[] columnFamily)
Delete a column family from a table.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.deleteColumnFamily(TableName tableName,
byte[] columnFamily) |
default void |
Admin.deleteColumnFamily(TableName tableName,
byte[] columnFamily)
Delete a column family from a table.
|
CompletableFuture<Void> |
AsyncHBaseAdmin.deleteColumnFamily(TableName tableName,
byte[] columnFamily) |
Future<Void> |
Admin.deleteColumnFamilyAsync(TableName tableName,
byte[] columnFamily)
Delete a column family from a table.
|
Future<Void> |
AdminOverAsyncAdmin.deleteColumnFamilyAsync(TableName tableName,
byte[] columnFamily) |
CompletableFuture<Void> |
AsyncAdmin.deleteTable(TableName tableName)
Deletes a table.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.deleteTable(TableName tableName) |
default void |
Admin.deleteTable(TableName tableName)
Deletes a table.
|
CompletableFuture<Void> |
AsyncHBaseAdmin.deleteTable(TableName tableName) |
Future<Void> |
Admin.deleteTableAsync(TableName tableName)
Deletes the table but does not block and wait for it to be completely removed.
|
Future<Void> |
AdminOverAsyncAdmin.deleteTableAsync(TableName tableName) |
CompletableFuture<Void> |
AsyncAdmin.disableTable(TableName tableName)
Disable a table.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.disableTable(TableName tableName) |
default void |
Admin.disableTable(TableName tableName)
Disable table and wait on completion.
|
CompletableFuture<Void> |
AsyncHBaseAdmin.disableTable(TableName tableName) |
Future<Void> |
Admin.disableTableAsync(TableName tableName)
Disable the table but does not block and wait for it to be completely disabled.
|
Future<Void> |
AdminOverAsyncAdmin.disableTableAsync(TableName tableName) |
CompletableFuture<Void> |
AsyncAdmin.disableTableReplication(TableName tableName)
Disable a table's replication switch.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.disableTableReplication(TableName tableName) |
void |
Admin.disableTableReplication(TableName tableName)
Disable a table's replication switch.
|
void |
AdminOverAsyncAdmin.disableTableReplication(TableName tableName) |
CompletableFuture<Void> |
AsyncHBaseAdmin.disableTableReplication(TableName tableName) |
CompletableFuture<Void> |
AsyncAdmin.enableTable(TableName tableName)
Enable a table.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.enableTable(TableName tableName) |
default void |
Admin.enableTable(TableName tableName)
Enable a table.
|
CompletableFuture<Void> |
AsyncHBaseAdmin.enableTable(TableName tableName) |
Future<Void> |
Admin.enableTableAsync(TableName tableName)
Enable the table but does not block and wait for it to be completely enabled.
|
Future<Void> |
AdminOverAsyncAdmin.enableTableAsync(TableName tableName) |
CompletableFuture<Void> |
AsyncAdmin.enableTableReplication(TableName tableName)
Enable a table's replication switch.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.enableTableReplication(TableName tableName) |
void |
Admin.enableTableReplication(TableName tableName)
Enable a table's replication switch.
|
void |
AdminOverAsyncAdmin.enableTableReplication(TableName tableName) |
CompletableFuture<Void> |
AsyncHBaseAdmin.enableTableReplication(TableName tableName) |
CompletableFuture<Void> |
AsyncAdmin.flush(TableName tableName)
Flush a table.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.flush(TableName tableName) |
void |
Admin.flush(TableName tableName)
Flush a table.
|
void |
AdminOverAsyncAdmin.flush(TableName tableName) |
CompletableFuture<Void> |
AsyncHBaseAdmin.flush(TableName tableName) |
CompletableFuture<Void> |
AsyncAdmin.flush(TableName tableName,
byte[] columnFamily)
Flush the specified column family stores on all regions of the passed table.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.flush(TableName tableName,
byte[] columnFamily) |
void |
Admin.flush(TableName tableName,
byte[] columnFamily)
Flush the specified column family stores on all regions of the passed table.
|
void |
AdminOverAsyncAdmin.flush(TableName tableName,
byte[] columnFamily) |
CompletableFuture<Void> |
AsyncHBaseAdmin.flush(TableName tableName,
byte[] columnFamily) |
CompletableFuture<Void> |
AsyncAdmin.flush(TableName tableName,
List<byte[]> columnFamilies)
Flush the specified column family stores on all regions of the passed table.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.flush(TableName tableName,
List<byte[]> columnFamilyList) |
void |
Admin.flush(TableName tableName,
List<byte[]> columnFamilies)
Flush the specified column family stores on all regions of the passed table.
|
void |
AdminOverAsyncAdmin.flush(TableName tableName,
List<byte[]> columnFamilies) |
CompletableFuture<Void> |
AsyncHBaseAdmin.flush(TableName tableName,
List<byte[]> columnFamilies) |
private static int |
MutableRegionInfo.generateHashCode(TableName tableName,
byte[] startKey,
byte[] endKey,
long regionId,
int replicaId,
boolean offLine,
byte[] regionName) |
BufferedMutator |
SharedConnection.getBufferedMutator(TableName tableName) |
default BufferedMutator |
Connection.getBufferedMutator(TableName tableName)
Retrieve a
BufferedMutator for performing client-side buffering of writes. |
default AsyncBufferedMutator |
AsyncConnection.getBufferedMutator(TableName tableName)
Retrieve an
AsyncBufferedMutator for performing client-side buffering of writes. |
default AsyncBufferedMutator |
AsyncConnection.getBufferedMutator(TableName tableName,
ExecutorService pool)
Retrieve an
AsyncBufferedMutator for performing client-side buffering of writes. |
AsyncBufferedMutatorBuilder |
SharedAsyncConnection.getBufferedMutatorBuilder(TableName tableName) |
AsyncBufferedMutatorBuilder |
AsyncConnectionImpl.getBufferedMutatorBuilder(TableName tableName) |
AsyncBufferedMutatorBuilder |
AsyncConnection.getBufferedMutatorBuilder(TableName tableName)
Returns an
AsyncBufferedMutatorBuilder for creating AsyncBufferedMutator . |
AsyncBufferedMutatorBuilder |
SharedAsyncConnection.getBufferedMutatorBuilder(TableName tableName,
ExecutorService pool) |
AsyncBufferedMutatorBuilder |
AsyncConnectionImpl.getBufferedMutatorBuilder(TableName tableName,
ExecutorService pool) |
AsyncBufferedMutatorBuilder |
AsyncConnection.getBufferedMutatorBuilder(TableName tableName,
ExecutorService pool)
Returns an
AsyncBufferedMutatorBuilder for creating AsyncBufferedMutator . |
default CompletableFuture<CompactionState> |
AsyncAdmin.getCompactionState(TableName tableName)
Get the current compaction state of a table.
|
CompactionState |
Admin.getCompactionState(TableName tableName)
Get the current compaction state of a table.
|
CompactionState |
AdminOverAsyncAdmin.getCompactionState(TableName tableName) |
CompletableFuture<CompactionState> |
AsyncAdmin.getCompactionState(TableName tableName,
CompactType compactType)
Get the current compaction state of a table.
|
CompletableFuture<CompactionState> |
RawAsyncHBaseAdmin.getCompactionState(TableName tableName,
CompactType compactType) |
CompactionState |
Admin.getCompactionState(TableName tableName,
CompactType compactType)
Get the current compaction state of a table.
|
CompactionState |
AdminOverAsyncAdmin.getCompactionState(TableName tableName,
CompactType compactType) |
CompletableFuture<CompactionState> |
AsyncHBaseAdmin.getCompactionState(TableName tableName,
CompactType compactType) |
CompletableFuture<? extends SpaceQuotaSnapshotView> |
AsyncAdmin.getCurrentSpaceQuotaSnapshot(TableName tableName)
Returns the Master's view of a quota on the given
tableName or null if the Master has
no quota information on that table. |
CompletableFuture<SpaceQuotaSnapshot> |
RawAsyncHBaseAdmin.getCurrentSpaceQuotaSnapshot(TableName tableName) |
SpaceQuotaSnapshotView |
Admin.getCurrentSpaceQuotaSnapshot(TableName tableName)
Returns the Master's view of a quota on the given
tableName or null if the Master has
no quota information on that table. |
SpaceQuotaSnapshotView |
AdminOverAsyncAdmin.getCurrentSpaceQuotaSnapshot(TableName tableName) |
CompletableFuture<SpaceQuotaSnapshot> |
AsyncHBaseAdmin.getCurrentSpaceQuotaSnapshot(TableName tableName) |
CompletableFuture<TableDescriptor> |
AsyncAdmin.getDescriptor(TableName tableName)
Method for getting the tableDescriptor
|
CompletableFuture<TableDescriptor> |
RawAsyncHBaseAdmin.getDescriptor(TableName tableName) |
TableDescriptor |
Admin.getDescriptor(TableName tableName)
Get a table descriptor.
|
TableDescriptor |
AdminOverAsyncAdmin.getDescriptor(TableName tableName) |
CompletableFuture<TableDescriptor> |
AsyncHBaseAdmin.getDescriptor(TableName tableName) |
CompletableFuture<Optional<Long>> |
AsyncAdmin.getLastMajorCompactionTimestamp(TableName tableName)
Get the timestamp of the last major compaction for the passed table.
|
CompletableFuture<Optional<Long>> |
RawAsyncHBaseAdmin.getLastMajorCompactionTimestamp(TableName tableName) |
long |
Admin.getLastMajorCompactionTimestamp(TableName tableName)
Get the timestamp of the last major compaction for the passed table The timestamp of the oldest
HFile resulting from a major compaction of that table, or 0 if no such HFile could be found.
|
long |
AdminOverAsyncAdmin.getLastMajorCompactionTimestamp(TableName tableName) |
CompletableFuture<Optional<Long>> |
AsyncHBaseAdmin.getLastMajorCompactionTimestamp(TableName tableName) |
(package private) int |
AsyncRegionLocator.getNumberOfCachedRegionLocations(TableName tableName) |
(package private) int |
AsyncNonMetaRegionLocator.getNumberOfCachedRegionLocations(TableName tableName) |
(package private) static int |
ConnectionUtils.getPriority(TableName tableName) |
(package private) CompletableFuture<HRegionLocation> |
AsyncRegionLocator.getRegionLocation(TableName tableName,
byte[] row,
int replicaId,
RegionLocateType type,
boolean reload,
long timeoutNs) |
(package private) CompletableFuture<HRegionLocation> |
AsyncRegionLocator.getRegionLocation(TableName tableName,
byte[] row,
int replicaId,
RegionLocateType type,
long timeoutNs) |
(package private) CompletableFuture<HRegionLocation> |
AsyncRegionLocator.getRegionLocation(TableName tableName,
byte[] row,
RegionLocateType type,
boolean reload,
long timeoutNs) |
(package private) CompletableFuture<HRegionLocation> |
AsyncRegionLocator.getRegionLocation(TableName tableName,
byte[] row,
RegionLocateType type,
long timeoutNs) |
(package private) RegionLocations |
AsyncRegionLocator.getRegionLocationInCache(TableName tableName,
byte[] row) |
(package private) RegionLocations |
AsyncNonMetaRegionLocator.getRegionLocationInCache(TableName tableName,
byte[] row) |
CompletableFuture<RegionLocations> |
AsyncClusterConnection.getRegionLocations(TableName tableName,
byte[] row,
boolean reload)
Return all the replicas for a region.
|
CompletableFuture<RegionLocations> |
AsyncClusterConnectionImpl.getRegionLocations(TableName tableName,
byte[] row,
boolean reload) |
(package private) CompletableFuture<RegionLocations> |
AsyncNonMetaRegionLocator.getRegionLocations(TableName tableName,
byte[] row,
int replicaId,
RegionLocateType locateType,
boolean reload) |
(package private) CompletableFuture<RegionLocations> |
AsyncRegionLocator.getRegionLocations(TableName tableName,
byte[] row,
RegionLocateType type,
boolean reload,
long timeoutNs) |
private CompletableFuture<RegionLocations> |
AsyncNonMetaRegionLocator.getRegionLocationsInternal(TableName tableName,
byte[] row,
int replicaId,
RegionLocateType locateType,
boolean reload) |
RegionLocator |
SharedConnection.getRegionLocator(TableName tableName) |
AsyncTableRegionLocator |
SharedAsyncConnection.getRegionLocator(TableName tableName) |
AsyncTableRegionLocator |
AsyncConnectionImpl.getRegionLocator(TableName tableName) |
RegionLocator |
Connection.getRegionLocator(TableName tableName)
Retrieve a RegionLocator implementation to inspect region information on a table.
|
RegionLocator |
ConnectionOverAsyncConnection.getRegionLocator(TableName tableName) |
AsyncTableRegionLocator |
AsyncConnection.getRegionLocator(TableName tableName)
Retrieve a AsyncRegionLocator implementation to inspect region information on a table.
|
CompletableFuture<List<RegionMetrics>> |
AsyncAdmin.getRegionMetrics(ServerName serverName,
TableName tableName)
Get a list of
RegionMetrics of all regions hosted on a region server for a table. |
CompletableFuture<List<RegionMetrics>> |
RawAsyncHBaseAdmin.getRegionMetrics(ServerName serverName,
TableName tableName) |
List<RegionMetrics> |
Admin.getRegionMetrics(ServerName serverName,
TableName tableName)
Get
RegionMetrics of all regions hosted on a regionserver for a table. |
List<RegionMetrics> |
AdminOverAsyncAdmin.getRegionMetrics(ServerName serverName,
TableName tableName) |
CompletableFuture<List<RegionMetrics>> |
AsyncHBaseAdmin.getRegionMetrics(ServerName serverName,
TableName tableName) |
CompletableFuture<List<RegionInfo>> |
AsyncAdmin.getRegions(TableName tableName)
Get the regions of a given table.
|
CompletableFuture<List<RegionInfo>> |
RawAsyncHBaseAdmin.getRegions(TableName tableName) |
List<RegionInfo> |
Admin.getRegions(TableName tableName)
Get the regions of a given table.
|
List<RegionInfo> |
AdminOverAsyncAdmin.getRegions(TableName tableName) |
CompletableFuture<List<RegionInfo>> |
AsyncHBaseAdmin.getRegions(TableName tableName) |
CompletableFuture<RSGroupInfo> |
AsyncAdmin.getRSGroup(TableName tableName)
Get group info for the given table
|
CompletableFuture<RSGroupInfo> |
RawAsyncHBaseAdmin.getRSGroup(TableName table) |
RSGroupInfo |
Admin.getRSGroup(TableName tableName)
Get group info for the given table
|
RSGroupInfo |
AdminOverAsyncAdmin.getRSGroup(TableName tableName) |
CompletableFuture<RSGroupInfo> |
AsyncHBaseAdmin.getRSGroup(TableName tableName) |
default Table |
Connection.getTable(TableName tableName)
Retrieve a Table implementation for accessing a table.
|
default AsyncTable<AdvancedScanResultConsumer> |
AsyncConnection.getTable(TableName tableName)
Retrieve an
AsyncTable implementation for accessing a table. |
default Table |
Connection.getTable(TableName tableName,
ExecutorService pool)
Retrieve a Table implementation for accessing a table.
|
default AsyncTable<ScanResultConsumer> |
AsyncConnection.getTable(TableName tableName,
ExecutorService pool)
Retrieve an
AsyncTable implementation for accessing a table. |
AsyncTableBuilder<AdvancedScanResultConsumer> |
SharedAsyncConnection.getTableBuilder(TableName tableName) |
AsyncTableBuilder<AdvancedScanResultConsumer> |
AsyncConnectionImpl.getTableBuilder(TableName tableName) |
AsyncTableBuilder<AdvancedScanResultConsumer> |
AsyncConnection.getTableBuilder(TableName tableName)
Returns an
AsyncTableBuilder for creating AsyncTable . |
TableBuilder |
SharedConnection.getTableBuilder(TableName tableName,
ExecutorService pool) |
AsyncTableBuilder<ScanResultConsumer> |
SharedAsyncConnection.getTableBuilder(TableName tableName,
ExecutorService pool) |
AsyncTableBuilder<ScanResultConsumer> |
AsyncConnectionImpl.getTableBuilder(TableName tableName,
ExecutorService pool) |
TableBuilder |
Connection.getTableBuilder(TableName tableName,
ExecutorService pool)
Returns an
TableBuilder for creating Table . |
TableBuilder |
ConnectionOverAsyncConnection.getTableBuilder(TableName tableName,
ExecutorService pool) |
AsyncTableBuilder<ScanResultConsumer> |
AsyncConnection.getTableBuilder(TableName tableName,
ExecutorService pool)
Returns an
AsyncTableBuilder for creating AsyncTable . |
private AsyncNonMetaRegionLocator.TableCache |
AsyncNonMetaRegionLocator.getTableCache(TableName tableName) |
private CompletableFuture<List<HRegionLocation>> |
RawAsyncHBaseAdmin.getTableHRegionLocations(TableName tableName)
List all region locations for the specific table.
|
private CompletableFuture<byte[][]> |
RawAsyncHBaseAdmin.getTableSplits(TableName tableName) |
private CompletableFuture<Void> |
RawAsyncHBaseAdmin.internalRestoreSnapshot(String snapshotName,
TableName tableName,
boolean restoreAcl,
String customSFT) |
private boolean |
AsyncRegionLocator.isMeta(TableName tableName) |
CompletableFuture<Boolean> |
AsyncAdmin.isTableAvailable(TableName tableName)
Check if a table is available.
|
CompletableFuture<Boolean> |
RawAsyncHBaseAdmin.isTableAvailable(TableName tableName) |
boolean |
Admin.isTableAvailable(TableName tableName)
Check if a table is available.
|
boolean |
AdminOverAsyncAdmin.isTableAvailable(TableName tableName) |
CompletableFuture<Boolean> |
AsyncHBaseAdmin.isTableAvailable(TableName tableName) |
CompletableFuture<Boolean> |
AsyncAdmin.isTableDisabled(TableName tableName)
Check if a table is disabled.
|
CompletableFuture<Boolean> |
RawAsyncHBaseAdmin.isTableDisabled(TableName tableName) |
boolean |
Admin.isTableDisabled(TableName tableName)
Check if a table is disabled.
|
boolean |
AdminOverAsyncAdmin.isTableDisabled(TableName tableName) |
CompletableFuture<Boolean> |
AsyncHBaseAdmin.isTableDisabled(TableName tableName) |
CompletableFuture<Boolean> |
AsyncAdmin.isTableEnabled(TableName tableName)
Check if a table is enabled.
|
CompletableFuture<Boolean> |
RawAsyncHBaseAdmin.isTableEnabled(TableName tableName) |
boolean |
Admin.isTableEnabled(TableName tableName)
Check if a table is enabled.
|
boolean |
AdminOverAsyncAdmin.isTableEnabled(TableName tableName) |
CompletableFuture<Boolean> |
AsyncHBaseAdmin.isTableEnabled(TableName tableName) |
private void |
RawAsyncHBaseAdmin.legacyFlush(CompletableFuture<Void> future,
TableName tableName,
List<byte[]> columnFamilies) |
private void |
AsyncNonMetaRegionLocator.locateInMeta(TableName tableName,
AsyncNonMetaRegionLocator.LocateRequest req) |
default CompletableFuture<Void> |
AsyncAdmin.majorCompact(TableName tableName)
Major compact a table.
|
void |
Admin.majorCompact(TableName tableName)
Major compact a table.
|
void |
AdminOverAsyncAdmin.majorCompact(TableName tableName) |
default CompletableFuture<Void> |
AsyncAdmin.majorCompact(TableName tableName,
byte[] columnFamily)
Major compact a column family within a table.
|
void |
Admin.majorCompact(TableName tableName,
byte[] columnFamily)
Major compact a column family within a table.
|
void |
AdminOverAsyncAdmin.majorCompact(TableName tableName,
byte[] columnFamily) |
CompletableFuture<Void> |
AsyncAdmin.majorCompact(TableName tableName,
byte[] columnFamily,
CompactType compactType)
Major compact a column family within a table.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.majorCompact(TableName tableName,
byte[] columnFamily,
CompactType compactType) |
void |
Admin.majorCompact(TableName tableName,
byte[] columnFamily,
CompactType compactType)
Major compact a column family within a table.
|
void |
AdminOverAsyncAdmin.majorCompact(TableName tableName,
byte[] columnFamily,
CompactType compactType) |
CompletableFuture<Void> |
AsyncHBaseAdmin.majorCompact(TableName tableName,
byte[] columnFamily,
CompactType compactType) |
CompletableFuture<Void> |
AsyncAdmin.majorCompact(TableName tableName,
CompactType compactType)
Major compact a table.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.majorCompact(TableName tableName,
CompactType compactType) |
void |
Admin.majorCompact(TableName tableName,
CompactType compactType)
Major compact a table.
|
void |
AdminOverAsyncAdmin.majorCompact(TableName tableName,
CompactType compactType) |
CompletableFuture<Void> |
AsyncHBaseAdmin.majorCompact(TableName tableName,
CompactType compactType) |
CompletableFuture<Void> |
AsyncAdmin.modifyColumnFamily(TableName tableName,
ColumnFamilyDescriptor columnFamily)
Modify an existing column family on a table.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.modifyColumnFamily(TableName tableName,
ColumnFamilyDescriptor columnFamily) |
default void |
Admin.modifyColumnFamily(TableName tableName,
ColumnFamilyDescriptor columnFamily)
Modify an existing column family on a table.
|
CompletableFuture<Void> |
AsyncHBaseAdmin.modifyColumnFamily(TableName tableName,
ColumnFamilyDescriptor columnFamily) |
Future<Void> |
Admin.modifyColumnFamilyAsync(TableName tableName,
ColumnFamilyDescriptor columnFamily)
Modify an existing column family on a table.
|
Future<Void> |
AdminOverAsyncAdmin.modifyColumnFamilyAsync(TableName tableName,
ColumnFamilyDescriptor columnFamily) |
CompletableFuture<Void> |
AsyncAdmin.modifyColumnFamilyStoreFileTracker(TableName tableName,
byte[] family,
String dstSFT)
Change the store file tracker of the given table's given family.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.modifyColumnFamilyStoreFileTracker(TableName tableName,
byte[] family,
String dstSFT) |
default void |
Admin.modifyColumnFamilyStoreFileTracker(TableName tableName,
byte[] family,
String dstSFT)
Change the store file tracker of the given table's given family.
|
CompletableFuture<Void> |
AsyncHBaseAdmin.modifyColumnFamilyStoreFileTracker(TableName tableName,
byte[] family,
String dstSFT) |
Future<Void> |
Admin.modifyColumnFamilyStoreFileTrackerAsync(TableName tableName,
byte[] family,
String dstSFT)
Change the store file tracker of the given table's given family.
|
Future<Void> |
AdminOverAsyncAdmin.modifyColumnFamilyStoreFileTrackerAsync(TableName tableName,
byte[] family,
String dstSFT) |
CompletableFuture<Void> |
AsyncAdmin.modifyTableStoreFileTracker(TableName tableName,
String dstSFT)
Change the store file tracker of the given table.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.modifyTableStoreFileTracker(TableName tableName,
String dstSFT) |
default void |
Admin.modifyTableStoreFileTracker(TableName tableName,
String dstSFT)
Change the store file tracker of the given table.
|
CompletableFuture<Void> |
AsyncHBaseAdmin.modifyTableStoreFileTracker(TableName tableName,
String dstSFT) |
Future<Void> |
Admin.modifyTableStoreFileTrackerAsync(TableName tableName,
String dstSFT)
Change the store file tracker of the given table.
|
Future<Void> |
AdminOverAsyncAdmin.modifyTableStoreFileTrackerAsync(TableName tableName,
String dstSFT) |
static TableDescriptorBuilder |
TableDescriptorBuilder.newBuilder(TableName name) |
static RegionInfoBuilder |
RegionInfoBuilder.newBuilder(TableName tableName) |
private boolean |
AsyncNonMetaRegionLocator.onScanNext(TableName tableName,
AsyncNonMetaRegionLocator.LocateRequest req,
Result result) |
static TableState |
TableState.parseFrom(TableName tableName,
byte[] bytes) |
CompletableFuture<String> |
AsyncClusterConnection.prepareBulkLoad(TableName tableName)
Return the token for this bulk load.
|
CompletableFuture<String> |
AsyncClusterConnectionImpl.prepareBulkLoad(TableName tableName) |
AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder<T> |
AsyncRpcRetryingCallerFactory.MasterRequestCallerBuilder.priority(TableName tableName) |
private <PREQ,PRESP> |
RawAsyncHBaseAdmin.procedureCall(TableName tableName,
PREQ preq,
RawAsyncHBaseAdmin.MasterRpcCall<PRESP,PREQ> rpcCall,
RawAsyncHBaseAdmin.Converter<Long,PRESP> respConverter,
RawAsyncHBaseAdmin.ProcedureBiConsumer consumer) |
(package private) static void |
ConnectionUtils.resetController(HBaseRpcController controller,
long timeoutNs,
int priority,
TableName tableName) |
private CompletableFuture<Void> |
RawAsyncHBaseAdmin.restoreSnapshot(String snapshotName,
TableName tableName,
boolean takeFailSafeSnapshot,
boolean restoreAcl) |
int |
CatalogReplicaLoadBalanceSelector.select(TableName tablename,
byte[] row,
RegionLocateType locateType)
Select a catalog replica region where client go to loop up the input row key.
|
int |
CatalogReplicaLoadBalanceSimpleSelector.select(TableName tableName,
byte[] row,
RegionLocateType locateType)
When it looks up a location, it will call this method to find a replica region to go.
|
private CompletableFuture<Void> |
RawAsyncHBaseAdmin.setTableReplication(TableName tableName,
boolean enableRep)
Set the table's replication switch if the table's replication switch is already not set.
|
default CompletableFuture<Void> |
AsyncAdmin.snapshot(String snapshotName,
TableName tableName)
Take a snapshot for the given table.
|
default void |
Admin.snapshot(String snapshotName,
TableName tableName)
Take a snapshot for the given table.
|
default void |
Admin.snapshot(String snapshotName,
TableName tableName,
Map<String,Object> snapshotProps)
Create typed snapshot of the table.
|
default CompletableFuture<Void> |
AsyncAdmin.snapshot(String snapshotName,
TableName tableName,
SnapshotType type)
Create typed snapshot of the table.
|
default void |
Admin.snapshot(String snapshotName,
TableName tableName,
SnapshotType type)
Create typed snapshot of the table.
|
default void |
Admin.snapshot(String snapshotName,
TableName tableName,
SnapshotType type,
Map<String,Object> snapshotProps)
Create typed snapshot of the table.
|
CompletableFuture<Void> |
AsyncAdmin.split(TableName tableName)
Split a table.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.split(TableName tableName) |
void |
Admin.split(TableName tableName)
Split a table.
|
void |
AdminOverAsyncAdmin.split(TableName tableName) |
CompletableFuture<Void> |
AsyncHBaseAdmin.split(TableName tableName) |
CompletableFuture<Void> |
AsyncAdmin.split(TableName tableName,
byte[] splitPoint)
Split a table.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.split(TableName tableName,
byte[] splitPoint) |
void |
Admin.split(TableName tableName,
byte[] splitPoint)
Split a table.
|
void |
AdminOverAsyncAdmin.split(TableName tableName,
byte[] splitPoint) |
CompletableFuture<Void> |
AsyncHBaseAdmin.split(TableName tableName,
byte[] splitPoint) |
AsyncRpcRetryingCallerFactory.SingleRequestCallerBuilder<T> |
AsyncRpcRetryingCallerFactory.SingleRequestCallerBuilder.table(TableName tableName) |
AsyncRpcRetryingCallerFactory.BatchCallerBuilder |
AsyncRpcRetryingCallerFactory.BatchCallerBuilder.table(TableName tableName) |
CompletableFuture<Boolean> |
AsyncAdmin.tableExists(TableName tableName)
Check if a table exists.
|
CompletableFuture<Boolean> |
RawAsyncHBaseAdmin.tableExists(TableName tableName) |
boolean |
Admin.tableExists(TableName tableName)
Check if a table exists.
|
boolean |
AdminOverAsyncAdmin.tableExists(TableName tableName) |
CompletableFuture<Boolean> |
AsyncHBaseAdmin.tableExists(TableName tableName) |
(package private) static <T> CompletableFuture<T> |
ConnectionUtils.timelineConsistentRead(AsyncRegionLocator locator,
TableName tableName,
Query query,
byte[] row,
RegionLocateType locateType,
Function<Integer,CompletableFuture<T>> requestReplica,
long rpcTimeoutNs,
long primaryCallTimeoutNs,
org.apache.hbase.thirdparty.io.netty.util.Timer retryTimer,
Optional<MetricsConnection> metrics) |
CompletableFuture<Void> |
AsyncAdmin.truncateTable(TableName tableName,
boolean preserveSplits)
Truncate a table.
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.truncateTable(TableName tableName,
boolean preserveSplits) |
default void |
Admin.truncateTable(TableName tableName,
boolean preserveSplits)
Truncate a table.
|
CompletableFuture<Void> |
AsyncHBaseAdmin.truncateTable(TableName tableName,
boolean preserveSplits) |
Future<Void> |
Admin.truncateTableAsync(TableName tableName,
boolean preserveSplits)
Truncate the table but does not block and wait for it to be completely enabled.
|
Future<Void> |
AdminOverAsyncAdmin.truncateTableAsync(TableName tableName,
boolean preserveSplits) |
private CompletableFuture<Void> |
RawAsyncHBaseAdmin.trySyncTableToPeerCluster(TableName tableName,
byte[][] splits,
ReplicationPeerDescription peer) |
void |
MetricsConnection.updateRpc(org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor method,
TableName tableName,
org.apache.hbase.thirdparty.com.google.protobuf.Message param,
MetricsConnection.CallStats stats,
Throwable e)
Report RPC context to metrics system.
|
private void |
MetricsConnection.updateTableMetric(String methodName,
TableName tableName,
MetricsConnection.CallStats stats,
Throwable e)
Report table rpc context to metrics system.
|
Modifier and Type | Method and Description |
---|---|
CompletableFuture<Void> |
AsyncAdmin.appendReplicationPeerTableCFs(String peerId,
Map<TableName,List<String>> tableCfs)
Append the replicable table-cf config of the specified peer
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.appendReplicationPeerTableCFs(String id,
Map<TableName,List<String>> tableCfs) |
default void |
Admin.appendReplicationPeerTableCFs(String id,
Map<TableName,List<String>> tableCfs)
Append the replicable table column family config from the specified peer.
|
CompletableFuture<Void> |
AsyncHBaseAdmin.appendReplicationPeerTableCFs(String peerId,
Map<TableName,List<String>> tableCfs) |
private void |
RawAsyncHBaseAdmin.checkAndGetTableName(byte[] encodeRegionName,
AtomicReference<TableName> tableName,
CompletableFuture<TableName> result) |
private void |
RawAsyncHBaseAdmin.checkAndGetTableName(byte[] encodeRegionName,
AtomicReference<TableName> tableName,
CompletableFuture<TableName> result) |
CompletableFuture<List<TableDescriptor>> |
AsyncAdmin.listTableDescriptors(List<TableName> tableNames)
List specific tables including system tables.
|
CompletableFuture<List<TableDescriptor>> |
RawAsyncHBaseAdmin.listTableDescriptors(List<TableName> tableNames) |
List<TableDescriptor> |
Admin.listTableDescriptors(List<TableName> tableNames)
Get tableDescriptors.
|
List<TableDescriptor> |
AdminOverAsyncAdmin.listTableDescriptors(List<TableName> tableNames) |
CompletableFuture<List<TableDescriptor>> |
AsyncHBaseAdmin.listTableDescriptors(List<TableName> tableNames) |
CompletableFuture<Void> |
AsyncAdmin.removeReplicationPeerTableCFs(String peerId,
Map<TableName,List<String>> tableCfs)
Remove some table-cfs from config of the specified peer
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.removeReplicationPeerTableCFs(String id,
Map<TableName,List<String>> tableCfs) |
default void |
Admin.removeReplicationPeerTableCFs(String id,
Map<TableName,List<String>> tableCfs)
Remove some table-cfs from config of the specified peer.
|
CompletableFuture<Void> |
AsyncHBaseAdmin.removeReplicationPeerTableCFs(String peerId,
Map<TableName,List<String>> tableCfs) |
CompletableFuture<Void> |
AsyncAdmin.setRSGroup(Set<TableName> tables,
String groupName)
Set the RegionServer group for tables
|
CompletableFuture<Void> |
RawAsyncHBaseAdmin.setRSGroup(Set<TableName> tables,
String groupName) |
void |
Admin.setRSGroup(Set<TableName> tables,
String groupName)
Set the RegionServer group for tables
|
void |
AdminOverAsyncAdmin.setRSGroup(Set<TableName> tables,
String groupName) |
CompletableFuture<Void> |
AsyncHBaseAdmin.setRSGroup(Set<TableName> tables,
String groupName) |
NormalizeTableFilterParams.Builder |
NormalizeTableFilterParams.Builder.tableNames(List<TableName> tableNames) |
Constructor and Description |
---|
AddColumnFamilyProcedureBiConsumer(TableName tableName) |
AsyncBatchRpcRetryingCaller(org.apache.hbase.thirdparty.io.netty.util.Timer retryTimer,
AsyncConnectionImpl conn,
TableName tableName,
List<? extends Row> actions,
long pauseNs,
long pauseNsForServerOverloaded,
int maxAttempts,
long operationTimeoutNs,
long rpcTimeoutNs,
int startLogErrorsCnt,
Map<String,byte[]> requestAttributes) |
AsyncClientScanner(Scan scan,
AdvancedScanResultConsumer consumer,
TableName tableName,
AsyncConnectionImpl conn,
org.apache.hbase.thirdparty.io.netty.util.Timer retryTimer,
long pauseNs,
long pauseNsForServerOverloaded,
int maxAttempts,
long scanTimeoutNs,
long rpcTimeoutNs,
int startLogErrorsCnt,
Map<String,byte[]> requestAttributes) |
AsyncRegionLocationCache(TableName tableName) |
AsyncSingleRequestRpcRetryingCaller(org.apache.hbase.thirdparty.io.netty.util.Timer retryTimer,
AsyncConnectionImpl conn,
TableName tableName,
byte[] row,
int replicaId,
RegionLocateType locateType,
AsyncSingleRequestRpcRetryingCaller.Callable<T> callable,
int priority,
long pauseNs,
long pauseNsForServerOverloaded,
int maxAttempts,
long operationTimeoutNs,
long rpcTimeoutNs,
int startLogErrorsCnt,
Map<String,byte[]> requestAttributes) |
AsyncTableBuilderBase(TableName tableName,
AsyncConnectionConfiguration connConf) |
AsyncTableRegionLocatorImpl(TableName tableName,
AsyncConnectionImpl conn) |
AsyncTableResultScanner(TableName tableName,
Scan scan,
long maxCacheSize) |
BufferedMutatorParams(TableName tableName) |
CatalogReplicaLoadBalanceSimpleSelector(TableName tableName,
AsyncConnectionImpl conn,
IntSupplier getNumOfReplicas) |
CreateTableProcedureBiConsumer(TableName tableName) |
DeleteColumnFamilyProcedureBiConsumer(TableName tableName) |
DeleteTableProcedureBiConsumer(TableName tableName) |
DisableTableProcedureBiConsumer(TableName tableName) |
EnableTableProcedureBiConsumer(TableName tableName) |
FlushTableProcedureBiConsumer(TableName tableName) |
MergeTableRegionProcedureBiConsumer(TableName tableName) |
ModifyableTableDescriptor(TableName name)
Construct a table descriptor specifying a TableName object
|
ModifyableTableDescriptor(TableName name,
Collection<ColumnFamilyDescriptor> families,
Map<Bytes,Bytes> values) |
ModifyableTableDescriptor(TableName name,
TableDescriptor desc)
Construct a table descriptor by cloning the descriptor passed as a parameter.
|
ModifyColumnFamilyProcedureBiConsumer(TableName tableName) |
ModifyColumnFamilyStoreFileTrackerProcedureBiConsumer(TableName tableName) |
ModifyTableProcedureBiConsumer(AsyncAdmin admin,
TableName tableName) |
ModifyTableStoreFileTrackerProcedureBiConsumer(AsyncAdmin admin,
TableName tableName) |
MutableRegionInfo(long regionId,
TableName tableName,
int replicaId)
Package private constructor used constructing MutableRegionInfo for the first meta regions
|
MutableRegionInfo(TableName tableName,
byte[] startKey,
byte[] endKey,
boolean split,
long regionId,
int replicaId,
boolean offLine) |
RegionCoprocessorRpcChannel(AsyncConnectionImpl conn,
TableName tableName,
RegionInfo region,
byte[] row,
long rpcTimeoutNs,
long operationTimeoutNs) |
RegionCoprocessorRpcChannelImpl(AsyncConnectionImpl conn,
TableName tableName,
RegionInfo region,
byte[] row,
long rpcTimeoutNs,
long operationTimeoutNs) |
RegionInfoBuilder(TableName tableName) |
SnapshotDescription(String name,
TableName table) |
SnapshotDescription(String name,
TableName table,
SnapshotType type) |
SnapshotDescription(String snapshotName,
TableName tableName,
SnapshotType type,
Map<String,Object> snapshotProps)
SnapshotDescription Parameterized Constructor
|
SnapshotDescription(String name,
TableName table,
SnapshotType type,
String owner) |
SnapshotDescription(String name,
TableName table,
SnapshotType type,
String owner,
long creationTime,
int version)
Deprecated.
since 2.3.0 and will be removed in 4.0.0. Use
SnapshotDescription(String, TableName, SnapshotType, String, long, int, Map) |
SnapshotDescription(String name,
TableName table,
SnapshotType type,
String owner,
long creationTime,
int version,
Map<String,Object> snapshotProps)
SnapshotDescription Parameterized Constructor
|
SnapshotProcedureBiConsumer(TableName tableName) |
SplitTableRegionProcedureBiConsumer(TableName tableName) |
TableBuilderBase(TableName tableName,
ConnectionConfiguration connConf) |
TableCache(TableName tableName) |
TableDescriptorBuilder(TableName name) |
TableProcedureBiConsumer(TableName tableName) |
TableState(TableName tableName,
TableState.State state)
Create instance of TableState.
|
TruncateRegionProcedureBiConsumer(TableName tableName) |
TruncateTableProcedureBiConsumer(TableName tableName) |
Constructor and Description |
---|
NormalizeTableFilterParams(List<TableName> tableNames,
String regex,
String namespace) |
Modifier and Type | Field and Description |
---|---|
private static TableName |
BufferedMutatorExample.TABLE |
private TableName |
MultiThreadedClientExample.WriteExampleCallable.tableName |
private TableName |
MultiThreadedClientExample.SingleWriteExampleCallable.tableName |
private TableName |
MultiThreadedClientExample.ReadExampleCallable.tableName |
Modifier and Type | Method and Description |
---|---|
void |
RefreshHFilesClient.refreshHFiles(TableName tableName) |
private void |
MultiThreadedClientExample.warmUpConnectionCache(Connection connection,
TableName tn) |
Constructor and Description |
---|
ReadExampleCallable(Connection connection,
TableName tableName) |
SingleWriteExampleCallable(Connection connection,
TableName tableName) |
WriteExampleCallable(Connection connection,
TableName tableName) |
Modifier and Type | Method and Description |
---|---|
static org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockRequest |
LockServiceClient.buildLockRequest(org.apache.hadoop.hbase.shaded.protobuf.generated.LockServiceProtos.LockType type,
String namespace,
TableName tableName,
List<RegionInfo> regionInfos,
String description,
long nonceGroup,
long nonce) |
EntityLock |
LockServiceClient.tableLock(TableName tableName,
boolean exclusive,
String description,
Abortable abort)
Create a new EntityLock object to acquire an exclusive or shared lock on a table.
|
Modifier and Type | Field and Description |
---|---|
private TableName |
TableCFs.table |
Modifier and Type | Method and Description |
---|---|
TableName |
TableCFs.getTable() |
Modifier and Type | Method and Description |
---|---|
static Map<TableName,List<String>> |
ReplicationPeerConfigUtil.convert2Map(org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF[] tableCFs)
Convert tableCFs Object to Map.
|
private static Map<TableName,List<String>> |
ReplicationPeerConfigUtil.copyTableCFsMap(Map<TableName,List<String>> preTableCfs) |
private static Map<TableName,List<String>> |
ReplicationPeerConfigUtil.mergeTableCFs(Map<TableName,List<String>> preTableCfs,
Map<TableName,List<String>> tableCfs) |
static Map<TableName,List<String>> |
ReplicationPeerConfigUtil.parseTableCFsFromConfig(String tableCFsConfig)
Convert tableCFs string into Map.
|
Modifier and Type | Method and Description |
---|---|
static ReplicationPeerConfig |
ReplicationPeerConfigUtil.appendExcludeTableCFsToReplicationPeerConfig(Map<TableName,List<String>> excludeTableCfs,
ReplicationPeerConfig peerConfig) |
static ReplicationPeerConfig |
ReplicationPeerConfigUtil.appendTableCFsToReplicationPeerConfig(Map<TableName,List<String>> tableCfs,
ReplicationPeerConfig peerConfig) |
static org.apache.hadoop.hbase.shaded.protobuf.generated.ReplicationProtos.TableCF[] |
ReplicationPeerConfigUtil.convert(Map<TableName,? extends Collection<String>> tableCfs)
convert map to TableCFs Object
|
static String |
ReplicationPeerConfigUtil.convertToString(Map<TableName,? extends Collection<String>> tableCfs) |
private static Map<TableName,List<String>> |
ReplicationPeerConfigUtil.copyTableCFsMap(Map<TableName,List<String>> preTableCfs) |
private static Map<TableName,List<String>> |
ReplicationPeerConfigUtil.mergeTableCFs(Map<TableName,List<String>> preTableCfs,
Map<TableName,List<String>> tableCfs) |
private static Map<TableName,List<String>> |
ReplicationPeerConfigUtil.mergeTableCFs(Map<TableName,List<String>> preTableCfs,
Map<TableName,List<String>> tableCfs) |
static ReplicationPeerConfig |
ReplicationPeerConfigUtil.removeExcludeTableCFsFromReplicationPeerConfig(Map<TableName,List<String>> excludeTableCfs,
ReplicationPeerConfig peerConfig,
String id) |
static ReplicationPeerConfig |
ReplicationPeerConfigUtil.removeTableCFsFromReplicationPeerConfig(Map<TableName,List<String>> tableCfs,
ReplicationPeerConfig peerConfig,
String id) |
Constructor and Description |
---|
TableCFs(TableName table,
Map<String,Integer> cfs) |
Modifier and Type | Field and Description |
---|---|
private TableName |
TableOperationSpanBuilder.tableName |
Modifier and Type | Method and Description |
---|---|
(package private) static void |
TableSpanBuilder.populateTableNameAttributes(Map<io.opentelemetry.api.common.AttributeKey<?>,Object> attributes,
TableName tableName)
Static utility method that performs the primary logic of this builder.
|
TableSpanBuilder |
TableSpanBuilder.setTableName(TableName tableName) |
TableOperationSpanBuilder |
TableOperationSpanBuilder.setTableName(TableName tableName) |
Modifier and Type | Method and Description |
---|---|
default void |
MasterObserver.postCompletedDeleteTableAction(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName)
Called after
HMaster deletes a table. |
default void |
MasterObserver.postCompletedDisableTableAction(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName)
Called after the disableTable operation has been requested.
|
default void |
MasterObserver.postCompletedEnableTableAction(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName)
Called after the enableTable operation has been requested.
|
default void |
MasterObserver.postCompletedModifyTableAction(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName,
TableDescriptor oldDescriptor,
TableDescriptor currentDescriptor)
Called after to modifying a table's properties.
|
default void |
MasterObserver.postCompletedTruncateTableAction(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName)
Called after
HMaster truncates a table. |
default void |
MasterObserver.postDeleteTable(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName)
Called after the deleteTable operation has been requested.
|
default void |
MasterObserver.postDisableTable(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName)
Called after the disableTable operation has been requested.
|
default void |
MasterObserver.postEnableTable(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName)
Called after the enableTable operation has been requested.
|
default void |
MasterObserver.postGetRSGroupInfoOfTable(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName)
Called after getting region server group info of the passed tableName.
|
default void |
MasterObserver.postGetUserPermissions(ObserverContext<MasterCoprocessorEnvironment> ctx,
String userName,
String namespace,
TableName tableName,
byte[] family,
byte[] qualifier)
Called after getting user permissions.
|
default void |
MasterObserver.postModifyColumnFamilyStoreFileTracker(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName,
byte[] family,
String dstSFT)
Called after modifying a family store file tracker.
|
default void |
MasterObserver.postModifyTable(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName,
TableDescriptor oldDescriptor,
TableDescriptor currentDescriptor)
Called after the modifyTable operation has been requested.
|
default void |
MasterObserver.postModifyTableStoreFileTracker(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName,
String dstSFT)
Called after modifying a table's store file tracker.
|
default void |
MasterObserver.postRequestLock(ObserverContext<MasterCoprocessorEnvironment> ctx,
String namespace,
TableName tableName,
RegionInfo[] regionInfos,
String description)
Called after new LockProcedure is queued.
|
default void |
MasterObserver.postSetTableQuota(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName,
GlobalQuotaSettings quotas)
Called after the quota for the table is stored.
|
default void |
MasterObserver.postSetUserQuota(ObserverContext<MasterCoprocessorEnvironment> ctx,
String userName,
TableName tableName,
GlobalQuotaSettings quotas)
Called after the quota for the user on the specified table is stored.
|
default void |
MasterObserver.postTableFlush(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName)
Called after the table memstore is flushed to disk.
|
default void |
MasterObserver.postTruncateTable(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName)
Called after the truncateTable operation has been requested.
|
default void |
MasterObserver.preDeleteTable(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName)
Called before
HMaster deletes a table. |
default void |
MasterObserver.preDeleteTableAction(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName)
Called before
HMaster deletes a table. |
default void |
MasterObserver.preDisableTable(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName)
Called prior to disabling a table.
|
default void |
MasterObserver.preDisableTableAction(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName)
Called prior to disabling a table.
|
default void |
MasterObserver.preEnableTable(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName)
Called prior to enabling a table.
|
default void |
MasterObserver.preEnableTableAction(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName)
Called prior to enabling a table.
|
default void |
MasterObserver.preGetRSGroupInfoOfTable(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName)
Called before getting region server group info of the passed tableName.
|
default void |
MasterObserver.preGetUserPermissions(ObserverContext<MasterCoprocessorEnvironment> ctx,
String userName,
String namespace,
TableName tableName,
byte[] family,
byte[] qualifier)
Called before getting user permissions.
|
default void |
MasterObserver.preLockHeartbeat(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tn,
String description)
Called before heartbeat to a lock.
|
default String |
MasterObserver.preModifyColumnFamilyStoreFileTracker(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName,
byte[] family,
String dstSFT)
Called prior to modifying a family's store file tracker.
|
default TableDescriptor |
MasterObserver.preModifyTable(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName,
TableDescriptor currentDescriptor,
TableDescriptor newDescriptor)
Called prior to modifying a table's properties.
|
default void |
MasterObserver.preModifyTableAction(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName,
TableDescriptor currentDescriptor,
TableDescriptor newDescriptor)
Called prior to modifying a table's properties.
|
default String |
MasterObserver.preModifyTableStoreFileTracker(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName,
String dstSFT)
Called prior to modifying a table's store file tracker.
|
default void |
MasterObserver.preRequestLock(ObserverContext<MasterCoprocessorEnvironment> ctx,
String namespace,
TableName tableName,
RegionInfo[] regionInfos,
String description)
Called before new LockProcedure is queued.
|
default void |
MasterObserver.preSetTableQuota(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName,
GlobalQuotaSettings quotas)
Called before the quota for the table is stored.
|
default void |
MasterObserver.preSetUserQuota(ObserverContext<MasterCoprocessorEnvironment> ctx,
String userName,
TableName tableName,
GlobalQuotaSettings quotas)
Called before the quota for the user on the specified table is stored.
|
default void |
MasterObserver.preSplitRegion(ObserverContext<MasterCoprocessorEnvironment> c,
TableName tableName,
byte[] splitRow)
Called before the split region procedure is called.
|
default void |
MasterObserver.preSplitRegionAction(ObserverContext<MasterCoprocessorEnvironment> c,
TableName tableName,
byte[] splitRow)
Called before the region is split.
|
default void |
MasterObserver.preTableFlush(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName)
Called before the table memstore is flushed to disk.
|
default void |
MasterObserver.preTruncateTable(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName)
Called before
HMaster truncates a table. |
default void |
MasterObserver.preTruncateTableAction(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName)
Called before
HMaster truncates a table. |
Modifier and Type | Method and Description |
---|---|
default void |
MasterObserver.postGetTableDescriptors(ObserverContext<MasterCoprocessorEnvironment> ctx,
List<TableName> tableNamesList,
List<TableDescriptor> descriptors,
String regex)
Called after a getTableDescriptors request has been processed.
|
default void |
MasterObserver.postMoveServersAndTables(ObserverContext<MasterCoprocessorEnvironment> ctx,
Set<Address> servers,
Set<TableName> tables,
String targetGroup)
Called after servers are moved to target region server group
|
default void |
MasterObserver.postMoveTables(ObserverContext<MasterCoprocessorEnvironment> ctx,
Set<TableName> tables,
String targetGroup)
Called after servers are moved to target region server group
|
default void |
MasterObserver.preGetTableDescriptors(ObserverContext<MasterCoprocessorEnvironment> ctx,
List<TableName> tableNamesList,
List<TableDescriptor> descriptors,
String regex)
Called before a getTableDescriptors request has been processed.
|
default void |
MasterObserver.preMoveServersAndTables(ObserverContext<MasterCoprocessorEnvironment> ctx,
Set<Address> servers,
Set<TableName> tables,
String targetGroup)
Called before servers are moved to target region server group
|
default void |
MasterObserver.preMoveTables(ObserverContext<MasterCoprocessorEnvironment> ctx,
Set<TableName> tables,
String targetGroup)
Called before tables are moved to target region server group
|
Modifier and Type | Method and Description |
---|---|
void |
ExampleMasterObserverWithMetrics.preDisableTable(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName) |
Modifier and Type | Method and Description |
---|---|
protected List<RegionPlan> |
FavoredNodeLoadBalancer.balanceTable(TableName tableName,
Map<ServerName,List<RegionInfo>> loadOfOneTable) |
Modifier and Type | Method and Description |
---|---|
static void |
ErasureCodingUtils.setPolicy(org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path rootDir,
TableName tableName,
String policy)
Sets the EC policy on the table directory for the specified table
|
static void |
ErasureCodingUtils.unsetPolicy(org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path rootDir,
TableName tableName)
Unsets any EC policy specified on the path.
|
Modifier and Type | Method and Description |
---|---|
static TableName |
HFileLink.getReferencedTableName(String fileName)
Get the Table name of the referenced link
|
Modifier and Type | Method and Description |
---|---|
static Pair<TableName,String> |
HFileLink.parseBackReferenceName(String name) |
Modifier and Type | Method and Description |
---|---|
static HFileLink |
HFileLink.build(org.apache.hadoop.conf.Configuration conf,
TableName table,
String region,
String family,
String hfile)
Create an HFileLink instance from table/region/family/hfile location
|
static String |
HFileLink.create(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path dstFamilyPath,
String familyName,
String dstTableName,
String dstRegionName,
TableName linkedTable,
String linkedRegion,
String hfileName,
boolean createBackRef)
Create a new HFileLink
|
static String |
HFileLink.create(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path dstFamilyPath,
TableName linkedTable,
String linkedRegion,
String hfileName)
Create a new HFileLink
|
static String |
HFileLink.create(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path dstFamilyPath,
TableName linkedTable,
String linkedRegion,
String hfileName,
boolean createBackRef)
Create a new HFileLink.
|
static String |
HFileLink.createHFileLinkName(TableName tableName,
String regionName,
String hfileName)
Create a new HFileLink name
|
static org.apache.hadoop.fs.Path |
HFileLink.createPath(TableName table,
String region,
String family,
String hfile)
Create an HFileLink relative path for the table/region/family/hfile location
|
Modifier and Type | Method and Description |
---|---|
private boolean |
HFilePrettyPrinter.mobFileExists(org.apache.hadoop.fs.FileSystem fs,
TableName tn,
String mobFileName,
String family,
Set<String> foundMobFiles,
Set<String> missingMobFiles)
Checks whether the referenced mob file exists.
|
Modifier and Type | Field and Description |
---|---|
private TableName |
HBaseRpcControllerImpl.tableName |
Modifier and Type | Method and Description |
---|---|
TableName |
HBaseRpcControllerImpl.getTableName() |
default TableName |
HBaseRpcController.getTableName()
Returns Region's table name or null if not available or pertinent.
|
TableName |
DelegatingHBaseRpcController.getTableName() |
Modifier and Type | Method and Description |
---|---|
void |
HBaseRpcControllerImpl.setPriority(TableName tn) |
void |
HBaseRpcController.setPriority(TableName tn)
Set the priority for this operation.
|
void |
DelegatingHBaseRpcController.setPriority(TableName tn) |
void |
HBaseRpcControllerImpl.setTableName(TableName tableName) |
default void |
HBaseRpcController.setTableName(TableName tableName)
Sets Region's table name.
|
void |
DelegatingHBaseRpcController.setTableName(TableName tableName) |
Modifier and Type | Field and Description |
---|---|
private TableName |
TableSplit.m_tableName |
Modifier and Type | Method and Description |
---|---|
TableName |
TableSplit.getTable()
Returns table name
|
Modifier and Type | Method and Description |
---|---|
private static int |
TableMapReduceUtil.getRegionCount(org.apache.hadoop.conf.Configuration conf,
TableName tableName) |
protected void |
TableInputFormatBase.initializeTable(Connection connection,
TableName tableName)
Allows subclasses to initialize the table information.
|
Constructor and Description |
---|
TableSplit(TableName tableName,
byte[] startRow,
byte[] endRow,
String location)
Constructor
|
Modifier and Type | Field and Description |
---|---|
private TableName |
TableSplit.tableName |
Modifier and Type | Field and Description |
---|---|
private Map<TableName,TableName> |
WALPlayer.WALMapper.tables |
private Map<TableName,TableName> |
WALPlayer.WALMapper.tables |
Modifier and Type | Method and Description |
---|---|
TableName |
TableSplit.getTable()
Returns the table name.
|
Modifier and Type | Method and Description |
---|---|
static Triple<TableName,Scan,org.apache.hadoop.fs.Path> |
ExportUtils.getArgumentsFromCommandLine(org.apache.hadoop.conf.Configuration conf,
String[] args) |
private List<TableName> |
WALPlayer.getTableNameList(String[] tables) |
Modifier and Type | Method and Description |
---|---|
static void |
TableInputFormat.configureSplitTable(org.apache.hadoop.mapreduce.Job job,
TableName tableName)
Sets split table in map-reduce job.
|
private static void |
ImportTsv.createTable(Admin admin,
TableName tableName,
String[] columns) |
private static int |
TableMapReduceUtil.getRegionCount(org.apache.hadoop.conf.Configuration conf,
TableName tableName) |
protected void |
TableInputFormatBase.initializeTable(Connection connection,
TableName tableName)
Allows subclasses to initialize the table information.
|
static void |
TableMapReduceUtil.initTableMapperJob(TableName table,
Scan scan,
Class<? extends TableMapper> mapper,
Class<?> outputKeyClass,
Class<?> outputValueClass,
org.apache.hadoop.mapreduce.Job job)
Use this before submitting a TableMap job.
|
Constructor and Description |
---|
TableSplit(TableName tableName,
byte[] startRow,
byte[] endRow,
String location)
Creates a new instance without a scanner.
|
TableSplit(TableName tableName,
byte[] startRow,
byte[] endRow,
String location,
long length)
Creates a new instance without a scanner.
|
TableSplit(TableName tableName,
Scan scan,
byte[] startRow,
byte[] endRow,
String location)
Creates a new instance while assigning all variables.
|
TableSplit(TableName tableName,
Scan scan,
byte[] startRow,
byte[] endRow,
String location,
long length)
Creates a new instance while assigning all variables.
|
TableSplit(TableName tableName,
Scan scan,
byte[] startRow,
byte[] endRow,
String location,
String encodedRegionName,
long length)
Creates a new instance while assigning all variables.
|
Modifier and Type | Field and Description |
---|---|
private TableName |
AssignmentVerificationReport.tableName |
Modifier and Type | Field and Description |
---|---|
private Set<TableName> |
SnapshotOfRegionAssignmentFromMeta.disabledTables |
private Map<TableName,AtomicInteger> |
HMaster.mobCompactionStates |
private ConcurrentMap<TableName,TableState.State> |
TableStateManager.tableName2State |
private Map<TableName,List<RegionInfo>> |
SnapshotOfRegionAssignmentFromMeta.tableToRegionMap
the table name to region map
|
private Set<TableName> |
RegionPlacementMaintainer.targetTableSet |
private IdReadWriteLock<TableName> |
TableStateManager.tnLock |
Modifier and Type | Method and Description |
---|---|
Map<TableName,Integer> |
RegionPlacementMaintainer.getRegionsMovement(FavoredNodesPlan newPlan)
Return how many regions will move per table since their primary RS will change
|
Set<TableName> |
SnapshotOfRegionAssignmentFromMeta.getTableSet()
Get the table set
|
(package private) Set<TableName> |
TableStateManager.getTablesInStates(TableState.State... states)
Return all tables in given states.
|
Map<TableName,List<RegionInfo>> |
SnapshotOfRegionAssignmentFromMeta.getTableToRegionMap()
Get regions for tables
|
private Map<TableName,List<byte[]>> |
RegionsRecoveryChore.getTableToRegionsByRefCount(Map<ServerName,ServerMetrics> serverMetricsMap) |
List<TableName> |
HMaster.listTableNames(String namespace,
String regex,
boolean includeSysTables)
Returns the list of table names that match the specified request
|
List<TableName> |
HMaster.listTableNamesByNamespace(String name) |
List<TableName> |
MasterServices.listTableNamesByNamespace(String name)
Get list of table names by namespace
|
Modifier and Type | Method and Description |
---|---|
long |
HMaster.addColumn(TableName tableName,
ColumnFamilyDescriptor column,
long nonceGroup,
long nonce) |
long |
MasterServices.addColumn(TableName tableName,
ColumnFamilyDescriptor column,
long nonceGroup,
long nonce)
Add a new column to an existing table
|
private void |
HMaster.checkTableExists(TableName tableName) |
void |
HMaster.checkTableModifiable(TableName tableName) |
void |
MasterServices.checkTableModifiable(TableName tableName)
Check table is modifiable; i.e.
|
long |
HMaster.deleteColumn(TableName tableName,
byte[] columnName,
long nonceGroup,
long nonce) |
long |
MasterServices.deleteColumn(TableName tableName,
byte[] columnName,
long nonceGroup,
long nonce)
Delete a column from an existing table
|
long |
HMaster.deleteTable(TableName tableName,
long nonceGroup,
long nonce) |
long |
MasterServices.deleteTable(TableName tableName,
long nonceGroup,
long nonce)
Delete a table
|
long |
HMaster.disableTable(TableName tableName,
long nonceGroup,
long nonce) |
long |
MasterServices.disableTable(TableName tableName,
long nonceGroup,
long nonce)
Disable an existing table
|
long |
HMaster.enableTable(TableName tableName,
long nonceGroup,
long nonce) |
long |
MasterServices.enableTable(TableName tableName,
long nonceGroup,
long nonce)
Enable an existing table
|
void |
AssignmentVerificationReport.fillUp(TableName tableName,
SnapshotOfRegionAssignmentFromMeta snapshot,
Map<String,Map<String,Float>> regionLocalityMap) |
void |
AssignmentVerificationReport.fillUpDispersion(TableName tableName,
SnapshotOfRegionAssignmentFromMeta snapshot,
FavoredNodesPlan newPlan)
Use this to project the dispersion scores
|
long |
HMaster.flushTable(TableName tableName,
List<byte[]> columnFamilies,
long nonceGroup,
long nonce) |
long |
MasterServices.flushTable(TableName tableName,
List<byte[]> columnFamilies,
long nonceGroup,
long nonce)
Flush an existing table
|
private void |
RegionPlacementMaintainer.genAssignmentPlan(TableName tableName,
SnapshotOfRegionAssignmentFromMeta assignmentSnapshot,
Map<String,Map<String,Float>> regionLocalityMap,
FavoredNodesPlan plan,
boolean munkresForSecondaryAndTertiary)
Generate the assignment plan for the existing table
|
CompactionState |
HMaster.getCompactionState(TableName tableName)
Get the compaction state of the table
|
long |
HMaster.getLastMajorCompactionTimestamp(TableName table) |
long |
MasterServices.getLastMajorCompactionTimestamp(TableName table) |
org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.GetRegionInfoResponse.CompactionState |
HMaster.getMobCompactionState(TableName tableName)
Gets the mob file compaction state for a specific table.
|
TableState |
TableStateManager.getTableState(TableName tableName) |
private static boolean |
HMaster.isCatalogTable(TableName tableName) |
boolean |
TableStateManager.isTablePresent(TableName tableName) |
boolean |
TableStateManager.isTableState(TableName tableName,
TableState.State... states) |
long |
HMaster.modifyColumn(TableName tableName,
ColumnFamilyDescriptor descriptor,
long nonceGroup,
long nonce) |
long |
MasterServices.modifyColumn(TableName tableName,
ColumnFamilyDescriptor descriptor,
long nonceGroup,
long nonce)
Modify the column descriptor of an existing column in an existing table
|
long |
HMaster.modifyColumnStoreFileTracker(TableName tableName,
byte[] family,
String dstSFT,
long nonceGroup,
long nonce) |
long |
MasterServices.modifyColumnStoreFileTracker(TableName tableName,
byte[] family,
String dstSFT,
long nonceGroup,
long nonce)
Modify the store file tracker of an existing column in an existing table
|
private long |
HMaster.modifyTable(TableName tableName,
HMaster.TableDescriptorGetter newDescriptorGetter,
long nonceGroup,
long nonce,
boolean shouldCheckDescriptor) |
private long |
HMaster.modifyTable(TableName tableName,
HMaster.TableDescriptorGetter newDescriptorGetter,
long nonceGroup,
long nonce,
boolean shouldCheckDescriptor,
boolean reopenRegions) |
default long |
MasterServices.modifyTable(TableName tableName,
TableDescriptor descriptor,
long nonceGroup,
long nonce)
Modify the descriptor of an existing table
|
long |
HMaster.modifyTable(TableName tableName,
TableDescriptor newDescriptor,
long nonceGroup,
long nonce,
boolean reopenRegions) |
long |
MasterServices.modifyTable(TableName tableName,
TableDescriptor descriptor,
long nonceGroup,
long nonce,
boolean reopenRegions)
Modify the descriptor of an existing table
|
long |
HMaster.modifyTableStoreFileTracker(TableName tableName,
String dstSFT,
long nonceGroup,
long nonce) |
long |
MasterServices.modifyTableStoreFileTracker(TableName tableName,
String dstSFT,
long nonceGroup,
long nonce)
Modify the store file tracker of an existing table
|
void |
MasterCoprocessorHost.postCompletedDeleteTableAction(TableName tableName,
User user) |
void |
MasterCoprocessorHost.postCompletedDisableTableAction(TableName tableName,
User user) |
void |
MasterCoprocessorHost.postCompletedEnableTableAction(TableName tableName,
User user) |
void |
MasterCoprocessorHost.postCompletedModifyTableAction(TableName tableName,
TableDescriptor oldDescriptor,
TableDescriptor currentDescriptor,
User user) |
void |
MasterCoprocessorHost.postCompletedTruncateTableAction(TableName tableName,
User user) |
void |
MasterCoprocessorHost.postDeleteTable(TableName tableName) |
void |
MasterCoprocessorHost.postDisableTable(TableName tableName) |
void |
MasterCoprocessorHost.postEnableTable(TableName tableName) |
void |
MasterCoprocessorHost.postGetRSGroupInfoOfTable(TableName tableName) |
void |
MasterCoprocessorHost.postGetUserPermissions(String userName,
String namespace,
TableName tableName,
byte[] family,
byte[] qualifier) |
void |
MasterCoprocessorHost.postModifyColumnFamilyStoreFileTracker(TableName tableName,
byte[] family,
String dstSFT) |
void |
MasterCoprocessorHost.postModifyTable(TableName tableName,
TableDescriptor oldDescriptor,
TableDescriptor currentDescriptor) |
void |
MasterCoprocessorHost.postModifyTableStoreFileTracker(TableName tableName,
String dstSFT) |
void |
MasterCoprocessorHost.postRequestLock(String namespace,
TableName tableName,
RegionInfo[] regionInfos,
LockType type,
String description) |
void |
MasterCoprocessorHost.postSetTableQuota(TableName table,
GlobalQuotaSettings quotas) |
void |
MasterCoprocessorHost.postSetUserQuota(String user,
TableName table,
GlobalQuotaSettings quotas) |
void |
MasterCoprocessorHost.postTableFlush(TableName tableName) |
void |
MasterCoprocessorHost.postTruncateTable(TableName tableName) |
void |
MasterCoprocessorHost.preDeleteTable(TableName tableName) |
void |
MasterCoprocessorHost.preDeleteTableAction(TableName tableName,
User user) |
void |
MasterCoprocessorHost.preDisableTable(TableName tableName) |
void |
MasterCoprocessorHost.preDisableTableAction(TableName tableName,
User user) |
void |
MasterCoprocessorHost.preEnableTable(TableName tableName) |
void |
MasterCoprocessorHost.preEnableTableAction(TableName tableName,
User user) |
void |
MasterCoprocessorHost.preGetRSGroupInfoOfTable(TableName tableName) |
void |
MasterCoprocessorHost.preGetUserPermissions(String userName,
String namespace,
TableName tableName,
byte[] family,
byte[] qualifier) |
String |
MasterCoprocessorHost.preModifyColumnFamilyStoreFileTracker(TableName tableName,
byte[] family,
String dstSFT) |
TableDescriptor |
MasterCoprocessorHost.preModifyTable(TableName tableName,
TableDescriptor currentDescriptor,
TableDescriptor newDescriptor) |
void |
MasterCoprocessorHost.preModifyTableAction(TableName tableName,
TableDescriptor currentDescriptor,
TableDescriptor newDescriptor,
User user) |
String |
MasterCoprocessorHost.preModifyTableStoreFileTracker(TableName tableName,
String dstSFT) |
void |
MasterCoprocessorHost.preRequestLock(String namespace,
TableName tableName,
RegionInfo[] regionInfos,
LockType type,
String description) |
void |
MasterCoprocessorHost.preSetTableQuota(TableName table,
GlobalQuotaSettings quotas) |
void |
MasterCoprocessorHost.preSetUserQuota(String user,
TableName table,
GlobalQuotaSettings quotas) |
void |
MasterCoprocessorHost.preSplitRegion(TableName tableName,
byte[] splitRow)
Invoked just before calling the split region procedure
|
void |
MasterCoprocessorHost.preSplitRegionAction(TableName tableName,
byte[] splitRow,
User user)
Invoked just before a split
|
void |
MasterCoprocessorHost.preTableFlush(TableName tableName) |
void |
MasterCoprocessorHost.preTruncateTable(TableName tableName) |
void |
MasterCoprocessorHost.preTruncateTableAction(TableName tableName,
User user) |
void |
RegionPlacementMaintainer.printDispersionScores(TableName table,
SnapshotOfRegionAssignmentFromMeta snapshot,
int numRegions,
FavoredNodesPlan newPlan,
boolean simplePrint) |
private TableState |
TableStateManager.readMetaState(TableName tableName) |
(package private) long |
HMaster.reopenRegions(TableName tableName,
List<byte[]> regionNames,
long nonceGroup,
long nonce)
Reopen regions provided in the argument
|
void |
HMaster.reportMobCompactionEnd(TableName tableName) |
void |
HMaster.reportMobCompactionStart(TableName tableName) |
void |
TableStateManager.setDeletedTable(TableName tableName) |
void |
TableStateManager.setTableState(TableName tableName,
TableState.State newState)
Set table state to provided.
|
long |
HMaster.truncateTable(TableName tableName,
boolean preserveSplits,
long nonceGroup,
long nonce) |
long |
MasterServices.truncateTable(TableName tableName,
boolean preserveSplits,
long nonceGroup,
long nonce)
Truncate a table
|
private void |
TableStateManager.updateMetaState(TableName tableName,
TableState.State newState) |
Modifier and Type | Method and Description |
---|---|
List<RegionPlan> |
LoadBalancer.balanceCluster(Map<TableName,Map<ServerName,List<RegionInfo>>> loadOfAllTable)
Perform the major balance operation for cluster.
|
void |
RegionPlacementMaintainer.checkDifferencesWithOldPlan(Map<TableName,Integer> movesPerTable,
Map<String,Map<String,Float>> regionLocalityMap,
FavoredNodesPlan newPlan)
Compares two plans and check whether the locality dropped or increased (prints the information
as a string) also prints the baseline locality
|
private List<TableDescriptor> |
HMaster.getTableDescriptors(List<TableDescriptor> htds,
String namespace,
String regex,
List<TableName> tableNameList,
boolean includeSysTables)
Return a list of table table descriptors after applying any provided filter parameters.
|
List<TableDescriptor> |
HMaster.listTableDescriptors(String namespace,
String regex,
List<TableName> tableNameList,
boolean includeSysTables)
Returns the list of table descriptors that match the specified request
|
void |
MasterCoprocessorHost.postGetTableDescriptors(List<TableName> tableNamesList,
List<TableDescriptor> descriptors,
String regex) |
void |
MasterCoprocessorHost.postMoveServersAndTables(Set<Address> servers,
Set<TableName> tables,
String targetGroup) |
void |
MasterCoprocessorHost.postMoveTables(Set<TableName> tables,
String targetGroup) |
void |
MasterCoprocessorHost.preGetTableDescriptors(List<TableName> tableNamesList,
List<TableDescriptor> descriptors,
String regex) |
void |
MasterCoprocessorHost.preMoveServersAndTables(Set<Address> servers,
Set<TableName> tables,
String targetGroup) |
void |
MasterCoprocessorHost.preMoveTables(Set<TableName> tables,
String targetGroup) |
private void |
RegionsRecoveryChore.prepareTableToReopenRegionsMap(Map<TableName,List<byte[]>> tableToReopenRegionsMap,
byte[] regionName,
int regionStoreRefCount) |
default void |
LoadBalancer.updateBalancerLoadInfo(Map<TableName,Map<ServerName,List<RegionInfo>>> loadOfAllTable)
In some scenarios, Balancer needs to update internal status or information according to the
current tables load
|
Constructor and Description |
---|
SnapshotOfRegionAssignmentFromMeta(Connection connection,
Set<TableName> disabledTables,
boolean excludeOfflinedSplitParents) |
Modifier and Type | Method and Description |
---|---|
TableName |
RegionStateNode.getTable() |
TableName |
MoveRegionProcedure.getTableName()
Deprecated.
|
TableName |
MergeTableRegionsProcedure.getTableName() |
TableName |
RegionRemoteProcedureBase.getTableName() |
TableName |
GCMultipleMergedRegionsProcedure.getTableName() |
TableName |
GCMergedRegionsProcedure.getTableName()
Deprecated.
|
TableName |
RegionTransitionProcedure.getTableName()
Deprecated.
|
Modifier and Type | Method and Description |
---|---|
Map<TableName,Map<ServerName,List<RegionInfo>>> |
RegionStates.getAssignmentsForBalancer(TableStateManager tableStateManager,
List<ServerName> onlineServers)
This is an EXPENSIVE clone.
|
Modifier and Type | Method and Description |
---|---|
TransitRegionStateProcedure[] |
AssignmentManager.createUnassignProceduresForClosingExcessRegionReplicas(TableName tableName,
int newReplicaCount)
Called by ModifyTableProcedures to unassign all the excess region replicas for a table.
|
TransitRegionStateProcedure[] |
AssignmentManager.createUnassignProceduresForDisabling(TableName tableName)
Called by DisableTableProcedure to unassign all the regions for a table.
|
void |
AssignmentManager.deleteTable(TableName tableName)
Delete the region states.
|
private TableDescriptor |
RegionStateStore.getDescriptor(TableName tableName) |
Map<RegionState.State,List<RegionInfo>> |
RegionStates.getRegionByStateOfTable(TableName tableName) |
List<RegionInfo> |
RegionStates.getRegionsOfTable(TableName table)
Returns Return online regions of table; does not include OFFLINE or SPLITTING regions.
|
private List<RegionInfo> |
RegionStates.getRegionsOfTable(TableName table,
Predicate<RegionStateNode> filter)
Returns Return the regions of the table and filter them.
|
List<RegionInfo> |
RegionStates.getRegionsOfTableForDeleting(TableName table)
Get the regions for deleting a table.
|
List<RegionInfo> |
RegionStates.getRegionsOfTableForEnabling(TableName table)
Get the regions for enabling a table.
|
List<HRegionLocation> |
RegionStates.getRegionsOfTableForReopen(TableName tableName)
Get the regions to be reopened when modifying a table.
|
private Stream<RegionStateNode> |
AssignmentManager.getRegionStateNodes(TableName tableName,
boolean excludeOfflinedSplitParents) |
RegionStatesCount |
AssignmentManager.getRegionStatesCount(TableName tableName)
Provide regions state count for given table.
|
Pair<Integer,Integer> |
AssignmentManager.getReopenStatus(TableName tableName)
Used by the client (via master) to identify if all regions have the schema updates
|
private Scan |
RegionStateStore.getScanForUpdateRegionReplicas(TableName tableName) |
List<RegionInfo> |
AssignmentManager.getTableRegions(TableName tableName,
boolean excludeOfflinedSplitParents) |
List<Pair<RegionInfo,ServerName>> |
AssignmentManager.getTableRegionsAndLocations(TableName tableName,
boolean excludeOfflinedSplitParents) |
(package private) ArrayList<RegionInfo> |
RegionStates.getTableRegionsInfo(TableName tableName) |
(package private) List<RegionStateNode> |
RegionStates.getTableRegionStateNodes(TableName tableName) |
(package private) ArrayList<RegionState> |
RegionStates.getTableRegionStates(TableName tableName) |
private boolean |
RegionStateStore.hasGlobalReplicationScope(TableName tableName) |
boolean |
RegionStates.hasTableRegionStates(TableName tableName) |
private boolean |
AssignmentManager.isTableDisabled(TableName tableName) |
private boolean |
RegionStates.isTableDisabled(TableStateManager tableStateManager,
TableName tableName) |
private boolean |
AssignmentManager.isTableEnabled(TableName tableName) |
void |
RegionStateStore.removeRegionReplicas(TableName tableName,
int oldReplicaCount,
int newReplicaCount) |
Modifier and Type | Method and Description |
---|---|
protected abstract List<RegionPlan> |
BaseLoadBalancer.balanceTable(TableName tableName,
Map<ServerName,List<RegionInfo>> loadOfOneTable)
Perform the major balance operation for table, all sub classes should override this method.
|
protected List<RegionPlan> |
SimpleLoadBalancer.balanceTable(TableName tableName,
Map<ServerName,List<RegionInfo>> loadOfOneTable)
Generate a global load balancing plan according to the specified map of server information to
the most loaded regions of each server.
|
protected List<RegionPlan> |
FavoredStochasticBalancer.balanceTable(TableName tableName,
Map<ServerName,List<RegionInfo>> loadOfOneTable)
For all regions correctly assigned to favored nodes, we just use the stochastic balancer
implementation.
|
protected List<RegionPlan> |
StochasticLoadBalancer.balanceTable(TableName tableName,
Map<ServerName,List<RegionInfo>> loadOfOneTable)
Given the cluster state this will try and approach an optimal balance.
|
private TableDescriptor |
RegionHDFSBlockLocationFinder.getDescriptor(TableName tableName)
return TableDescriptor for a given tableName
|
TableDescriptor |
ClusterInfoProvider.getTableDescriptor(TableName tableName)
Get the table descriptor for the given table.
|
TableDescriptor |
MasterClusterInfoProvider.getTableDescriptor(TableName tableName) |
(package private) boolean |
StochasticLoadBalancer.needsBalance(TableName tableName,
BalancerClusterState cluster) |
private void |
StochasticLoadBalancer.updateBalancerTableLoadInfo(TableName tableName,
Map<ServerName,List<RegionInfo>> loadOfOneTable) |
private void |
StochasticLoadBalancer.updateStochasticCosts(TableName tableName,
double overall,
double[] subCosts)
update costs to JMX
|
Modifier and Type | Method and Description |
---|---|
List<RegionPlan> |
BaseLoadBalancer.balanceCluster(Map<TableName,Map<ServerName,List<RegionInfo>>> loadOfAllTable)
Perform the major balance operation for cluster, will invoke
BaseLoadBalancer.balanceTable(TableName, Map) to do actual balance. |
List<RegionPlan> |
MaintenanceLoadBalancer.balanceCluster(Map<TableName,Map<ServerName,List<RegionInfo>>> loadOfAllTable) |
protected void |
BaseLoadBalancer.preBalanceCluster(Map<TableName,Map<ServerName,List<RegionInfo>>> loadOfAllTable)
Called before actually executing balanceCluster.
|
protected void |
SimpleLoadBalancer.preBalanceCluster(Map<TableName,Map<ServerName,List<RegionInfo>>> loadOfAllTable) |
(package private) void |
SimpleLoadBalancer.setClusterLoad(Map<TableName,Map<ServerName,List<RegionInfo>>> clusterLoad)
Pass RegionStates and allow balancer to set the current cluster load.
|
protected Map<ServerName,List<RegionInfo>> |
BaseLoadBalancer.toEnsumbleTableLoad(Map<TableName,Map<ServerName,List<RegionInfo>>> LoadOfAllTable) |
void |
StochasticLoadBalancer.updateBalancerLoadInfo(Map<TableName,Map<ServerName,List<RegionInfo>>> loadOfAllTable) |
Modifier and Type | Field and Description |
---|---|
private TableName |
MetaBrowser.scanTable |
private TableName |
RegionVisualizer.RegionDetails.tableName |
Modifier and Type | Method and Description |
---|---|
TableName |
MetaBrowser.getScanTable() |
TableName |
RegionVisualizer.RegionDetails.getTableName() |
private static TableName |
MetaBrowser.resolveScanTable(javax.servlet.http.HttpServletRequest request) |
Modifier and Type | Method and Description |
---|---|
private static Filter |
MetaBrowser.buildTableFilter(TableName tableName) |
Constructor and Description |
---|
RegionDetails(ServerName serverName,
TableName tableName,
RegionMetrics regionMetrics) |
Modifier and Type | Method and Description |
---|---|
private static RegionInfo |
MetaFixer.buildRegionInfo(TableName tn,
byte[] start,
byte[] end) |
private static Pair<Boolean,Boolean> |
CatalogJanitor.checkRegionReferences(MasterServices services,
TableName tableName,
RegionInfo region)
Checks if a region still holds references to parent.
|
Modifier and Type | Field and Description |
---|---|
private TableName |
LockProcedure.tableName |
private TableName |
LockManager.MasterLock.tableName |
Modifier and Type | Method and Description |
---|---|
TableName |
LockProcedure.getTableName() |
Modifier and Type | Method and Description |
---|---|
LockManager.MasterLock |
LockManager.createMasterLock(TableName tableName,
LockType type,
String description) |
long |
LockManager.RemoteLocks.requestTableLock(TableName tableName,
LockType type,
String description,
NonceKey nonceKey) |
Constructor and Description |
---|
LockProcedure(org.apache.hadoop.conf.Configuration conf,
TableName tableName,
LockType type,
String description,
CountDownLatch lockAcquireLatch)
Constructor for table lock.
|
MasterLock(TableName tableName,
LockType type,
String description) |
Modifier and Type | Field and Description |
---|---|
private TableName |
SimpleRegionNormalizer.NormalizeContext.tableName |
Modifier and Type | Field and Description |
---|---|
private RegionNormalizerWorkQueue<TableName> |
RegionNormalizerManager.workQueue |
private RegionNormalizerWorkQueue<TableName> |
RegionNormalizerWorker.workQueue |
Modifier and Type | Method and Description |
---|---|
TableName |
SimpleRegionNormalizer.NormalizeContext.getTableName() |
Modifier and Type | Method and Description |
---|---|
private List<NormalizationPlan> |
RegionNormalizerWorker.calculatePlans(TableName tableName) |
Modifier and Type | Method and Description |
---|---|
boolean |
RegionNormalizerManager.normalizeRegions(List<TableName> tables,
boolean isHighPriority)
Submit tables for normalization.
|
Constructor and Description |
---|
RegionNormalizerManager(RegionNormalizerStateStore regionNormalizerStateStore,
RegionNormalizerChore regionNormalizerChore,
RegionNormalizerWorkQueue<TableName> workQueue,
RegionNormalizerWorker worker) |
RegionNormalizerWorker(org.apache.hadoop.conf.Configuration configuration,
MasterServices masterServices,
RegionNormalizer regionNormalizer,
RegionNormalizerWorkQueue<TableName> workQueue) |
Modifier and Type | Field and Description |
---|---|
static TableName |
TableProcedureInterface.DUMMY_NAMESPACE_TABLE_NAME
Used for acquire/release lock for namespace related operations, just a place holder as we do
not have namespace table any more.
|
private TableName |
SnapshotProcedure.snapshotTable |
private TableName |
EnableTableProcedure.tableName |
private TableName |
FlushTableProcedure.tableName |
private TableName |
DisableTableProcedure.tableName |
private TableName |
ReopenTableRegionsProcedure.tableName |
private TableName |
TruncateTableProcedure.tableName |
private TableName |
DeleteTableProcedure.tableName |
private TableName |
ModifyTableDescriptorProcedure.tableName |
Modifier and Type | Field and Description |
---|---|
private FairQueue<TableName> |
MasterProcedureScheduler.metaRunQueue |
private Map<TableName,LockAndQueue> |
SchemaLocking.tableLocks |
private FairQueue<TableName> |
MasterProcedureScheduler.tableRunQueue |
Modifier and Type | Method and Description |
---|---|
TableName |
FlushRegionProcedure.getTableName() |
TableName |
EnableTableProcedure.getTableName() |
TableName |
ModifyTableProcedure.getTableName() |
TableName |
CreateTableProcedure.getTableName() |
TableName |
FlushTableProcedure.getTableName() |
TableName |
SnapshotRegionProcedure.getTableName() |
TableName |
RestoreSnapshotProcedure.getTableName() |
TableName |
DisableTableProcedure.getTableName() |
TableName |
InitMetaProcedure.getTableName() |
TableName |
AbstractStateMachineRegionProcedure.getTableName() |
TableName |
ReopenTableRegionsProcedure.getTableName() |
TableName |
CloneSnapshotProcedure.getTableName() |
TableName |
AbstractStateMachineNamespaceProcedure.getTableName() |
TableName |
SnapshotVerifyProcedure.getTableName() |
TableName |
TruncateTableProcedure.getTableName() |
TableName |
DeleteTableProcedure.getTableName() |
TableName |
ModifyTableDescriptorProcedure.getTableName() |
TableName |
TableProcedureInterface.getTableName()
Returns the name of the table the procedure is operating on
|
abstract TableName |
AbstractStateMachineTableProcedure.getTableName() |
TableName |
SnapshotProcedure.getTableName() |
private static TableName |
MasterProcedureScheduler.getTableName(Procedure<?> proc) |
Modifier and Type | Method and Description |
---|---|
private static void |
DeleteTableProcedure.cleanRegionsInMeta(MasterProcedureEnv env,
TableName tableName)
There may be items for this table still up in hbase:meta in the case where the info:regioninfo
column was empty because of some write error.
|
List<RegionInfo> |
CreateTableProcedure.CreateHdfsRegions.createHdfsRegions(MasterProcedureEnv env,
org.apache.hadoop.fs.Path tableRootDir,
TableName tableName,
List<RegionInfo> newRegions) |
protected static void |
DeleteTableProcedure.deleteAssignmentState(MasterProcedureEnv env,
TableName tableName) |
static void |
MasterDDLOperationHelper.deleteColumnFamilyFromFileSystem(MasterProcedureEnv env,
TableName tableName,
List<RegionInfo> regionInfoList,
byte[] familyName,
boolean hasMob)
Remove the column family from the file system
|
protected static void |
DeleteTableProcedure.deleteFromFs(MasterProcedureEnv env,
TableName tableName,
List<RegionInfo> regions,
boolean archive) |
protected static void |
DeleteTableProcedure.deleteFromMeta(MasterProcedureEnv env,
TableName tableName,
List<RegionInfo> regions) |
protected static void |
DeleteTableProcedure.deleteTableDescriptorCache(MasterProcedureEnv env,
TableName tableName) |
protected static void |
DeleteTableProcedure.deleteTableStates(MasterProcedureEnv env,
TableName tableName) |
(package private) LockAndQueue |
SchemaLocking.getTableLock(TableName tableName) |
static int |
MasterProcedureUtil.getTablePriority(TableName tableName)
Return the priority for the given table.
|
private TableQueue |
MasterProcedureScheduler.getTableQueue(TableName tableName) |
(package private) boolean |
MasterProcedureScheduler.markTableAsDeleted(TableName table,
Procedure<?> procedure)
Tries to remove the queue and the table-lock of the specified table.
|
(package private) LockAndQueue |
SchemaLocking.removeTableLock(TableName tableName) |
private void |
MasterProcedureScheduler.removeTableQueue(TableName tableName) |
protected static void |
CreateTableProcedure.setEnabledState(MasterProcedureEnv env,
TableName tableName) |
protected static void |
CreateTableProcedure.setEnablingState(MasterProcedureEnv env,
TableName tableName) |
protected static void |
DisableTableProcedure.setTableStateToDisabled(MasterProcedureEnv env,
TableName tableName)
Mark table state to Disabled
|
private static void |
DisableTableProcedure.setTableStateToDisabling(MasterProcedureEnv env,
TableName tableName)
Mark table state to Disabling
|
protected static void |
EnableTableProcedure.setTableStateToEnabled(MasterProcedureEnv env,
TableName tableName)
Mark table state to Enabled
|
protected static void |
EnableTableProcedure.setTableStateToEnabling(MasterProcedureEnv env,
TableName tableName)
Mark table state to Enabling
|
boolean |
MasterProcedureScheduler.waitRegions(Procedure<?> procedure,
TableName table,
RegionInfo... regionInfos)
Suspend the procedure if the specified set of regions are already locked.
|
boolean |
MasterProcedureScheduler.waitTableExclusiveLock(Procedure<?> procedure,
TableName table)
Suspend the procedure if the specified table is already locked.
|
private TableQueue |
MasterProcedureScheduler.waitTableQueueSharedLock(Procedure<?> procedure,
TableName table) |
boolean |
MasterProcedureScheduler.waitTableSharedLock(Procedure<?> procedure,
TableName table)
Suspend the procedure if the specified table is already locked.
|
void |
MasterProcedureScheduler.wakeRegions(Procedure<?> procedure,
TableName table,
RegionInfo... regionInfos)
Wake the procedures waiting for the specified regions
|
void |
MasterProcedureScheduler.wakeTableExclusiveLock(Procedure<?> procedure,
TableName table)
Wake the procedures waiting for the specified table
|
void |
MasterProcedureScheduler.wakeTableSharedLock(Procedure<?> procedure,
TableName table)
Wake the procedures waiting for the specified table
|
Modifier and Type | Field and Description |
---|---|
static TableName |
MasterRegionFactory.TABLE_NAME |
Modifier and Type | Method and Description |
---|---|
List<String> |
ReplicationPeerManager.getSerialPeerIdsBelongsTo(TableName tableName) |
private void |
OfflineTableReplicationQueueStorage.loadReplicationQueueData(org.apache.hadoop.conf.Configuration conf,
TableName tableName) |
private boolean |
ModifyPeerProcedure.needReopen(TableStateManager tsm,
TableName tn) |
private boolean |
AbstractPeerProcedure.needSetLastPushedSequenceId(TableStateManager tsm,
TableName tn) |
protected void |
AbstractPeerProcedure.setLastPushedSequenceIdForTable(MasterProcedureEnv env,
TableName tableName,
Map<String,Long> lastSeqIds) |
Modifier and Type | Method and Description |
---|---|
private void |
ReplicationPeerManager.checkNamespacesAndTableCfsConfigConflict(Set<String> namespaces,
Map<TableName,? extends Collection<String>> tableCfs)
Set a namespace in the peer config means that all tables in this namespace will be replicated
to the peer cluster.
|
Constructor and Description |
---|
OfflineTableReplicationQueueStorage(org.apache.hadoop.conf.Configuration conf,
TableName tableName) |
Modifier and Type | Field and Description |
---|---|
protected TableName |
TakeSnapshotHandler.snapshotTable |
private TableName |
MasterSnapshotVerifier.tableName |
Modifier and Type | Field and Description |
---|---|
private Map<TableName,Long> |
SnapshotManager.restoreTableToProcIdMap |
private Map<TableName,SnapshotSentinel> |
SnapshotManager.snapshotHandlers |
Modifier and Type | Method and Description |
---|---|
private long |
SnapshotManager.cloneSnapshot(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription reqSnapshot,
TableName tableName,
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot,
TableDescriptor snapshotTableDesc,
NonceKey nonceKey,
boolean restoreAcl,
String customSFT)
Clone the specified snapshot.
|
private boolean |
SnapshotManager.isRestoringTable(TableName tableName)
Verify if the restore of the specified table is in progress.
|
boolean |
SnapshotManager.isTableTakingAnySnapshot(TableName tableName) |
boolean |
SnapshotManager.isTakingSnapshot(TableName tableName)
Check to see if the specified table has a snapshot in progress.
|
private boolean |
SnapshotManager.isTakingSnapshot(TableName tableName,
boolean checkProcedure)
Check to see if the specified table has a snapshot in progress.
|
private long |
SnapshotManager.restoreSnapshot(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription reqSnapshot,
TableName tableName,
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot,
TableDescriptor snapshotTableDesc,
NonceKey nonceKey,
boolean restoreAcl)
Restore the specified snapshot.
|
void |
SnapshotManager.setSnapshotHandlerForTesting(TableName tableName,
SnapshotSentinel handler)
Set the handler for the current snapshot
|
Modifier and Type | Method and Description |
---|---|
private void |
SnapshotManager.cleanupSentinels(Map<TableName,SnapshotSentinel> sentinels)
Remove the sentinels that are marked as finished and the completion time has exceeded the
removal timeout.
|
private SnapshotSentinel |
SnapshotManager.removeSentinelIfFinished(Map<TableName,SnapshotSentinel> sentinels,
org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot)
Return the handler if it is currently live and has the same snapshot target name.
|
Modifier and Type | Field and Description |
---|---|
private static ConcurrentMap<TableName,String> |
ManualMobMaintHFileCleaner.MOB_REGIONS |
(package private) static ThreadLocal<org.apache.hbase.thirdparty.com.google.common.collect.SetMultimap<TableName,String>> |
DefaultMobStoreCompactor.mobRefSet |
Modifier and Type | Method and Description |
---|---|
static org.apache.hbase.thirdparty.com.google.common.collect.ImmutableSetMultimap.Builder<TableName,String> |
MobUtils.deserializeMobFileRefs(byte[] bytes)
Deserialize the set of referenced mob hfiles from store file metadata.
|
static Optional<TableName> |
MobUtils.getTableName(Cell cell)
Get the table name from when this cell was written into a mob hfile as a TableName.
|
Modifier and Type | Method and Description |
---|---|
private static void |
MobFileCleanupUtil.archiveMobFiles(org.apache.hadoop.conf.Configuration conf,
TableName tableName,
Admin admin,
byte[] family,
List<org.apache.hadoop.fs.Path> storeFiles)
Archives the mob files.
|
void |
RSMobFileCleanerChore.archiveMobFiles(org.apache.hadoop.conf.Configuration conf,
TableName tableName,
byte[] family,
List<org.apache.hadoop.fs.Path> storeFiles)
Archives the mob files.
|
private static void |
MobFileCleanupUtil.checkColumnFamilyDescriptor(org.apache.hadoop.conf.Configuration conf,
TableName table,
org.apache.hadoop.fs.FileSystem fs,
Admin admin,
ColumnFamilyDescriptor hcd,
Set<String> regionNames,
long maxCreationTimeToArchive) |
static void |
MobUtils.cleanExpiredMobFiles(org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.conf.Configuration conf,
TableName tableName,
ColumnFamilyDescriptor columnDescriptor,
CacheConfig cacheConfig,
long current)
Cleans the expired mob files.
|
static void |
MobFileCleanupUtil.cleanupObsoleteMobFiles(org.apache.hadoop.conf.Configuration conf,
TableName table,
Admin admin)
Performs housekeeping file cleaning (called by MOB Cleaner chore)
|
static org.apache.hadoop.fs.Path |
MobUtils.getMobFamilyPath(org.apache.hadoop.conf.Configuration conf,
TableName tableName,
String familyName)
Gets the family dir of the mob files.
|
static RegionInfo |
MobUtils.getMobRegionInfo(TableName tableName)
Gets the RegionInfo of the mob files.
|
static org.apache.hadoop.fs.Path |
MobUtils.getMobRegionPath(org.apache.hadoop.conf.Configuration conf,
TableName tableName)
Gets the region dir of the mob files.
|
static org.apache.hadoop.fs.Path |
MobUtils.getMobRegionPath(org.apache.hadoop.fs.Path rootDir,
TableName tableName)
Gets the region dir of the mob files under the specified root dir.
|
static org.apache.hadoop.fs.Path |
MobUtils.getMobTableDir(org.apache.hadoop.fs.Path rootDir,
TableName tableName)
Gets the table dir of the mob files under the qualified HBase root dir.
|
static boolean |
MobUtils.isMobRegionName(TableName tableName,
byte[] regionName)
Gets whether the current region name follows the pattern of a mob region name.
|
static boolean |
MobUtils.removeMobFiles(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
TableName tableName,
org.apache.hadoop.fs.Path tableDir,
byte[] family,
Collection<HStoreFile> storeFiles)
Archives the mob files.
|
private void |
MobFileCompactionChore.startCompaction(Admin admin,
TableName table,
RegionInfo region,
byte[] cf) |
Modifier and Type | Method and Description |
---|---|
private void |
DefaultMobStoreCompactor.calculateMobLengthMap(org.apache.hbase.thirdparty.com.google.common.collect.SetMultimap<TableName,String> mobRefs) |
static byte[] |
MobUtils.serializeMobFileRefs(org.apache.hbase.thirdparty.com.google.common.collect.SetMultimap<TableName,String> mobRefSet)
Serialize a set of referenced mob hfiles
|
Modifier and Type | Field and Description |
---|---|
(package private) TableName |
MobRefReporter.MobRefReducer.table |
Modifier and Type | Field and Description |
---|---|
static TableName |
WALEventTrackerTableAccessor.WAL_EVENT_TRACKER_TABLE_NAME
WALEventTrackerTableAccessor.WAL_EVENT_TRACKER_TABLE_NAME_STR table name - can be enabled with config -
hbase.regionserver.wal.event.tracker.enabled |
Modifier and Type | Field and Description |
---|---|
private Map<TableName,AtomicInteger> |
NamespaceTableAndRegionInfo.tableAndRegionInfo |
Modifier and Type | Method and Description |
---|---|
(package private) Set<TableName> |
NamespaceTableAndRegionInfo.getTables()
Gets the set of table names belonging to namespace.
|
Modifier and Type | Method and Description |
---|---|
private void |
NamespaceStateManager.addTable(TableName tableName,
int regionCount) |
(package private) void |
NamespaceTableAndRegionInfo.addTable(TableName tableName,
int regionCount) |
(package private) boolean |
NamespaceStateManager.checkAndUpdateNamespaceRegionCount(TableName name,
byte[] regionName,
int incr)
Check if adding a region violates namespace quota, if not update namespace cache.
|
(package private) void |
NamespaceStateManager.checkAndUpdateNamespaceRegionCount(TableName name,
int incr)
Check and update region count for an existing table.
|
(package private) void |
NamespaceStateManager.checkAndUpdateNamespaceTableCount(TableName table,
int numRegions) |
void |
NamespaceAuditor.checkQuotaToCreateTable(TableName tName,
int regions)
Check quota to create table.
|
void |
NamespaceAuditor.checkQuotaToUpdateRegion(TableName tName,
int regions)
Check and update region count quota for an existing table.
|
private void |
NamespaceAuditor.checkTableTypeAndThrowException(TableName name) |
(package private) boolean |
NamespaceTableAndRegionInfo.containsTable(TableName tableName) |
(package private) int |
NamespaceTableAndRegionInfo.decrementRegionCountForTable(TableName tableName,
int count) |
int |
NamespaceAuditor.getRegionCountOfTable(TableName tName)
Get region count for table
|
(package private) int |
NamespaceTableAndRegionInfo.getRegionCountOfTable(TableName tableName) |
(package private) int |
NamespaceTableAndRegionInfo.incRegionCountForTable(TableName tableName,
int count) |
void |
NamespaceAuditor.removeFromNamespaceUsage(TableName tableName) |
(package private) void |
NamespaceStateManager.removeTable(TableName tableName) |
(package private) void |
NamespaceTableAndRegionInfo.removeTable(TableName tableName) |
Modifier and Type | Field and Description |
---|---|
private Map<TableName,Procedure> |
MasterFlushTableProcedureManager.procMap |
Modifier and Type | Field and Description |
---|---|
static TableName |
QuotaTableUtil.QUOTA_TABLE_NAME
System table for quotas
|
private TableName |
QuotaSettings.tableName |
private TableName |
FileArchiverNotifierImpl.tn |
(package private) TableName |
FileArchiverNotifierFactoryImpl.CacheKey.tn |
Modifier and Type | Method and Description |
---|---|
protected static TableName |
QuotaTableUtil.getTableFromRowKey(byte[] key) |
TableName |
QuotaSettings.getTableName() |
Modifier and Type | Method and Description |
---|---|
(package private) Map<TableName,SpaceViolationPolicyEnforcement> |
RegionServerSpaceQuotaManager.copyActiveEnforcements()
Returns the collection of tables which have quota violation policies enforced on this
RegionServer.
|
Map<TableName,SpaceQuotaSnapshot> |
RegionServerSpaceQuotaManager.copyQuotaSnapshots()
Copies the last
SpaceQuotaSnapshot s that were recorded. |
Map<TableName,SpaceQuotaSnapshot> |
SpaceQuotaRefresherChore.fetchSnapshotsFromQuotaTable()
Reads all quota snapshots from the quota table.
|
static Map<TableName,QuotaState> |
QuotaUtil.fetchTableQuotas(Connection connection,
List<Get> gets,
Map<TableName,Double> tableMachineFactors) |
Set<TableName> |
QuotaObserverChore.TablesWithQuotas.filterInsufficientlyReportedTables(QuotaSnapshotStore<TableName> tableStore)
Filters out all tables for which the Master currently doesn't have enough region space
reports received from RegionServers yet.
|
Map<TableName,SpaceQuotaSnapshot> |
RegionServerSpaceQuotaManager.getActivePoliciesAsMap()
Converts a map of table to
SpaceViolationPolicyEnforcement s into
SpaceViolationPolicy s. |
(package private) Map<TableName,SpaceViolationPolicyEnforcement> |
ActivePolicyEnforcement.getLocallyCachedPolicies()
Returns an unmodifiable version of the policy enforcements that were cached because they are
not in violation of their quota.
|
Set<TableName> |
QuotaObserverChore.TablesWithQuotas.getNamespaceQuotaTables()
Returns an unmodifiable view of all tables in namespaces that have namespace quotas.
|
Map<TableName,SpaceViolationPolicyEnforcement> |
ActivePolicyEnforcement.getPolicies()
Returns an unmodifiable version of the active
SpaceViolationPolicyEnforcement s. |
static Map<TableName,SpaceQuotaSnapshot> |
QuotaTableUtil.getSnapshots(Connection conn)
Fetches all
SpaceQuotaSnapshot objects from the hbase:quota table. |
(package private) org.apache.hbase.thirdparty.com.google.common.collect.Multimap<TableName,String> |
SnapshotQuotaObserverChore.getSnapshotsFromTables(Admin admin,
Set<TableName> tablesToFetchSnapshotsFrom)
Computes a mapping of originating
TableName to snapshots, when the TableName
exists in the provided Set . |
(package private) org.apache.hbase.thirdparty.com.google.common.collect.Multimap<TableName,String> |
SnapshotQuotaObserverChore.getSnapshotsToComputeSize()
Fetches each table with a quota (table or namespace quota), and then fetch the name of each
snapshot which was created from that table.
|
(package private) Map<TableName,QuotaState> |
QuotaCache.getTableQuotaCache() |
Map<TableName,SpaceQuotaSnapshot> |
QuotaObserverChore.getTableQuotaSnapshots()
Returns an unmodifiable view over the current
SpaceQuotaSnapshot objects for each HBase
table with a quota defined. |
Set<TableName> |
QuotaObserverChore.TablesWithQuotas.getTableQuotaTables()
Returns an unmodifiable view of all tables with table quotas.
|
org.apache.hbase.thirdparty.com.google.common.collect.Multimap<String,TableName> |
QuotaObserverChore.TablesWithQuotas.getTablesByNamespace()
Returns a view of all tables that reside in a namespace with a namespace quota, grouped by
the namespace itself.
|
static org.apache.hbase.thirdparty.com.google.common.collect.Multimap<TableName,String> |
QuotaTableUtil.getTableSnapshots(Connection conn)
Returns a multimap for all existing table snapshot entries.
|
(package private) QuotaSnapshotStore<TableName> |
QuotaObserverChore.getTableSnapshotStore() |
Modifier and Type | Method and Description |
---|---|
void |
QuotaObserverChore.TablesWithQuotas.addNamespaceQuotaTable(TableName tn)
Adds a table with a namespace quota.
|
static void |
QuotaUtil.addTableQuota(Connection connection,
TableName table,
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas data) |
void |
QuotaObserverChore.TablesWithQuotas.addTableQuotaTable(TableName tn)
Adds a table with a table quota.
|
static void |
QuotaUtil.addUserQuota(Connection connection,
String user,
TableName table,
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas data) |
boolean |
RegionServerSpaceQuotaManager.areCompactionsDisabled(TableName tableName)
Returns whether or not compactions should be disabled for the given
tableName per
a space quota violation policy. |
org.apache.hadoop.hbase.shaded.protobuf.generated.RegionServerStatusProtos.FileArchiveNotificationRequest |
RegionServerSpaceQuotaManager.buildFileArchiveRequest(TableName tn,
Collection<Map.Entry<String,Long>> archivedFiles)
Builds the protobuf message to inform the Master of files being archived.
|
void |
MasterQuotaManager.checkAndUpdateNamespaceRegionQuota(TableName tName,
int regions) |
void |
MasterQuotaManager.checkNamespaceTableAndRegionQuota(TableName tName,
int regions) |
SpaceViolationPolicyEnforcement |
SpaceViolationPolicyEnforcementFactory.create(RegionServerServices rss,
TableName tableName,
SpaceQuotaSnapshot snapshot)
Constructs the appropriate
SpaceViolationPolicyEnforcement for tables that are in
violation of their space quota. |
(package private) static Put |
QuotaTableUtil.createPutForSnapshotSize(TableName tableName,
String snapshot,
long size)
|
(package private) static Put |
QuotaTableUtil.createPutForSpaceSnapshot(TableName tableName,
SpaceQuotaSnapshot snapshot)
|
(package private) static Scan |
QuotaTableUtil.createScanForSpaceSnapshotSizes(TableName table) |
SpaceViolationPolicyEnforcement |
SpaceViolationPolicyEnforcementFactory.createWithoutViolation(RegionServerServices rss,
TableName tableName,
SpaceQuotaSnapshot snapshot)
Creates the "default"
SpaceViolationPolicyEnforcement for a table that isn't in
violation. |
static void |
QuotaUtil.deleteTableQuota(Connection connection,
TableName table) |
static void |
QuotaUtil.deleteUserQuota(Connection connection,
String user,
TableName table) |
static void |
QuotaUtil.disableTableIfNotDisabled(Connection conn,
TableName tableName)
Method to disable a table, if not already disabled.
|
void |
RegionServerSpaceQuotaManager.disableViolationPolicyEnforcement(TableName tableName)
Disables enforcement on any violation policy on the given
tableName . |
static void |
QuotaUtil.enableTableIfNotEnabled(Connection conn,
TableName tableName)
Method to enable a table, if not already enabled.
|
void |
RegionServerSpaceQuotaManager.enforceViolationPolicy(TableName tableName,
SpaceQuotaSnapshot snapshot)
Enforces the given violationPolicy on the given table in this RegionServer.
|
Iterable<Map.Entry<RegionInfo,Long>> |
TableQuotaSnapshotStore.filterBySubject(TableName table) |
private static List<QuotaSettings> |
QuotaSettingsFactory.fromQuotas(String userName,
TableName tableName,
String namespace,
String regionServer,
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas quotas) |
(package private) static QuotaSettings |
QuotaSettingsFactory.fromSpace(TableName table,
String namespace,
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceQuota protoQuota) |
(package private) static SpaceLimitSettings |
SpaceLimitSettings.fromSpaceQuota(TableName tableName,
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceQuota proto)
Constructs a
SpaceLimitSettings from the provided protobuf message and tablename. |
(package private) static List<QuotaSettings> |
QuotaSettingsFactory.fromTableQuotas(TableName tableName,
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas quotas) |
static List<ThrottleSettings> |
QuotaSettingsFactory.fromTableThrottles(TableName tableName,
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Throttle throttle) |
protected static List<ThrottleSettings> |
QuotaSettingsFactory.fromThrottle(String userName,
TableName tableName,
String namespace,
String regionServer,
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Throttle throttle) |
(package private) static ThrottleSettings |
ThrottleSettings.fromTimedQuota(String userName,
TableName tableName,
String namespace,
String regionServer,
ThrottleType type,
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.TimedQuota timedQuota) |
(package private) static List<QuotaSettings> |
QuotaSettingsFactory.fromUserQuotas(String userName,
TableName tableName,
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas quotas) |
FileArchiverNotifier |
FileArchiverNotifierFactory.get(Connection conn,
org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
TableName tn)
Creates or obtains a
FileArchiverNotifier instance for the given args. |
FileArchiverNotifier |
FileArchiverNotifierFactoryImpl.get(Connection conn,
org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
TableName tn)
Returns the
FileArchiverNotifier instance for the given TableName . |
static SpaceQuotaSnapshot |
QuotaTableUtil.getCurrentSnapshotFromQuotaTable(Connection conn,
TableName tableName)
Returns the current space quota snapshot of the given
tableName from
QuotaTableUtil.QUOTA_TABLE_NAME or null if the no quota information is available for
that tableName. |
SpaceQuotaSnapshot |
TableQuotaSnapshotStore.getCurrentState(TableName table) |
(package private) FileArchiverNotifier |
SnapshotQuotaObserverChore.getNotifierForTable(TableName tn)
Returns the correct instance of
FileArchiverNotifier for the given table name. |
(package private) int |
QuotaObserverChore.TablesWithQuotas.getNumRegions(TableName table)
Computes the total number of regions in a table.
|
(package private) int |
QuotaObserverChore.TablesWithQuotas.getNumReportedRegions(TableName table,
QuotaSnapshotStore<TableName> tableStore)
Computes the number of regions reported for a table.
|
SpaceViolationPolicyEnforcement |
ActivePolicyEnforcement.getPolicyEnforcement(TableName tableName)
Returns the proper
SpaceViolationPolicyEnforcement implementation for the given table. |
OperationQuota |
RegionServerRpcQuotaManager.getQuota(org.apache.hadoop.security.UserGroupInformation ugi,
TableName table,
int blockSizeBytes)
Returns the quota for an operation.
|
(package private) org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas |
TableQuotaSnapshotStore.getQuotaForTable(TableName table)
Fetches the table quota.
|
int |
MasterQuotaManager.getRegionCountOfTable(TableName tName)
Returns cached region count, or -1 if quota manager is disabled or table status not found
|
protected static byte[] |
QuotaTableUtil.getSettingsQualifierForUserTable(TableName tableName) |
(package private) long |
FileArchiverNotifierImpl.getSizeOfStoreFile(TableName tn,
FileArchiverNotifierImpl.StoreFileReference storeFileName)
Computes the size of the store files for a single region.
|
(package private) long |
FileArchiverNotifierImpl.getSizeOfStoreFile(TableName tn,
String regionName,
String family,
String storeFile)
Computes the size of the store file given its name, region and family name in the archive
directory.
|
(package private) long |
FileArchiverNotifierImpl.getSizeOfStoreFiles(TableName tn,
Set<FileArchiverNotifierImpl.StoreFileReference> storeFileNames)
Computes the size of each store file in
storeFileNames |
(package private) long |
TableQuotaSnapshotStore.getSnapshotSizesForTable(TableName tn)
Fetches any serialized snapshot sizes from the quota table for the
tn provided. |
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceQuota |
TableQuotaSnapshotStore.getSpaceQuota(TableName subject) |
QuotaLimiter |
QuotaCache.getTableLimiter(TableName table)
Returns the limiter associated to the specified table.
|
QuotaLimiter |
UserQuotaState.getTableLimiter(TableName table)
Return the limiter for the specified table associated with this quota.
|
static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas |
QuotaTableUtil.getTableQuota(Connection connection,
TableName table) |
(package private) SpaceQuotaSnapshot |
QuotaObserverChore.getTableQuotaSnapshot(TableName table)
Fetches the
SpaceQuotaSnapshot for the given table. |
protected static byte[] |
QuotaTableUtil.getTableRowKey(TableName table) |
SpaceQuotaSnapshot |
TableQuotaSnapshotStore.getTargetState(TableName table,
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceQuota spaceQuota) |
QuotaLimiter |
QuotaCache.getUserLimiter(org.apache.hadoop.security.UserGroupInformation ugi,
TableName table)
Returns the limiter associated to the specified user/table.
|
static org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas |
QuotaTableUtil.getUserQuota(Connection connection,
String user,
TableName table) |
boolean |
QuotaObserverChore.TablesWithQuotas.hasNamespaceQuota(TableName tn)
Returns true if the table exists in a namespace with a namespace quota.
|
boolean |
QuotaObserverChore.TablesWithQuotas.hasTableQuota(TableName tn)
Returns true if the given table has a table quota.
|
void |
SpaceViolationPolicyEnforcement.initialize(RegionServerServices rss,
TableName tableName,
SpaceQuotaSnapshot snapshot)
Initializes this policy instance.
|
private boolean |
MasterQuotaManager.isInViolationAndPolicyDisable(TableName tableName,
QuotaObserverChore quotaObserverChore)
Method to check if a table is in violation and policy set on table is DISABLE.
|
static QuotaSettings |
QuotaSettingsFactory.limitTableSpace(TableName tableName,
long sizeLimit,
SpaceViolationPolicy violationPolicy)
Creates a
QuotaSettings object to limit the FileSystem space usage for the given table
to the given size in bytes. |
(package private) static Get |
QuotaTableUtil.makeGetForSnapshotSize(TableName tn,
String snapshot)
Creates a
Get for the HBase snapshot's size against the given table. |
static Get |
QuotaTableUtil.makeGetForTableQuotas(TableName table) |
static Get |
QuotaTableUtil.makeQuotaSnapshotGetForTable(TableName tn)
Creates a
Get which returns only SpaceQuotaSnapshot from the quota table for a
specific table. |
static Scan |
QuotaTableUtil.makeQuotaSnapshotScanForTable(TableName tn)
Creates a
Scan which returns only SpaceQuotaSnapshot from the quota table for a
specific table. |
protected static void |
QuotaTableUtil.parseTableResult(TableName table,
Result result,
QuotaTableUtil.TableQuotasVisitor visitor) |
void |
MasterQuotasObserver.postDeleteTable(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName) |
void |
MasterQuotaManager.removeRegionSizesForTable(TableName tableName)
Removes each region size entry where the RegionInfo references the provided TableName.
|
void |
MasterQuotaManager.removeTableFromNamespaceQuota(TableName tName)
Remove table from namespace quota.
|
static QuotaSettings |
QuotaSettingsFactory.removeTableSpaceLimit(TableName tableName)
Creates a
QuotaSettings object to remove the FileSystem space quota for the given
table. |
void |
TableQuotaSnapshotStore.setCurrentState(TableName table,
SpaceQuotaSnapshot snapshot) |
void |
UserQuotaState.setQuotas(TableName table,
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas quotas)
Add the quota information of the specified table.
|
void |
MasterQuotaManager.setTableQuota(TableName table,
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest req) |
(package private) void |
QuotaObserverChore.setTableQuotaSnapshot(TableName table,
SpaceQuotaSnapshot snapshot)
Stores the quota state for the given table.
|
void |
MasterQuotaManager.setUserQuota(String userName,
TableName table,
org.apache.hadoop.hbase.shaded.protobuf.generated.MasterProtos.SetQuotaRequest req) |
private static QuotaSettings |
QuotaSettingsFactory.throttle(String userName,
TableName tableName,
String namespace,
String regionServer,
ThrottleType type,
long limit,
TimeUnit timeUnit,
QuotaScope scope) |
static QuotaSettings |
QuotaSettingsFactory.throttleTable(TableName tableName,
ThrottleType type,
long limit,
TimeUnit timeUnit)
Throttle the specified table.
|
static QuotaSettings |
QuotaSettingsFactory.throttleTable(TableName tableName,
ThrottleType type,
long limit,
TimeUnit timeUnit,
QuotaScope scope)
Throttle the specified table.
|
static QuotaSettings |
QuotaSettingsFactory.throttleUser(String userName,
TableName tableName,
ThrottleType type,
long limit,
TimeUnit timeUnit)
Throttle the specified user on the specified table.
|
static QuotaSettings |
QuotaSettingsFactory.throttleUser(String userName,
TableName tableName,
ThrottleType type,
long limit,
TimeUnit timeUnit,
QuotaScope scope)
Throttle the specified user on the specified table.
|
void |
SpaceQuotaSnapshotNotifier.transitionTable(TableName tableName,
SpaceQuotaSnapshot snapshot)
Informs the cluster of the current state of a space quota for a table.
|
void |
TableSpaceQuotaSnapshotNotifier.transitionTable(TableName tableName,
SpaceQuotaSnapshot snapshot) |
static QuotaSettings |
QuotaSettingsFactory.unthrottleTable(TableName tableName)
Remove the throttling for the specified table.
|
static QuotaSettings |
QuotaSettingsFactory.unthrottleTableByThrottleType(TableName tableName,
ThrottleType type)
Remove the throttling for the specified table.
|
static QuotaSettings |
QuotaSettingsFactory.unthrottleUser(String userName,
TableName tableName)
Remove the throttling for the specified user on the specified table.
|
static QuotaSettings |
QuotaSettingsFactory.unthrottleUserByThrottleType(String userName,
TableName tableName,
ThrottleType type)
Remove the throttling for the specified user on the specified table.
|
(package private) void |
QuotaObserverChore.updateTableQuota(TableName table,
SpaceQuotaSnapshot currentSnapshot,
SpaceQuotaSnapshot targetSnapshot)
Updates the hbase:quota table with the new quota policy for this
table if
necessary. |
void |
QuotaTableUtil.TableQuotasVisitor.visitTableQuotas(TableName tableName,
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas quotas) |
void |
QuotaTableUtil.UserQuotasVisitor.visitUserQuotas(String userName,
TableName table,
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas quotas) |
Modifier and Type | Method and Description |
---|---|
(package private) Map<String,Long> |
SnapshotQuotaObserverChore.computeSnapshotSizes(org.apache.hbase.thirdparty.com.google.common.collect.Multimap<TableName,String> snapshotsToComputeSize)
Computes the size of each snapshot provided given the current files referenced by the table.
|
(package private) static List<Delete> |
QuotaTableUtil.createDeletesForExistingTableSnapshotSizes(org.apache.hbase.thirdparty.com.google.common.collect.Multimap<TableName,String> snapshotEntriesToRemove)
Returns a list of
Delete to remove given table snapshot entries to remove from quota
table |
(package private) void |
SpaceQuotaRefresherChore.extractQuotaSnapshot(Result result,
Map<TableName,SpaceQuotaSnapshot> snapshots)
Wrapper around
QuotaTableUtil.extractQuotaSnapshot(Result, Map) for testing. |
static void |
QuotaTableUtil.extractQuotaSnapshot(Result result,
Map<TableName,SpaceQuotaSnapshot> snapshots)
Extracts the
SpaceViolationPolicy and TableName from the provided
Result and adds them to the given Map . |
static Map<TableName,QuotaState> |
QuotaUtil.fetchTableQuotas(Connection connection,
List<Get> gets,
Map<TableName,Double> tableMachineFactors) |
static Map<String,UserQuotaState> |
QuotaUtil.fetchUserQuotas(Connection connection,
List<Get> gets,
Map<TableName,Double> tableMachineQuotaFactors,
double factor) |
Set<TableName> |
QuotaObserverChore.TablesWithQuotas.filterInsufficientlyReportedTables(QuotaSnapshotStore<TableName> tableStore)
Filters out all tables for which the Master currently doesn't have enough region space
reports received from RegionServers yet.
|
(package private) int |
QuotaObserverChore.TablesWithQuotas.getNumReportedRegions(TableName table,
QuotaSnapshotStore<TableName> tableStore)
Computes the number of regions reported for a table.
|
(package private) org.apache.hbase.thirdparty.com.google.common.collect.Multimap<TableName,String> |
SnapshotQuotaObserverChore.getSnapshotsFromTables(Admin admin,
Set<TableName> tablesToFetchSnapshotsFrom)
Computes a mapping of originating
TableName to snapshots, when the TableName
exists in the provided Set . |
static Get |
QuotaTableUtil.makeGetForUserQuotas(String user,
Iterable<TableName> tables,
Iterable<String> namespaces) |
(package private) void |
QuotaObserverChore.processNamespacesWithQuotas(Set<String> namespacesWithQuotas,
org.apache.hbase.thirdparty.com.google.common.collect.Multimap<String,TableName> tablesByNamespace)
Processes each namespace which has a quota defined and moves all of the tables contained in
that namespace into or out of violation of the quota.
|
(package private) void |
QuotaObserverChore.processTablesWithQuotas(Set<TableName> tablesWithTableQuotas)
Processes each
TableName which has a quota defined and moves it in or out of violation
based on the space use. |
(package private) void |
SnapshotQuotaObserverChore.pruneNamespaceSnapshots(org.apache.hbase.thirdparty.com.google.common.collect.Multimap<TableName,String> snapshotsToComputeSize)
Removes the snapshot entries that are present in Quota table but not in snapshotsToComputeSize
|
(package private) void |
SnapshotQuotaObserverChore.pruneTableSnapshots(org.apache.hbase.thirdparty.com.google.common.collect.Multimap<TableName,String> snapshotsToComputeSize)
Removes the snapshot entries that are present in Quota table but not in snapshotsToComputeSize
|
(package private) void |
SnapshotQuotaObserverChore.removeExistingTableSnapshotSizes(org.apache.hbase.thirdparty.com.google.common.collect.Multimap<TableName,String> snapshotEntriesToRemove) |
(package private) void |
QuotaObserverChore.updateNamespaceQuota(String namespace,
SpaceQuotaSnapshot currentSnapshot,
SpaceQuotaSnapshot targetSnapshot,
org.apache.hbase.thirdparty.com.google.common.collect.Multimap<String,TableName> tablesByNamespace)
Updates the hbase:quota table with the target quota policy for this
namespace if
necessary. |
void |
RegionServerSpaceQuotaManager.updateQuotaSnapshot(Map<TableName,SpaceQuotaSnapshot> newSnapshots)
Updates the current
SpaceQuotaSnapshot s for the RegionServer. |
Constructor and Description |
---|
CacheKey(Connection conn,
org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
TableName tn) |
FileArchiverNotifierImpl(Connection conn,
org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
TableName tn) |
GlobalQuotaSettings(String userName,
TableName tableName,
String namespace,
String regionServer) |
GlobalQuotaSettingsImpl(String username,
TableName tableName,
String namespace,
String regionServer,
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Quotas quotas) |
GlobalQuotaSettingsImpl(String userName,
TableName tableName,
String namespace,
String regionServer,
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.Throttle throttleProto,
Boolean bypassGlobals,
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceQuota spaceProto) |
QuotaGlobalsSettingsBypass(String userName,
TableName tableName,
String namespace,
String regionServer,
boolean bypassGlobals) |
QuotaSettings(String userName,
TableName tableName,
String namespace,
String regionServer) |
SpaceLimitSettings(TableName tableName)
Constructs a
SpaceLimitSettings to remove a space quota on the given tableName . |
SpaceLimitSettings(TableName tableName,
long sizeLimit,
SpaceViolationPolicy violationPolicy) |
SpaceLimitSettings(TableName tableName,
String namespace,
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.SpaceLimitRequest req) |
ThrottleSettings(String userName,
TableName tableName,
String namespace,
String regionServer,
org.apache.hadoop.hbase.shaded.protobuf.generated.QuotaProtos.ThrottleRequest proto) |
Constructor and Description |
---|
ActivePolicyEnforcement(Map<TableName,SpaceViolationPolicyEnforcement> activePolicies,
Map<TableName,SpaceQuotaSnapshot> snapshots,
RegionServerServices rss) |
ActivePolicyEnforcement(Map<TableName,SpaceViolationPolicyEnforcement> activePolicies,
Map<TableName,SpaceQuotaSnapshot> snapshots,
RegionServerServices rss) |
ActivePolicyEnforcement(Map<TableName,SpaceViolationPolicyEnforcement> activePolicies,
Map<TableName,SpaceQuotaSnapshot> snapshots,
RegionServerServices rss,
SpaceViolationPolicyEnforcementFactory factory) |
ActivePolicyEnforcement(Map<TableName,SpaceViolationPolicyEnforcement> activePolicies,
Map<TableName,SpaceQuotaSnapshot> snapshots,
RegionServerServices rss,
SpaceViolationPolicyEnforcementFactory factory) |
Modifier and Type | Field and Description |
---|---|
(package private) TableName |
AbstractViolationPolicyEnforcement.tableName |
Modifier and Type | Method and Description |
---|---|
TableName |
AbstractViolationPolicyEnforcement.getTableName() |
Modifier and Type | Method and Description |
---|---|
void |
AbstractViolationPolicyEnforcement.initialize(RegionServerServices rss,
TableName tableName,
SpaceQuotaSnapshot snapshot) |
void |
AbstractViolationPolicyEnforcement.setTableName(TableName tableName) |
Modifier and Type | Field and Description |
---|---|
private TableName |
MetricsTableSourceImpl.tableName |
Modifier and Type | Field and Description |
---|---|
private Map<TableName,List<org.apache.hadoop.fs.Path>> |
HMobStore.map |
private ConcurrentHashMap<TableName,MetricsTableWrapperAggregateImpl.MetricsTableValues> |
MetricsTableWrapperAggregateImpl.metricsTableMap |
Modifier and Type | Method and Description |
---|---|
TableName |
StoreContext.getTableName() |
TableName |
Store.getTableName() |
TableName |
HStore.getTableName() |
Modifier and Type | Method and Description |
---|---|
Set<TableName> |
HRegionServer.getOnlineTables()
Gets the online tables in this RS.
|
Modifier and Type | Method and Description |
---|---|
private org.apache.hadoop.fs.Path |
SecureBulkLoadManager.createStagingDir(org.apache.hadoop.fs.Path baseDir,
User user,
TableName tableName) |
List<org.apache.hadoop.fs.Path> |
HMobStore.getLocations(TableName tableName) |
List<? extends Region> |
OnlineRegions.getRegions(TableName tableName)
Get all online regions of a table in this RS.
|
List<HRegion> |
RegionServerServices.getRegions(TableName tableName) |
List<HRegion> |
HRegionServer.getRegions(TableName tableName)
Gets the online regions of the specified table.
|
boolean |
RegionServerServices.reportFileArchivalForQuotas(TableName tableName,
Collection<Map.Entry<String,Long>> archivedFiles)
Reports a collection of files, and their sizes, that belonged to the given
table were
just moved to the archive directory. |
boolean |
HRegionServer.reportFileArchivalForQuotas(TableName tableName,
Collection<Map.Entry<String,Long>> archivedFiles) |
Modifier and Type | Method and Description |
---|---|
void |
StoreFileWriter.appendMobMetadata(org.apache.hbase.thirdparty.com.google.common.collect.SetMultimap<TableName,String> mobRefSet)
Appends MOB - specific metadata (even if it is empty)
|
private void |
RSRpcServices.executeOpenRegionProcedures(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.OpenRegionRequest request,
Map<TableName,TableDescriptor> tdCache) |
Modifier and Type | Field and Description |
---|---|
private TableName |
MetricsTableRequests.tableName |
Modifier and Type | Method and Description |
---|---|
private void |
MetricsTableRequests.init(TableName tableName,
org.apache.hadoop.conf.Configuration conf) |
private static String |
MetricsTableRequests.qualifyMetrics(String prefix,
TableName tableName) |
Constructor and Description |
---|
MetricsTableRequests(TableName tableName,
org.apache.hadoop.conf.Configuration conf) |
Modifier and Type | Field and Description |
---|---|
private TableName |
ModifyStoreFileTrackerProcedure.tableName |
Modifier and Type | Method and Description |
---|---|
TableName |
ModifyStoreFileTrackerProcedure.getTableName() |
Constructor and Description |
---|
InitializeStoreFileTrackerProcedure(MasterProcedureEnv env,
TableName tableName) |
ModifyColumnFamilyStoreFileTrackerProcedure(MasterProcedureEnv env,
TableName tableName,
byte[] family,
String dstSFT) |
ModifyStoreFileTrackerProcedure(MasterProcedureEnv env,
TableName tableName,
String dstSFT) |
ModifyTableStoreFileTrackerProcedure(MasterProcedureEnv env,
TableName tableName,
String dstSFT) |
Modifier and Type | Field and Description |
---|---|
private ConcurrentMap<TableName,MutableFastCounter> |
MetricsWALSourceImpl.perTableAppendCount |
private ConcurrentMap<TableName,MutableFastCounter> |
MetricsWALSourceImpl.perTableAppendSize |
Modifier and Type | Method and Description |
---|---|
void |
MetricsWALSource.incrementAppendCount(TableName tableName)
Increment the count of wal appends
|
void |
MetricsWALSourceImpl.incrementAppendCount(TableName tableName) |
void |
MetricsWALSource.incrementAppendSize(TableName tableName,
long size)
Add the append size.
|
void |
MetricsWALSourceImpl.incrementAppendSize(TableName tableName,
long size) |
Modifier and Type | Field and Description |
---|---|
static TableName |
ReplicationStorageFactory.REPLICATION_QUEUE_TABLE_NAME_DEFAULT |
private TableName |
TableReplicationQueueStorage.tableName |
Modifier and Type | Field and Description |
---|---|
private Map<TableName,? extends Collection<String>> |
ReplicationPeerConfig.excludeTableCFsMap |
private Map<TableName,List<String>> |
ReplicationPeerConfig.ReplicationPeerConfigBuilderImpl.excludeTableCFsMap |
private Map<TableName,? extends Collection<String>> |
ReplicationPeerConfig.tableCFsMap |
private Map<TableName,List<String>> |
ReplicationPeerConfig.ReplicationPeerConfigBuilderImpl.tableCFsMap |
Modifier and Type | Method and Description |
---|---|
Map<TableName,List<String>> |
ReplicationPeerConfig.getExcludeTableCFsMap() |
Map<TableName,List<String>> |
ReplicationPeerImpl.getTableCFs() |
Map<TableName,List<String>> |
ReplicationPeer.getTableCFs()
Get replicable (table, cf-list) map of this peer
|
Map<TableName,List<String>> |
ReplicationPeerConfig.getTableCFsMap() |
private Map<TableName,List<String>> |
ReplicationPeerConfig.unmodifiableTableCFsMap(Map<TableName,List<String>> tableCFsMap) |
Modifier and Type | Method and Description |
---|---|
static TableDescriptor |
ReplicationStorageFactory.createReplicationQueueTableDescriptor(TableName tableName) |
static ReplicationBarrierFamilyFormat.ReplicationBarrierResult |
ReplicationBarrierFamilyFormat.getReplicationBarrierResult(Connection conn,
TableName tableName,
byte[] row,
byte[] encodedRegionName) |
static ReplicationQueueStorage |
ReplicationStorageFactory.getReplicationQueueStorage(Connection conn,
org.apache.hadoop.conf.Configuration conf,
TableName tableName)
Create a new
ReplicationQueueStorage . |
static List<Pair<String,Long>> |
ReplicationBarrierFamilyFormat.getTableEncodedRegionNameAndLastBarrier(Connection conn,
TableName tableName) |
static List<String> |
ReplicationBarrierFamilyFormat.getTableEncodedRegionNamesForSerialReplication(Connection conn,
TableName tableName) |
static boolean |
ReplicationStorageFactory.isReplicationQueueTable(org.apache.hadoop.conf.Configuration conf,
TableName tableName) |
boolean |
ReplicationPeerConfig.needToReplicate(TableName table)
Decide whether the table need replicate to the peer cluster
|
boolean |
ReplicationPeerConfig.needToReplicate(TableName table,
byte[] family)
Decide whether the passed family of the table need replicate to the peer cluster according to
this peer config.
|
Modifier and Type | Method and Description |
---|---|
private static boolean |
ReplicationUtils.isTableCFsEqual(Map<TableName,List<String>> tableCFs1,
Map<TableName,List<String>> tableCFs2) |
private static boolean |
ReplicationUtils.isTableCFsEqual(Map<TableName,List<String>> tableCFs1,
Map<TableName,List<String>> tableCFs2) |
ReplicationPeerConfigBuilder |
ReplicationPeerConfig.ReplicationPeerConfigBuilderImpl.setExcludeTableCFsMap(Map<TableName,List<String>> excludeTableCFsMap) |
ReplicationPeerConfigBuilder |
ReplicationPeerConfigBuilder.setExcludeTableCFsMap(Map<TableName,List<String>> tableCFsMap)
Sets the mapping of table name to column families which should not be replicated.
|
ReplicationPeerConfigBuilder |
ReplicationPeerConfig.ReplicationPeerConfigBuilderImpl.setTableCFsMap(Map<TableName,List<String>> tableCFsMap) |
ReplicationPeerConfigBuilder |
ReplicationPeerConfigBuilder.setTableCFsMap(Map<TableName,List<String>> tableCFsMap)
Sets an explicit map of tables and column families in those tables that should be replicated to
the given peer.
|
private Map<TableName,List<String>> |
ReplicationPeerConfig.unmodifiableTableCFsMap(Map<TableName,List<String>> tableCFsMap) |
Constructor and Description |
---|
TableReplicationQueueStorage(Connection conn,
TableName tableName) |
Modifier and Type | Field and Description |
---|---|
static TableName |
ReplicationSinkTrackerTableCreator.REPLICATION_SINK_TRACKER_TABLE_NAME
ReplicationSinkTrackerTableCreator.REPLICATION_SINK_TRACKER_TABLE_NAME_STR table name - can be enabled with config -
hbase.regionserver.replication.sink.tracker.enabled |
Modifier and Type | Field and Description |
---|---|
private ConcurrentMap<TableName,String> |
SyncReplicationPeerMappingManager.table2PeerId |
Modifier and Type | Method and Description |
---|---|
void |
ReplicationSourceInterface.addHFileRefs(TableName tableName,
byte[] family,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs)
Add hfile names to the queue to be replicated.
|
void |
ReplicationSourceManager.addHFileRefs(TableName tableName,
byte[] family,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs) |
void |
ReplicationSource.addHFileRefs(TableName tableName,
byte[] family,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs) |
(package private) void |
Replication.addHFileRefsToQueue(TableName tableName,
byte[] family,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs) |
private void |
ReplicationSink.batch(TableName tableName,
Collection<List<Row>> allRows,
int batchRowSizeThreshold)
Do the changes and handle the pool
|
private void |
ReplicationSink.buildBulkLoadHFileMap(Map<String,List<Pair<byte[],List<String>>>> bulkLoadHFileMap,
TableName table,
org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.BulkLoadDescriptor bld) |
boolean |
SyncReplicationPeerInfoProvider.checkState(TableName table,
BiPredicate<SyncReplicationState,SyncReplicationState> checker)
Check whether the given table is contained in a sync replication peer which can pass the state
checker.
|
boolean |
SyncReplicationPeerInfoProviderImpl.checkState(TableName table,
BiPredicate<SyncReplicationState,SyncReplicationState> checker) |
private org.apache.hadoop.fs.Path |
HFileReplicator.createStagingDir(org.apache.hadoop.fs.Path baseDir,
User user,
TableName tableName) |
private void |
HFileReplicator.doBulkLoad(org.apache.hadoop.conf.Configuration conf,
TableName tableName,
org.apache.hadoop.fs.Path stagingDir,
Deque<BulkLoadHFiles.LoadQueueItem> queue,
int maxRetries) |
boolean |
WALEntrySinkFilter.filter(TableName table,
long writeTime) |
private String |
ReplicationSink.getHFilePath(TableName table,
org.apache.hadoop.hbase.shaded.protobuf.generated.WALProtos.BulkLoadDescriptor bld,
String storeFile,
byte[] family) |
(package private) String |
SyncReplicationPeerMappingManager.getPeerId(TableName tableName) |
Optional<Pair<String,String>> |
SyncReplicationPeerInfoProvider.getPeerIdAndRemoteWALDir(TableName table)
Return the peer id and remote WAL directory if the table is synchronously replicated and the
state is
SyncReplicationState.ACTIVE . |
Optional<Pair<String,String>> |
SyncReplicationPeerInfoProviderImpl.getPeerIdAndRemoteWALDir(TableName table) |
Modifier and Type | Method and Description |
---|---|
private org.apache.hbase.thirdparty.javax.ws.rs.core.Response |
SchemaResource.replace(TableName name,
TableSchemaModel model,
org.apache.hbase.thirdparty.javax.ws.rs.core.UriInfo uriInfo,
Admin admin) |
private org.apache.hbase.thirdparty.javax.ws.rs.core.Response |
SchemaResource.update(TableName name,
TableSchemaModel model,
org.apache.hbase.thirdparty.javax.ws.rs.core.UriInfo uriInfo,
Admin admin) |
Modifier and Type | Field and Description |
---|---|
(package private) static TableName |
RSGroupInfoManagerImpl.RSGROUP_TABLE_NAME |
Modifier and Type | Field and Description |
---|---|
(package private) org.apache.hbase.thirdparty.com.google.common.collect.ImmutableMap<TableName,RSGroupInfo> |
RSGroupInfoManagerImpl.RSGroupInfoHolder.tableName2Group |
private SortedSet<TableName> |
RSGroupInfo.tables
Deprecated.
Since 3.0.0, will be removed in 4.0.0. The rsgroup information will be stored in
the configuration of a table so this will be removed.
|
Modifier and Type | Method and Description |
---|---|
private Pair<Map<TableName,Map<ServerName,List<RegionInfo>>>,List<RegionPlan>> |
RSGroupBasedLoadBalancer.correctAssignments(Map<TableName,Map<ServerName,List<RegionInfo>>> existingAssignments) |
(package private) Map<TableName,Map<ServerName,List<RegionInfo>>> |
RSGroupInfoManagerImpl.getRSGroupAssignmentsByTable(TableStateManager tableStateManager,
String groupName)
This is an EXPENSIVE clone.
|
SortedSet<TableName> |
RSGroupInfo.getTables()
Deprecated.
Since 3.0.0, will be removed in 4.0.0. The rsgroup information will be stored in
the configuration of a table so this will be removed.
|
static List<TableName> |
RSGroupUtil.listTablesInRSGroup(MasterServices master,
String groupName) |
Modifier and Type | Method and Description |
---|---|
void |
RSGroupInfo.addTable(TableName table)
Deprecated.
Since 3.0.0, will be removed in 4.0.0. The rsgroup information will be stored in
the configuration of a table so this will be removed.
|
boolean |
RSGroupInfo.containsTable(TableName table)
Deprecated.
Since 3.0.0, will be removed in 4.0.0. The rsgroup information will be stored in
the configuration of a table so this will be removed.
|
String |
RSGroupInfoManager.determineRSGroupInfoForTable(TableName tableName)
Determine
RSGroupInfo for the given table. |
String |
RSGroupInfoManagerImpl.determineRSGroupInfoForTable(TableName tableName) |
String |
DisabledRSGroupInfoManager.determineRSGroupInfoForTable(TableName tableName) |
RSGroupInfo |
RSGroupInfoManager.getRSGroupForTable(TableName tableName)
Get
RSGroupInfo for the given table. |
RSGroupInfo |
RSGroupInfoManagerImpl.getRSGroupForTable(TableName tableName) |
RSGroupInfo |
DisabledRSGroupInfoManager.getRSGroupForTable(TableName tableName) |
static Optional<RSGroupInfo> |
RSGroupUtil.getRSGroupInfo(MasterServices master,
RSGroupInfoManager manager,
TableName tableName)
Will try to get the rsgroup from
TableDescriptor first, and then try to get the rsgroup
from the NamespaceDescriptor . |
RSGroupInfo |
RSGroupAdminClient.getRSGroupInfoOfTable(TableName tableName)
Deprecated.
Gets
RSGroupInfo for the given table's group. |
private boolean |
RSGroupInfoManagerImpl.isTableInGroup(TableName tableName,
String groupName,
Set<TableName> tablesInGroupCache) |
boolean |
RSGroupInfo.removeTable(TableName table)
Deprecated.
Since 3.0.0, will be removed in 4.0.0. The rsgroup information will be stored in
the configuration of a table so this will be removed.
|
Modifier and Type | Method and Description |
---|---|
void |
RSGroupInfo.addAllTables(Collection<TableName> arg)
Deprecated.
Since 3.0.0, will be removed in 4.0.0. The rsgroup information will be stored in
the configuration of a table so this will be removed.
|
List<RegionPlan> |
RSGroupBasedLoadBalancer.balanceCluster(Map<TableName,Map<ServerName,List<RegionInfo>>> loadOfAllTable)
Balance by RSGroup.
|
private Pair<Map<TableName,Map<ServerName,List<RegionInfo>>>,List<RegionPlan>> |
RSGroupBasedLoadBalancer.correctAssignments(Map<TableName,Map<ServerName,List<RegionInfo>>> existingAssignments) |
private boolean |
RSGroupInfoManagerImpl.isTableInGroup(TableName tableName,
String groupName,
Set<TableName> tablesInGroupCache) |
void |
RSGroupAdminClient.moveServersAndTables(Set<Address> servers,
Set<TableName> tables,
String targetGroup)
Deprecated.
Move given set of servers and tables to the specified target RegionServer group.
|
void |
RSGroupAdminClient.moveTables(Set<TableName> tables,
String targetGroup)
Deprecated.
Move given set of tables to the specified target RegionServer group.
|
private void |
RSGroupAdminServiceImpl.moveTablesAndWait(Set<TableName> tables,
String targetGroup)
Deprecated.
|
private void |
RSGroupInfoManagerImpl.moveTablesAndWait(Set<TableName> tables,
String targetGroup) |
void |
RSGroupInfoManager.setRSGroup(Set<TableName> tables,
String groupName)
Set group for tables.
|
void |
RSGroupInfoManagerImpl.setRSGroup(Set<TableName> tables,
String groupName) |
void |
DisabledRSGroupInfoManager.setRSGroup(Set<TableName> tables,
String groupName) |
void |
RSGroupBasedLoadBalancer.updateBalancerLoadInfo(Map<TableName,Map<ServerName,List<RegionInfo>>> loadOfAllTable) |
Constructor and Description |
---|
MigrateRSGroupProcedure(MasterProcedureEnv env,
TableName tableName) |
Constructor and Description |
---|
RSGroupInfo(String name,
Set<Address> servers,
Set<TableName> tables)
Deprecated.
Since 3.0.0, will be removed in 4.0.0. The rsgroup information for a table will be
stored in the configuration of a table so this will be removed.
|
Modifier and Type | Field and Description |
---|---|
static TableName |
PermissionStorage.ACL_TABLE_NAME
Internal storage table for access control lists
|
static TableName |
AccessControlClient.ACL_TABLE_NAME |
private TableName |
AuthResult.table |
private TableName |
AccessControlFilter.table |
private TableName |
TablePermission.table |
private TableName |
AuthResult.Params.tableName |
private TableName |
Permission.Builder.tableName |
private TableName |
GetUserPermissionsRequest.tableName |
private TableName |
GetUserPermissionsRequest.Builder.tableName |
Modifier and Type | Field and Description |
---|---|
private Map<TableName,List<UserPermission>> |
AccessController.tableAcls |
private ConcurrentHashMap<TableName,AuthManager.PermissionCache<TablePermission>> |
AuthManager.tableCache
Cache for table permission.
|
Modifier and Type | Method and Description |
---|---|
TableName |
AuthResult.getTableName() |
TableName |
GetUserPermissionsRequest.getTableName() |
TableName |
TablePermission.getTableName() |
private TableName |
AccessController.getTableName(Region region) |
private TableName |
AccessController.getTableName(RegionCoprocessorEnvironment e) |
static TableName |
ShadedAccessControlUtil.toTableName(org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName tableNamePB) |
Modifier and Type | Method and Description |
---|---|
(package private) static Pair<Set<String>,Set<TableName>> |
SnapshotScannerHDFSAclController.SnapshotScannerHDFSAclStorage.getUserNamespaceAndTable(Table aclTable,
String userName) |
Modifier and Type | Method and Description |
---|---|
boolean |
AuthManager.accessUserTable(User user,
TableName table,
Permission.Action action)
Checks if the user has access to the full table or at least a family/qualifier for the
specified action.
|
boolean |
SnapshotScannerHDFSAclHelper.addTableAcl(TableName tableName,
Set<String> users,
String operation)
Add table user acls
|
(package private) static void |
SnapshotScannerHDFSAclController.SnapshotScannerHDFSAclStorage.addUserTableHdfsAcl(Connection connection,
Set<String> users,
TableName tableName) |
(package private) static void |
SnapshotScannerHDFSAclController.SnapshotScannerHDFSAclStorage.addUserTableHdfsAcl(Connection connection,
String user,
TableName tableName) |
(package private) static void |
SnapshotScannerHDFSAclController.SnapshotScannerHDFSAclStorage.addUserTableHdfsAcl(Table aclTable,
String user,
TableName tableName) |
static AuthResult |
AuthResult.allow(String request,
String reason,
User user,
Permission.Action action,
TableName table,
byte[] family,
byte[] qualifier) |
static AuthResult |
AuthResult.allow(String request,
String reason,
User user,
Permission.Action action,
TableName table,
Map<byte[],? extends Collection<?>> families) |
boolean |
AuthManager.authorizeCell(User user,
TableName table,
Cell cell,
Permission.Action action)
Check if user has given action privilige in cell scope.
|
private boolean |
AuthManager.authorizeFamily(Set<TablePermission> permissions,
TableName table,
byte[] family,
Permission.Action action) |
private boolean |
AuthManager.authorizeTable(Set<TablePermission> permissions,
TableName table,
byte[] family,
byte[] qualifier,
Permission.Action action) |
boolean |
AuthManager.authorizeUserFamily(User user,
TableName table,
byte[] family,
Permission.Action action)
Check if user has given action privilige in table:family scope.
|
boolean |
AuthManager.authorizeUserTable(User user,
TableName table,
byte[] family,
byte[] qualifier,
Permission.Action action)
Check if user has given action privilige in table:family:qualifier scope.
|
boolean |
AuthManager.authorizeUserTable(User user,
TableName table,
byte[] family,
Permission.Action action)
Check if user has given action privilige in table:family scope.
|
boolean |
AuthManager.authorizeUserTable(User user,
TableName table,
Permission.Action action)
Check if user has given action privilige in table scope.
|
static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.GrantRequest |
AccessControlUtil.buildGrantRequest(String username,
TableName tableName,
byte[] family,
byte[] qualifier,
boolean mergeExistingPermissions,
org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Action... actions)
Create a request to grant user table permissions.
|
static org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.RevokeRequest |
AccessControlUtil.buildRevokeRequest(String username,
TableName tableName,
byte[] family,
byte[] qualifier,
org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.Permission.Action... actions)
Create a request to revoke user table permissions.
|
void |
AccessController.checkLockPermissions(ObserverContext<?> ctx,
String namespace,
TableName tableName,
RegionInfo[] regionInfos,
String reason) |
void |
AccessChecker.checkLockPermissions(User user,
String namespace,
TableName tableName,
RegionInfo[] regionInfos,
String reason) |
void |
NoopAccessChecker.checkLockPermissions(User user,
String namespace,
TableName tableName,
RegionInfo[] regionInfos,
String reason) |
(package private) void |
SnapshotScannerHDFSAclHelper.createTableDirectories(TableName tableName) |
void |
ZKPermissionWatcher.deleteTableACLNode(TableName tableName)
Delete the acl notify node of table
|
(package private) static void |
SnapshotScannerHDFSAclController.SnapshotScannerHDFSAclStorage.deleteTableHdfsAcl(Table aclTable,
TableName tableName) |
(package private) static void |
SnapshotScannerHDFSAclController.SnapshotScannerHDFSAclStorage.deleteUserTableHdfsAcl(Connection connection,
Set<String> users,
TableName tableName) |
(package private) static void |
SnapshotScannerHDFSAclController.SnapshotScannerHDFSAclStorage.deleteUserTableHdfsAcl(Table aclTable,
String user,
TableName tableName) |
static AuthResult |
AuthResult.deny(String request,
String reason,
User user,
Permission.Action action,
TableName table,
byte[] family,
byte[] qualifier) |
static AuthResult |
AuthResult.deny(String request,
String reason,
User user,
Permission.Action action,
TableName table,
Map<byte[],? extends Collection<?>> families) |
private boolean |
TablePermission.failCheckTable(TableName table) |
private Set<String> |
SnapshotScannerHDFSAclController.filterUsersToRemoveNsAccessAcl(Table aclTable,
TableName tableName,
Set<String> tablesUsers)
Remove table user access HDFS acl from namespace directory if the user has no permissions of
global, ns of the table or other tables of the ns, eg: Bob has 'ns1:t1' read permission, when
delete 'ns1:t1', if Bob has global read permission, '@ns1' read permission or
'ns1:other_tables' read permission, then skip remove Bob access acl in ns1Dirs, otherwise,
remove Bob access acl.
|
(package private) org.apache.hadoop.fs.Path |
SnapshotScannerHDFSAclHelper.PathHelper.getArchiveTableDir(TableName tableName) |
(package private) org.apache.hadoop.fs.Path |
SnapshotScannerHDFSAclHelper.PathHelper.getDataTableDir(TableName tableName) |
(package private) org.apache.hadoop.fs.Path |
SnapshotScannerHDFSAclHelper.PathHelper.getMobTableDir(TableName tableName) |
static org.apache.hbase.thirdparty.com.google.common.collect.ListMultimap<String,UserPermission> |
PermissionStorage.getTablePermissions(org.apache.hadoop.conf.Configuration conf,
TableName tableName) |
(package private) List<org.apache.hadoop.fs.Path> |
SnapshotScannerHDFSAclHelper.getTableRootPaths(TableName tableName,
boolean includeSnapshotPath)
return paths that user will table permission will visit
|
private List<org.apache.hadoop.fs.Path> |
SnapshotScannerHDFSAclHelper.getTableSnapshotPaths(TableName tableName) |
(package private) static Set<String> |
SnapshotScannerHDFSAclController.SnapshotScannerHDFSAclStorage.getTableUsers(Table aclTable,
TableName tableName) |
(package private) org.apache.hadoop.fs.Path |
SnapshotScannerHDFSAclHelper.PathHelper.getTmpTableDir(TableName tableName) |
static List<UserPermission> |
AccessControlUtil.getUserPermissions(org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterface protocol,
TableName t)
Deprecated.
Use
Admin.getUserPermissions(GetUserPermissionsRequest) instead. |
static List<UserPermission> |
AccessControlUtil.getUserPermissions(org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterface protocol,
TableName t,
byte[] columnFamily,
byte[] columnQualifier,
String userName)
Deprecated.
Use
Admin.getUserPermissions(GetUserPermissionsRequest) instead. |
(package private) Set<String> |
SnapshotScannerHDFSAclHelper.getUsersWithTableReadAction(TableName tableName,
boolean includeNamespace,
boolean includeGlobal)
Return users with table read permission
|
private UserPermission |
SnapshotScannerHDFSAclController.getUserTablePermission(org.apache.hadoop.conf.Configuration conf,
String userName,
TableName tableName) |
static List<UserPermission> |
PermissionStorage.getUserTablePermissions(org.apache.hadoop.conf.Configuration conf,
TableName tableName,
byte[] cf,
byte[] cq,
String userName,
boolean hasFilterUser)
Returns the currently granted permissions for a given table as the specified user plus
associated permissions.
|
private static void |
AccessControlClient.grant(Connection connection,
TableName tableName,
String userName,
byte[] family,
byte[] qual,
boolean mergeExistingPermissions,
Permission.Action... actions)
Grants permission on the specified table for the specified user
|
static void |
AccessControlClient.grant(Connection connection,
TableName tableName,
String userName,
byte[] family,
byte[] qual,
Permission.Action... actions)
Grants permission on the specified table for the specified user.
|
static void |
AccessControlUtil.grant(org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterface protocol,
String userShortName,
TableName tableName,
byte[] f,
byte[] q,
boolean mergeExistingPermissions,
Permission.Action... actions)
Deprecated.
Use
Admin.grant(UserPermission, boolean) instead. |
static boolean |
AccessControlUtil.hasPermission(org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterface protocol,
TableName tableName,
byte[] columnFamily,
byte[] columnQualifier,
String userName,
Permission.Action[] actions)
Deprecated.
Use
Admin.hasUserPermissions(String, List) instead. |
(package private) static boolean |
SnapshotScannerHDFSAclController.SnapshotScannerHDFSAclStorage.hasUserTableHdfsAcl(Table aclTable,
String user,
TableName tableName) |
boolean |
TablePermission.implies(TableName table,
byte[] family,
byte[] qualifier,
Permission.Action action)
Check if given action can performs on given table:family:qualifier.
|
boolean |
TablePermission.implies(TableName table,
byte[] family,
Permission.Action action)
Check if given action can performs on given table:family.
|
boolean |
TablePermission.implies(TableName table,
KeyValue kv,
Permission.Action action)
Checks if this permission grants access to perform the given action on the given table and key
value.
|
private boolean |
SnapshotScannerHDFSAclController.isHdfsAclSet(Table aclTable,
String userName,
String namespace,
TableName tableName)
Check if user global/namespace/table HDFS acls is already set
|
private boolean |
SnapshotScannerHDFSAclController.isHdfsAclSet(Table aclTable,
String userName,
TableName tableName) |
private boolean |
SnapshotScannerHDFSAclController.needHandleTableHdfsAcl(TableName tableName,
String operation) |
static Permission.Builder |
Permission.newBuilder(TableName tableName)
Build a table permission
|
static GetUserPermissionsRequest.Builder |
GetUserPermissionsRequest.newBuilder(TableName tableName)
Build a get table permission request
|
private AuthResult |
AccessChecker.permissionGranted(String request,
User user,
Permission.Action permRequest,
TableName tableName,
byte[] family,
byte[] qualifier) |
AuthResult |
AccessChecker.permissionGranted(String request,
User user,
Permission.Action permRequest,
TableName tableName,
Map<byte[],? extends Collection<?>> families)
Check the current user for authorization to perform a specific action against the given set of
row data.
|
AuthResult |
NoopAccessChecker.permissionGranted(String request,
User user,
Permission.Action permRequest,
TableName tableName,
Map<byte[],? extends Collection<?>> families) |
void |
SnapshotScannerHDFSAclController.postCompletedDeleteTableAction(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName) |
void |
SnapshotScannerHDFSAclController.postCompletedTruncateTableAction(ObserverContext<MasterCoprocessorEnvironment> c,
TableName tableName) |
void |
AccessController.postDeleteTable(ObserverContext<MasterCoprocessorEnvironment> c,
TableName tableName) |
void |
SnapshotScannerHDFSAclController.postModifyTable(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName,
TableDescriptor oldDescriptor,
TableDescriptor currentDescriptor) |
void |
AccessController.postModifyTable(ObserverContext<MasterCoprocessorEnvironment> c,
TableName tableName,
TableDescriptor oldDesc,
TableDescriptor currentDesc) |
void |
AccessController.postTruncateTable(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName) |
void |
AccessController.preDeleteTable(ObserverContext<MasterCoprocessorEnvironment> c,
TableName tableName) |
void |
AccessController.preDisableTable(ObserverContext<MasterCoprocessorEnvironment> c,
TableName tableName) |
void |
AccessController.preEnableTable(ObserverContext<MasterCoprocessorEnvironment> c,
TableName tableName) |
void |
AccessController.preGetRSGroupInfoOfTable(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName) |
void |
AccessController.preGetUserPermissions(ObserverContext<MasterCoprocessorEnvironment> ctx,
String userName,
String namespace,
TableName tableName,
byte[] family,
byte[] qualifier) |
private void |
AccessController.preGetUserPermissions(User caller,
String userName,
String namespace,
TableName tableName,
byte[] family,
byte[] qualifier) |
void |
AccessController.preLockHeartbeat(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName,
String description) |
String |
AccessController.preModifyColumnFamilyStoreFileTracker(ObserverContext<MasterCoprocessorEnvironment> c,
TableName tableName,
byte[] family,
String dstSFT) |
TableDescriptor |
CoprocessorWhitelistMasterObserver.preModifyTable(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName,
TableDescriptor currentDesc,
TableDescriptor newDesc) |
TableDescriptor |
AccessController.preModifyTable(ObserverContext<MasterCoprocessorEnvironment> c,
TableName tableName,
TableDescriptor currentDesc,
TableDescriptor newDesc) |
String |
AccessController.preModifyTableStoreFileTracker(ObserverContext<MasterCoprocessorEnvironment> c,
TableName tableName,
String dstSFT) |
void |
AccessController.preRequestLock(ObserverContext<MasterCoprocessorEnvironment> ctx,
String namespace,
TableName tableName,
RegionInfo[] regionInfos,
String description) |
void |
AccessController.preSetTableQuota(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName,
GlobalQuotaSettings quotas) |
void |
AccessController.preSetUserQuota(ObserverContext<MasterCoprocessorEnvironment> ctx,
String userName,
TableName tableName,
GlobalQuotaSettings quotas) |
void |
AccessController.preSplitRegion(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName,
byte[] splitRow) |
void |
AccessController.preTableFlush(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName) |
void |
AccessController.preTruncateTable(ObserverContext<MasterCoprocessorEnvironment> c,
TableName tableName) |
void |
AuthManager.refreshTableCacheFromWritable(TableName table,
byte[] data)
Update acl info for table.
|
boolean |
SnapshotScannerHDFSAclHelper.removeNamespaceAccessAcl(TableName tableName,
Set<String> removeUsers,
String operation)
Remove table access acl from namespace dir when delete table
|
void |
AuthManager.removeTable(TableName table)
Remove given table from AuthManager's table cache.
|
boolean |
SnapshotScannerHDFSAclHelper.removeTableAcl(TableName tableName,
Set<String> users)
Remove table acls when modify table
|
boolean |
SnapshotScannerHDFSAclHelper.removeTableDefaultAcl(TableName tableName,
Set<String> removeUsers)
Remove default acl from table archive dir when delete table
|
(package private) static void |
PermissionStorage.removeTablePermissions(org.apache.hadoop.conf.Configuration conf,
TableName tableName,
byte[] column,
Table t)
Remove specified table column from the acl table.
|
(package private) static void |
PermissionStorage.removeTablePermissions(org.apache.hadoop.conf.Configuration conf,
TableName tableName,
Table t)
Remove specified table from the _acl_ table.
|
private static void |
PermissionStorage.removeTablePermissions(TableName tableName,
byte[] column,
Table table,
boolean closeTable) |
private void |
SnapshotScannerHDFSAclController.removeUserTableHdfsAcl(Table aclTable,
String userName,
TableName tableName,
UserPermission userPermission) |
void |
AccessController.requireAccess(ObserverContext<?> ctx,
String request,
TableName tableName,
Permission.Action... permissions) |
void |
AccessChecker.requireAccess(User user,
String request,
TableName tableName,
Permission.Action... permissions)
Authorizes that the current user has any of the given permissions to access the table.
|
void |
NoopAccessChecker.requireAccess(User user,
String request,
TableName tableName,
Permission.Action... permissions) |
void |
AccessController.requireGlobalPermission(ObserverContext<?> ctx,
String request,
Permission.Action perm,
TableName tableName,
Map<byte[],? extends Collection<byte[]>> familyMap) |
void |
AccessChecker.requireGlobalPermission(User user,
String request,
Permission.Action perm,
TableName tableName,
Map<byte[],? extends Collection<byte[]>> familyMap,
String filterUser)
Checks that the user has the given global permission.
|
void |
NoopAccessChecker.requireGlobalPermission(User user,
String request,
Permission.Action perm,
TableName tableName,
Map<byte[],? extends Collection<byte[]>> familyMap,
String filterUser) |
void |
AccessController.requireNamespacePermission(ObserverContext<?> ctx,
String request,
String namespace,
TableName tableName,
Map<byte[],? extends Collection<byte[]>> familyMap,
Permission.Action... permissions) |
void |
AccessChecker.requireNamespacePermission(User user,
String request,
String namespace,
TableName tableName,
Map<byte[],? extends Collection<byte[]>> familyMap,
Permission.Action... permissions)
Checks that the user has the given global or namespace permission.
|
void |
NoopAccessChecker.requireNamespacePermission(User user,
String request,
String namespace,
TableName tableName,
Map<byte[],? extends Collection<byte[]>> familyMap,
Permission.Action... permissions) |
void |
AccessController.requirePermission(ObserverContext<?> ctx,
String request,
TableName tableName,
byte[] family,
byte[] qualifier,
Permission.Action... permissions) |
void |
AccessChecker.requirePermission(User user,
String request,
TableName tableName,
byte[] family,
byte[] qualifier,
String filterUser,
Permission.Action... permissions)
Authorizes that the current user has any of the given permissions for the given table, column
family and column qualifier.
|
void |
NoopAccessChecker.requirePermission(User user,
String request,
TableName tableName,
byte[] family,
byte[] qualifier,
String filterUser,
Permission.Action... permissions) |
void |
AccessController.requireTablePermission(ObserverContext<?> ctx,
String request,
TableName tableName,
byte[] family,
byte[] qualifier,
Permission.Action... permissions) |
void |
AccessChecker.requireTablePermission(User user,
String request,
TableName tableName,
byte[] family,
byte[] qualifier,
Permission.Action... permissions)
Authorizes that the current user has any of the given permissions for the given table, column
family and column qualifier.
|
void |
NoopAccessChecker.requireTablePermission(User user,
String request,
TableName tableName,
byte[] family,
byte[] qualifier,
Permission.Action... permissions) |
static void |
AccessControlClient.revoke(Connection connection,
TableName tableName,
String username,
byte[] family,
byte[] qualifier,
Permission.Action... actions)
Revokes the permission on the table
|
static void |
AccessControlUtil.revoke(org.apache.hbase.thirdparty.com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.shaded.protobuf.generated.AccessControlProtos.AccessControlService.BlockingInterface protocol,
String userShortName,
TableName tableName,
byte[] f,
byte[] q,
Permission.Action... actions)
Deprecated.
Use
Admin.revoke(UserPermission) instead. |
AuthResult.Params |
AuthResult.Params.setTableName(TableName table) |
private boolean |
SnapshotScannerHDFSAclCleaner.tableExists(TableName tableName) |
static org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.TableName |
ShadedAccessControlUtil.toProtoTableName(TableName tableName) |
private void |
AuthManager.updateTableCache(TableName table,
org.apache.hbase.thirdparty.com.google.common.collect.ListMultimap<String,Permission> tablePerms)
Updates the internal table permissions cache for specified table.
|
Modifier and Type | Method and Description |
---|---|
boolean |
SnapshotScannerHDFSAclHelper.grantAcl(UserPermission userPermission,
Set<String> skipNamespaces,
Set<TableName> skipTables)
Set acl when grant user permission
|
private void |
SnapshotScannerHDFSAclHelper.handleGlobalAcl(Set<String> users,
Set<String> skipNamespaces,
Set<TableName> skipTables,
SnapshotScannerHDFSAclHelper.HDFSAclOperation.OperationType operationType) |
private void |
SnapshotScannerHDFSAclHelper.handleGrantOrRevokeAcl(UserPermission userPermission,
SnapshotScannerHDFSAclHelper.HDFSAclOperation.OperationType operationType,
Set<String> skipNamespaces,
Set<TableName> skipTables) |
private void |
SnapshotScannerHDFSAclHelper.handleNamespaceAcl(Set<String> namespaces,
Set<String> users,
Set<String> skipNamespaces,
Set<TableName> skipTables,
SnapshotScannerHDFSAclHelper.HDFSAclOperation.OperationType operationType) |
private void |
SnapshotScannerHDFSAclHelper.handleTableAcl(Set<TableName> tableNames,
Set<String> users,
Set<String> skipNamespaces,
Set<TableName> skipTables,
SnapshotScannerHDFSAclHelper.HDFSAclOperation.OperationType operationType) |
private void |
SnapshotScannerHDFSAclHelper.handleTableAcl(Set<TableName> tableNames,
Set<String> users,
Set<String> skipNamespaces,
Set<TableName> skipTables,
SnapshotScannerHDFSAclHelper.HDFSAclOperation.OperationType operationType) |
void |
AccessController.postGetTableDescriptors(ObserverContext<MasterCoprocessorEnvironment> ctx,
List<TableName> tableNamesList,
List<TableDescriptor> descriptors,
String regex) |
void |
AccessController.preGetTableDescriptors(ObserverContext<MasterCoprocessorEnvironment> ctx,
List<TableName> tableNamesList,
List<TableDescriptor> descriptors,
String regex) |
void |
AccessController.preMoveServersAndTables(ObserverContext<MasterCoprocessorEnvironment> ctx,
Set<Address> servers,
Set<TableName> tables,
String targetGroup) |
void |
AccessController.preMoveTables(ObserverContext<MasterCoprocessorEnvironment> ctx,
Set<TableName> tables,
String targetGroup) |
boolean |
SnapshotScannerHDFSAclHelper.revokeAcl(UserPermission userPermission,
Set<String> skipNamespaces,
Set<TableName> skipTables)
Remove acl when grant or revoke user permission
|
Constructor and Description |
---|
AccessControlFilter(AuthManager mgr,
User ugi,
TableName tableName,
AccessControlFilter.Strategy strategy,
Map<ByteRange,Integer> cfVsMaxVersions) |
AuthResult(boolean allowed,
String request,
String reason,
User user,
Permission.Action action,
TableName table,
byte[] family,
byte[] qualifier) |
AuthResult(boolean allowed,
String request,
String reason,
User user,
Permission.Action action,
TableName table,
Map<byte[],? extends Collection<?>> families) |
Builder(TableName tableName) |
Builder(TableName tableName) |
GetUserPermissionsRequest(String userName,
String namespace,
TableName tableName,
byte[] family,
byte[] qualifier) |
TablePermission(TableName table,
byte[] family,
byte[] qualifier,
Permission.Action... assigned)
Construct a table:family:qualifier permission.
|
Modifier and Type | Field and Description |
---|---|
static TableName |
VisibilityConstants.LABELS_TABLE_NAME
Internal storage table for visibility labels
|
Modifier and Type | Method and Description |
---|---|
void |
VisibilityController.preDisableTable(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName) |
TableDescriptor |
VisibilityController.preModifyTable(ObserverContext<MasterCoprocessorEnvironment> ctx,
TableName tableName,
TableDescriptor currentDescriptor,
TableDescriptor newDescriptor) |
Modifier and Type | Field and Description |
---|---|
static TableName |
SlowLogTableAccessor.SLOW_LOG_TABLE_NAME
hbase:slowlog table name - can be enabled with config -
hbase.regionserver.slowlog.systable.enabled
|
Modifier and Type | Field and Description |
---|---|
private TableName |
SnapshotInfo.SnapshotStats.snapshotTable |
private TableName |
RestoreSnapshotHelper.snapshotTable |
private TableName |
CreateSnapshot.tableName |
Modifier and Type | Method and Description |
---|---|
static RegionInfo |
RestoreSnapshotHelper.cloneRegionInfo(TableName tableName,
RegionInfo snapshotRegionInfo) |
private static Pair<org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo,Long> |
ExportSnapshot.getSnapshotFileAndSize(org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.conf.Configuration conf,
TableName table,
String region,
String family,
String hfile,
long size) |
static void |
RestoreSnapshotHelper.restoreSnapshotAcl(org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotDescription snapshot,
TableName newTableName,
org.apache.hadoop.conf.Configuration conf) |
Constructor and Description |
---|
TablePartiallyOpenException(TableName tableName) |
Modifier and Type | Method and Description |
---|---|
private static TableName |
ThriftHBaseServiceHandler.getTableName(ByteBuffer buffer) |
Modifier and Type | Method and Description |
---|---|
static TableName |
ThriftUtilities.tableNameFromThrift(org.apache.hadoop.hbase.thrift2.generated.TTableName tableName) |
static TableName[] |
ThriftUtilities.tableNamesArrayFromThrift(List<org.apache.hadoop.hbase.thrift2.generated.TTableName> tableNames) |
Modifier and Type | Method and Description |
---|---|
static List<TableName> |
ThriftUtilities.tableNamesFromThrift(List<org.apache.hadoop.hbase.thrift2.generated.TTableName> tableNames) |
Modifier and Type | Method and Description |
---|---|
static org.apache.hadoop.hbase.thrift2.generated.TTableName |
ThriftUtilities.tableNameFromHBase(TableName table) |
static List<org.apache.hadoop.hbase.thrift2.generated.TTableName> |
ThriftUtilities.tableNamesFromHBase(TableName[] in) |
Modifier and Type | Method and Description |
---|---|
static List<org.apache.hadoop.hbase.thrift2.generated.TTableName> |
ThriftUtilities.tableNamesFromHBase(List<TableName> in) |
Modifier and Type | Field and Description |
---|---|
private TableName |
ThriftTable.tableName |
Modifier and Type | Method and Description |
---|---|
TableName |
ThriftTable.getName() |
TableName[] |
ThriftAdmin.listTableNames() |
TableName[] |
ThriftAdmin.listTableNames(Pattern pattern) |
TableName[] |
ThriftAdmin.listTableNames(Pattern pattern,
boolean includeSysTables) |
TableName[] |
ThriftAdmin.listTableNamesByNamespace(String name) |
Modifier and Type | Method and Description |
---|---|
Pair<List<String>,List<TableName>> |
ThriftAdmin.getConfiguredNamespacesAndTablesInRSGroup(String groupName) |
Map<TableName,SpaceQuotaSnapshot> |
ThriftAdmin.getRegionServerSpaceQuotaSnapshots(ServerName serverName) |
Map<TableName,Long> |
ThriftAdmin.getSpaceQuotaTableSizes() |
List<TableName> |
ThriftAdmin.listTableNamesByState(boolean isEnabled) |
List<TableName> |
ThriftAdmin.listTablesInRSGroup(String groupName) |
Modifier and Type | Method and Description |
---|---|
void |
ThriftAdmin.addColumnFamily(TableName tableName,
ColumnFamilyDescriptor columnFamily) |
Future<Void> |
ThriftAdmin.addColumnFamilyAsync(TableName tableName,
ColumnFamilyDescriptor columnFamily) |
CacheEvictionStats |
ThriftAdmin.clearBlockCache(TableName tableName) |
Future<Void> |
ThriftAdmin.cloneSnapshotAsync(String snapshotName,
TableName tableName,
boolean cloneAcl,
String customSFT) |
void |
ThriftAdmin.cloneTableSchema(TableName tableName,
TableName newTableName,
boolean preserveSplits) |
void |
ThriftAdmin.compact(TableName tableName) |
void |
ThriftAdmin.compact(TableName tableName,
byte[] columnFamily) |
void |
ThriftAdmin.compact(TableName tableName,
byte[] columnFamily,
CompactType compactType) |
void |
ThriftAdmin.compact(TableName tableName,
CompactType compactType) |
void |
ThriftAdmin.deleteColumnFamily(TableName tableName,
byte[] columnFamily) |
Future<Void> |
ThriftAdmin.deleteColumnFamilyAsync(TableName tableName,
byte[] columnFamily) |
void |
ThriftAdmin.deleteTable(TableName tableName) |
Future<Void> |
ThriftAdmin.deleteTableAsync(TableName tableName) |
void |
ThriftAdmin.disableTable(TableName tableName) |
Future<Void> |
ThriftAdmin.disableTableAsync(TableName tableName) |
void |
ThriftAdmin.disableTableReplication(TableName tableName) |
void |
ThriftAdmin.enableTable(TableName tableName) |
Future<Void> |
ThriftAdmin.enableTableAsync(TableName tableName) |
void |
ThriftAdmin.enableTableReplication(TableName tableName) |
void |
ThriftAdmin.flush(TableName tableName) |
void |
ThriftAdmin.flush(TableName tableName,
byte[] columnFamily) |
void |
ThriftAdmin.flush(TableName tableName,
List<byte[]> columnFamilies) |
BufferedMutator |
ThriftConnection.getBufferedMutator(TableName tableName) |
CompactionState |
ThriftAdmin.getCompactionState(TableName tableName) |
CompactionState |
ThriftAdmin.getCompactionState(TableName tableName,
CompactType compactType) |
SpaceQuotaSnapshot |
ThriftAdmin.getCurrentSpaceQuotaSnapshot(TableName tableName) |
TableDescriptor |
ThriftAdmin.getDescriptor(TableName tableName) |
long |
ThriftAdmin.getLastMajorCompactionTimestamp(TableName tableName) |
RegionLocator |
ThriftConnection.getRegionLocator(TableName tableName) |
List<RegionMetrics> |
ThriftAdmin.getRegionMetrics(ServerName serverName,
TableName tableName) |
List<RegionInfo> |
ThriftAdmin.getRegions(TableName tableName) |
RSGroupInfo |
ThriftAdmin.getRSGroup(TableName tableName) |
TableBuilder |
ThriftConnection.getTableBuilder(TableName tableName,
ExecutorService pool)
Get a TableBuider to build ThriftTable, ThriftTable is NOT thread safe
|
boolean |
ThriftAdmin.isTableAvailable(TableName tableName) |
boolean |
ThriftAdmin.isTableDisabled(TableName tableName) |
boolean |
ThriftAdmin.isTableEnabled(TableName tableName) |
void |
ThriftAdmin.majorCompact(TableName tableName) |
void |
ThriftAdmin.majorCompact(TableName tableName,
byte[] columnFamily) |
void |
ThriftAdmin.majorCompact(TableName tableName,
byte[] columnFamily,
CompactType compactType) |
void |
ThriftAdmin.majorCompact(TableName tableName,
CompactType compactType) |
void |
ThriftAdmin.modifyColumnFamily(TableName tableName,
ColumnFamilyDescriptor columnFamily) |
Future<Void> |
ThriftAdmin.modifyColumnFamilyAsync(TableName tableName,
ColumnFamilyDescriptor columnFamily) |
Future<Void> |
ThriftAdmin.modifyColumnFamilyStoreFileTrackerAsync(TableName tableName,
byte[] family,
String dstSFT) |
Future<Void> |
ThriftAdmin.modifyTableStoreFileTrackerAsync(TableName tableName,
String dstSFT) |
void |
ThriftAdmin.snapshot(String snapshotName,
TableName tableName) |
void |
ThriftAdmin.snapshot(String snapshotName,
TableName tableName,
SnapshotType type) |
void |
ThriftAdmin.split(TableName tableName) |
void |
ThriftAdmin.split(TableName tableName,
byte[] splitPoint) |
boolean |
ThriftAdmin.tableExists(TableName tableName) |
void |
ThriftAdmin.truncateTable(TableName tableName,
boolean preserveSplits) |
Future<Void> |
ThriftAdmin.truncateTableAsync(TableName tableName,
boolean preserveSplits) |
Modifier and Type | Method and Description |
---|---|
List<TableDescriptor> |
ThriftAdmin.listTableDescriptors(List<TableName> tableNames) |
void |
ThriftAdmin.setRSGroup(Set<TableName> tables,
String groupName) |
Constructor and Description |
---|
ThriftTable(TableName tableName,
org.apache.hadoop.hbase.thrift2.generated.THBaseService.Client client,
org.apache.thrift.transport.TTransport tTransport,
org.apache.hadoop.conf.Configuration conf) |
Modifier and Type | Field and Description |
---|---|
static TableName |
CanaryTool.DEFAULT_WRITE_TABLE_NAME |
private TableName |
CanaryTool.RegionTaskResult.tableName |
private TableName |
CanaryTool.RegionMonitor.writeTableName |
Modifier and Type | Method and Description |
---|---|
TableName |
CanaryTool.RegionTaskResult.getTableName() |
Modifier and Type | Method and Description |
---|---|
Map<BulkLoadHFiles.LoadQueueItem,ByteBuffer> |
BulkLoadHFilesTool.bulkLoad(TableName tableName,
Map<byte[],List<org.apache.hadoop.fs.Path>> family2Files) |
Map<BulkLoadHFiles.LoadQueueItem,ByteBuffer> |
BulkLoadHFiles.bulkLoad(TableName tableName,
Map<byte[],List<org.apache.hadoop.fs.Path>> family2Files)
Perform a bulk load of the given directory into the given pre-existing table.
|
Map<BulkLoadHFiles.LoadQueueItem,ByteBuffer> |
BulkLoadHFilesTool.bulkLoad(TableName tableName,
org.apache.hadoop.fs.Path dir) |
Map<BulkLoadHFiles.LoadQueueItem,ByteBuffer> |
BulkLoadHFiles.bulkLoad(TableName tableName,
org.apache.hadoop.fs.Path dir)
Perform a bulk load of the given directory into the given pre-existing table.
|
protected void |
BulkLoadHFilesTool.bulkLoadPhase(AsyncClusterConnection conn,
TableName tableName,
Deque<BulkLoadHFiles.LoadQueueItem> queue,
org.apache.hbase.thirdparty.com.google.common.collect.Multimap<ByteBuffer,BulkLoadHFiles.LoadQueueItem> regionGroups,
boolean copyFiles,
Map<BulkLoadHFiles.LoadQueueItem,ByteBuffer> item2RegionMap)
This takes the LQI's grouped by likely regions and attempts to bulk load them.
|
private void |
BulkLoadHFilesTool.checkRegionIndexValid(int idx,
List<Pair<byte[],byte[]>> startEndKeys,
TableName tableName)
we can consider there is a region hole or overlap in following conditions.
|
private void |
BulkLoadHFilesTool.cleanup(AsyncClusterConnection conn,
TableName tableName,
Deque<BulkLoadHFiles.LoadQueueItem> queue,
ExecutorService pool) |
private void |
BulkLoadHFilesTool.createTable(TableName tableName,
org.apache.hadoop.fs.Path hfofDir,
AsyncAdmin admin)
If the table is created for the first time, then "completebulkload" reads the files twice.
|
private Map<BulkLoadHFiles.LoadQueueItem,ByteBuffer> |
BulkLoadHFilesTool.doBulkLoad(AsyncClusterConnection conn,
TableName tableName,
Map<byte[],List<org.apache.hadoop.fs.Path>> map,
boolean silence,
boolean copyFile)
Perform a bulk load of the given map of families to hfiles into the given pre-existing table.
|
private Map<BulkLoadHFiles.LoadQueueItem,ByteBuffer> |
BulkLoadHFilesTool.doBulkLoad(AsyncClusterConnection conn,
TableName tableName,
org.apache.hadoop.fs.Path hfofDir,
boolean silence,
boolean copyFile)
Perform a bulk load of the given directory into the given pre-existing table.
|
protected Pair<List<BulkLoadHFiles.LoadQueueItem>,String> |
BulkLoadHFilesTool.groupOrSplit(AsyncClusterConnection conn,
TableName tableName,
org.apache.hbase.thirdparty.com.google.common.collect.Multimap<ByteBuffer,BulkLoadHFiles.LoadQueueItem> regionGroups,
BulkLoadHFiles.LoadQueueItem item,
List<Pair<byte[],byte[]>> startEndKeys)
Attempt to assign the given load queue item into its target region group.
|
private Pair<org.apache.hbase.thirdparty.com.google.common.collect.Multimap<ByteBuffer,BulkLoadHFiles.LoadQueueItem>,Set<String>> |
BulkLoadHFilesTool.groupOrSplitPhase(AsyncClusterConnection conn,
TableName tableName,
ExecutorService pool,
Deque<BulkLoadHFiles.LoadQueueItem> queue,
List<Pair<byte[],byte[]>> startEndKeys) |
void |
BulkLoadHFilesTool.loadHFileQueue(AsyncClusterConnection conn,
TableName tableName,
Deque<BulkLoadHFiles.LoadQueueItem> queue,
boolean copyFiles)
Used by the replication sink to load the hfiles from the source cluster.
|
private Map<BulkLoadHFiles.LoadQueueItem,ByteBuffer> |
BulkLoadHFilesTool.performBulkLoad(AsyncClusterConnection conn,
TableName tableName,
Deque<BulkLoadHFiles.LoadQueueItem> queue,
ExecutorService pool,
boolean copyFile) |
static void |
BulkLoadHFilesTool.prepareHFileQueue(AsyncClusterConnection conn,
TableName tableName,
Map<byte[],List<org.apache.hadoop.fs.Path>> map,
Deque<BulkLoadHFiles.LoadQueueItem> queue,
boolean silence)
Prepare a collection of
LoadQueueItem from list of source hfiles contained in the
passed directory and validates whether the prepared queue has all the valid table column
families in it. |
static void |
BulkLoadHFilesTool.prepareHFileQueue(org.apache.hadoop.conf.Configuration conf,
AsyncClusterConnection conn,
TableName tableName,
org.apache.hadoop.fs.Path hfilesDir,
Deque<BulkLoadHFiles.LoadQueueItem> queue,
boolean validateHFile,
boolean silence)
Prepare a collection of
LoadQueueItem from list of source hfiles contained in the
passed directory and validates whether the prepared queue has all the valid table column
families in it. |
private void |
BulkLoadHFilesTool.tableExists(AsyncClusterConnection conn,
TableName tableName) |
private void |
BulkLoadHFilesTool.throwAndLogTableNotFoundException(TableName tn) |
protected CompletableFuture<Collection<BulkLoadHFiles.LoadQueueItem>> |
BulkLoadHFilesTool.tryAtomicRegionLoad(AsyncClusterConnection conn,
TableName tableName,
boolean copyFiles,
byte[] first,
Collection<BulkLoadHFiles.LoadQueueItem> lqis)
Attempts to do an atomic load of many hfiles into a region.
|
Constructor and Description |
---|
RegionMonitor(Connection connection,
String[] monitorTargets,
boolean useRegExp,
CanaryTool.Sink sink,
ExecutorService executor,
boolean writeSniffing,
TableName writeTableName,
boolean treatFailureAsError,
HashMap<String,Long> configuredReadTableTimeouts,
long configuredWriteTableTimeout,
long allowedFailures) |
RegionTaskResult(RegionInfo region,
TableName tableName,
ServerName serverName,
ColumnFamilyDescriptor column) |
Modifier and Type | Field and Description |
---|---|
private TableName |
HBaseFsck.cleanReplicationBarrierTable
Deprecated.
|
(package private) TableName |
HbckTableInfo.tableName |
Modifier and Type | Field and Description |
---|---|
private Map<TableName,TableDescriptor> |
FSTableDescriptors.cache |
private Map<TableName,Set<String>> |
HBaseFsck.orphanTableDirs
Deprecated.
|
private Map<TableName,Set<String>> |
HBaseFsck.skippedRegions
Deprecated.
|
private Set<TableName> |
HBaseFsck.tablesIncluded
Deprecated.
|
private SortedMap<TableName,HbckTableInfo> |
HBaseFsck.tablesInfo
Deprecated.
This map from Tablename -> TableInfo contains the structures necessary to detect table
consistency problems (holes, dupes, overlaps).
|
private Map<TableName,TableState> |
HBaseFsck.tableStates
Deprecated.
|
Modifier and Type | Method and Description |
---|---|
TableName |
HbckTableInfo.getName() |
TableName |
HbckRegionInfo.getTableName() |
static TableName |
HFileArchiveUtil.getTableName(org.apache.hadoop.fs.Path archivePath) |
static TableName |
CommonFSUtils.getTableName(org.apache.hadoop.fs.Path tablePath)
Returns the
TableName object representing the table directory
under path rootdir |
Modifier and Type | Method and Description |
---|---|
private SortedMap<TableName,HbckTableInfo> |
HBaseFsck.checkHdfsIntegrity(boolean fixHoles,
boolean fixOverlaps)
Deprecated.
|
(package private) SortedMap<TableName,HbckTableInfo> |
HBaseFsck.checkIntegrity()
Deprecated.
Checks tables integrity.
|
(package private) Set<TableName> |
HBaseFsck.getIncludedTables()
Deprecated.
|
private SortedMap<TableName,HbckTableInfo> |
HBaseFsck.loadHdfsRegionInfos()
Deprecated.
Populate hbi's from regionInfos loaded from file system.
|
static Map<TableName,TableState.State> |
ZKDataMigrator.queryForTableStates(ZKWatcher zkw)
Deprecated.
Since 2.0.0. To be removed in hbase-3.0.0.
|
Modifier and Type | Method and Description |
---|---|
(package private) static void |
RegionSplitter.createPresplitTable(TableName tableName,
RegionSplitter.SplitAlgorithm splitAlgo,
String[] columnFamilies,
org.apache.hadoop.conf.Configuration conf) |
private boolean |
HBaseFsck.fabricateTableInfo(FSTableDescriptors fstd,
TableName tableName,
Set<String> columns)
Deprecated.
To fabricate a .tableinfo file with following contents
1. |
TableDescriptor |
FSTableDescriptors.get(TableName tableName)
Get the current table descriptor for the given table, or null if none exists.
|
static org.apache.hadoop.fs.Path |
HFileArchiveUtil.getRegionArchiveDir(org.apache.hadoop.fs.Path rootDir,
TableName tableName,
org.apache.hadoop.fs.Path regiondir)
Get the archive directory for a given region under the specified table
|
static org.apache.hadoop.fs.Path |
HFileArchiveUtil.getRegionArchiveDir(org.apache.hadoop.fs.Path rootDir,
TableName tableName,
String encodedRegionName)
Get the archive directory for a given region under the specified table
|
static org.apache.hadoop.fs.Path |
CommonFSUtils.getRegionDir(org.apache.hadoop.fs.Path rootdir,
TableName tableName,
String regionName)
Returns the
Path object representing the region directory under
path rootdir |
(package private) static LinkedList<Pair<byte[],byte[]>> |
RegionSplitter.getSplits(Connection connection,
TableName tableName,
RegionSplitter.SplitAlgorithm splitAlgo) |
static org.apache.hadoop.fs.Path |
HFileArchiveUtil.getStoreArchivePath(org.apache.hadoop.conf.Configuration conf,
TableName tableName,
String regionName,
String familyName)
Get the directory to archive a store directory
|
static org.apache.hadoop.fs.Path |
HFileArchiveUtil.getTableArchivePath(org.apache.hadoop.conf.Configuration conf,
TableName tableName)
Get the path to the table archive directory based on the configured archive directory.
|
static org.apache.hadoop.fs.Path |
HFileArchiveUtil.getTableArchivePath(org.apache.hadoop.fs.Path rootdir,
TableName tableName)
Get the path to the table archive directory based on the configured archive directory.
|
static TableDescriptor |
FSTableDescriptors.getTableDescriptorFromFs(org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path hbaseRootDir,
TableName tableName)
Returns the latest table descriptor for the given table directly from the file system if it
exists, bypassing the local cache.
|
static org.apache.hadoop.fs.Path |
CommonFSUtils.getTableDir(org.apache.hadoop.fs.Path rootdir,
TableName tableName)
Returns the
Path object representing the table directory under
path rootdir |
private org.apache.hadoop.fs.Path |
FSTableDescriptors.getTableDir(TableName tableName)
Return the table directory in HDFS
|
private static Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path> |
RegionSplitter.getTableDirAndSplitFile(org.apache.hadoop.conf.Configuration conf,
TableName tableName) |
private static org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.DeprecatedTableState.State |
ZKDataMigrator.getTableState(ZKWatcher zkw,
TableName tableName)
Deprecated.
Since 2.0.0. To be removed in hbase-3.0.0.
|
static Map<String,org.apache.hadoop.fs.Path> |
FSUtils.getTableStoreFilePathMap(Map<String,org.apache.hadoop.fs.Path> map,
org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path hbaseRootDir,
TableName tableName)
Runs through the HBase rootdir/tablename and creates a reverse lookup map for table StoreFile
names to the full Path.
|
static Map<String,org.apache.hadoop.fs.Path> |
FSUtils.getTableStoreFilePathMap(Map<String,org.apache.hadoop.fs.Path> resultMap,
org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path hbaseRootDir,
TableName tableName,
org.apache.hadoop.fs.PathFilter sfFilter,
ExecutorService executor,
FSUtils.ProgressReporter progressReporter)
Runs through the HBase rootdir/tablename and creates a reverse lookup map for table StoreFile
names to the full Path.
|
static Map<String,org.apache.hadoop.fs.Path> |
FSUtils.getTableStoreFilePathMap(Map<String,org.apache.hadoop.fs.Path> resultMap,
org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path hbaseRootDir,
TableName tableName,
org.apache.hadoop.fs.PathFilter sfFilter,
ExecutorService executor,
HbckErrorReporter progressReporter)
Deprecated.
Since 2.3.0. For removal in hbase4. Use ProgressReporter override instead.
|
static org.apache.hadoop.fs.Path |
CommonFSUtils.getWALRegionDir(org.apache.hadoop.conf.Configuration conf,
TableName tableName,
String encodedRegionName)
Returns the WAL region directory based on the given table name and region name
|
static org.apache.hadoop.fs.Path |
CommonFSUtils.getWALTableDir(org.apache.hadoop.conf.Configuration conf,
TableName tableName)
Returns the Table directory under the WALRootDir for the specified table name
|
static org.apache.hadoop.fs.Path |
CommonFSUtils.getWrongWALRegionDir(org.apache.hadoop.conf.Configuration conf,
TableName tableName,
String encodedRegionName)
Deprecated.
For compatibility, will be removed in 4.0.0.
|
void |
HBaseFsck.includeTable(TableName table)
Deprecated.
|
static boolean |
ServerRegionReplicaUtil.isMetaRegionReplicaReplicationEnabled(org.apache.hadoop.conf.Configuration conf,
TableName tn)
Returns True if hbase:meta Region Read Replica is enabled.
|
static boolean |
ServerRegionReplicaUtil.isRegionReplicaReplicationEnabled(org.apache.hadoop.conf.Configuration conf,
TableName tn) |
(package private) boolean |
HBaseFsck.isTableDisabled(TableName tableName)
Deprecated.
Check if the specified region's table is disabled.
|
(package private) boolean |
HBaseFsck.isTableIncluded(TableName table)
Deprecated.
Only check/fix tables specified by the list, Empty list means all tables are included.
|
TableDescriptor |
FSTableDescriptors.remove(TableName tablename)
Removes the table descriptor from the local cache and returns it.
|
(package private) static void |
RegionSplitter.rollingSplit(TableName tableName,
RegionSplitter.SplitAlgorithm splitAlgo,
org.apache.hadoop.conf.Configuration conf) |
(package private) static LinkedList<Pair<byte[],byte[]>> |
RegionSplitter.splitScan(LinkedList<Pair<byte[],byte[]>> regionList,
Connection connection,
TableName tableName,
RegionSplitter.SplitAlgorithm splitAlgo) |
Modifier and Type | Method and Description |
---|---|
(package private) TableDescriptor[] |
HBaseFsck.getTableDescriptors(List<TableName> tableNames)
Deprecated.
|
private void |
HBaseFsck.printTableSummary(SortedMap<TableName,HbckTableInfo> tablesInfo)
Deprecated.
Prints summary of all tables found on the system.
|
Constructor and Description |
---|
HbckTableInfo(TableName name,
HBaseFsck hbck) |
Modifier and Type | Field and Description |
---|---|
protected TableName |
MajorCompactor.tableName |
Constructor and Description |
---|
MajorCompactor(org.apache.hadoop.conf.Configuration conf,
TableName tableName,
Set<String> storesToCompact,
int concurrency,
long timestamp,
long sleepForMs) |
Modifier and Type | Field and Description |
---|---|
private TableName |
WALKeyImpl.tablename |
(package private) TableName |
EntryBuffers.RegionEntryBuffer.tableName |
Modifier and Type | Method and Description |
---|---|
TableName |
WALKey.getTableName()
Returns table name
|
TableName |
WALKeyImpl.getTableName()
Returns table name
|
TableName |
EntryBuffers.RegionEntryBuffer.getTableName() |
Modifier and Type | Method and Description |
---|---|
protected AbstractRecoveredEditsOutputSink.RecoveredEditsWriter |
AbstractRecoveredEditsOutputSink.createRecoveredEditsWriter(TableName tableName,
byte[] region,
long seqId)
Returns a writer that wraps a
WALProvider.Writer and its Path. |
private StoreFileWriter |
BoundedRecoveredHFilesOutputSink.createRecoveredHFileWriter(TableName tableName,
String regionName,
long seqId,
String familyName,
boolean isMetaTable) |
private AbstractRecoveredEditsOutputSink.RecoveredEditsWriter |
RecoveredEditsOutputSink.getRecoveredEditsWriter(TableName tableName,
byte[] region,
long seqId)
Get a writer and path for a log starting at the given entry.
|
(package private) static org.apache.hadoop.fs.Path |
WALSplitUtil.getRegionSplitEditsPath(TableName tableName,
byte[] encodedRegionName,
long seqId,
String fileNameBeingSplit,
String tmpDirName,
org.apache.hadoop.conf.Configuration conf)
Path to a file under RECOVERED_EDITS_DIR directory of the region found in
logEntry
named for the sequenceid in the passed logEntry : e.g. |
protected void |
WALKeyImpl.init(byte[] encodedRegionName,
TableName tablename,
long logSeqNum,
long now,
List<UUID> clusterIds,
long nonceGroup,
long nonce,
MultiVersionConcurrencyControl mvcc,
NavigableMap<byte[],Integer> replicationScope,
Map<String,byte[]> extendedAttributes) |
(package private) void |
WALKeyImpl.internTableName(TableName tablename)
Drop this instance's tablename byte array and instead hold a reference to the provided
tablename.
|
private boolean |
WALSplitter.isRegionDirPresentUnderRoot(TableName tn,
String region) |
(package private) static org.apache.hadoop.fs.Path |
WALSplitUtil.tryCreateRecoveredHFilesDir(org.apache.hadoop.fs.FileSystem rootFS,
org.apache.hadoop.conf.Configuration conf,
TableName tableName,
String encodedRegionName,
String familyName)
Return path to recovered.hfiles directory of the region's column family: e.g.
|
Constructor and Description |
---|
RegionEntryBuffer(TableName tableName,
byte[] region) |
WALKeyImpl(byte[] encodedRegionName,
TableName tablename,
long now) |
WALKeyImpl(byte[] encodedRegionName,
TableName tablename,
long now,
List<UUID> clusterIds,
long nonceGroup,
long nonce,
MultiVersionConcurrencyControl mvcc)
Create the log key for writing to somewhere.
|
WALKeyImpl(byte[] encodedRegionName,
TableName tablename,
long now,
List<UUID> clusterIds,
long nonceGroup,
long nonce,
MultiVersionConcurrencyControl mvcc,
NavigableMap<byte[],Integer> replicationScope)
Create the log key for writing to somewhere.
|
WALKeyImpl(byte[] encodedRegionName,
TableName tablename,
long now,
List<UUID> clusterIds,
long nonceGroup,
long nonce,
MultiVersionConcurrencyControl mvcc,
NavigableMap<byte[],Integer> replicationScope,
Map<String,byte[]> extendedAttributes) |
WALKeyImpl(byte[] encodedRegionName,
TableName tablename,
long logSeqNum,
long now,
List<UUID> clusterIds,
long nonceGroup,
long nonce,
MultiVersionConcurrencyControl mvcc)
Create the log key for writing to somewhere.
|
WALKeyImpl(byte[] encodedRegionName,
TableName tablename,
long logSeqNum,
long now,
List<UUID> clusterIds,
long nonceGroup,
long nonce,
MultiVersionConcurrencyControl mvcc,
NavigableMap<byte[],Integer> replicationScope)
Create the log key for writing to somewhere.
|
WALKeyImpl(byte[] encodedRegionName,
TableName tablename,
long logSeqNum,
long nonceGroup,
long nonce,
MultiVersionConcurrencyControl mvcc)
Create the log key for writing to somewhere.
|
WALKeyImpl(byte[] encodedRegionName,
TableName tablename,
long logSeqNum,
long now,
UUID clusterId) |
WALKeyImpl(byte[] encodedRegionName,
TableName tablename,
long logSeqNum,
long now,
UUID clusterId,
MultiVersionConcurrencyControl mvcc) |
WALKeyImpl(byte[] encodedRegionName,
TableName tablename,
long now,
MultiVersionConcurrencyControl mvcc) |
WALKeyImpl(byte[] encodedRegionName,
TableName tablename,
long now,
MultiVersionConcurrencyControl mvcc,
NavigableMap<byte[],Integer> replicationScope) |
WALKeyImpl(byte[] encodedRegionName,
TableName tablename,
long now,
MultiVersionConcurrencyControl mvcc,
NavigableMap<byte[],Integer> replicationScope,
Map<String,byte[]> extendedAttributes) |
WALKeyImpl(byte[] encodedRegionName,
TableName tablename,
long now,
NavigableMap<byte[],Integer> replicationScope) |
Modifier and Type | Field and Description |
---|---|
(package private) static TableName |
HelloHBase.MY_TABLE_NAME |
Modifier and Type | Field and Description |
---|---|
(package private) static TableName |
HelloHBase.MY_TABLE_NAME |
Copyright © 2007–2020 The Apache Software Foundation. All rights reserved.