Modifier and Type | Method and Description |
---|---|
static Pair<RegionInfo,ServerName> |
MetaTableAccessor.getRegion(Connection connection,
byte[] regionName)
Deprecated.
|
static Pair<Integer,Integer> |
TagUtil.readVIntValuePart(Tag tag,
int offset)
Reads an int value stored as a VInt at tag's given offset.
|
Modifier and Type | Method and Description |
---|---|
private static CompletableFuture<List<Pair<RegionInfo,ServerName>>> |
AsyncMetaTableAccessor.getTableRegionsAndLocations(AsyncTable<AdvancedScanResultConsumer> metaTable,
Optional<TableName> tableName,
boolean excludeOfflinedSplitParents)
Used to get table regions' info and server.
|
static List<Pair<RegionInfo,ServerName>> |
MetaTableAccessor.getTableRegionsAndLocations(Connection connection,
TableName tableName)
Do not use this method to get meta table regions, use methods in MetaTableLocator instead.
|
static List<Pair<RegionInfo,ServerName>> |
MetaTableAccessor.getTableRegionsAndLocations(Connection connection,
TableName tableName,
boolean excludeOfflinedSplitParents)
Do not use this method to get meta table regions, use methods in MetaTableLocator instead.
|
Modifier and Type | Method and Description |
---|---|
private static List<RegionInfo> |
MetaTableAccessor.getListOfRegionInfos(List<Pair<RegionInfo,ServerName>> pairs) |
Modifier and Type | Method and Description |
---|---|
Pair<Result[],ScannerCallable> |
ScannerCallableWithReplicas.RetryingRPC.call(int callTimeout) |
Pair<Integer,Integer> |
Admin.getAlterStatus(byte[] tableName)
Deprecated.
Since 2.0.0. Will be removed in 3.0.0. No longer needed now you get a Future
on an operation.
|
Pair<Integer,Integer> |
HBaseAdmin.getAlterStatus(byte[] tableName) |
Pair<Integer,Integer> |
Admin.getAlterStatus(TableName tableName)
Deprecated.
Since 2.0.0. Will be removed in 3.0.0. No longer needed now you get a Future
on an operation.
|
Pair<Integer,Integer> |
HBaseAdmin.getAlterStatus(TableName tableName) |
private Pair<List<byte[]>,List<HRegionLocation>> |
HTable.getKeysAndRegionsInRange(byte[] startKey,
byte[] endKey,
boolean includeEndKey)
Get the corresponding start keys and regions for an arbitrary range of
keys.
|
private Pair<List<byte[]>,List<HRegionLocation>> |
HTable.getKeysAndRegionsInRange(byte[] startKey,
byte[] endKey,
boolean includeEndKey,
boolean reload)
Get the corresponding start keys and regions for an arbitrary range of
keys.
|
(package private) Pair<RegionInfo,ServerName> |
HBaseAdmin.getRegion(byte[] regionName) |
Pair<byte[][],byte[][]> |
RegionLocator.getStartEndKeys()
Gets the starting and ending row keys for every region in the currently
open table.
|
Pair<byte[][],byte[][]> |
HRegionLocator.getStartEndKeys()
Gets the starting and ending row keys for every region in the currently
open table.
|
(package private) Pair<byte[][],byte[][]> |
HRegionLocator.getStartEndKeys(List<RegionLocations> regions) |
private Pair<RegionState.State,ServerName> |
ZKAsyncRegistry.getStateAndServerName(org.apache.hadoop.hbase.shaded.protobuf.generated.ZooKeeperProtos.MetaRegionServer proto) |
Modifier and Type | Method and Description |
---|---|
private void |
ScannerCallableWithReplicas.addCallsForCurrentReplica(ResultBoundedCompletionService<Pair<Result[],ScannerCallable>> cs) |
private void |
ScannerCallableWithReplicas.addCallsForOtherReplicas(ResultBoundedCompletionService<Pair<Result[],ScannerCallable>> cs,
int min,
int max) |
boolean |
SecureBulkLoadClient.secureBulkLoadHFiles(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ClientService.BlockingInterface client,
List<Pair<byte[],String>> familyPaths,
byte[] regionName,
boolean assignSeqNum,
org.apache.hadoop.security.token.Token<?> userToken,
String bulkToken)
Securely bulk load a list of HFiles using client protocol.
|
boolean |
SecureBulkLoadClient.secureBulkLoadHFiles(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ClientService.BlockingInterface client,
List<Pair<byte[],String>> familyPaths,
byte[] regionName,
boolean assignSeqNum,
org.apache.hadoop.security.token.Token<?> userToken,
String bulkToken,
boolean copyFiles)
Securely bulk load a list of HFiles using client protocol.
|
Modifier and Type | Method and Description |
---|---|
private <R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message> |
AggregationClient.getAvgArgs(Table table,
ColumnInterpreter<R,S,P,Q,T> ci,
Scan scan)
It computes average while fetching sum and row count from all the
corresponding regions.
|
private <R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message> |
AggregationClient.getAvgArgs(TableName tableName,
ColumnInterpreter<R,S,P,Q,T> ci,
Scan scan)
It computes average while fetching sum and row count from all the
corresponding regions.
|
private <R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message> |
AggregationClient.getMedianArgs(Table table,
ColumnInterpreter<R,S,P,Q,T> ci,
Scan scan)
It helps locate the region with median for a given column whose weight
is specified in an optional column.
|
private <R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message> |
AggregationClient.getStdArgs(Table table,
ColumnInterpreter<R,S,P,Q,T> ci,
Scan scan)
It computes a global standard deviation for a given column and its value.
|
Modifier and Type | Method and Description |
---|---|
private static Pair<String,String> |
Constraints.getKeyValueForClass(HTableDescriptor desc,
Class<? extends Constraint> clazz)
Get the kv
Map.Entry in the descriptor for the specified class |
Modifier and Type | Method and Description |
---|---|
static void |
Constraints.add(HTableDescriptor desc,
Pair<Class<? extends Constraint>,org.apache.hadoop.conf.Configuration>... constraints)
Add constraints and their associated configurations to the table.
|
Modifier and Type | Method and Description |
---|---|
default void |
RegionObserver.postBulkLoadHFile(ObserverContext<RegionCoprocessorEnvironment> ctx,
List<Pair<byte[],String>> stagingFamilyPaths,
Map<byte[],List<org.apache.hadoop.fs.Path>> finalPaths)
Called after bulkLoadHFile.
|
default void |
RegionObserver.preBulkLoadHFile(ObserverContext<RegionCoprocessorEnvironment> ctx,
List<Pair<byte[],String>> familyPaths)
Called before bulkLoadHFile.
|
default void |
RegionObserver.preCommitStoreFile(ObserverContext<RegionCoprocessorEnvironment> ctx,
byte[] family,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs)
Called before moving bulk loaded hfile to region directory.
|
Modifier and Type | Method and Description |
---|---|
private Pair<Map<ServerName,List<RegionInfo>>,List<RegionInfo>> |
FavoredNodeLoadBalancer.segregateRegionsAndAssignRegionsWithFavoredNodes(List<RegionInfo> regions,
List<ServerName> availableServers) |
Modifier and Type | Field and Description |
---|---|
private List<Pair<byte[],byte[]>> |
FuzzyRowFilter.fuzzyKeysData |
private PriorityQueue<Pair<byte[],Pair<byte[],byte[]>>> |
FuzzyRowFilter.RowTracker.nextRows |
private PriorityQueue<Pair<byte[],Pair<byte[],byte[]>>> |
FuzzyRowFilter.RowTracker.nextRows |
Modifier and Type | Method and Description |
---|---|
private void |
FuzzyRowFilter.preprocessSearchKey(Pair<byte[],byte[]> p) |
(package private) void |
FuzzyRowFilter.RowTracker.updateWith(Cell currentCell,
Pair<byte[],byte[]> fuzzyData) |
Constructor and Description |
---|
FuzzyRowFilter(List<Pair<byte[],byte[]>> fuzzyKeysData) |
Modifier and Type | Method and Description |
---|---|
(package private) static Pair<TableName,String> |
HFileLink.parseBackReferenceName(String name) |
Modifier and Type | Field and Description |
---|---|
(package private) static Map<Pair<String,String>,KeyProvider> |
Encryption.keyProviderCache |
Modifier and Type | Method and Description |
---|---|
static Pair<Long,MemoryType> |
MemorySizeUtil.getGlobalMemStoreSize(org.apache.hadoop.conf.Configuration conf) |
static Pair<Integer,Integer> |
StreamUtils.readRawVarint32(byte[] input,
int offset)
Reads a varInt value stored in an array.
|
static Pair<Integer,Integer> |
StreamUtils.readRawVarint32(ByteBuffer input,
int offset) |
Modifier and Type | Field and Description |
---|---|
static Pair<org.apache.hbase.thirdparty.io.netty.channel.EventLoopGroup,Class<? extends org.apache.hbase.thirdparty.io.netty.channel.Channel>> |
DefaultNettyEventLoopConfig.GROUP_AND_CHANNEL_CLASS |
Modifier and Type | Field and Description |
---|---|
private static Map<String,Pair<org.apache.hbase.thirdparty.io.netty.channel.EventLoopGroup,Class<? extends org.apache.hbase.thirdparty.io.netty.channel.Channel>>> |
NettyRpcClientConfigHelper.EVENT_LOOP_CONFIG_MAP |
Modifier and Type | Method and Description |
---|---|
(package private) static Pair<ByteBuff,RpcServer.CallCleanup> |
RpcServer.allocateByteBuffToReadInto(ByteBufferPool pool,
int minSizeForPoolUse,
int reqLen)
This is extracted to a static method for better unit testing.
|
Pair<org.apache.hbase.thirdparty.com.google.protobuf.Message,CellScanner> |
SimpleRpcServer.call(org.apache.hbase.thirdparty.com.google.protobuf.BlockingService service,
org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor md,
org.apache.hbase.thirdparty.com.google.protobuf.Message param,
CellScanner cellScanner,
long receiveTime,
MonitoredRPCHandler status) |
Pair<org.apache.hbase.thirdparty.com.google.protobuf.Message,CellScanner> |
NettyRpcServer.call(org.apache.hbase.thirdparty.com.google.protobuf.BlockingService service,
org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor md,
org.apache.hbase.thirdparty.com.google.protobuf.Message param,
CellScanner cellScanner,
long receiveTime,
MonitoredRPCHandler status) |
Pair<org.apache.hbase.thirdparty.com.google.protobuf.Message,CellScanner> |
RpcServerInterface.call(org.apache.hbase.thirdparty.com.google.protobuf.BlockingService service,
org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor md,
org.apache.hbase.thirdparty.com.google.protobuf.Message param,
CellScanner cellScanner,
long receiveTime,
MonitoredRPCHandler status)
Deprecated.
As of release 1.3, this will be removed in HBase 3.0
|
Pair<org.apache.hbase.thirdparty.com.google.protobuf.Message,CellScanner> |
SimpleRpcServer.call(org.apache.hbase.thirdparty.com.google.protobuf.BlockingService service,
org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor md,
org.apache.hbase.thirdparty.com.google.protobuf.Message param,
CellScanner cellScanner,
long receiveTime,
MonitoredRPCHandler status,
long startTime,
int timeout) |
Pair<org.apache.hbase.thirdparty.com.google.protobuf.Message,CellScanner> |
NettyRpcServer.call(org.apache.hbase.thirdparty.com.google.protobuf.BlockingService service,
org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor md,
org.apache.hbase.thirdparty.com.google.protobuf.Message param,
CellScanner cellScanner,
long receiveTime,
MonitoredRPCHandler status,
long startTime,
int timeout) |
Pair<org.apache.hbase.thirdparty.com.google.protobuf.Message,CellScanner> |
RpcServerInterface.call(org.apache.hbase.thirdparty.com.google.protobuf.BlockingService service,
org.apache.hbase.thirdparty.com.google.protobuf.Descriptors.MethodDescriptor md,
org.apache.hbase.thirdparty.com.google.protobuf.Message param,
CellScanner cellScanner,
long receiveTime,
MonitoredRPCHandler status,
long startTime,
int timeout)
Deprecated.
As of release 2.0, this will be removed in HBase 3.0
|
Pair<org.apache.hbase.thirdparty.com.google.protobuf.Message,CellScanner> |
RpcServerInterface.call(RpcCall call,
MonitoredRPCHandler status) |
Pair<org.apache.hbase.thirdparty.com.google.protobuf.Message,CellScanner> |
RpcServer.call(RpcCall call,
MonitoredRPCHandler status)
This is a server side method, which is invoked over RPC.
|
(package private) static Pair<org.apache.hbase.thirdparty.io.netty.channel.EventLoopGroup,Class<? extends org.apache.hbase.thirdparty.io.netty.channel.Channel>> |
NettyRpcClientConfigHelper.getEventLoopConfig(org.apache.hadoop.conf.Configuration conf) |
Modifier and Type | Method and Description |
---|---|
protected Pair<byte[][],byte[][]> |
TableInputFormatBase.getStartEndKeys() |
protected Pair<byte[][],byte[][]> |
TableInputFormat.getStartEndKeys() |
Pair<Integer,Integer> |
ImportTsv.TsvParser.parseRowKey(byte[] lineBytes,
int length)
Return starting position and length of row key from the specified line bytes.
|
Modifier and Type | Method and Description |
---|---|
(package private) void |
HashTable.TableHash.selectPartitions(Pair<byte[][],byte[][]> regionStartEndKeys)
Choose partitions between row ranges to hash to a single output file
Selects region boundaries that fall within the scan range, and groups them
into the desired number of partitions.
|
Modifier and Type | Method and Description |
---|---|
private static Pair<ReplicationPeerConfig,org.apache.hadoop.conf.Configuration> |
VerifyReplication.getPeerQuorumConfig(org.apache.hadoop.conf.Configuration conf,
String peerId) |
Modifier and Type | Field and Description |
---|---|
private List<Pair<RegionInfo,RegionInfo>> |
CatalogJanitor.Report.holes |
private Map<String,Pair<ServerName,List<ServerName>>> |
HbckChore.inconsistentRegions
The inconsistent regions.
|
private Map<String,Pair<ServerName,List<ServerName>>> |
HbckChore.inconsistentRegionsSnapshot |
private List<Pair<RegionInfo,RegionInfo>> |
CatalogJanitor.Report.overlaps |
private static Comparator<Pair<ServerName,Long>> |
DeadServer.ServerNameDeathDateComparator |
private List<Pair<RegionInfo,ServerName>> |
CatalogJanitor.Report.unknownServers |
Modifier and Type | Method and Description |
---|---|
private Pair<Boolean,Boolean> |
CatalogJanitor.checkDaughterInFs(RegionInfo parent,
RegionInfo daughter)
Checks if a daughter region -- either splitA or splitB -- still holds
references to parent.
|
private Pair<ServerName,org.apache.hadoop.hbase.shaded.protobuf.generated.HBaseProtos.RegionServerInfo> |
RegionServerTracker.getServerInfo(String name) |
Modifier and Type | Method and Description |
---|---|
List<Pair<ServerName,Long>> |
DeadServer.copyDeadServersSince(long ts)
Extract all the servers dead since a given time, and sort them.
|
protected List<Pair<ServerName,Long>> |
ClusterStatusPublisher.getDeadServers(long since)
Get the servers which died since a given timestamp.
|
List<Pair<RegionInfo,RegionInfo>> |
CatalogJanitor.Report.getHoles() |
Map<String,Pair<ServerName,List<ServerName>>> |
HbckChore.getInconsistentRegions()
Found the inconsistent regions.
|
List<Pair<RegionInfo,RegionInfo>> |
CatalogJanitor.Report.getOverlaps() |
List<Pair<RegionInfo,ServerName>> |
CatalogJanitor.Report.getUnknownServers() |
Modifier and Type | Method and Description |
---|---|
private RegionInfo |
MetaFixer.getHoleCover(Pair<RegionInfo,RegionInfo> hole) |
private boolean |
CatalogJanitor.hasNoReferences(Pair<Boolean,Boolean> p) |
Modifier and Type | Method and Description |
---|---|
Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path> |
SplitTableRegionProcedure.StoreFileSplitter.call() |
Pair<Integer,Integer> |
AssignmentManager.getReopenStatus(TableName tableName)
Used by the client (via master) to identify if all regions have the schema updates
|
private Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path> |
SplitTableRegionProcedure.splitStoreFile(HRegionFileSystem regionFs,
byte[] family,
HStoreFile sf) |
private Pair<Integer,Integer> |
SplitTableRegionProcedure.splitStoreFiles(MasterProcedureEnv env,
HRegionFileSystem regionFs)
Create Split directory
|
Modifier and Type | Method and Description |
---|---|
private Pair<Map<ServerName,List<RegionInfo>>,List<RegionInfo>> |
FavoredStochasticBalancer.segregateRegionsAndAssignRegionsWithFavoredNodes(Collection<RegionInfo> regions,
List<ServerName> onlineServers) |
Modifier and Type | Field and Description |
---|---|
private Map<String,Pair<String,String>> |
CloneSnapshotProcedure.parentsToChildrenPairMap |
private Map<String,Pair<String,String>> |
RestoreSnapshotProcedure.parentsToChildrenPairMap |
Modifier and Type | Method and Description |
---|---|
protected void |
EnabledTableSnapshotHandler.snapshotRegions(List<Pair<RegionInfo,ServerName>> regions)
This method kicks off a snapshot procedure.
|
void |
DisabledTableSnapshotHandler.snapshotRegions(List<Pair<RegionInfo,ServerName>> regionsAndLocations) |
protected abstract void |
TakeSnapshotHandler.snapshotRegions(List<Pair<RegionInfo,ServerName>> regions)
Snapshot the specified regions
|
Modifier and Type | Method and Description |
---|---|
private Pair<Long,Long> |
PartitionedMobCompactor.getFileInfo(List<HStoreFile> storeFiles)
Gets the max seqId and number of cells of the store files.
|
Modifier and Type | Method and Description |
---|---|
(package private) Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path> |
HRegionFileSystem.bulkLoadStoreFile(String familyName,
org.apache.hadoop.fs.Path srcPath,
long seqNum)
Bulk load: Add a specified store file to the specified family.
|
Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path> |
HStore.preBulkLoadHFile(String srcPathStr,
long seqNum)
This method should only be called from Region.
|
Modifier and Type | Method and Description |
---|---|
List<Pair<NonceKey,WALEdit>> |
HRegion.BatchOperation.buildWALEdits(MiniBatchOperationInProgress<Mutation> miniBatchOp)
Builds separate WALEdit per nonce by applying input mutations.
|
List<Pair<NonceKey,WALEdit>> |
HRegion.MutationBatchOperation.buildWALEdits(MiniBatchOperationInProgress<Mutation> miniBatchOp) |
Modifier and Type | Method and Description |
---|---|
Map<byte[],List<org.apache.hadoop.fs.Path>> |
HRegion.bulkLoadHFiles(Collection<Pair<byte[],String>> familyPaths,
boolean assignSeqId,
HRegion.BulkLoadListener bulkLoadListener)
Attempts to atomically load a group of hfiles.
|
Map<byte[],List<org.apache.hadoop.fs.Path>> |
HRegion.bulkLoadHFiles(Collection<Pair<byte[],String>> familyPaths,
boolean assignSeqId,
HRegion.BulkLoadListener bulkLoadListener,
boolean copyFile)
Attempts to atomically load a group of hfiles.
|
private static boolean |
HRegion.hasMultipleColumnFamilies(Collection<Pair<byte[],String>> familyPaths)
Determines whether multiple column families are present
Precondition: familyPaths is not null
|
void |
RegionCoprocessorHost.postBulkLoadHFile(List<Pair<byte[],String>> familyPaths,
Map<byte[],List<org.apache.hadoop.fs.Path>> map) |
void |
RegionCoprocessorHost.preBulkLoadHFile(List<Pair<byte[],String>> familyPaths) |
boolean |
RegionCoprocessorHost.preCommitStoreFile(byte[] family,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs) |
Modifier and Type | Method and Description |
---|---|
private Pair<Long,Integer> |
StripeCompactionPolicy.estimateTargetKvs(Collection<HStoreFile> files,
double splitCount) |
Modifier and Type | Method and Description |
---|---|
protected static Pair<DeleteTracker,ColumnTracker> |
ScanQueryMatcher.getTrackers(RegionCoprocessorHost host,
NavigableSet<byte[]> columns,
ScanInfo scanInfo,
long oldestUnexpiredTS,
Scan userScan) |
Modifier and Type | Method and Description |
---|---|
Pair<String,SortedSet<String>> |
ReplicationQueuesZKImpl.claimQueue(String regionserver,
String queueId) |
Pair<String,SortedSet<String>> |
ReplicationQueues.claimQueue(String regionserver,
String queueId)
Take ownership for the queue identified by queueId and belongs to a dead region server.
|
Pair<ReplicationPeerConfig,org.apache.hadoop.conf.Configuration> |
ReplicationPeers.getPeerConf(String peerId)
Returns the configuration needed to talk to the remote slave cluster.
|
Pair<ReplicationPeerConfig,org.apache.hadoop.conf.Configuration> |
ReplicationPeersZKImpl.getPeerConf(String peerId) |
private Pair<String,SortedSet<String>> |
ReplicationQueuesZKImpl.moveQueueUsingMulti(String znode,
String peerId)
It "atomically" copies one peer's wals queue from another dead region server and returns them
all sorted.
|
Modifier and Type | Method and Description |
---|---|
void |
ReplicationQueuesZKImpl.addHFileRefs(String peerId,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs) |
void |
ReplicationQueues.addHFileRefs(String peerId,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs)
Add new hfile references to the queue.
|
Modifier and Type | Field and Description |
---|---|
private Map<String,List<Pair<byte[],List<String>>>> |
HFileReplicator.bulkLoadHFileMap |
Modifier and Type | Method and Description |
---|---|
private Pair<Integer,Integer> |
ReplicationSourceWALReader.countDistinctRowKeysAndHFiles(WALEdit edit)
Count the number of different row keys in the given edit because of mini-batching.
|
Modifier and Type | Method and Description |
---|---|
private void |
ReplicationSink.addFamilyAndItsHFilePathToTableInMap(byte[] family,
String pathToHfileFromNS,
List<Pair<byte[],List<String>>> familyHFilePathsList) |
void |
ReplicationSourceManager.addHFileRefs(TableName tableName,
byte[] family,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs) |
void |
ReplicationSourceInterface.addHFileRefs(TableName tableName,
byte[] family,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs)
Add hfile names to the queue to be replicated.
|
void |
ReplicationSource.addHFileRefs(TableName tableName,
byte[] family,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs) |
(package private) void |
Replication.addHFileRefsToQueue(TableName tableName,
byte[] family,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs) |
private void |
ReplicationSink.addNewTableEntryInMap(Map<String,List<Pair<byte[],List<String>>>> bulkLoadHFileMap,
byte[] family,
String pathToHfileFromNS,
String tableName) |
private void |
ReplicationSink.buildBulkLoadHFileMap(Map<String,List<Pair<byte[],List<String>>>> bulkLoadHFileMap,
TableName table,
Cell cell) |
void |
ReplicationObserver.preCommitStoreFile(ObserverContext<RegionCoprocessorEnvironment> ctx,
byte[] family,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs) |
Constructor and Description |
---|
HFileReplicator(org.apache.hadoop.conf.Configuration sourceClusterConf,
String sourceBaseNamespaceDirPath,
String sourceHFileArchiveDirPath,
Map<String,List<Pair<byte[],List<String>>>> tableQueueMap,
org.apache.hadoop.conf.Configuration conf,
Connection connection) |
Modifier and Type | Method and Description |
---|---|
private static Pair<org.eclipse.jetty.servlet.FilterHolder,Class<? extends org.glassfish.jersey.servlet.ServletContainer>> |
RESTServer.loginServerPrincipal(UserProvider userProvider,
org.apache.hadoop.conf.Configuration conf) |
Modifier and Type | Method and Description |
---|---|
private static Pair<String,TablePermission> |
AccessControlLists.parsePermissionRecord(byte[] entryName,
Cell kv) |
Modifier and Type | Method and Description |
---|---|
void |
AccessController.preBulkLoadHFile(ObserverContext<RegionCoprocessorEnvironment> ctx,
List<Pair<byte[],String>> familyPaths)
Verifies user has CREATE privileges on
the Column Families involved in the bulkLoadHFile
request.
|
Modifier and Type | Field and Description |
---|---|
private List<Pair<List<Tag>,Byte>> |
VisibilityScanDeleteTracker.visibilityTagsDeleteColumns |
private List<Pair<List<Tag>,Byte>> |
VisibilityScanDeleteTracker.visiblityTagsDeleteColumnVersion |
Modifier and Type | Method and Description |
---|---|
private Pair<Boolean,Tag> |
VisibilityController.checkForReservedVisibilityTagPresence(Cell cell,
Pair<Boolean,Tag> pair)
Checks whether cell contains any tag with type as VISIBILITY_TAG_TYPE.
|
protected Pair<Map<String,Integer>,Map<String,List<Integer>>> |
DefaultVisibilityLabelServiceImpl.extractLabelsAndAuths(List<List<Cell>> labelDetails) |
Modifier and Type | Method and Description |
---|---|
private Pair<Boolean,Tag> |
VisibilityController.checkForReservedVisibilityTagPresence(Cell cell,
Pair<Boolean,Tag> pair)
Checks whether cell contains any tag with type as VISIBILITY_TAG_TYPE.
|
Modifier and Type | Field and Description |
---|---|
private List<Pair<org.apache.hadoop.io.BytesWritable,Long>> |
ExportSnapshot.ExportSnapshotInputFormat.ExportSnapshotInputSplit.files |
private List<Pair<org.apache.hadoop.io.BytesWritable,Long>> |
ExportSnapshot.ExportSnapshotInputFormat.ExportSnapshotRecordReader.files |
private Map<String,Pair<String,String>> |
RestoreSnapshotHelper.parentsMap |
private Map<String,Pair<String,String>> |
RestoreSnapshotHelper.RestoreMetaChanges.parentsMap |
Modifier and Type | Method and Description |
---|---|
(package private) static List<List<Pair<org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo,Long>>> |
ExportSnapshot.getBalancedSplits(List<Pair<org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo,Long>> files,
int ngroups)
Given a list of file paths and sizes, create around ngroups in as balanced a way as possible.
|
Map<String,Pair<String,String>> |
RestoreSnapshotHelper.RestoreMetaChanges.getParentToChildrenPairMap()
Returns the map of parent-children_pair.
|
private static List<Pair<org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo,Long>> |
ExportSnapshot.getSnapshotFiles(org.apache.hadoop.conf.Configuration conf,
org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path snapshotDir)
Extract the list of files (HFiles/WALs) to copy using Map-Reduce.
|
private List<Pair<org.apache.hadoop.io.BytesWritable,Long>> |
ExportSnapshot.ExportSnapshotInputFormat.ExportSnapshotInputSplit.getSplitKeys() |
Modifier and Type | Method and Description |
---|---|
(package private) static List<List<Pair<org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo,Long>>> |
ExportSnapshot.getBalancedSplits(List<Pair<org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo,Long>> files,
int ngroups)
Given a list of file paths and sizes, create around ngroups in as balanced a way as possible.
|
Constructor and Description |
---|
ExportSnapshotInputSplit(List<Pair<org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo,Long>> snapshotFiles) |
ExportSnapshotRecordReader(List<Pair<org.apache.hadoop.io.BytesWritable,Long>> files) |
RestoreMetaChanges(TableDescriptor htd,
Map<String,Pair<String,String>> parentsMap) |
Modifier and Type | Method and Description |
---|---|
protected Pair<List<LoadIncrementalHFiles.LoadQueueItem>,String> |
LoadIncrementalHFiles.groupOrSplit(org.apache.hbase.thirdparty.com.google.common.collect.Multimap<ByteBuffer,LoadIncrementalHFiles.LoadQueueItem> regionGroups,
LoadIncrementalHFiles.LoadQueueItem item,
Table table,
Pair<byte[][],byte[][]> startEndKeys)
Attempt to assign the given load queue item into its target region group.
|
private Pair<org.apache.hbase.thirdparty.com.google.common.collect.Multimap<ByteBuffer,LoadIncrementalHFiles.LoadQueueItem>,Set<String>> |
LoadIncrementalHFiles.groupOrSplitPhase(Table table,
ExecutorService pool,
Deque<LoadIncrementalHFiles.LoadQueueItem> queue,
Pair<byte[][],byte[][]> startEndKeys) |
Modifier and Type | Method and Description |
---|---|
protected Pair<List<LoadIncrementalHFiles.LoadQueueItem>,String> |
LoadIncrementalHFiles.groupOrSplit(org.apache.hbase.thirdparty.com.google.common.collect.Multimap<ByteBuffer,LoadIncrementalHFiles.LoadQueueItem> regionGroups,
LoadIncrementalHFiles.LoadQueueItem item,
Table table,
Pair<byte[][],byte[][]> startEndKeys)
Attempt to assign the given load queue item into its target region group.
|
private Pair<org.apache.hbase.thirdparty.com.google.common.collect.Multimap<ByteBuffer,LoadIncrementalHFiles.LoadQueueItem>,Set<String>> |
LoadIncrementalHFiles.groupOrSplitPhase(Table table,
ExecutorService pool,
Deque<LoadIncrementalHFiles.LoadQueueItem> queue,
Pair<byte[][],byte[][]> startEndKeys) |
void |
LoadIncrementalHFiles.loadHFileQueue(Table table,
Connection conn,
Deque<LoadIncrementalHFiles.LoadQueueItem> queue,
Pair<byte[][],byte[][]> startEndKeys)
Used by the replication sink to load the hfiles from the source cluster.
|
void |
LoadIncrementalHFiles.loadHFileQueue(Table table,
Connection conn,
Deque<LoadIncrementalHFiles.LoadQueueItem> queue,
Pair<byte[][],byte[][]> startEndKeys,
boolean copyFile)
Used by the replication sink to load the hfiles from the source cluster.
|
Modifier and Type | Method and Description |
---|---|
private String |
LoadIncrementalHFiles.toString(List<Pair<byte[],String>> list) |
Modifier and Type | Field and Description |
---|---|
private Deque<Pair<Integer,Integer>> |
MunkresAssignment.path |
Modifier and Type | Method and Description |
---|---|
static Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.FSDataOutputStream> |
HBaseFsck.checkAndMarkRunningHbck(org.apache.hadoop.conf.Configuration conf,
RetryCounter retryCounter)
This method maintains a lock using a file.
|
private Pair<Integer,Integer> |
MunkresAssignment.findUncoveredZero()
Find a zero cost assignment which is not covered.
|
private static Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path> |
RegionSplitter.getTableDirAndSplitFile(org.apache.hadoop.conf.Configuration conf,
TableName tableName) |
static <T1,T2> Pair<T1,T2> |
Pair.newPair(T1 a,
T2 b)
Constructs a new pair, inferring the type via the passed arguments
|
private Pair<Integer,Integer> |
MunkresAssignment.primeInRow(int r)
Find a primed zero in the specified row.
|
private Pair<Integer,Integer> |
MunkresAssignment.starInCol(int c)
Find a starred zero in the specified column.
|
private Pair<Integer,Integer> |
MunkresAssignment.starInRow(int r)
Find a starred zero in a specified row.
|
Modifier and Type | Method and Description |
---|---|
(package private) static LinkedList<Pair<byte[],byte[]>> |
RegionSplitter.getSplits(Connection connection,
TableName tableName,
RegionSplitter.SplitAlgorithm splitAlgo) |
(package private) static LinkedList<Pair<byte[],byte[]>> |
RegionSplitter.splitScan(LinkedList<Pair<byte[],byte[]>> regionList,
Connection connection,
TableName tableName,
RegionSplitter.SplitAlgorithm splitAlgo) |
Modifier and Type | Method and Description |
---|---|
(package private) static LinkedList<Pair<byte[],byte[]>> |
RegionSplitter.splitScan(LinkedList<Pair<byte[],byte[]>> regionList,
Connection connection,
TableName tableName,
RegionSplitter.SplitAlgorithm splitAlgo) |
Modifier and Type | Field and Description |
---|---|
private static Map<String,Pair<org.apache.hbase.thirdparty.io.netty.channel.EventLoopGroup,Class<? extends org.apache.hbase.thirdparty.io.netty.channel.Channel>>> |
NettyAsyncFSWALConfigHelper.EVENT_LOOP_CONFIG_MAP |
Modifier and Type | Method and Description |
---|---|
(package private) static Pair<org.apache.hbase.thirdparty.io.netty.channel.EventLoopGroup,Class<? extends org.apache.hbase.thirdparty.io.netty.channel.Channel>> |
NettyAsyncFSWALConfigHelper.getEventLoopConfig(org.apache.hadoop.conf.Configuration conf) |
Modifier and Type | Method and Description |
---|---|
static List<WALSplitter.MutationReplay> |
WALSplitter.getMutationsFromWALEntry(org.apache.hadoop.hbase.shaded.protobuf.generated.AdminProtos.WALEntry entry,
CellScanner cells,
Pair<WALKey,WALEdit> logEntry,
Durability durability)
This function is used to construct mutations from a WALEntry.
|
Modifier and Type | Method and Description |
---|---|
List<Pair<RegionInfo,ServerName>> |
MetaTableLocator.getMetaRegionsAndLocations(ZKWatcher zkw) |
List<Pair<RegionInfo,ServerName>> |
MetaTableLocator.getMetaRegionsAndLocations(ZKWatcher zkw,
int replicaId)
Gets the meta regions and their locations for the given path and replica ID.
|
Modifier and Type | Method and Description |
---|---|
private List<RegionInfo> |
MetaTableLocator.getListOfRegionInfos(List<Pair<RegionInfo,ServerName>> pairs) |
Copyright © 2007–2019 The Apache Software Foundation. All rights reserved.