Modifier and Type | Method and Description |
---|---|
static Pair<HRegionInfo,ServerName> |
HRegionInfo.getHRegionInfoAndServerName(Result r)
Deprecated.
use MetaTableAccessor methods for interacting with meta layouts
|
static Pair<HRegionInfo,ServerName> |
MetaTableAccessor.getRegion(Connection connection,
byte[] regionName)
Deprecated.
|
static Pair<HRegionInfo,HRegionInfo> |
MetaTableAccessor.getRegionsFromMergeQualifier(Connection connection,
byte[] regionName)
Get regions from the merge qualifier of the specified merged region
|
Modifier and Type | Method and Description |
---|---|
static List<Pair<HRegionInfo,ServerName>> |
MetaTableAccessor.getTableRegionsAndLocations(ZooKeeperWatcher zkw,
Connection connection,
TableName tableName) |
static List<Pair<HRegionInfo,ServerName>> |
MetaTableAccessor.getTableRegionsAndLocations(ZooKeeperWatcher zkw,
Connection connection,
TableName tableName,
boolean excludeOfflinedSplitParents) |
static List<Pair<HRegionInfo,ServerName>> |
MetaTableAccessor.getTableRegionsAndLocations(ZooKeeperWatcher zkw,
Connection connection,
TableName tableName,
boolean excludeOfflinedSplitParents,
boolean excludeReplicaRegions) |
Modifier and Type | Method and Description |
---|---|
Pair<Integer,Integer> |
Admin.getAlterStatus(byte[] tableName)
Get the status of alter command - indicates how many regions have received the updated schema
Asynchronous operation.
|
Pair<Integer,Integer> |
HBaseAdmin.getAlterStatus(byte[] tableName)
Get the status of alter command - indicates how many regions have received
the updated schema Asynchronous operation.
|
Pair<Integer,Integer> |
Admin.getAlterStatus(TableName tableName)
Get the status of alter command - indicates how many regions have received the updated schema
Asynchronous operation.
|
Pair<Integer,Integer> |
HBaseAdmin.getAlterStatus(TableName tableName)
Get the status of alter command - indicates how many regions have received
the updated schema Asynchronous operation.
|
Pair<byte[][],byte[][]> |
HRegionLocator.getStartEndKeys()
Gets the starting and ending row keys for every region in the currently
open table.
|
Pair<byte[][],byte[][]> |
HTable.getStartEndKeys()
Deprecated.
Since 1.1.0. Use
RegionLocator.getStartEndKeys() instead; |
Pair<byte[][],byte[][]> |
RegionLocator.getStartEndKeys()
Gets the starting and ending row keys for every region in the currently
open table.
|
Modifier and Type | Method and Description |
---|---|
boolean |
SecureBulkLoadClient.bulkLoadHFiles(List<Pair<byte[],String>> familyPaths,
org.apache.hadoop.security.token.Token<?> userToken,
String bulkToken,
byte[] startRow)
Deprecated.
|
boolean |
SecureBulkLoadClient.bulkLoadHFiles(List<Pair<byte[],String>> familyPaths,
org.apache.hadoop.security.token.Token<?> userToken,
String bulkToken,
byte[] startRow,
List<String> clusterIds) |
Modifier and Type | Method and Description |
---|---|
static void |
Constraints.add(HTableDescriptor desc,
Pair<Class<? extends Constraint>,org.apache.hadoop.conf.Configuration>... constraints)
Add constraints and their associated configurations to the table.
|
Modifier and Type | Method and Description |
---|---|
boolean |
BaseRegionObserver.postBulkLoadHFile(ObserverContext<RegionCoprocessorEnvironment> ctx,
List<Pair<byte[],String>> familyPaths,
boolean hasLoaded) |
boolean |
RegionObserver.postBulkLoadHFile(ObserverContext<RegionCoprocessorEnvironment> ctx,
List<Pair<byte[],String>> familyPaths,
boolean hasLoaded)
Called after bulkLoadHFile.
|
void |
BaseRegionObserver.preBulkLoadHFile(ObserverContext<RegionCoprocessorEnvironment> ctx,
List<Pair<byte[],String>> familyPaths) |
void |
RegionObserver.preBulkLoadHFile(ObserverContext<RegionCoprocessorEnvironment> ctx,
List<Pair<byte[],String>> familyPaths)
Called before bulkLoadHFile.
|
void |
BaseRegionObserver.preCommitStoreFile(ObserverContext<RegionCoprocessorEnvironment> ctx,
byte[] family,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs) |
void |
RegionObserver.preCommitStoreFile(ObserverContext<RegionCoprocessorEnvironment> ctx,
byte[] family,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs)
Called before moving bulk loaded hfile to region directory.
|
Constructor and Description |
---|
FuzzyRowFilter(List<Pair<byte[],byte[]>> fuzzyKeysData) |
Modifier and Type | Method and Description |
---|---|
static Pair<Integer,Integer> |
StreamUtils.readRawVarint32(byte[] input,
int offset)
Reads a varInt value stored in an array.
|
Modifier and Type | Method and Description |
---|---|
Pair<com.google.protobuf.Message,CellScanner> |
RpcServerInterface.call(com.google.protobuf.BlockingService service,
com.google.protobuf.Descriptors.MethodDescriptor md,
com.google.protobuf.Message param,
CellScanner cellScanner,
long receiveTime,
MonitoredRPCHandler status)
Deprecated.
As of release 1.3, this will be removed in HBase 3.0
|
Pair<com.google.protobuf.Message,CellScanner> |
RpcServer.call(com.google.protobuf.BlockingService service,
com.google.protobuf.Descriptors.MethodDescriptor md,
com.google.protobuf.Message param,
CellScanner cellScanner,
long receiveTime,
MonitoredRPCHandler status) |
Pair<com.google.protobuf.Message,CellScanner> |
RpcServerInterface.call(com.google.protobuf.BlockingService service,
com.google.protobuf.Descriptors.MethodDescriptor md,
com.google.protobuf.Message param,
CellScanner cellScanner,
long receiveTime,
MonitoredRPCHandler status,
long startTime,
int timeout) |
Pair<com.google.protobuf.Message,CellScanner> |
RpcServer.call(com.google.protobuf.BlockingService service,
com.google.protobuf.Descriptors.MethodDescriptor md,
com.google.protobuf.Message param,
CellScanner cellScanner,
long receiveTime,
MonitoredRPCHandler status,
long startTime,
int timeout)
This is a server side method, which is invoked over RPC.
|
Modifier and Type | Method and Description |
---|---|
protected Pair<byte[][],byte[][]> |
TableInputFormatBase.getStartEndKeys() |
protected Pair<byte[][],byte[][]> |
TableInputFormat.getStartEndKeys() |
Pair<Integer,Integer> |
ImportTsv.TsvParser.parseRowKey(byte[] lineBytes,
int length)
Return starting position and length of row key from the specified line bytes.
|
Modifier and Type | Method and Description |
---|---|
protected List<LoadIncrementalHFiles.LoadQueueItem> |
LoadIncrementalHFiles.groupOrSplit(com.google.common.collect.Multimap<ByteBuffer,LoadIncrementalHFiles.LoadQueueItem> regionGroups,
LoadIncrementalHFiles.LoadQueueItem item,
Table table,
Pair<byte[][],byte[][]> startEndKeys)
Attempt to assign the given load queue item into its target region group.
|
void |
LoadIncrementalHFiles.loadHFileQueue(Table table,
Connection conn,
Deque<LoadIncrementalHFiles.LoadQueueItem> queue,
Pair<byte[][],byte[][]> startEndKeys)
Used by the replication sink to load the hfiles from the source cluster.
|
Modifier and Type | Method and Description |
---|---|
Pair<Integer,Integer> |
AssignmentManager.getReopenStatus(TableName tableName)
Used by the client to identify if all regions have the schema updates
|
Modifier and Type | Method and Description |
---|---|
List<Pair<ServerName,Long>> |
DeadServer.copyDeadServersSince(long ts)
Extract all the servers dead since a given time, and sort them.
|
protected List<Pair<ServerName,Long>> |
ClusterStatusPublisher.getDeadServers(long since)
Get the servers which died since a given timestamp.
|
Modifier and Type | Method and Description |
---|---|
protected void |
EnabledTableSnapshotHandler.snapshotRegions(List<Pair<HRegionInfo,ServerName>> regions)
This method kicks off a snapshot procedure.
|
protected abstract void |
TakeSnapshotHandler.snapshotRegions(List<Pair<HRegionInfo,ServerName>> regions)
Snapshot the specified regions
|
void |
DisabledTableSnapshotHandler.snapshotRegions(List<Pair<HRegionInfo,ServerName>> regionsAndLocations) |
Modifier and Type | Method and Description |
---|---|
Pair<ProcedureInfo,Procedure> |
ProcedureExecutor.getResultOrProcedure(long procId) |
Modifier and Type | Method and Description |
---|---|
static Pair<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ReplicateWALEntryRequest,CellScanner> |
ReplicationProtbufUtil.buildReplicateWALEntryRequest(WAL.Entry[] entries)
Create a new ReplicateWALEntryRequest from a list of WAL entries
|
static Pair<org.apache.hadoop.hbase.protobuf.generated.AdminProtos.ReplicateWALEntryRequest,CellScanner> |
ReplicationProtbufUtil.buildReplicateWALEntryRequest(WAL.Entry[] entries,
byte[] encodedRegionName,
String replicationClusterId,
org.apache.hadoop.fs.Path sourceBaseNamespaceDir,
org.apache.hadoop.fs.Path sourceHFileArchiveDir)
Create a new ReplicateWALEntryRequest from a list of HLog entries
|
Modifier and Type | Method and Description |
---|---|
static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest |
RequestConverter.buildBulkLoadHFileRequest(List<Pair<byte[],String>> familyPaths,
byte[] regionName,
boolean assignSeqNum)
Deprecated.
use buildBulkLoadHFileRequest(final List
|
static org.apache.hadoop.hbase.protobuf.generated.ClientProtos.BulkLoadHFileRequest |
RequestConverter.buildBulkLoadHFileRequest(List<Pair<byte[],String>> familyPaths,
byte[] regionName,
boolean assignSeqNum,
List<String> clusterIds)
Create a protocol buffer bulk load request
|
static org.apache.hadoop.hbase.protobuf.generated.AdminProtos.UpdateFavoredNodesRequest |
RequestConverter.buildUpdateFavoredNodesRequest(List<Pair<HRegionInfo,List<ServerName>>> updateRegionInfos)
Create a protocol buffer UpdateFavoredNodesRequest to update a list of favorednode mappings
|
static boolean |
ProtobufUtil.bulkLoadHFile(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ClientService.BlockingInterface client,
List<Pair<byte[],String>> familyPaths,
byte[] regionName,
boolean assignSeqNum)
Deprecated.
use bulkLoadHFile(final ClientService.BlockingInterface client,
final List
|
static boolean |
ProtobufUtil.bulkLoadHFile(org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ClientService.BlockingInterface client,
List<Pair<byte[],String>> familyPaths,
byte[] regionName,
boolean assignSeqNum,
List<String> clusterIds)
A helper to bulk load a list of HFiles using client protocol.
|
Modifier and Type | Method and Description |
---|---|
Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path> |
HStore.preBulkLoadHFile(String srcPathStr,
long seqNum) |
Modifier and Type | Method and Description |
---|---|
boolean |
HRegion.bulkLoadHFiles(Collection<Pair<byte[],String>> familyPaths,
boolean assignSeqId,
Region.BulkLoadListener bulkLoadListener)
Deprecated.
|
boolean |
Region.bulkLoadHFiles(Collection<Pair<byte[],String>> familyPaths,
boolean assignSeqId,
Region.BulkLoadListener bulkLoadListener)
Deprecated.
Do not use, see HBASE-22380
|
boolean |
HRegion.bulkLoadHFiles(Collection<Pair<byte[],String>> familyPaths,
boolean assignSeqId,
Region.BulkLoadListener bulkLoadListener,
List<String> clusterIds) |
boolean |
Region.bulkLoadHFiles(Collection<Pair<byte[],String>> familyPaths,
boolean assignSeqId,
Region.BulkLoadListener bulkLoadListener,
List<String> clusterIds)
Attempts to atomically load a group of hfiles.
|
boolean |
RegionCoprocessorHost.postBulkLoadHFile(List<Pair<byte[],String>> familyPaths,
boolean hasLoaded) |
boolean |
RegionCoprocessorHost.preBulkLoadHFile(List<Pair<byte[],String>> familyPaths) |
boolean |
RegionCoprocessorHost.preCommitStoreFile(byte[] family,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs) |
List<CompactionRequest> |
CompactionRequestor.requestCompaction(Region r,
String why,
int pri,
List<Pair<CompactionRequest,Store>> requests,
User user) |
List<CompactionRequest> |
CompactSplitThread.requestCompaction(Region r,
String why,
int p,
List<Pair<CompactionRequest,Store>> requests,
User user) |
List<CompactionRequest> |
CompactionRequestor.requestCompaction(Region r,
String why,
List<Pair<CompactionRequest,Store>> requests) |
List<CompactionRequest> |
CompactSplitThread.requestCompaction(Region r,
String why,
List<Pair<CompactionRequest,Store>> requests) |
Modifier and Type | Method and Description |
---|---|
void |
WALEditsReplaySink.replayEntries(List<Pair<HRegionLocation,WAL.Entry>> entries)
Replay an array of actions of the same region directly into the newly assigned Region Server
|
Modifier and Type | Method and Description |
---|---|
Pair<String,SortedSet<String>> |
ReplicationQueues.claimQueue(String regionserver,
String queueId)
Take ownership for the queue identified by queueId and belongs to a dead region server.
|
Pair<String,SortedSet<String>> |
ReplicationQueuesZKImpl.claimQueue(String regionserver,
String queueId) |
Pair<ReplicationPeerConfig,org.apache.hadoop.conf.Configuration> |
ReplicationPeersZKImpl.getPeerConf(String peerId) |
Pair<ReplicationPeerConfig,org.apache.hadoop.conf.Configuration> |
ReplicationPeers.getPeerConf(String peerId)
Returns the configuration needed to talk to the remote slave cluster.
|
Modifier and Type | Method and Description |
---|---|
void |
ReplicationQueues.addHFileRefs(String peerId,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs)
Add new hfile references to the queue.
|
void |
ReplicationQueuesZKImpl.addHFileRefs(String peerId,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs) |
Modifier and Type | Method and Description |
---|---|
void |
ReplicationSource.addHFileRefs(TableName tableName,
byte[] family,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs) |
void |
ReplicationSourceManager.addHFileRefs(TableName tableName,
byte[] family,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs) |
void |
ReplicationSourceInterface.addHFileRefs(TableName tableName,
byte[] family,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs)
Add hfile names to the queue to be replicated.
|
void |
ReplicationObserver.preCommitStoreFile(ObserverContext<RegionCoprocessorEnvironment> ctx,
byte[] family,
List<Pair<org.apache.hadoop.fs.Path,org.apache.hadoop.fs.Path>> pairs) |
Constructor and Description |
---|
HFileReplicator(org.apache.hadoop.conf.Configuration sourceClusterConf,
String sourceBaseNamespaceDirPath,
String sourceHFileArchiveDirPath,
Map<String,List<Pair<byte[],List<String>>>> tableQueueMap,
org.apache.hadoop.conf.Configuration conf,
Connection connection,
List<String> sourceClusterIds) |
Modifier and Type | Method and Description |
---|---|
void |
AccessController.preBulkLoadHFile(ObserverContext<RegionCoprocessorEnvironment> ctx,
List<Pair<byte[],String>> familyPaths)
Verifies user has CREATE privileges on
the Column Families involved in the bulkLoadHFile
request.
|
Modifier and Type | Method and Description |
---|---|
protected Pair<Map<String,Integer>,Map<String,List<Integer>>> |
DefaultVisibilityLabelServiceImpl.extractLabelsAndAuths(List<List<Cell>> labelDetails) |
Modifier and Type | Method and Description |
---|---|
static <T1,T2> Pair<T1,T2> |
Pair.newPair(T1 a,
T2 b)
Constructs a new pair, inferring the type via the passed arguments
|
Modifier and Type | Method and Description |
---|---|
static List<WALSplitter.MutationReplay> |
WALSplitter.getMutationsFromWALEntry(org.apache.hadoop.hbase.protobuf.generated.AdminProtos.WALEntry entry,
CellScanner cells,
Pair<WALKey,WALEdit> logEntry,
Durability durability)
This function is used to construct mutations from a WALEntry.
|
Modifier and Type | Method and Description |
---|---|
List<Pair<HRegionInfo,ServerName>> |
MetaTableLocator.getMetaRegionsAndLocations(ZooKeeperWatcher zkw) |
List<Pair<HRegionInfo,ServerName>> |
MetaTableLocator.getMetaRegionsAndLocations(ZooKeeperWatcher zkw,
int replicaId) |
Copyright © 2007–2019 The Apache Software Foundation. All rights reserved.