Package | Description |
---|---|
org.apache.hadoop.hbase | |
org.apache.hadoop.hbase.client |
Provides HBase Client
|
org.apache.hadoop.hbase.ipc |
Tools to help define network clients and servers.
|
org.apache.hadoop.hbase.master | |
org.apache.hadoop.hbase.master.handler | |
org.apache.hadoop.hbase.master.procedure | |
org.apache.hadoop.hbase.procedure2 | |
org.apache.hadoop.hbase.protobuf |
Holds classes generated from protobuf
src/main/protobuf definition files. |
org.apache.hadoop.hbase.regionserver | |
org.apache.hadoop.hbase.regionserver.compactions | |
org.apache.hadoop.hbase.security | |
org.apache.hadoop.hbase.security.access | |
org.apache.hadoop.hbase.security.token | |
org.apache.hadoop.hbase.security.visibility | |
org.apache.hadoop.hbase.snapshot |
Modifier and Type | Method and Description |
---|---|
JVMClusterUtil.MasterThread |
LocalHBaseCluster.addMaster(org.apache.hadoop.conf.Configuration c,
int index,
User user) |
JVMClusterUtil.RegionServerThread |
LocalHBaseCluster.addRegionServer(org.apache.hadoop.conf.Configuration config,
int index,
User user) |
static boolean |
ProcedureInfo.isProcedureOwner(ProcedureInfo procInfo,
User user)
Check if the user is this procedure's owner
|
HTableDescriptor |
HTableDescriptor.setOwner(User owner)
Deprecated.
since 0.94.1
|
Modifier and Type | Method and Description |
---|---|
static HConnection |
HConnectionManager.createConnection(org.apache.hadoop.conf.Configuration conf,
ExecutorService pool,
User user)
Deprecated.
|
static Connection |
ConnectionFactory.createConnection(org.apache.hadoop.conf.Configuration conf,
ExecutorService pool,
User user)
Create a new Connection instance using the passed
conf instance. |
static HConnection |
HConnectionManager.createConnection(org.apache.hadoop.conf.Configuration conf,
User user)
Deprecated.
|
static Connection |
ConnectionFactory.createConnection(org.apache.hadoop.conf.Configuration conf,
User user)
Create a new Connection instance using the passed
conf instance. |
static ClusterConnection |
ConnectionUtils.createShortCircuitConnection(org.apache.hadoop.conf.Configuration conf,
ExecutorService pool,
User user,
ServerName serverName,
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.AdminService.BlockingInterface admin,
org.apache.hadoop.hbase.protobuf.generated.ClientProtos.ClientService.BlockingInterface client)
Creates a short-circuit connection that can bypass the RPC layer (serialization,
deserialization, networking, etc..) when talking to a local server.
|
Modifier and Type | Field and Description |
---|---|
protected User |
RpcServer.Connection.user |
Modifier and Type | Method and Description |
---|---|
User |
RpcCallContext.getRequestUser()
Returns the user credentials associated with the current RPC request or
null if no credentials were provided. |
static User |
RpcServer.getRequestUser()
Returns the user credentials associated with the current RPC request or
null if no credentials were provided. |
User |
RpcServer.Call.getRequestUser() |
User |
ConnectionId.getTicket() |
Modifier and Type | Method and Description |
---|---|
com.google.protobuf.BlockingRpcChannel |
AbstractRpcClient.createBlockingRpcChannel(ServerName sn,
User ticket,
int rpcTimeout) |
com.google.protobuf.BlockingRpcChannel |
RpcClient.createBlockingRpcChannel(ServerName sn,
User user,
int rpcTimeout)
Creates a "channel" that can be used by a blocking protobuf service.
|
com.google.protobuf.RpcChannel |
AbstractRpcClient.createRpcChannel(ServerName sn,
User user,
int rpcTimeout) |
com.google.protobuf.RpcChannel |
RpcClient.createRpcChannel(ServerName sn,
User user,
int rpcTimeout)
Creates a "channel" that can be used by a protobuf service.
|
int |
PriorityFunction.getPriority(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RequestHeader header,
com.google.protobuf.Message param,
User user)
Returns the 'priority type' of the specified request.
|
static int |
ConnectionId.hashCode(User ticket,
String serviceName,
InetSocketAddress address) |
Constructor and Description |
---|
AbstractRpcClient.BlockingRpcChannelImplementation(AbstractRpcClient<?> rpcClient,
InetSocketAddress addr,
User ticket,
int rpcTimeout) |
AbstractRpcClient.RpcChannelImplementation(AbstractRpcClient<?> rpcClient,
InetSocketAddress addr,
User ticket,
int rpcTimeout) |
ConnectionId(User ticket,
String serviceName,
InetSocketAddress address) |
Modifier and Type | Method and Description |
---|---|
void |
HMaster.dispatchMergingRegions(HRegionInfo region_a,
HRegionInfo region_b,
boolean forcible,
User user) |
void |
MasterServices.dispatchMergingRegions(HRegionInfo region_a,
HRegionInfo region_b,
boolean forcible,
User user)
Merge two regions.
|
int |
MasterAnnotationReadingPriorityFunction.getPriority(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RequestHeader header,
com.google.protobuf.Message param,
User user) |
void |
ServerManager.sendRegionsMerge(ServerName server,
HRegionInfo region_a,
HRegionInfo region_b,
boolean forcible,
User user)
Sends an MERGE REGIONS RPC to the specified server to merge the specified
regions.
|
Constructor and Description |
---|
DispatchMergingRegionHandler(MasterServices services,
CatalogJanitor catalogJanitor,
HRegionInfo region_a,
HRegionInfo region_b,
boolean forcible,
User user) |
Modifier and Type | Method and Description |
---|---|
User |
MasterProcedureEnv.getRequestUser() |
Modifier and Type | Method and Description |
---|---|
boolean |
ProcedureExecutor.isProcedureOwner(long procId,
User user)
Check if the user is this procedure's owner
|
void |
ProcedureExecutor.setFailureResultForNonce(NonceKey nonceKey,
String procName,
User procOwner,
IOException exception)
If the failure failed before submitting it, we may want to give back the
same error to the requests with the same nonceKey.
|
Modifier and Type | Method and Description |
---|---|
static void |
ProtobufUtil.mergeRegions(com.google.protobuf.RpcController controller,
org.apache.hadoop.hbase.protobuf.generated.AdminProtos.AdminService.BlockingInterface admin,
HRegionInfo region_a,
HRegionInfo region_b,
boolean forcible,
User user)
A helper to merge regions using admin protocol.
|
Modifier and Type | Method and Description |
---|---|
boolean |
HRegion.compact(CompactionContext compaction,
Store store,
ThroughputController throughputController,
User user) |
List<StoreFile> |
Store.compact(CompactionContext compaction,
ThroughputController throughputController,
User user) |
List<StoreFile> |
HStore.compact(CompactionContext compaction,
ThroughputController throughputController,
User user) |
HRegion |
RegionMergeTransactionImpl.execute(Server server,
RegionServerServices services,
User user) |
PairOfSameType<Region> |
SplitTransactionImpl.execute(Server server,
RegionServerServices services,
User user)
Run the transaction.
|
Region |
RegionMergeTransaction.execute(Server server,
RegionServerServices services,
User user)
Run the transaction.
|
PairOfSameType<Region> |
SplitTransaction.execute(Server server,
RegionServerServices services,
User user)
Run the transaction.
|
int |
AnnotationReadingPriorityFunction.getPriority(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RequestHeader header,
com.google.protobuf.Message param,
User user)
Returns a 'priority' based on the request type.
|
int |
RSRpcServices.getPriority(org.apache.hadoop.hbase.protobuf.generated.RPCProtos.RequestHeader header,
com.google.protobuf.Message param,
User user) |
CompactionContext |
Store.requestCompaction(int priority,
CompactionRequest baseRequest,
User user) |
CompactionContext |
HStore.requestCompaction(int priority,
CompactionRequest baseRequest,
User user) |
CompactionRequest |
CompactionRequestor.requestCompaction(Region r,
Store s,
String why,
int pri,
CompactionRequest request,
User user) |
CompactionRequest |
CompactSplitThread.requestCompaction(Region r,
Store s,
String why,
int priority,
CompactionRequest request,
User user) |
List<CompactionRequest> |
CompactionRequestor.requestCompaction(Region r,
String why,
int pri,
List<Pair<CompactionRequest,Store>> requests,
User user) |
List<CompactionRequest> |
CompactSplitThread.requestCompaction(Region r,
String why,
int p,
List<Pair<CompactionRequest,Store>> requests,
User user) |
void |
CompactSplitThread.requestRegionsMerge(Region a,
Region b,
boolean forcible,
long masterSystemTime,
User user) |
void |
CompactSplitThread.requestSplit(Region r,
byte[] midKey,
User user) |
boolean |
RegionMergeTransactionImpl.rollback(Server server,
RegionServerServices services,
User user) |
boolean |
SplitTransactionImpl.rollback(Server server,
RegionServerServices services,
User user) |
boolean |
RegionMergeTransaction.rollback(Server server,
RegionServerServices services,
User user)
Roll back a failed transaction
|
boolean |
SplitTransaction.rollback(Server server,
RegionServerServices services,
User user)
Roll back a failed transaction
|
void |
RegionMergeTransactionImpl.stepsAfterPONR(Server server,
RegionServerServices services,
HRegion mergedRegion,
User user) |
PairOfSameType<Region> |
SplitTransactionImpl.stepsAfterPONR(Server server,
RegionServerServices services,
PairOfSameType<Region> regions,
User user) |
Modifier and Type | Method and Description |
---|---|
protected List<org.apache.hadoop.fs.Path> |
Compactor.compact(CompactionRequest request,
Compactor.InternalScannerFactory scannerFactory,
Compactor.CellSinkFactory<T> sinkFactory,
ThroughputController throughputController,
User user) |
List<org.apache.hadoop.fs.Path> |
StripeCompactor.compact(CompactionRequest request,
int targetCount,
long targetSize,
byte[] left,
byte[] right,
byte[] majorRangeFromRow,
byte[] majorRangeToRow,
ThroughputController throughputController,
User user) |
List<org.apache.hadoop.fs.Path> |
StripeCompactor.compact(CompactionRequest request,
List<byte[]> targetBoundaries,
byte[] majorRangeFromRow,
byte[] majorRangeToRow,
ThroughputController throughputController,
User user) |
List<org.apache.hadoop.fs.Path> |
DateTieredCompactor.compact(CompactionRequest request,
List<Long> lowerBoundaries,
ThroughputController throughputController,
User user) |
List<org.apache.hadoop.fs.Path> |
DefaultCompactor.compact(CompactionRequest request,
ThroughputController throughputController,
User user)
Do a minor/major compaction on an explicit set of storefiles from a Store.
|
abstract List<org.apache.hadoop.fs.Path> |
CompactionContext.compact(ThroughputController throughputController,
User user) |
abstract List<org.apache.hadoop.fs.Path> |
StripeCompactionPolicy.StripeCompactionRequest.execute(StripeCompactor compactor,
ThroughputController throughputController,
User user)
Executes the request against compactor (essentially, just calls correct overload of
compact method), to simulate more dynamic dispatch.
|
protected InternalScanner |
Compactor.postCreateCoprocScanner(CompactionRequest request,
ScanType scanType,
InternalScanner scanner,
User user)
Calls coprocessor, if any, to create scanners - after normal scanner creation.
|
protected InternalScanner |
Compactor.preCreateCoprocScanner(CompactionRequest request,
ScanType scanType,
long earliestPutTs,
List<StoreFileScanner> scanners,
User user,
long readPoint)
Calls coprocessor, if any, to create compaction scanner - before normal scanner creation.
|
Modifier and Type | Class and Description |
---|---|
static class |
User.SecureHadoopUser
Bridges
User invocations to underlying calls to
UserGroupInformation for secure Hadoop
0.20 and versions 0.21 and above. |
Modifier and Type | Method and Description |
---|---|
static User |
User.create(org.apache.hadoop.security.UserGroupInformation ugi)
Wraps an underlying
UserGroupInformation instance. |
User |
UserProvider.create(org.apache.hadoop.security.UserGroupInformation ugi)
Wraps an underlying
UserGroupInformation instance. |
static User |
User.createUserForTesting(org.apache.hadoop.conf.Configuration conf,
String name,
String[] groups)
Generates a new
User instance specifically for use in test code. |
static User |
User.SecureHadoopUser.createUserForTesting(org.apache.hadoop.conf.Configuration conf,
String name,
String[] groups) |
static User |
User.getCurrent()
Returns the
User instance within current execution context. |
User |
UserProvider.getCurrent() |
static User |
Superusers.getSystemUser() |
Modifier and Type | Method and Description |
---|---|
static boolean |
Superusers.isSuperUser(User user) |
Modifier and Type | Method and Description |
---|---|
User |
AuthResult.getUser() |
Modifier and Type | Method and Description |
---|---|
static AuthResult |
AuthResult.allow(String request,
String reason,
User user,
Permission.Action action,
String namespace) |
static AuthResult |
AuthResult.allow(String request,
String reason,
User user,
Permission.Action action,
TableName table,
byte[] family,
byte[] qualifier) |
static AuthResult |
AuthResult.allow(String request,
String reason,
User user,
Permission.Action action,
TableName table,
Map<byte[],? extends Collection<?>> families) |
boolean |
TableAuthManager.authorize(User user,
Permission.Action action)
Authorize a global permission based on ACLs for the given user and the
user's groups.
|
boolean |
TableAuthManager.authorize(User user,
String namespace,
Permission.Action action) |
boolean |
TableAuthManager.authorize(User user,
TableName table,
byte[] family,
byte[] qualifier,
Permission.Action action) |
boolean |
TableAuthManager.authorize(User user,
TableName table,
byte[] family,
Permission.Action action) |
boolean |
TableAuthManager.authorize(User user,
TableName table,
Cell cell,
Permission.Action action)
Authorize a user for a given KV.
|
boolean |
TableAuthManager.authorizeUser(User user,
TableName table,
byte[] family,
byte[] qualifier,
Permission.Action action) |
boolean |
TableAuthManager.authorizeUser(User user,
TableName table,
byte[] family,
Permission.Action action)
Checks authorization to a given table and column family for a user, based on the
stored user permissions.
|
static AuthResult |
AuthResult.deny(String request,
String reason,
User user,
Permission.Action action,
String namespace) |
static AuthResult |
AuthResult.deny(String request,
String reason,
User user,
Permission.Action action,
TableName table,
byte[] family,
byte[] qualifier) |
static AuthResult |
AuthResult.deny(String request,
String reason,
User user,
Permission.Action action,
TableName table,
Map<byte[],? extends Collection<?>> families) |
static List<Permission> |
AccessControlLists.getCellPermissionsForUser(User user,
Cell cell) |
boolean |
TableAuthManager.hasAccess(User user,
TableName table,
Permission.Action action) |
boolean |
TableAuthManager.matchPermission(User user,
TableName table,
byte[] family,
byte[] qualifier,
Permission.Action action) |
boolean |
TableAuthManager.matchPermission(User user,
TableName table,
byte[] family,
Permission.Action action)
Returns true if the given user has a
TablePermission matching up
to the column family portion of a permission. |
void |
AccessChecker.requireAccess(User user,
String request,
TableName tableName,
Permission.Action... permissions)
Authorizes that the current user has any of the given permissions to access the table.
|
void |
AccessChecker.requireGlobalPermission(User user,
String request,
Permission.Action perm,
String namespace)
Checks that the user has the given global permission.
|
void |
AccessChecker.requireGlobalPermission(User user,
String request,
Permission.Action perm,
TableName tableName,
Map<byte[],? extends Collection<byte[]>> familyMap)
Checks that the user has the given global permission.
|
void |
AccessChecker.requireNamespacePermission(User user,
String request,
String namespace,
Permission.Action... permissions)
Checks that the user has the given global or namespace permission.
|
void |
AccessChecker.requireNamespacePermission(User user,
String request,
String namespace,
TableName tableName,
Map<byte[],? extends Collection<byte[]>> familyMap,
Permission.Action... permissions)
Checks that the user has the given global or namespace permission.
|
void |
AccessChecker.requirePermission(User user,
String request,
Permission.Action perm)
Authorizes that the current user has global privileges for the given action.
|
void |
AccessChecker.requirePermission(User user,
String request,
TableName tableName,
byte[] family,
byte[] qualifier,
Permission.Action... permissions)
Authorizes that the current user has any of the given permissions for the
given table, column family and column qualifier.
|
void |
AccessChecker.requireTablePermission(User user,
String request,
TableName tableName,
byte[] family,
byte[] qualifier,
Permission.Action... permissions)
Authorizes that the current user has any of the given permissions for the
given table, column family and column qualifier.
|
boolean |
TableAuthManager.userHasAccess(User user,
TableName table,
Permission.Action action)
Checks if the user has access to the full table or at least a family/qualifier
for the specified action.
|
Constructor and Description |
---|
AuthResult(boolean allowed,
String request,
String reason,
User user,
Permission.Action action,
String namespace) |
AuthResult(boolean allowed,
String request,
String reason,
User user,
Permission.Action action,
TableName table,
byte[] family,
byte[] qualifier) |
AuthResult(boolean allowed,
String request,
String reason,
User user,
Permission.Action action,
TableName table,
Map<byte[],? extends Collection<?>> families) |
Modifier and Type | Method and Description |
---|---|
static void |
TokenUtil.addTokenForJob(Connection conn,
org.apache.hadoop.mapred.JobConf job,
User user)
Checks for an authentication token for the given user, obtaining a new token if necessary,
and adds it to the credentials for the given map reduce job.
|
static void |
TokenUtil.addTokenForJob(Connection conn,
User user,
org.apache.hadoop.mapreduce.Job job)
Checks for an authentication token for the given user, obtaining a new token if necessary,
and adds it to the credentials for the given map reduce job.
|
static boolean |
TokenUtil.addTokenIfMissing(Connection conn,
User user)
Checks if an authentication tokens exists for the connected cluster,
obtaining one if needed and adding it to the user's credentials.
|
static void |
TokenUtil.obtainAndCacheToken(Connection conn,
User user)
Obtain an authentication token for the given user and add it to the
user's credentials.
|
static org.apache.hadoop.security.token.Token<AuthenticationTokenIdentifier> |
TokenUtil.obtainToken(Connection conn,
User user)
Obtain and return an authentication token for the current user.
|
static void |
TokenUtil.obtainTokenForJob(Connection conn,
org.apache.hadoop.mapred.JobConf job,
User user)
Obtain an authentication token on behalf of the given user and add it to
the credentials for the given map reduce job.
|
static void |
TokenUtil.obtainTokenForJob(Connection conn,
User user,
org.apache.hadoop.mapreduce.Job job)
Obtain an authentication token on behalf of the given user and add it to
the credentials for the given map reduce job.
|
Modifier and Type | Method and Description |
---|---|
static User |
VisibilityUtils.getActiveUser() |
Modifier and Type | Method and Description |
---|---|
List<String> |
ScanLabelGenerator.getLabels(User user,
Authorizations authorizations)
Helps to get a list of lables associated with an UGI
|
List<String> |
SimpleScanLabelGenerator.getLabels(User user,
Authorizations authorizations) |
List<String> |
EnforcingScanLabelGenerator.getLabels(User user,
Authorizations authorizations) |
List<String> |
FeedUserAuthScanLabelGenerator.getLabels(User user,
Authorizations authorizations) |
List<String> |
DefinedSetFilterScanLabelGenerator.getLabels(User user,
Authorizations authorizations) |
boolean |
DefaultVisibilityLabelServiceImpl.havingSystemAuth(User user) |
boolean |
VisibilityLabelService.havingSystemAuth(User user)
System checks for user auth during admin operations.
|
Modifier and Type | Method and Description |
---|---|
static boolean |
SnapshotDescriptionUtils.isSnapshotOwner(org.apache.hadoop.hbase.protobuf.generated.HBaseProtos.SnapshotDescription snapshot,
User user)
Check if the user is this table snapshot's owner
|
Copyright © 2007–2019 The Apache Software Foundation. All rights reserved.