org.apache.hadoop.hbase.client.Put.add(byte[], byte[], byte[])
|
org.apache.hadoop.hbase.client.Put.add(byte[], byte[], long, byte[])
|
org.apache.hadoop.hbase.client.Put.add(byte[], ByteBuffer, long, ByteBuffer)
|
org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil.addDependencyJars(Configuration, Class>...)
|
org.apache.hadoop.hbase.security.visibility.VisibilityClient.addLabel(Configuration, String)
|
org.apache.hadoop.hbase.security.visibility.VisibilityClient.addLabels(Configuration, String[])
|
org.apache.hadoop.hbase.client.replication.ReplicationAdmin.addPeer(String, String)
Use addPeer(String, ReplicationPeerConfig, Map) instead.
|
org.apache.hadoop.hbase.client.replication.ReplicationAdmin.addPeer(String, String, String) |
org.apache.hadoop.hbase.client.Result.addResults(ClientProtos.RegionLoadStats)
|
org.apache.hadoop.hbase.client.MetaScanner.allTableRegions(Configuration, Connection, TableName, boolean)
|
org.apache.hadoop.hbase.client.HTableWrapper.batch(List extends Row>)
|
org.apache.hadoop.hbase.client.Table.batch(List extends Row>)
If any exception is thrown by one of the actions, there is no way to
retrieve the partially executed results. Use Table.batch(List, Object[]) instead.
|
org.apache.hadoop.hbase.client.HTable.batch(List extends Row>)
If any exception is thrown by one of the actions, there is no way to
retrieve the partially executed results. Use HTable.batch(List, Object[]) instead.
|
org.apache.hadoop.hbase.client.HTableWrapper.batchCallback(List extends Row>, Batch.Callback)
|
org.apache.hadoop.hbase.client.Table.batchCallback(List extends Row>, Batch.Callback)
|
org.apache.hadoop.hbase.client.HTable.batchCallback(List extends Row>, Batch.Callback)
|
org.apache.hadoop.hbase.protobuf.RequestConverter.buildBulkLoadHFileRequest(List>, byte[], boolean)
use buildBulkLoadHFileRequest(final List> familyPaths,
final byte[] regionName, boolean assignSeqNum, List clusterIds)
|
org.apache.hadoop.hbase.quotas.QuotaSettings.buildSetQuotaRequestProto(QuotaSettings)
Removed in HBase 2.0+ as a part of removing protobuf from our API
|
org.apache.hadoop.hbase.protobuf.ProtobufUtil.bulkLoadHFile(ClientProtos.ClientService.BlockingInterface, List>, byte[], boolean)
use bulkLoadHFile(final ClientService.BlockingInterface client,
final List> familyPaths, final byte[] regionName, boolean assignSeqNum,
List clusterIds) instead.
|
org.apache.hadoop.hbase.regionserver.HRegion.bulkLoadHFiles(Collection>, boolean, Region.BulkLoadListener) |
org.apache.hadoop.hbase.regionserver.Region.bulkLoadHFiles(Collection>, boolean, Region.BulkLoadListener)
Do not use, see HBASE-22380
|
org.apache.hadoop.hbase.client.coprocessor.SecureBulkLoadClient.bulkLoadHFiles(List>, Token>, String, byte[]) |
org.apache.hadoop.hbase.mapreduce.TableInputFormatBase.calculateRebalancedSplits(List, JobContext, long) |
org.apache.hadoop.hbase.ipc.RpcServerInterface.call(BlockingService, Descriptors.MethodDescriptor, Message, CellScanner, long, MonitoredRPCHandler)
As of release 1.3, this will be removed in HBase 3.0
|
org.apache.hadoop.hbase.security.visibility.VisibilityClient.clearAuths(Configuration, String[], String)
|
org.apache.hadoop.hbase.client.HConnection.clearCaches(ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.clearRegionCache()
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.clearRegionCache(byte[]) |
org.apache.hadoop.hbase.client.HConnection.clearRegionCache(TableName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HBaseAdmin.compact(byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.compact(byte[], byte[])
|
org.apache.hadoop.hbase.regionserver.Store.compact(CompactionContext, ThroughputController)
see compact(CompactionContext, ThroughputController, User)
|
org.apache.hadoop.hbase.client.HBaseAdmin.compact(String)
|
org.apache.hadoop.hbase.client.HBaseAdmin.compact(String, String)
|
org.apache.hadoop.hbase.KeyValue.RawBytesComparator.compareFlatKey(byte[], int, int, byte[], int, int)
Since 0.99.2.
|
org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.configureIncrementalLoad(Job, HTable)
|
org.apache.hadoop.hbase.HTableDescriptor.convert() |
org.apache.hadoop.hbase.HColumnDescriptor.convert() |
org.apache.hadoop.hbase.ClusterStatus.convert() |
org.apache.hadoop.hbase.ClusterStatus.convert(ClusterStatusProtos.ClusterStatus) |
org.apache.hadoop.hbase.HColumnDescriptor.convert(HBaseProtos.ColumnFamilySchema) |
org.apache.hadoop.hbase.HTableDescriptor.convert(HBaseProtos.TableSchema) |
org.apache.hadoop.hbase.ProcedureInfo.convert(ProcedureProtos.Procedure) |
org.apache.hadoop.hbase.ProcedureInfo.convertToProcedureProto(ProcedureInfo) |
org.apache.hadoop.hbase.mapreduce.CellCreator.create(byte[], int, int, byte[], int, int, byte[], int, int, long, byte[], int, int, String) |
org.apache.hadoop.hbase.regionserver.RpcSchedulerFactory.create(Configuration, PriorityFunction)
since 1.0.0.
|
org.apache.hadoop.hbase.regionserver.SimpleRpcSchedulerFactory.create(Configuration, PriorityFunction)
since 1.0.0.
|
org.apache.hadoop.hbase.regionserver.FifoRpcSchedulerFactory.create(Configuration, PriorityFunction) |
org.apache.hadoop.hbase.client.Result.createCompleteResult(List) |
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(Configuration) |
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(Configuration, ExecutorService) |
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(Configuration, ExecutorService, User) |
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(Configuration, User) |
org.apache.hadoop.hbase.KeyValue.createFirstOnRow(byte[])
|
org.apache.hadoop.hbase.KeyValue.createFirstOnRow(byte[], byte[], byte[])
|
org.apache.hadoop.hbase.KeyValue.createFirstOnRow(byte[], int, int, byte[], int, int, byte[], int, int)
|
org.apache.hadoop.hbase.client.ConnectionUtils.createShortCircuitHConnection(Connection, ServerName, AdminProtos.AdminService.BlockingInterface, ClientProtos.ClientService.BlockingInterface) |
org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteException)
Use RemoteException.unwrapRemoteException() instead.
In fact we should look into deprecating this whole class - St.Ack 2010929
|
org.apache.hadoop.hbase.client.HConnectionManager.deleteAllConnections()
kept for backward compatibility, but the behavior is broken. HBASE-8983
|
org.apache.hadoop.hbase.client.HConnectionManager.deleteAllConnections(boolean) |
org.apache.hadoop.hbase.client.HConnection.deleteCachedRegionLocation(HRegionLocation)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.Delete.deleteColumn(byte[], byte[])
|
org.apache.hadoop.hbase.client.Delete.deleteColumn(byte[], byte[], long)
|
org.apache.hadoop.hbase.client.Delete.deleteColumns(byte[], byte[])
|
org.apache.hadoop.hbase.client.Delete.deleteColumns(byte[], byte[], long)
|
org.apache.hadoop.hbase.client.HConnectionManager.deleteConnection(Configuration) |
org.apache.hadoop.hbase.client.Delete.deleteFamily(byte[])
|
org.apache.hadoop.hbase.client.Delete.deleteFamily(byte[], long)
|
org.apache.hadoop.hbase.client.Delete.deleteFamilyVersion(byte[], long)
|
org.apache.hadoop.hbase.client.HConnectionManager.deleteStaleConnection(HConnection) |
org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.doBulkLoad(Path, HTable)
|
org.apache.hadoop.hbase.KeyValueUtil.ensureKeyValue(Cell)
without any replacement.
|
org.apache.hadoop.hbase.KeyValueUtil.ensureKeyValues(List) | |
org.apache.hadoop.hbase.KeyValueUtil.ensureKeyValueTypeForMR(Cell)
without any replacement.
|
org.apache.hadoop.hbase.CellUtil.estimatedHeapSizeOfWithoutTags(Cell) |
org.apache.hadoop.hbase.CellUtil.estimatedSizeOf(Cell)
please use estimatedSerializedSizeOf(Cell)
|
org.apache.hadoop.hbase.client.HConnectionManager.execute(HConnectable)
Internal method, do not use thru HConnectionManager.
|
org.apache.hadoop.hbase.regionserver.RegionMergeTransaction.execute(Server, RegionServerServices)
use #execute(Server, RegionServerServices, User)
|
org.apache.hadoop.hbase.regionserver.SplitTransaction.execute(Server, RegionServerServices)
use #execute(Server, RegionServerServices, User)
|
org.apache.hadoop.hbase.client.HTableWrapper.exists(List) |
org.apache.hadoop.hbase.client.HTableInterface.exists(List)
|
org.apache.hadoop.hbase.client.HTable.exists(List) |
org.apache.hadoop.hbase.rest.client.RemoteHTable.exists(List) |
org.apache.hadoop.hbase.client.HBaseAdmin.flush(byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.flush(String)
|
org.apache.hadoop.hbase.client.HTableInterface.flushCommits()
|
org.apache.hadoop.hbase.client.HConnection.getAdmin(ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.getAdmin(ServerName, boolean)
You can pass master flag but nothing special is done.
|
org.apache.hadoop.hbase.client.HTable.getAllRegionLocations()
|
org.apache.hadoop.hbase.security.visibility.DefaultVisibilityLabelServiceImpl.getAuths(byte[], boolean) |
org.apache.hadoop.hbase.security.visibility.VisibilityLabelService.getAuths(byte[], boolean)
|
org.apache.hadoop.hbase.security.visibility.VisibilityClient.getAuths(Configuration, String)
|
org.apache.hadoop.hbase.KeyValue.getBuffer()
Since 0.98.0. Use Cell Interface instead. Do not presume single backing buffer.
|
org.apache.hadoop.hbase.zookeeper.ZKUtil.getChildDataAndWatchForNewChildren(ZooKeeperWatcher, String)
Unused
|
org.apache.hadoop.hbase.client.HConnection.getClient(ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.Result.getColumn(byte[], byte[])
|
org.apache.hadoop.hbase.client.Result.getColumnLatest(byte[], byte[])
|
org.apache.hadoop.hbase.client.Result.getColumnLatest(byte[], int, int, byte[], int, int)
|
org.apache.hadoop.hbase.client.HBaseAdmin.getCompactionState(byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.getCompactionState(String)
|
org.apache.hadoop.hbase.client.Admin.getCompactionStateForRegion(byte[]) |
org.apache.hadoop.hbase.client.HTable.getConnection()
This method will be changed from public to package protected.
|
org.apache.hadoop.hbase.client.HConnectionManager.getConnection(Configuration) |
org.apache.hadoop.hbase.client.HConnection.getCurrentNrHRS()
This method will be changed from public to package protected.
|
org.apache.hadoop.hbase.HColumnDescriptor.getDataBlockEncodingOnDisk()
|
org.apache.hadoop.hbase.HRegionInfo.getDaughterRegions(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.wal.WAL.getEarliestMemstoreSeqNum(byte[])
|
org.apache.hadoop.hbase.client.HTable.getEndKeys()
|
org.apache.hadoop.hbase.TagRewriteCell.getFamily() |
org.apache.hadoop.hbase.KeyValue.getFamily() |
org.apache.hadoop.hbase.Cell.getFamily()
|
org.apache.hadoop.hbase.client.Mutation.getFamilyMap()
|
org.apache.hadoop.hbase.ProcedureInfo.getForeignExceptionMessage() |
org.apache.hadoop.hbase.HRegionInfo.getHRegionInfo(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.client.MetaScanner.getHRegionInfo(Result)
|
org.apache.hadoop.hbase.HRegionInfo.getHRegionInfo(Result, byte[])
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.HRegionInfo.getHRegionInfoAndServerName(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.mapreduce.TableInputFormatBase.getHTable()
|
org.apache.hadoop.hbase.mapred.TableInputFormatBase.getHTable()
|
org.apache.hadoop.hbase.client.HConnection.getHTableDescriptor(byte[]) |
org.apache.hadoop.hbase.client.HConnection.getHTableDescriptor(TableName) |
org.apache.hadoop.hbase.client.HConnection.getHTableDescriptors(List) |
org.apache.hadoop.hbase.client.HConnection.getHTableDescriptorsByTableName(List)
|
org.apache.hadoop.hbase.client.ClusterConnection.getKeepAliveMasterService() |
org.apache.hadoop.hbase.client.HConnection.getKeepAliveMasterService()
Since 0.96.0
|
org.apache.hadoop.hbase.client.HConnection.getMaster()
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.HRegionInfo.getMergeRegions(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.TagRewriteCell.getMvccVersion() |
org.apache.hadoop.hbase.Cell.getMvccVersion()
as of 1.0, use Cell.getSequenceId()
Internal use only. A region-specific sequence ID given to each operation. It always exists for
cells in the memstore but is not retained forever. It may survive several flushes, but
generally becomes irrelevant after the cell's row is no longer involved in any operations that
require strict consistency.
|
org.apache.hadoop.hbase.HTableDescriptor.getName()
|
org.apache.hadoop.hbase.filter.FilterWrapper.getNextKeyHint(KeyValue) |
org.apache.hadoop.hbase.filter.FilterList.getNextKeyHint(KeyValue) |
org.apache.hadoop.hbase.filter.Filter.getNextKeyHint(KeyValue) |
org.apache.hadoop.hbase.filter.FilterBase.getNextKeyHint(KeyValue) |
org.apache.hadoop.hbase.client.HConnection.getNonceGenerator()
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.regionserver.Region.getOldestSeqIdOfStore(byte[])
Since version 1.2.0. Exposes too much about our internals; shutting it down.
Do not use.
|
org.apache.hadoop.hbase.HTableDescriptor.getOwnerString()
since 0.94.1
|
org.apache.hadoop.hbase.http.InfoServer.getPort() |
org.apache.hadoop.hbase.http.HttpServer.getPort() |
org.apache.hadoop.hbase.TagRewriteCell.getQualifier() |
org.apache.hadoop.hbase.KeyValue.getQualifier() |
org.apache.hadoop.hbase.Cell.getQualifier()
|
org.apache.hadoop.hbase.MetaTableAccessor.getRegion(Connection, byte[])
|
org.apache.hadoop.hbase.client.HConnection.getRegionCachePrefetch(byte[])
always return false since 0.99
|
org.apache.hadoop.hbase.client.HTable.getRegionCachePrefetch(byte[])
always return false since 0.99
|
org.apache.hadoop.hbase.client.HTable.getRegionCachePrefetch(Configuration, byte[])
always return false since 0.99
|
org.apache.hadoop.hbase.client.HTable.getRegionCachePrefetch(Configuration, TableName)
always return false since 0.99
|
org.apache.hadoop.hbase.client.HConnection.getRegionCachePrefetch(TableName)
always return false since 0.99
|
org.apache.hadoop.hbase.client.HTable.getRegionCachePrefetch(TableName)
always return false since 0.99
|
org.apache.hadoop.hbase.MetaTableAccessor.getRegionCount(Configuration, String) |
org.apache.hadoop.hbase.client.HTable.getRegionLocation(byte[])
|
org.apache.hadoop.hbase.client.HTable.getRegionLocation(byte[], boolean)
|
org.apache.hadoop.hbase.client.HConnection.getRegionLocation(byte[], byte[], boolean) |
org.apache.hadoop.hbase.client.HTable.getRegionLocation(String)
|
org.apache.hadoop.hbase.client.HConnection.getRegionLocation(TableName, byte[], boolean)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HTable.getRegionLocations()
|
org.apache.hadoop.hbase.client.HTable.getRegionsInRange(byte[], byte[])
This is no longer a public API
|
org.apache.hadoop.hbase.client.HTable.getRegionsInRange(byte[], byte[], boolean)
This is no longer a public API
|
org.apache.hadoop.hbase.TagRewriteCell.getRow() |
org.apache.hadoop.hbase.KeyValue.getRow() |
org.apache.hadoop.hbase.Cell.getRow()
|
org.apache.hadoop.hbase.TableName.getRowComparator()
The comparator is an internal property of the table. Should
not have been exposed here
|
org.apache.hadoop.hbase.client.HTableWrapper.getRowOrBefore(byte[], byte[]) |
org.apache.hadoop.hbase.client.HTableInterface.getRowOrBefore(byte[], byte[])
As of version 0.92 this method is deprecated without
replacement. Since version 0.96+, you can use reversed scan.
getRowOrBefore is used internally to find entries in hbase:meta and makes
various assumptions about the table (which are true for hbase:meta but not
in general) to be efficient.
|
org.apache.hadoop.hbase.client.HTable.getRowOrBefore(byte[], byte[])
Use reversed scan instead.
|
org.apache.hadoop.hbase.thrift.ThriftServerRunner.HBaseHandler.getRowOrBefore(ByteBuffer, ByteBuffer, ByteBuffer) |
org.apache.hadoop.hbase.protobuf.ProtobufUtil.getRowOrBefore(ClientProtos.ClientService.BlockingInterface, byte[], byte[], byte[])
since 0.99 - use reversed scanner instead.
|
org.apache.hadoop.hbase.client.HTableWrapper.getRpcTimeout() |
org.apache.hadoop.hbase.client.Table.getRpcTimeout()
Use getReadRpcTimeout or getWriteRpcTimeout instead
|
org.apache.hadoop.hbase.client.HTable.getRpcTimeout() |
org.apache.hadoop.hbase.rest.client.RemoteHTable.getRpcTimeout() |
org.apache.hadoop.hbase.client.Scan.getScanMetrics()
|
org.apache.hadoop.hbase.regionserver.StoreFile.Reader.getScanner(boolean, boolean)
Do not write further code which depends on this call. Instead
use getStoreFileScanner() which uses the StoreFileScanner class/interface
which is the preferred way to scan a store with higher level concepts.
|
org.apache.hadoop.hbase.regionserver.StoreFile.Reader.getScanner(boolean, boolean, boolean)
Do not write further code which depends on this call. Instead
use getStoreFileScanner() which uses the StoreFileScanner class/interface
which is the preferred way to scan a store with higher level concepts.
|
org.apache.hadoop.hbase.client.HTable.getScannerCaching()
|
org.apache.hadoop.hbase.HRegionInfo.getSeqNumDuringOpen(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.ClusterStatus.getServerInfo()
|
org.apache.hadoop.hbase.HRegionInfo.getServerName(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.KeyValue.KVComparator.getShortMidpointKey(byte[], byte[])
Since 0.99.2;
|
org.apache.hadoop.hbase.io.ImmutableBytesWritable.getSize()
|
org.apache.hadoop.hbase.snapshot.HBaseSnapshotException.getSnapshotDescription()
since 1.3.0, will be removed in 3.0.0
|
org.apache.hadoop.hbase.mapreduce.TableInputFormatBase.getSplitKey(byte[], byte[], boolean) |
org.apache.hadoop.hbase.client.HTable.getStartEndKeys()
|
org.apache.hadoop.hbase.client.HTable.getStartKeys()
|
org.apache.hadoop.hbase.client.Result.getStats() |
org.apache.hadoop.hbase.regionserver.HStore.getStoreHomedir(Path, HRegionInfo, byte[]) |
org.apache.hadoop.hbase.regionserver.HStore.getStoreHomedir(Path, String, byte[]) |
org.apache.hadoop.hbase.HTableDescriptor.getTableDir(Path, byte[]) |
org.apache.hadoop.hbase.HRegionInfo.getTableName()
|
org.apache.hadoop.hbase.client.ClientScanner.getTableName()
|
org.apache.hadoop.hbase.client.HTableInterface.getTableName()
|
org.apache.hadoop.hbase.HRegionInfo.getTableName(byte[])
|
org.apache.hadoop.hbase.client.HConnection.getTableNames()
|
org.apache.hadoop.hbase.client.HBaseAdmin.getTableNames()
|
org.apache.hadoop.hbase.client.HBaseAdmin.getTableNames(Pattern)
|
org.apache.hadoop.hbase.client.HBaseAdmin.getTableNames(String)
|
org.apache.hadoop.hbase.KeyValue.getType() |
org.apache.hadoop.hbase.security.access.AccessControlClient.getUserPermissions(Configuration, String)
|
org.apache.hadoop.hbase.TagRewriteCell.getValue() |
org.apache.hadoop.hbase.KeyValue.getValue() |
org.apache.hadoop.hbase.Cell.getValue()
|
org.apache.hadoop.hbase.HRegionInfo.getVersion()
HRI is no longer a VersionedWritable
|
org.apache.hadoop.hbase.client.BufferedMutatorImpl.getWriteBuffer()
|
org.apache.hadoop.hbase.client.HTable.getWriteBuffer()
since 0.96. This is an internal buffer that should not be read nor write.
|
org.apache.hadoop.hbase.client.Table.getWriteBufferSize()
|
org.apache.hadoop.hbase.client.HTableInterface.getWriteBufferSize()
|
org.apache.hadoop.hbase.client.Mutation.getWriteToWAL()
|
org.apache.hadoop.hbase.util.FSUtils.getWrongWALRegionDir(Configuration, TableName, String)
For compatibility, will be removed in 4.0.0.
|
org.apache.hadoop.hbase.security.access.AccessControlClient.grant(Configuration, String, Permission.Action...)
|
org.apache.hadoop.hbase.security.access.AccessControlClient.grant(Configuration, String, String, Permission.Action...)
|
org.apache.hadoop.hbase.security.access.AccessControlClient.grant(Configuration, TableName, String, byte[], byte[], Permission.Action...)
|
org.apache.hadoop.hbase.security.visibility.DefaultVisibilityLabelServiceImpl.havingSystemAuth(byte[]) |
org.apache.hadoop.hbase.security.visibility.VisibilityLabelService.havingSystemAuth(byte[])
|
org.apache.hadoop.hbase.KeyValue.heapSizeWithoutTags() |
org.apache.hadoop.hbase.client.HTableInterface.incrementColumnValue(byte[], byte[], byte[], long, boolean)
|
org.apache.hadoop.hbase.client.HTable.incrementColumnValue(byte[], byte[], byte[], long, boolean)
|
org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil.initCredentialsForCluster(Job, String)
|
org.apache.hadoop.hbase.regionserver.HRegion.initialize()
use HRegion.createHRegion() or HRegion.openHRegion()
|
org.apache.hadoop.hbase.ipc.BalancedQueueRpcExecutor.initializeQueues(int, Class extends BlockingQueue>, Object...) |
org.apache.hadoop.hbase.security.access.AccessControlClient.isAccessControllerRunning(Configuration)
|
org.apache.hadoop.hbase.client.HTableInterface.isAutoFlush()
|
org.apache.hadoop.hbase.KeyValue.iscreate(InputStream)
|
org.apache.hadoop.hbase.client.ClusterConnection.isDeadServer(ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.isDeadServer(ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.KeyValue.isDelete() |
org.apache.hadoop.hbase.master.cleaner.BaseLogCleanerDelegate.isLogDeletable(FileStatus) |
org.apache.hadoop.hbase.client.ClusterConnection.isMasterRunning()
this has been deprecated without a replacement
|
org.apache.hadoop.hbase.client.HConnection.isMasterRunning()
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HBaseAdmin.isMasterRunning()
this has been deprecated without a replacement
|
org.apache.hadoop.hbase.client.Result.isPartial()
|
org.apache.hadoop.hbase.client.Admin.isSnapshotFinished(HBaseProtos.SnapshotDescription) |
org.apache.hadoop.hbase.security.Superusers.isSuperUser(String)
|
org.apache.hadoop.hbase.client.HConnection.isTableAvailable(byte[]) |
org.apache.hadoop.hbase.client.HConnection.isTableAvailable(byte[], byte[][]) |
org.apache.hadoop.hbase.client.HConnection.isTableAvailable(TableName, byte[][])
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.isTableDisabled(byte[]) |
org.apache.hadoop.hbase.client.HConnection.isTableEnabled(byte[]) |
org.apache.hadoop.hbase.client.HTable.isTableEnabled(byte[])
|
org.apache.hadoop.hbase.client.HTable.isTableEnabled(Configuration, byte[])
|
org.apache.hadoop.hbase.client.HTable.isTableEnabled(Configuration, String)
|
org.apache.hadoop.hbase.client.HTable.isTableEnabled(Configuration, TableName)
|
org.apache.hadoop.hbase.client.HTable.isTableEnabled(String)
|
org.apache.hadoop.hbase.client.HTable.isTableEnabled(TableName)
|
org.apache.hadoop.hbase.client.Result.list()
|
org.apache.hadoop.hbase.security.visibility.VisibilityClient.listLabels(Configuration, String)
|
org.apache.hadoop.hbase.client.replication.ReplicationAdmin.listPeers()
|
org.apache.hadoop.hbase.client.Admin.listSnapshots() |
org.apache.hadoop.hbase.client.Admin.listSnapshots(Pattern) |
org.apache.hadoop.hbase.client.Admin.listSnapshots(String) |
org.apache.hadoop.hbase.client.HConnection.listTableNames()
|
org.apache.hadoop.hbase.client.HConnection.listTables()
|
org.apache.hadoop.hbase.client.Admin.listTableSnapshots(Pattern, Pattern) |
org.apache.hadoop.hbase.client.Admin.listTableSnapshots(String, String) |
org.apache.hadoop.hbase.client.HConnection.locateRegion(byte[])
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.locateRegion(byte[], byte[]) |
org.apache.hadoop.hbase.client.HConnection.locateRegion(TableName, byte[])
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.locateRegions(byte[]) |
org.apache.hadoop.hbase.client.HConnection.locateRegions(byte[], boolean, boolean) |
org.apache.hadoop.hbase.client.HConnection.locateRegions(TableName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.locateRegions(TableName, boolean, boolean)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HBaseAdmin.majorCompact(byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.majorCompact(byte[], byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.majorCompact(String)
|
org.apache.hadoop.hbase.client.HBaseAdmin.majorCompact(String, String)
|
org.apache.hadoop.hbase.security.token.TokenUtil.obtainAndCacheToken(Configuration, UserGroupInformation)
|
org.apache.hadoop.hbase.security.User.obtainAuthTokenForJob(Configuration, Job)
Use TokenUtil.obtainAuthTokenForJob(Connection,User,Job)
instead.
|
org.apache.hadoop.hbase.security.User.obtainAuthTokenForJob(JobConf)
Use TokenUtil.obtainAuthTokenForJob(Connection,JobConf,User)
instead.
|
org.apache.hadoop.hbase.security.token.TokenUtil.obtainToken(Configuration)
|
org.apache.hadoop.hbase.security.token.TokenUtil.obtainTokenForJob(Configuration, UserGroupInformation, Job)
|
org.apache.hadoop.hbase.security.token.TokenUtil.obtainTokenForJob(JobConf, UserGroupInformation)
|
org.apache.hadoop.hbase.KeyValue.oswrite(KeyValue, OutputStream)
|
org.apache.hadoop.hbase.zookeeper.ZKConfig.parseZooCfg(Configuration, InputStream)
in 0.96 onwards. HBase will no longer rely on zoo.cfg
availability.
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.postCompact(ObserverContext, Store, StoreFile)
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.postCompactSelection(ObserverContext, Store, ImmutableList)
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.postFlush(ObserverContext)
|
org.apache.hadoop.hbase.master.MasterCoprocessorHost.postGetTableDescriptors(List) |
org.apache.hadoop.hbase.coprocessor.MasterObserver.postGetTableDescriptors(ObserverContext, List)
Use postGetTableDescriptors with regex instead.
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.postIncrementColumnValue(ObserverContext, byte[], byte[], byte[], long, boolean, long)
This hook is no longer called by the RegionServer
|
org.apache.hadoop.hbase.regionserver.RegionServerServices.postOpenDeployTasks(Region)
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.postSplit(ObserverContext, Region, Region)
Use postCompleteSplit() instead
|
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.postWALRestore(HRegionInfo, HLogKey, WALEdit)
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.postWALRestore(ObserverContext, HRegionInfo, HLogKey, WALEdit)
|
org.apache.hadoop.hbase.coprocessor.WALObserver.postWALWrite(ObserverContext, HRegionInfo, HLogKey, WALEdit)
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.preCompact(ObserverContext, Store, InternalScanner, ScanType)
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.preCompactScannerOpen(ObserverContext, Store, List extends KeyValueScanner>, ScanType, long, InternalScanner)
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.preCompactScannerOpen(ObserverContext, Store, List extends KeyValueScanner>, ScanType, long, InternalScanner, CompactionRequest)
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.preCompactSelection(ObserverContext, Store, List)
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.preFlush(ObserverContext)
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.preFlushScannerOpen(ObserverContext, Store, KeyValueScanner, InternalScanner)
|
org.apache.hadoop.hbase.master.MasterCoprocessorHost.preGetTableDescriptors(List, List) |
org.apache.hadoop.hbase.coprocessor.MasterObserver.preGetTableDescriptors(ObserverContext, List, List)
Use preGetTableDescriptors with regex instead.
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.preIncrementColumnValue(ObserverContext, byte[], byte[], byte[], long, boolean)
This hook is no longer called by the RegionServer
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.preSplit(ObserverContext)
Use preSplit(
final ObserverContext<RegionCoprocessorEnvironment> c, byte[] splitRow)
|
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preWALRestore(HRegionInfo, HLogKey, WALEdit)
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.preWALRestore(ObserverContext, HRegionInfo, HLogKey, WALEdit)
|
org.apache.hadoop.hbase.coprocessor.WALObserver.preWALWrite(ObserverContext, HRegionInfo, HLogKey, WALEdit)
|
org.apache.hadoop.hbase.client.HConnection.processBatch(List extends Row>, byte[], ExecutorService, Object[]) |
org.apache.hadoop.hbase.client.HConnection.processBatch(List extends Row>, TableName, ExecutorService, Object[])
|
org.apache.hadoop.hbase.client.HConnection.processBatchCallback(List extends Row>, byte[], ExecutorService, Object[], Batch.Callback) |
org.apache.hadoop.hbase.client.HConnection.processBatchCallback(List extends Row>, TableName, ExecutorService, Object[], Batch.Callback)
|
org.apache.hadoop.hbase.client.HTableMultiplexer.put(byte[], List) |
org.apache.hadoop.hbase.client.HTableMultiplexer.put(byte[], Put) |
org.apache.hadoop.hbase.client.HTableMultiplexer.put(byte[], Put, int) |
org.apache.hadoop.hbase.client.HTablePool.putTable(HTableInterface) |
org.apache.hadoop.hbase.client.Result.raw()
|
org.apache.hadoop.hbase.HRegionInfo.readFields(DataInput)
Use protobuf deserialization instead.
|
org.apache.hadoop.hbase.HTableDescriptor.readFields(DataInput)
|
org.apache.hadoop.hbase.HColumnDescriptor.readFields(DataInput)
|
org.apache.hadoop.hbase.io.Reference.readFields(DataInput)
Writables are going away. Use the pb serialization methods instead.
Remove in a release after 0.96 goes out. This is here only to migrate
old Reference files written with Writables before 0.96.
|
org.apache.hadoop.hbase.util.Bytes.readVLong(byte[], int)
|
org.apache.hadoop.hbase.client.HConnection.relocateRegion(byte[], byte[]) |
org.apache.hadoop.hbase.client.HConnection.relocateRegion(TableName, byte[])
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.regionserver.RegionServerServices.reportRegionStateTransition(RegionServerStatusProtos.RegionStateTransition.TransitionCode, HRegionInfo...)
|
org.apache.hadoop.hbase.regionserver.RegionServerServices.reportRegionStateTransition(RegionServerStatusProtos.RegionStateTransition.TransitionCode, long, HRegionInfo...)
|
org.apache.hadoop.hbase.regionserver.Store.requestCompaction(int, CompactionRequest)
see requestCompaction(int, CompactionRequest, User)
|
org.apache.hadoop.hbase.io.hfile.HFileScanner.reseekTo(byte[]) |
org.apache.hadoop.hbase.io.hfile.HFileScanner.reseekTo(byte[], int, int) |
org.apache.hadoop.hbase.master.snapshot.SnapshotManager.restoreSnapshot(HBaseProtos.SnapshotDescription) |
org.apache.hadoop.hbase.mapreduce.TableInputFormatBase.reverseDNS(InetAddress)
mistakenly made public in 0.98.7. scope will change to package-private
|
org.apache.hadoop.hbase.security.access.AccessControlClient.revoke(Configuration, String, Permission.Action...)
|
org.apache.hadoop.hbase.security.access.AccessControlClient.revoke(Configuration, String, String, Permission.Action...)
|
org.apache.hadoop.hbase.security.access.AccessControlClient.revoke(Configuration, TableName, String, byte[], byte[], Permission.Action...)
|
org.apache.hadoop.hbase.regionserver.RegionMergeTransaction.rollback(Server, RegionServerServices)
use #rollback(Server, RegionServerServices, User)
|
org.apache.hadoop.hbase.regionserver.SplitTransaction.rollback(Server, RegionServerServices)
use #rollback(Server, RegionServerServices, User)
|
org.apache.hadoop.hbase.client.HBaseAdmin.rollHLogWriter(String)
|
org.apache.hadoop.hbase.io.hfile.HFileScanner.seekBefore(byte[]) |
org.apache.hadoop.hbase.io.hfile.HFileScanner.seekBefore(byte[], int, int) |
org.apache.hadoop.hbase.io.hfile.HFileScanner.seekTo(byte[]) |
org.apache.hadoop.hbase.io.hfile.HFileScanner.seekTo(byte[], int, int) |
org.apache.hadoop.hbase.io.encoding.DataBlockEncoder.EncodedSeeker.seekToKeyInBlock(byte[], int, int, boolean) |
org.apache.hadoop.hbase.security.visibility.VisibilityClient.setAuths(Configuration, String[], String)
|
org.apache.hadoop.hbase.client.HTableInterface.setAutoFlush(boolean)
in 0.96. When called with setAutoFlush(false), this function also
set clearBufferOnFail to true, which is unexpected but kept for historical reasons.
Replace it with setAutoFlush(false, false) if this is exactly what you want, though
this is the method you want for most cases.
|
org.apache.hadoop.hbase.client.HTable.setAutoFlush(boolean) |
org.apache.hadoop.hbase.client.HTableInterface.setAutoFlush(boolean, boolean)
|
org.apache.hadoop.hbase.client.HTableInterface.setAutoFlushTo(boolean)
|
org.apache.hadoop.hbase.http.HttpServer.Builder.setBindAddress(String) |
org.apache.hadoop.hbase.HColumnDescriptor.setEncodeOnDisk(boolean)
As of release 0.98
(HBASE-9870).
This will be removed in HBase 2.0.0. This method does nothing now.
|
org.apache.hadoop.hbase.client.Mutation.setFamilyMap(NavigableMap>)
|
org.apache.hadoop.hbase.client.Put.setFamilyMap(NavigableMap>) |
org.apache.hadoop.hbase.client.Delete.setFamilyMap(NavigableMap>) |
org.apache.hadoop.hbase.client.Increment.setFamilyMap(NavigableMap>) |
org.apache.hadoop.hbase.client.Append.setFamilyMap(NavigableMap>) |
org.apache.hadoop.hbase.mapreduce.TableInputFormatBase.setHTable(HTable)
|
org.apache.hadoop.hbase.mapred.TableInputFormatBase.setHTable(HTable)
|
org.apache.hadoop.hbase.mapreduce.TableRecordReader.setHTable(Table)
Use setTable() instead.
|
org.apache.hadoop.hbase.HColumnDescriptor.setKeepDeletedCells(boolean)
|
org.apache.hadoop.hbase.HTableDescriptor.setName(byte[]) |
org.apache.hadoop.hbase.http.HttpServer.Builder.setName(String) |
org.apache.hadoop.hbase.HTableDescriptor.setName(TableName) |
org.apache.hadoop.hbase.HTableDescriptor.setOwner(User)
since 0.94.1
|
org.apache.hadoop.hbase.HTableDescriptor.setOwnerString(String)
since 0.94.1
|
org.apache.hadoop.hbase.client.replication.ReplicationAdmin.setPeerTableCFs(String, String)
|
org.apache.hadoop.hbase.http.HttpServer.Builder.setPort(int) |
org.apache.hadoop.hbase.client.HConnection.setRegionCachePrefetch(byte[], boolean)
does nothing since 0.99
|
org.apache.hadoop.hbase.client.HTable.setRegionCachePrefetch(byte[], boolean)
does nothing since 0.99
|
org.apache.hadoop.hbase.client.HTable.setRegionCachePrefetch(Configuration, byte[], boolean)
does nothing since 0.99
|
org.apache.hadoop.hbase.client.HTable.setRegionCachePrefetch(Configuration, TableName, boolean)
does nothing since 0.99
|
org.apache.hadoop.hbase.client.HConnection.setRegionCachePrefetch(TableName, boolean)
does nothing since since 0.99
|
org.apache.hadoop.hbase.client.HTable.setRegionCachePrefetch(TableName, boolean)
does nothing since 0.99
|
org.apache.hadoop.hbase.client.HTableWrapper.setRpcTimeout(int) |
org.apache.hadoop.hbase.client.Table.setRpcTimeout(int)
Use setReadRpcTimeout or setWriteRpcTimeout instead
|
org.apache.hadoop.hbase.client.HTable.setRpcTimeout(int) |
org.apache.hadoop.hbase.rest.client.RemoteHTable.setRpcTimeout(int) |
org.apache.hadoop.hbase.client.HTable.setScannerCaching(int)
|
org.apache.hadoop.hbase.client.HConnectionManager.setServerSideHConnectionRetries(Configuration, String, Log)
Internal method, do not use.
|
org.apache.hadoop.hbase.client.Scan.setStartRow(byte[])
use Scan.withStartRow(byte[]) instead. This method may change the inclusive of
the stop row to keep compatible with the old behavior.
|
org.apache.hadoop.hbase.client.Result.setStatistics(ClientProtos.RegionLoadStats) |
org.apache.hadoop.hbase.client.Scan.setStopRow(byte[])
use Scan.withStartRow(byte[]) instead. This method may change the inclusive of
the stop row to keep compatible with the old behavior.
|
org.apache.hadoop.hbase.quotas.QuotaSettings.setupSetQuotaRequest(MasterProtos.SetQuotaRequest.Builder)
Removed in HBase 2.0+ as a part of removing protobuf from our API
|
org.apache.hadoop.hbase.client.BufferedMutatorImpl.setWriteBufferSize(long)
|
org.apache.hadoop.hbase.client.Table.setWriteBufferSize(long)
|
org.apache.hadoop.hbase.client.HTableInterface.setWriteBufferSize(long)
|
org.apache.hadoop.hbase.client.Mutation.setWriteToWAL(boolean)
|
org.apache.hadoop.hbase.client.Put.setWriteToWAL(boolean) |
org.apache.hadoop.hbase.client.Delete.setWriteToWAL(boolean) |
org.apache.hadoop.hbase.client.Increment.setWriteToWAL(boolean) |
org.apache.hadoop.hbase.client.Append.setWriteToWAL(boolean) |
org.apache.hadoop.hbase.HColumnDescriptor.shouldCacheBloomsOnWrite()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldCacheDataInL1()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldCacheDataOnWrite()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldCacheIndexesOnWrite()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldCompressTags()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldEvictBlocksOnClose()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldPrefetchBlocksOnOpen()
|
org.apache.hadoop.hbase.regionserver.RegionSplitPolicy.skipStoreFileRangeCheck()
|
org.apache.hadoop.hbase.client.Admin.snapshot(HBaseProtos.SnapshotDescription) |
org.apache.hadoop.hbase.client.Admin.snapshot(String, TableName, HBaseProtos.SnapshotDescription.Type) |
org.apache.hadoop.hbase.client.HBaseAdmin.split(byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.split(byte[], byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.split(String)
|
org.apache.hadoop.hbase.client.HBaseAdmin.split(String, String)
|
org.apache.hadoop.hbase.regionserver.RegionMergeTransactionImpl.stepsAfterPONR(Server, RegionServerServices, HRegion) |
org.apache.hadoop.hbase.regionserver.SplitTransactionImpl.stepsAfterPONR(Server, RegionServerServices, PairOfSameType) |
org.apache.hadoop.hbase.client.Admin.takeSnapshotAsync(HBaseProtos.SnapshotDescription) |
org.apache.hadoop.hbase.protobuf.ProtobufUtil.toBulkLoadDescriptor(TableName, ByteString, Map>, Map, long)
use toBulkLoadDescriptor(TableName tableName, ByteString encodedRegionName,
Map> storeFiles, Map storeFilesSize, long bulkloadSeqId,
List clusterIds) instead.
|
org.apache.hadoop.hbase.filter.FilterList.toString(int) |
org.apache.hadoop.hbase.filter.FilterWrapper.transform(KeyValue) |
org.apache.hadoop.hbase.filter.FilterList.transform(KeyValue) |
org.apache.hadoop.hbase.filter.Filter.transform(KeyValue) |
org.apache.hadoop.hbase.filter.FilterBase.transform(KeyValue) |
org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.tryAtomicRegionLoad(HConnection, byte[], byte[], Collection)
|
org.apache.hadoop.hbase.client.HConnection.updateCachedLocations(byte[], byte[], Object, HRegionLocation) |
org.apache.hadoop.hbase.client.HConnection.updateCachedLocations(TableName, byte[], byte[], Object, ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.updateCachedLocations(TableName, byte[], Object, HRegionLocation) |
org.apache.hadoop.hbase.zookeeper.ZKUtil.updateExistingNodeData(ZooKeeperWatcher, String, byte[], int)
Unused
|
org.apache.hadoop.hbase.HRegionInfo.write(DataOutput)
|
org.apache.hadoop.hbase.HTableDescriptor.write(DataOutput)
Writables are going away.
Use MessageLite.toByteArray() instead.
|
org.apache.hadoop.hbase.HColumnDescriptor.write(DataOutput)
|
org.apache.hadoop.hbase.regionserver.wal.HLogKey.write(DataOutput)
Don't use these Writables methods. Use PB instead.
|