org.apache.hadoop.hbase.client.Put.add(byte[], byte[], byte[])
|
org.apache.hadoop.hbase.client.Put.add(byte[], byte[], long, byte[])
|
org.apache.hadoop.hbase.client.Put.add(byte[], ByteBuffer, long, ByteBuffer)
|
org.apache.hadoop.hbase.security.visibility.VisibilityClient.addLabel(Configuration, String)
|
org.apache.hadoop.hbase.security.visibility.VisibilityClient.addLabels(Configuration, String[])
|
org.apache.hadoop.hbase.client.replication.ReplicationAdmin.addPeer(String, String)
Use addPeer(String, ReplicationPeerConfig, Map) instead.
|
org.apache.hadoop.hbase.client.replication.ReplicationAdmin.addPeer(String, String, String) |
org.apache.hadoop.hbase.client.Result.addResults(ClientProtos.RegionLoadStats)
|
org.apache.hadoop.hbase.client.MetaScanner.allTableRegions(Configuration, Connection, TableName, boolean)
|
org.apache.hadoop.hbase.client.HTable.batch(List extends Row>)
If any exception is thrown by one of the actions, there is no way to
retrieve the partially executed results. Use HTable.batch(List, Object[]) instead.
|
org.apache.hadoop.hbase.client.Table.batch(List extends Row>)
If any exception is thrown by one of the actions, there is no way to
retrieve the partially executed results. Use Table.batch(List, Object[]) instead.
|
org.apache.hadoop.hbase.client.HTablePool.PooledHTable.batch(List extends Row>)
|
org.apache.hadoop.hbase.client.HTableWrapper.batch(List extends Row>)
|
org.apache.hadoop.hbase.client.HTable.batchCallback(List extends Row>, Batch.Callback)
|
org.apache.hadoop.hbase.client.Table.batchCallback(List extends Row>, Batch.Callback)
|
org.apache.hadoop.hbase.client.HTablePool.PooledHTable.batchCallback(List extends Row>, Batch.Callback)
|
org.apache.hadoop.hbase.client.HTableWrapper.batchCallback(List extends Row>, Batch.Callback)
|
org.apache.hadoop.hbase.quotas.QuotaSettings.buildSetQuotaRequestProto(QuotaSettings)
Removed in HBase 2.0+ as a part of removing protobuf from our API
|
org.apache.hadoop.hbase.security.visibility.VisibilityClient.clearAuths(Configuration, String[], String)
|
org.apache.hadoop.hbase.client.HConnection.clearCaches(ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.clearRegionCache()
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.clearRegionCache(byte[]) |
org.apache.hadoop.hbase.client.HConnection.clearRegionCache(TableName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HBaseAdmin.compact(byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.compact(byte[], byte[])
|
org.apache.hadoop.hbase.regionserver.Store.compact(CompactionContext, CompactionThroughputController)
see compact(CompactionContext, CompactionThroughputController, User)
|
org.apache.hadoop.hbase.client.HBaseAdmin.compact(String)
|
org.apache.hadoop.hbase.client.HBaseAdmin.compact(String, String)
|
org.apache.hadoop.hbase.KeyValue.RawBytesComparator.compareFlatKey(byte[], int, int, byte[], int, int)
Since 0.99.2.
|
org.apache.hadoop.hbase.mapreduce.HFileOutputFormat2.configureIncrementalLoad(Job, HTable)
|
org.apache.hadoop.hbase.mapreduce.CellCreator.create(byte[], int, int, byte[], int, int, byte[], int, int, long, byte[], int, int, String) |
org.apache.hadoop.hbase.regionserver.RpcSchedulerFactory.create(Configuration, PriorityFunction) |
org.apache.hadoop.hbase.regionserver.SimpleRpcSchedulerFactory.create(Configuration, PriorityFunction) |
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(Configuration) |
org.apache.hadoop.hbase.client.ConnectionManager.createConnection(Configuration, boolean) |
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(Configuration, boolean) |
org.apache.hadoop.hbase.client.ConnectionManager.createConnection(Configuration, boolean, ExecutorService, User) |
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(Configuration, boolean, ExecutorService, User) |
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(Configuration, ExecutorService) |
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(Configuration, ExecutorService, User) |
org.apache.hadoop.hbase.client.HConnectionManager.createConnection(Configuration, User) |
org.apache.hadoop.hbase.KeyValue.createFirstOnRow(byte[])
|
org.apache.hadoop.hbase.KeyValue.createFirstOnRow(byte[], byte[], byte[])
|
org.apache.hadoop.hbase.KeyValue.createFirstOnRow(byte[], int, int, byte[], int, int, byte[], int, int)
|
org.apache.hadoop.hbase.client.ConnectionUtils.createShortCircuitHConnection(Connection, ServerName, AdminProtos.AdminService.BlockingInterface, ClientProtos.ClientService.BlockingInterface) |
org.apache.hadoop.hbase.RemoteExceptionHandler.decodeRemoteException(RemoteException)
Use RemoteException.unwrapRemoteException() instead.
In fact we should look into deprecating this whole class - St.Ack 2010929
|
org.apache.hadoop.hbase.client.ConnectionManager.deleteAllConnections()
kept for backward compatibility, but the behavior is broken. HBASE-8983
|
org.apache.hadoop.hbase.client.HConnectionManager.deleteAllConnections()
kept for backward compatibility, but the behavior is broken. HBASE-8983
|
org.apache.hadoop.hbase.client.ConnectionManager.deleteAllConnections(boolean) |
org.apache.hadoop.hbase.client.HConnectionManager.deleteAllConnections(boolean) |
org.apache.hadoop.hbase.client.HConnection.deleteCachedRegionLocation(HRegionLocation)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.Delete.deleteColumn(byte[], byte[])
|
org.apache.hadoop.hbase.client.Delete.deleteColumn(byte[], byte[], long)
|
org.apache.hadoop.hbase.client.Delete.deleteColumns(byte[], byte[])
|
org.apache.hadoop.hbase.client.Delete.deleteColumns(byte[], byte[], long)
|
org.apache.hadoop.hbase.client.ConnectionManager.deleteConnection(Configuration) |
org.apache.hadoop.hbase.client.HConnectionManager.deleteConnection(Configuration) |
org.apache.hadoop.hbase.client.ConnectionManager.deleteConnection(HConnection, boolean) |
org.apache.hadoop.hbase.client.ConnectionManager.deleteConnection(HConnectionKey, boolean) |
org.apache.hadoop.hbase.client.Delete.deleteFamily(byte[])
|
org.apache.hadoop.hbase.client.Delete.deleteFamily(byte[], long)
|
org.apache.hadoop.hbase.client.Delete.deleteFamilyVersion(byte[], long)
|
org.apache.hadoop.hbase.client.ConnectionManager.deleteStaleConnection(HConnection) |
org.apache.hadoop.hbase.client.HConnectionManager.deleteStaleConnection(HConnection) |
org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.doBulkLoad(Path, HTable)
|
org.apache.hadoop.hbase.KeyValueUtil.ensureKeyValue(Cell)
without any replacement.
|
org.apache.hadoop.hbase.KeyValueUtil.ensureKeyValues(List) | |
org.apache.hadoop.hbase.KeyValueUtil.ensureKeyValueTypeForMR(Cell)
without any replacement.
|
org.apache.hadoop.hbase.CellUtil.estimatedHeapSizeOfWithoutTags(Cell) |
org.apache.hadoop.hbase.CellUtil.estimatedSizeOf(Cell)
please use estimatedSerializedSizeOf(Cell)
|
org.apache.hadoop.hbase.client.HConnectionManager.execute(HConnectable)
Internal method, do not use thru HConnectionManager.
|
org.apache.hadoop.hbase.regionserver.RegionMergeTransaction.execute(Server, RegionServerServices)
use #execute(Server, RegionServerServices, User)
|
org.apache.hadoop.hbase.regionserver.SplitTransaction.execute(Server, RegionServerServices)
use #execute(Server, RegionServerServices, User)
|
org.apache.hadoop.hbase.client.HTable.exists(List) |
org.apache.hadoop.hbase.client.HTableInterface.exists(List)
|
org.apache.hadoop.hbase.client.HTableWrapper.exists(List) |
org.apache.hadoop.hbase.rest.client.RemoteHTable.exists(List) |
org.apache.hadoop.hbase.client.HBaseAdmin.flush(byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.flush(String)
|
org.apache.hadoop.hbase.client.HTableInterface.flushCommits()
|
org.apache.hadoop.hbase.client.HConnection.getAdmin(ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.getAdmin(ServerName, boolean)
You can pass master flag but nothing special is done.
|
org.apache.hadoop.hbase.client.HTable.getAllRegionLocations()
|
org.apache.hadoop.hbase.security.visibility.VisibilityLabelService.getAuths(byte[], boolean)
|
org.apache.hadoop.hbase.security.visibility.DefaultVisibilityLabelServiceImpl.getAuths(byte[], boolean) |
org.apache.hadoop.hbase.security.visibility.VisibilityClient.getAuths(Configuration, String)
|
org.apache.hadoop.hbase.KeyValue.getBuffer()
Since 0.98.0. Use Cell Interface instead. Do not presume single backing buffer.
|
org.apache.hadoop.hbase.zookeeper.ZKUtil.getChildDataAndWatchForNewChildren(ZooKeeperWatcher, String)
Unused
|
org.apache.hadoop.hbase.client.HConnection.getClient(ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.Result.getColumn(byte[], byte[])
|
org.apache.hadoop.hbase.client.Result.getColumnLatest(byte[], byte[])
|
org.apache.hadoop.hbase.client.Result.getColumnLatest(byte[], int, int, byte[], int, int)
|
org.apache.hadoop.hbase.client.HBaseAdmin.getCompactionState(byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.getCompactionState(String)
|
org.apache.hadoop.hbase.client.HTable.getConnection()
This method will be changed from public to package protected.
|
org.apache.hadoop.hbase.client.ConnectionManager.getConnection(Configuration) |
org.apache.hadoop.hbase.client.HConnectionManager.getConnection(Configuration) |
org.apache.hadoop.hbase.client.HConnection.getCurrentNrHRS()
This method will be changed from public to package protected.
|
org.apache.hadoop.hbase.HColumnDescriptor.getDataBlockEncodingOnDisk()
|
org.apache.hadoop.hbase.HRegionInfo.getDaughterRegions(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.wal.WAL.getEarliestMemstoreSeqNum(byte[])
|
org.apache.hadoop.hbase.client.HTable.getEndKeys()
|
org.apache.hadoop.hbase.io.encoding.BufferedDataBlockEncoder.SeekerState.getFamily() |
org.apache.hadoop.hbase.io.encoding.BufferedDataBlockEncoder.ClonedSeekerState.getFamily() |
org.apache.hadoop.hbase.KeyValue.getFamily() |
org.apache.hadoop.hbase.Cell.getFamily()
|
org.apache.hadoop.hbase.codec.prefixtree.PrefixTreeSeeker.ClonedPrefixTreeCell.getFamily() |
org.apache.hadoop.hbase.TagRewriteCell.getFamily() |
org.apache.hadoop.hbase.client.Mutation.getFamilyMap()
|
org.apache.hadoop.hbase.HRegionInfo.getHRegionInfo(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.client.MetaScanner.getHRegionInfo(Result)
|
org.apache.hadoop.hbase.HRegionInfo.getHRegionInfo(Result, byte[])
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.HRegionInfo.getHRegionInfoAndServerName(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.mapreduce.TableInputFormatBase.getHTable()
|
org.apache.hadoop.hbase.mapred.TableInputFormatBase.getHTable()
|
org.apache.hadoop.hbase.client.ConnectionManager.HConnectionImplementation.getHTableDescriptor(byte[])
|
org.apache.hadoop.hbase.client.HConnection.getHTableDescriptor(byte[]) |
org.apache.hadoop.hbase.client.ConnectionManager.HConnectionImplementation.getHTableDescriptor(TableName)
|
org.apache.hadoop.hbase.client.HConnection.getHTableDescriptor(TableName) |
org.apache.hadoop.hbase.client.ConnectionManager.HConnectionImplementation.getHTableDescriptors(List)
|
org.apache.hadoop.hbase.client.HConnection.getHTableDescriptors(List) |
org.apache.hadoop.hbase.client.ConnectionManager.HConnectionImplementation.getHTableDescriptorsByTableName(List)
|
org.apache.hadoop.hbase.client.HConnection.getHTableDescriptorsByTableName(List)
|
org.apache.hadoop.hbase.client.HConnection.getKeepAliveMasterService()
Since 0.96.0
|
org.apache.hadoop.hbase.client.ClusterConnection.getKeepAliveMasterService() |
org.apache.hadoop.hbase.client.HTable.getKeysAndRegionsInRange(byte[], byte[], boolean)
This is no longer a public API
|
org.apache.hadoop.hbase.client.HTable.getKeysAndRegionsInRange(byte[], byte[], boolean, boolean)
This is no longer a public API
|
org.apache.hadoop.hbase.client.HConnection.getMaster()
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.HRegionInfo.getMergeRegions(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.io.encoding.BufferedDataBlockEncoder.ClonedSeekerState.getMvccVersion() |
org.apache.hadoop.hbase.Cell.getMvccVersion()
as of 1.0, use Cell.getSequenceId()
Internal use only. A region-specific sequence ID given to each operation. It always exists for
cells in the memstore but is not retained forever. It may survive several flushes, but
generally becomes irrelevant after the cell's row is no longer involved in any operations that
require strict consistency.
|
org.apache.hadoop.hbase.codec.prefixtree.PrefixTreeSeeker.ClonedPrefixTreeCell.getMvccVersion() |
org.apache.hadoop.hbase.TagRewriteCell.getMvccVersion() |
org.apache.hadoop.hbase.HTableDescriptor.getName()
|
org.apache.hadoop.hbase.filter.FilterWrapper.getNextKeyHint(KeyValue) |
org.apache.hadoop.hbase.filter.FilterList.getNextKeyHint(KeyValue) |
org.apache.hadoop.hbase.filter.FilterBase.getNextKeyHint(KeyValue) |
org.apache.hadoop.hbase.filter.Filter.getNextKeyHint(KeyValue) |
org.apache.hadoop.hbase.client.HConnection.getNonceGenerator()
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.regionserver.Region.getOldestSeqIdOfStore(byte[])
Since version 1.2.0. Exposes too much about our internals; shutting it down.
Do not use.
|
org.apache.hadoop.hbase.HTableDescriptor.getOwnerString() |
org.apache.hadoop.hbase.http.HttpServer.getPort() |
org.apache.hadoop.hbase.http.InfoServer.getPort() |
org.apache.hadoop.hbase.io.encoding.BufferedDataBlockEncoder.SeekerState.getQualifier() |
org.apache.hadoop.hbase.io.encoding.BufferedDataBlockEncoder.ClonedSeekerState.getQualifier() |
org.apache.hadoop.hbase.KeyValue.getQualifier() |
org.apache.hadoop.hbase.Cell.getQualifier()
|
org.apache.hadoop.hbase.codec.prefixtree.PrefixTreeSeeker.ClonedPrefixTreeCell.getQualifier() |
org.apache.hadoop.hbase.TagRewriteCell.getQualifier() |
org.apache.hadoop.hbase.MetaTableAccessor.getRegion(Connection, byte[])
|
org.apache.hadoop.hbase.client.HTable.getRegionCachePrefetch(byte[])
always return false since 0.99
|
org.apache.hadoop.hbase.client.ConnectionManager.HConnectionImplementation.getRegionCachePrefetch(byte[]) |
org.apache.hadoop.hbase.client.HConnection.getRegionCachePrefetch(byte[])
always return false since 0.99
|
org.apache.hadoop.hbase.client.HTable.getRegionCachePrefetch(Configuration, byte[])
always return false since 0.99
|
org.apache.hadoop.hbase.client.HTable.getRegionCachePrefetch(Configuration, TableName)
always return false since 0.99
|
org.apache.hadoop.hbase.client.HTable.getRegionCachePrefetch(TableName)
always return false since 0.99
|
org.apache.hadoop.hbase.client.ConnectionManager.HConnectionImplementation.getRegionCachePrefetch(TableName) |
org.apache.hadoop.hbase.client.HConnection.getRegionCachePrefetch(TableName)
always return false since 0.99
|
org.apache.hadoop.hbase.MetaTableAccessor.getRegionCount(Configuration, String) |
org.apache.hadoop.hbase.regionserver.HRegion.getRegionDir(Path, HRegionInfo)
For tests only; to be removed.
|
org.apache.hadoop.hbase.regionserver.HRegion.getRegionDir(Path, String)
For tests only; to be removed.
|
org.apache.hadoop.hbase.client.HTable.getRegionLocation(byte[])
|
org.apache.hadoop.hbase.client.HTable.getRegionLocation(byte[], boolean)
|
org.apache.hadoop.hbase.client.HConnection.getRegionLocation(byte[], byte[], boolean) |
org.apache.hadoop.hbase.client.HTable.getRegionLocation(String)
|
org.apache.hadoop.hbase.client.HConnection.getRegionLocation(TableName, byte[], boolean)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HTable.getRegionLocations()
|
org.apache.hadoop.hbase.client.HTable.getRegionsInRange(byte[], byte[])
This is no longer a public API
|
org.apache.hadoop.hbase.client.HTable.getRegionsInRange(byte[], byte[], boolean)
This is no longer a public API
|
org.apache.hadoop.hbase.io.encoding.BufferedDataBlockEncoder.SeekerState.getRow() |
org.apache.hadoop.hbase.io.encoding.BufferedDataBlockEncoder.ClonedSeekerState.getRow() |
org.apache.hadoop.hbase.KeyValue.getRow() |
org.apache.hadoop.hbase.Cell.getRow()
|
org.apache.hadoop.hbase.codec.prefixtree.PrefixTreeSeeker.ClonedPrefixTreeCell.getRow() |
org.apache.hadoop.hbase.TagRewriteCell.getRow() |
org.apache.hadoop.hbase.TableName.getRowComparator()
The comparator is an internal property of the table. Should
not have been exposed here
|
org.apache.hadoop.hbase.client.HTable.getRowOrBefore(byte[], byte[])
Use reversed scan instead.
|
org.apache.hadoop.hbase.client.HTablePool.PooledHTable.getRowOrBefore(byte[], byte[]) |
org.apache.hadoop.hbase.client.HTableInterface.getRowOrBefore(byte[], byte[])
As of version 0.92 this method is deprecated without
replacement. Since version 0.96+, you can use reversed scan.
getRowOrBefore is used internally to find entries in hbase:meta and makes
various assumptions about the table (which are true for hbase:meta but not
in general) to be efficient.
|
org.apache.hadoop.hbase.client.HTableWrapper.getRowOrBefore(byte[], byte[]) |
org.apache.hadoop.hbase.thrift.ThriftServerRunner.HBaseHandler.getRowOrBefore(ByteBuffer, ByteBuffer, ByteBuffer) |
org.apache.hadoop.hbase.regionserver.StoreFile.Reader.getScanner(boolean, boolean) |
org.apache.hadoop.hbase.regionserver.StoreFile.Reader.getScanner(boolean, boolean, boolean) |
org.apache.hadoop.hbase.client.HTable.getScannerCaching()
|
org.apache.hadoop.hbase.HRegionInfo.getSeqNumDuringOpen(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.ClusterStatus.getServerInfo()
|
org.apache.hadoop.hbase.HRegionInfo.getServerName(Result)
use MetaTableAccessor methods for interacting with meta layouts
|
org.apache.hadoop.hbase.KeyValue.KVComparator.getShortMidpointKey(byte[], byte[])
Since 0.99.2;
|
org.apache.hadoop.hbase.io.ImmutableBytesWritable.getSize()
|
org.apache.hadoop.hbase.client.HTable.getStartEndKeys()
|
org.apache.hadoop.hbase.client.HTable.getStartKeys()
|
org.apache.hadoop.hbase.regionserver.HStore.getStoreHomedir(Path, HRegionInfo, byte[]) |
org.apache.hadoop.hbase.regionserver.HStore.getStoreHomedir(Path, String, byte[]) |
org.apache.hadoop.hbase.HTableDescriptor.getTableDir(Path, byte[]) |
org.apache.hadoop.hbase.HRegionInfo.getTableName()
|
org.apache.hadoop.hbase.client.ClientScanner.getTableName()
|
org.apache.hadoop.hbase.client.HTableInterface.getTableName()
|
org.apache.hadoop.hbase.HRegionInfo.getTableName(byte[])
|
org.apache.hadoop.hbase.client.ConnectionManager.HConnectionImplementation.getTableNames()
|
org.apache.hadoop.hbase.client.HBaseAdmin.getTableNames()
|
org.apache.hadoop.hbase.client.HConnection.getTableNames()
|
org.apache.hadoop.hbase.client.HBaseAdmin.getTableNames(Pattern)
|
org.apache.hadoop.hbase.client.HBaseAdmin.getTableNames(String)
|
org.apache.hadoop.hbase.KeyValue.getType() |
org.apache.hadoop.hbase.security.access.AccessControlClient.getUserPermissions(Configuration, String)
|
org.apache.hadoop.hbase.io.encoding.BufferedDataBlockEncoder.SeekerState.getValue() |
org.apache.hadoop.hbase.io.encoding.BufferedDataBlockEncoder.ClonedSeekerState.getValue() |
org.apache.hadoop.hbase.KeyValue.getValue() |
org.apache.hadoop.hbase.Cell.getValue()
|
org.apache.hadoop.hbase.codec.prefixtree.PrefixTreeSeeker.ClonedPrefixTreeCell.getValue() |
org.apache.hadoop.hbase.TagRewriteCell.getValue() |
org.apache.hadoop.hbase.HRegionInfo.getVersion()
HRI is no longer a VersionedWritable
|
org.apache.hadoop.hbase.client.HTable.getWriteBuffer()
since 0.96. This is an internal buffer that should not be read nor write.
|
org.apache.hadoop.hbase.client.BufferedMutatorImpl.getWriteBuffer()
|
org.apache.hadoop.hbase.client.Table.getWriteBufferSize()
|
org.apache.hadoop.hbase.client.HTableInterface.getWriteBufferSize()
|
org.apache.hadoop.hbase.client.Mutation.getWriteToWAL()
|
org.apache.hadoop.hbase.security.access.AccessControlClient.grant(Configuration, String, Permission.Action...)
|
org.apache.hadoop.hbase.security.access.AccessControlClient.grant(Configuration, String, String, Permission.Action...)
|
org.apache.hadoop.hbase.security.access.AccessControlClient.grant(Configuration, TableName, String, byte[], byte[], Permission.Action...)
|
org.apache.hadoop.hbase.security.visibility.VisibilityLabelService.havingSystemAuth(byte[])
|
org.apache.hadoop.hbase.security.visibility.DefaultVisibilityLabelServiceImpl.havingSystemAuth(byte[]) |
org.apache.hadoop.hbase.KeyValue.heapSizeWithoutTags() |
org.apache.hadoop.hbase.client.HTable.incrementColumnValue(byte[], byte[], byte[], long, boolean)
|
org.apache.hadoop.hbase.client.HTableInterface.incrementColumnValue(byte[], byte[], byte[], long, boolean)
|
org.apache.hadoop.hbase.mapreduce.TableMapReduceUtil.initCredentialsForCluster(Job, String)
|
org.apache.hadoop.hbase.regionserver.HRegion.initialize()
use HRegion.createHRegion() or HRegion.openHRegion()
|
org.apache.hadoop.hbase.security.access.AccessControlClient.isAccessControllerRunning(Configuration)
|
org.apache.hadoop.hbase.client.HTableInterface.isAutoFlush()
|
org.apache.hadoop.hbase.KeyValue.iscreate(InputStream)
|
org.apache.hadoop.hbase.client.HConnection.isDeadServer(ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.ClusterConnection.isDeadServer(ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.KeyValue.isDelete() |
org.apache.hadoop.hbase.master.cleaner.BaseLogCleanerDelegate.isLogDeletable(FileStatus) |
org.apache.hadoop.hbase.client.ConnectionManager.HConnectionImplementation.isMasterRunning() |
org.apache.hadoop.hbase.client.HBaseAdmin.isMasterRunning()
this has been deprecated without a replacement
|
org.apache.hadoop.hbase.client.HConnection.isMasterRunning()
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.ClusterConnection.isMasterRunning()
this has been deprecated without a replacement
|
org.apache.hadoop.hbase.security.Superusers.isSuperUser(String)
|
org.apache.hadoop.hbase.client.HConnection.isTableAvailable(byte[]) |
org.apache.hadoop.hbase.client.HConnection.isTableAvailable(byte[], byte[][]) |
org.apache.hadoop.hbase.client.HConnection.isTableAvailable(TableName, byte[][])
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.isTableDisabled(byte[]) |
org.apache.hadoop.hbase.client.HTable.isTableEnabled(byte[])
|
org.apache.hadoop.hbase.client.HConnection.isTableEnabled(byte[]) |
org.apache.hadoop.hbase.client.HTable.isTableEnabled(Configuration, byte[])
|
org.apache.hadoop.hbase.client.HTable.isTableEnabled(Configuration, String)
|
org.apache.hadoop.hbase.client.HTable.isTableEnabled(Configuration, TableName)
|
org.apache.hadoop.hbase.client.HTable.isTableEnabled(String)
|
org.apache.hadoop.hbase.client.HTable.isTableEnabled(TableName)
|
org.apache.hadoop.hbase.client.Result.list()
|
org.apache.hadoop.hbase.security.visibility.VisibilityClient.listLabels(Configuration, String)
|
org.apache.hadoop.hbase.client.replication.ReplicationAdmin.listPeers()
|
org.apache.hadoop.hbase.client.ConnectionManager.HConnectionImplementation.listTableNames()
|
org.apache.hadoop.hbase.client.HConnection.listTableNames()
|
org.apache.hadoop.hbase.client.ConnectionManager.HConnectionImplementation.listTables()
|
org.apache.hadoop.hbase.client.HConnection.listTables()
|
org.apache.hadoop.hbase.client.HConnection.locateRegion(byte[])
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.locateRegion(byte[], byte[]) |
org.apache.hadoop.hbase.client.HConnection.locateRegion(TableName, byte[])
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.locateRegions(byte[]) |
org.apache.hadoop.hbase.client.HConnection.locateRegions(byte[], boolean, boolean) |
org.apache.hadoop.hbase.client.HConnection.locateRegions(TableName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.locateRegions(TableName, boolean, boolean)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HBaseAdmin.majorCompact(byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.majorCompact(byte[], byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.majorCompact(String)
|
org.apache.hadoop.hbase.client.HBaseAdmin.majorCompact(String, String)
|
org.apache.hadoop.hbase.security.token.TokenUtil.obtainAndCacheToken(Configuration, UserGroupInformation)
|
org.apache.hadoop.hbase.security.User.obtainAuthTokenForJob(Configuration, Job)
Use TokenUtil.obtainAuthTokenForJob(Connection,User,Job)
instead.
|
org.apache.hadoop.hbase.security.User.obtainAuthTokenForJob(JobConf)
Use TokenUtil.obtainAuthTokenForJob(Connection,JobConf,User)
instead.
|
org.apache.hadoop.hbase.security.token.TokenUtil.obtainToken(Configuration)
|
org.apache.hadoop.hbase.security.token.TokenUtil.obtainTokenForJob(Configuration, UserGroupInformation, Job)
|
org.apache.hadoop.hbase.security.token.TokenUtil.obtainTokenForJob(JobConf, UserGroupInformation)
|
org.apache.hadoop.hbase.KeyValue.oswrite(KeyValue, OutputStream)
|
org.apache.hadoop.hbase.zookeeper.ZKConfig.parseZooCfg(Configuration, InputStream)
in 0.96 onwards. HBase will no longer rely on zoo.cfg
availability.
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.postCompact(ObserverContext, Store, StoreFile)
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.postCompactSelection(ObserverContext, Store, ImmutableList)
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.postFlush(ObserverContext)
|
org.apache.hadoop.hbase.master.MasterCoprocessorHost.postGetTableDescriptors(List) |
org.apache.hadoop.hbase.coprocessor.MasterObserver.postGetTableDescriptors(ObserverContext, List)
Use postGetTableDescriptors with regex instead.
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.postIncrementColumnValue(ObserverContext, byte[], byte[], byte[], long, boolean, long)
This hook is no longer called by the RegionServer
|
org.apache.hadoop.hbase.regionserver.RegionServerServices.postOpenDeployTasks(Region)
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.postSplit(ObserverContext, Region, Region)
Use postCompleteSplit() instead
|
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.postWALRestore(HRegionInfo, HLogKey, WALEdit)
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.postWALRestore(ObserverContext, HRegionInfo, HLogKey, WALEdit)
|
org.apache.hadoop.hbase.coprocessor.WALObserver.postWALWrite(ObserverContext, HRegionInfo, HLogKey, WALEdit)
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.preCompact(ObserverContext, Store, InternalScanner, ScanType)
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.preCompactScannerOpen(ObserverContext, Store, List extends KeyValueScanner>, ScanType, long, InternalScanner)
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.preCompactSelection(ObserverContext, Store, List)
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.preFlush(ObserverContext)
|
org.apache.hadoop.hbase.master.MasterCoprocessorHost.preGetTableDescriptors(List, List) |
org.apache.hadoop.hbase.coprocessor.MasterObserver.preGetTableDescriptors(ObserverContext, List, List)
Use preGetTableDescriptors with regex instead.
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.preIncrementColumnValue(ObserverContext, byte[], byte[], byte[], long, boolean)
This hook is no longer called by the RegionServer
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.preSplit(ObserverContext)
Use preSplit(
final ObserverContext<RegionCoprocessorEnvironment> c, byte[] splitRow)
|
org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.preWALRestore(HRegionInfo, HLogKey, WALEdit)
|
org.apache.hadoop.hbase.coprocessor.RegionObserver.preWALRestore(ObserverContext, HRegionInfo, HLogKey, WALEdit)
|
org.apache.hadoop.hbase.coprocessor.WALObserver.preWALWrite(ObserverContext, HRegionInfo, HLogKey, WALEdit)
|
org.apache.hadoop.hbase.client.ConnectionManager.HConnectionImplementation.processBatch(List extends Row>, byte[], ExecutorService, Object[]) |
org.apache.hadoop.hbase.client.HConnection.processBatch(List extends Row>, byte[], ExecutorService, Object[]) |
org.apache.hadoop.hbase.client.ConnectionManager.HConnectionImplementation.processBatch(List extends Row>, TableName, ExecutorService, Object[]) |
org.apache.hadoop.hbase.client.HConnection.processBatch(List extends Row>, TableName, ExecutorService, Object[])
|
org.apache.hadoop.hbase.client.ConnectionManager.HConnectionImplementation.processBatchCallback(List extends Row>, byte[], ExecutorService, Object[], Batch.Callback) |
org.apache.hadoop.hbase.client.HConnection.processBatchCallback(List extends Row>, byte[], ExecutorService, Object[], Batch.Callback) |
org.apache.hadoop.hbase.client.ConnectionManager.HConnectionImplementation.processBatchCallback(List extends Row>, TableName, ExecutorService, Object[], Batch.Callback)
|
org.apache.hadoop.hbase.client.HConnection.processBatchCallback(List extends Row>, TableName, ExecutorService, Object[], Batch.Callback)
|
org.apache.hadoop.hbase.client.HTableMultiplexer.put(byte[], List) |
org.apache.hadoop.hbase.client.HTableMultiplexer.put(byte[], Put) |
org.apache.hadoop.hbase.client.HTableMultiplexer.put(byte[], Put, int) |
org.apache.hadoop.hbase.client.HTablePool.putTable(HTableInterface) |
org.apache.hadoop.hbase.client.Result.raw()
|
org.apache.hadoop.hbase.regionserver.wal.Compressor.readCompressed(DataInput, Dictionary) |
org.apache.hadoop.hbase.HRegionInfo.readFields(byte[], int, int) |
org.apache.hadoop.hbase.HTableDescriptor.readFields(DataInput)
|
org.apache.hadoop.hbase.HColumnDescriptor.readFields(DataInput)
|
org.apache.hadoop.hbase.HRegionInfo.readFields(DataInput)
Use protobuf deserialization instead.
|
org.apache.hadoop.hbase.io.Reference.readFields(DataInput)
Writables are going away. Use the pb serialization methods instead.
Remove in a release after 0.96 goes out. This is here only to migrate
old Reference files written with Writables before 0.96.
|
org.apache.hadoop.hbase.util.Bytes.readVLong(byte[], int)
|
org.apache.hadoop.hbase.client.HConnection.relocateRegion(byte[], byte[]) |
org.apache.hadoop.hbase.client.HConnection.relocateRegion(TableName, byte[])
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.regionserver.RegionServerServices.reportRegionStateTransition(RegionServerStatusProtos.RegionStateTransition.TransitionCode, HRegionInfo...)
|
org.apache.hadoop.hbase.regionserver.RegionServerServices.reportRegionStateTransition(RegionServerStatusProtos.RegionStateTransition.TransitionCode, long, HRegionInfo...)
|
org.apache.hadoop.hbase.regionserver.Store.requestCompaction(int, CompactionRequest)
see requestCompaction(int, CompactionRequest, User)
|
org.apache.hadoop.hbase.io.hfile.HFileScanner.reseekTo(byte[]) |
org.apache.hadoop.hbase.io.hfile.HFileScanner.reseekTo(byte[], int, int) |
org.apache.hadoop.hbase.mapreduce.TableInputFormatBase.reverseDNS(InetAddress)
mistakenly made public in 0.98.7. scope will change to package-private
|
org.apache.hadoop.hbase.security.access.AccessControlClient.revoke(Configuration, String, Permission.Action...)
|
org.apache.hadoop.hbase.security.access.AccessControlClient.revoke(Configuration, String, String, Permission.Action...)
|
org.apache.hadoop.hbase.security.access.AccessControlClient.revoke(Configuration, TableName, String, byte[], byte[], Permission.Action...)
|
org.apache.hadoop.hbase.regionserver.RegionMergeTransaction.rollback(Server, RegionServerServices)
use #rollback(Server, RegionServerServices, User)
|
org.apache.hadoop.hbase.regionserver.SplitTransaction.rollback(Server, RegionServerServices)
use #rollback(Server, RegionServerServices, User)
|
org.apache.hadoop.hbase.client.HBaseAdmin.rollHLogWriter(String)
|
org.apache.hadoop.hbase.io.hfile.HFileScanner.seekBefore(byte[]) |
org.apache.hadoop.hbase.io.hfile.HFileScanner.seekBefore(byte[], int, int) |
org.apache.hadoop.hbase.io.hfile.HFileScanner.seekTo(byte[]) |
org.apache.hadoop.hbase.io.hfile.HFileScanner.seekTo(byte[], int, int) |
org.apache.hadoop.hbase.io.encoding.DataBlockEncoder.EncodedSeeker.seekToKeyInBlock(byte[], int, int, boolean) |
org.apache.hadoop.hbase.security.visibility.VisibilityClient.setAuths(Configuration, String[], String)
|
org.apache.hadoop.hbase.client.HTable.setAutoFlush(boolean) |
org.apache.hadoop.hbase.client.HTableInterface.setAutoFlush(boolean)
in 0.96. When called with setAutoFlush(false), this function also
set clearBufferOnFail to true, which is unexpected but kept for historical reasons.
Replace it with setAutoFlush(false, false) if this is exactly what you want, though
this is the method you want for most cases.
|
org.apache.hadoop.hbase.client.HTableInterface.setAutoFlush(boolean, boolean)
|
org.apache.hadoop.hbase.client.HTableInterface.setAutoFlushTo(boolean)
|
org.apache.hadoop.hbase.http.HttpServer.Builder.setBindAddress(String) |
org.apache.hadoop.hbase.HColumnDescriptor.setEncodeOnDisk(boolean)
As of release 0.98
(HBASE-9870).
This will be removed in HBase 2.0.0. This method does nothing now.
|
org.apache.hadoop.hbase.client.Delete.setFamilyMap(NavigableMap>) |
org.apache.hadoop.hbase.client.Put.setFamilyMap(NavigableMap>) |
org.apache.hadoop.hbase.client.Increment.setFamilyMap(NavigableMap>) |
org.apache.hadoop.hbase.client.Mutation.setFamilyMap(NavigableMap>)
|
org.apache.hadoop.hbase.client.Append.setFamilyMap(NavigableMap>) |
org.apache.hadoop.hbase.mapreduce.TableInputFormatBase.setHTable(HTable)
|
org.apache.hadoop.hbase.mapred.TableInputFormatBase.setHTable(HTable)
|
org.apache.hadoop.hbase.mapreduce.TableRecordReader.setHTable(Table)
Use setTable() instead.
|
org.apache.hadoop.hbase.HColumnDescriptor.setKeepDeletedCells(boolean)
|
org.apache.hadoop.hbase.HTableDescriptor.setName(byte[]) |
org.apache.hadoop.hbase.http.HttpServer.Builder.setName(String) |
org.apache.hadoop.hbase.HTableDescriptor.setName(TableName) |
org.apache.hadoop.hbase.HTableDescriptor.setOwner(User) |
org.apache.hadoop.hbase.HTableDescriptor.setOwnerString(String) |
org.apache.hadoop.hbase.client.replication.ReplicationAdmin.setPeerTableCFs(String, String)
|
org.apache.hadoop.hbase.http.HttpServer.Builder.setPort(int) |
org.apache.hadoop.hbase.client.HTable.setRegionCachePrefetch(byte[], boolean)
does nothing since 0.99
|
org.apache.hadoop.hbase.client.ConnectionManager.HConnectionImplementation.setRegionCachePrefetch(byte[], boolean) |
org.apache.hadoop.hbase.client.HConnection.setRegionCachePrefetch(byte[], boolean)
does nothing since 0.99
|
org.apache.hadoop.hbase.client.HTable.setRegionCachePrefetch(Configuration, byte[], boolean)
does nothing since 0.99
|
org.apache.hadoop.hbase.client.HTable.setRegionCachePrefetch(Configuration, TableName, boolean)
does nothing since 0.99
|
org.apache.hadoop.hbase.client.HTable.setRegionCachePrefetch(TableName, boolean)
does nothing since 0.99
|
org.apache.hadoop.hbase.client.ConnectionManager.HConnectionImplementation.setRegionCachePrefetch(TableName, boolean) |
org.apache.hadoop.hbase.client.HConnection.setRegionCachePrefetch(TableName, boolean)
does nothing since since 0.99
|
org.apache.hadoop.hbase.client.HTable.setScannerCaching(int)
|
org.apache.hadoop.hbase.client.HConnectionManager.setServerSideHConnectionRetries(Configuration, String, Log)
Internal method, do not use.
|
org.apache.hadoop.hbase.quotas.QuotaSettings.setupSetQuotaRequest(MasterProtos.SetQuotaRequest.Builder)
Removed in HBase 2.0+ as a part of removing protobuf from our API
|
org.apache.hadoop.hbase.client.Table.setWriteBufferSize(long)
|
org.apache.hadoop.hbase.client.BufferedMutatorImpl.setWriteBufferSize(long)
|
org.apache.hadoop.hbase.client.HTableInterface.setWriteBufferSize(long)
|
org.apache.hadoop.hbase.client.Delete.setWriteToWAL(boolean) |
org.apache.hadoop.hbase.client.Put.setWriteToWAL(boolean) |
org.apache.hadoop.hbase.client.Increment.setWriteToWAL(boolean) |
org.apache.hadoop.hbase.client.Mutation.setWriteToWAL(boolean)
|
org.apache.hadoop.hbase.client.Append.setWriteToWAL(boolean) |
org.apache.hadoop.hbase.HColumnDescriptor.shouldCacheBloomsOnWrite()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldCacheDataInL1()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldCacheDataOnWrite()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldCacheIndexesOnWrite()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldCompressTags()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldEvictBlocksOnClose()
|
org.apache.hadoop.hbase.HColumnDescriptor.shouldPrefetchBlocksOnOpen()
|
org.apache.hadoop.hbase.regionserver.RegionSplitPolicy.skipStoreFileRangeCheck()
|
org.apache.hadoop.hbase.client.HBaseAdmin.split(byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.split(byte[], byte[])
|
org.apache.hadoop.hbase.client.HBaseAdmin.split(String)
|
org.apache.hadoop.hbase.client.HBaseAdmin.split(String, String)
|
org.apache.hadoop.hbase.regionserver.RegionMergeTransactionImpl.stepsAfterPONR(Server, RegionServerServices, HRegion) |
org.apache.hadoop.hbase.regionserver.SplitTransactionImpl.stepsAfterPONR(Server, RegionServerServices, PairOfSameType) |
org.apache.hadoop.hbase.filter.FilterWrapper.transform(KeyValue) |
org.apache.hadoop.hbase.filter.FilterList.transform(KeyValue) |
org.apache.hadoop.hbase.filter.FilterBase.transform(KeyValue) |
org.apache.hadoop.hbase.filter.Filter.transform(KeyValue) |
org.apache.hadoop.hbase.mapreduce.LoadIncrementalHFiles.tryAtomicRegionLoad(HConnection, byte[], byte[], Collection)
|
org.apache.hadoop.hbase.regionserver.wal.Compressor.uncompressIntoArray(byte[], int, DataInput, Dictionary) |
org.apache.hadoop.hbase.client.HConnection.updateCachedLocations(byte[], byte[], Object, HRegionLocation) |
org.apache.hadoop.hbase.client.HConnection.updateCachedLocations(TableName, byte[], byte[], Object, ServerName)
internal method, do not use thru HConnection
|
org.apache.hadoop.hbase.client.HConnection.updateCachedLocations(TableName, byte[], Object, HRegionLocation) |
org.apache.hadoop.hbase.zookeeper.ZKUtil.updateExistingNodeData(ZooKeeperWatcher, String, byte[], int)
Unused
|
org.apache.hadoop.hbase.HTableDescriptor.write(DataOutput)
Writables are going away.
Use MessageLite.toByteArray() instead.
|
org.apache.hadoop.hbase.HColumnDescriptor.write(DataOutput)
|
org.apache.hadoop.hbase.HRegionInfo.write(DataOutput)
|
org.apache.hadoop.hbase.regionserver.wal.HLogKey.write(DataOutput)
Don't use these Writables methods. Use PB instead.
|
org.apache.hadoop.hbase.regionserver.wal.Compressor.writeCompressed(byte[], int, int, DataOutput, Dictionary) |