Uses of Class
org.apache.hadoop.hbase.client.Scan
Packages that use Scan
Package
Description
Provides HBase Client
Table of Contents
Provides HBase MapReduce
Input/OutputFormats, a table indexing MapReduce job, and utility methods.
Provides HBase MapReduce
Input/OutputFormats, a table indexing MapReduce job, and utility methods.
Provides an HBase Thrift
service.
-
Uses of Scan in org.apache.hadoop.hbase
Methods in org.apache.hadoop.hbase that return ScanModifier and TypeMethodDescriptionprivate static Scan
AsyncMetaTableAccessor.getMetaScan
(AsyncTable<?> metaTable, int rowUpperLimit) private static Scan
MetaTableAccessor.getMetaScan
(org.apache.hadoop.conf.Configuration conf, int rowUpperLimit) static Scan
MetaTableAccessor.getScanForTableName
(org.apache.hadoop.conf.Configuration conf, TableName tableName) This method creates a Scan object that will only scan catalog rows that belong to the specified table. -
Uses of Scan in org.apache.hadoop.hbase.backup.impl
Methods in org.apache.hadoop.hbase.backup.impl that return ScanModifier and TypeMethodDescriptionprivate Scan
BackupSystemTable.createScanForBackupHistory()
Creates Scan operation to load backup historyprivate Scan
BackupSystemTable.createScanForBackupSetList()
Creates Scan operation to load backup set list(package private) static Scan
BackupSystemTable.createScanForBulkLoadedFiles
(String backupId) (package private) static Scan
BackupSystemTable.createScanForOrigBulkLoadedFiles
(TableName table) private Scan
BackupSystemTable.createScanForReadLogTimestampMap
(String backupRoot) Creates Scan to load table-> { RS -> ts} map of mapsprivate Scan
BackupSystemTable.createScanForReadRegionServerLastLogRollResult
(String backupRoot) Creates Scan operation to load last RS log roll results -
Uses of Scan in org.apache.hadoop.hbase.client
Subclasses of Scan in org.apache.hadoop.hbase.clientFields in org.apache.hadoop.hbase.client declared as ScanModifier and TypeFieldDescriptionprivate final Scan
ImmutableScan.delegateScan
private final Scan
AsyncClientScanner.scan
private Scan
AsyncRpcRetryingCallerFactory.ScanSingleRegionCallerBuilder.scan
private final Scan
AsyncScanSingleRegionRpcRetryingCaller.scan
private final Scan
AsyncTableResultScanner.scan
protected final Scan
ClientScanner.scan
private Scan
OnlineLogRecord.OnlineLogRecordBuilder.scan
protected final Scan
ScannerCallable.scan
private final Scan
ScannerCallableWithReplicas.scan
private Scan
TableSnapshotScanner.scan
private final Scan
ClientScanner.scanForMetrics
Fields in org.apache.hadoop.hbase.client with type parameters of type ScanMethods in org.apache.hadoop.hbase.client that return ScanModifier and TypeMethodDescriptionImmutableScan.addColumn
(byte[] family, byte[] qualifier) Scan.addColumn
(byte[] family, byte[] qualifier) Get the column from the specified family with the specified qualifier.ImmutableScan.addFamily
(byte[] family) Scan.addFamily
(byte[] family) Get all columns from the specified family.static Scan
Scan.createScanFromCursor
(Cursor cursor) Create a new Scan with a cursor.protected Scan
ClientScanner.getScan()
protected Scan
ScannerCallable.getScan()
ImmutableScan.readAllVersions()
Scan.readAllVersions()
Get all available versions.ImmutableScan.readVersions
(int versions) Scan.readVersions
(int versions) Get up to the specified number of versions of each column.(package private) Scan
ImmutableScan.resetMvccReadPoint()
(package private) Scan
Scan.resetMvccReadPoint()
Set the mvcc read point to -1 which means do not use it.ImmutableScan.setACL
(String user, Permission perms) ImmutableScan.setACL
(Map<String, Permission> perms) Scan.setACL
(String user, Permission perms) Scan.setACL
(Map<String, Permission> perms) ImmutableScan.setAllowPartialResults
(boolean allowPartialResults) Scan.setAllowPartialResults
(boolean allowPartialResults) Setting whether the caller wants to see the partial results when server returns less-than-expected cells.ImmutableScan.setAsyncPrefetch
(boolean asyncPrefetch) Deprecated.Scan.setAsyncPrefetch
(boolean asyncPrefetch) ImmutableScan.setAttribute
(String name, byte[] value) Scan.setAttribute
(String name, byte[] value) ImmutableScan.setAuthorizations
(Authorizations authorizations) Scan.setAuthorizations
(Authorizations authorizations) ImmutableScan.setBatch
(int batch) Scan.setBatch
(int batch) Set the maximum number of cells to return for each call to next().ImmutableScan.setCacheBlocks
(boolean cacheBlocks) Scan.setCacheBlocks
(boolean cacheBlocks) Set whether blocks should be cached for this Scan.ImmutableScan.setCaching
(int caching) Scan.setCaching
(int caching) Set the number of rows for caching that will be passed to scanners.ImmutableScan.setColumnFamilyTimeRange
(byte[] cf, long minStamp, long maxStamp) Scan.setColumnFamilyTimeRange
(byte[] cf, long minStamp, long maxStamp) ImmutableScan.setConsistency
(Consistency consistency) Scan.setConsistency
(Consistency consistency) private Scan
RawAsyncTableImpl.setDefaultScanConfig
(Scan scan) ImmutableScan.setFamilyMap
(Map<byte[], NavigableSet<byte[]>> familyMap) Scan.setFamilyMap
(Map<byte[], NavigableSet<byte[]>> familyMap) Setting the familyMapImmutableScan.setIsolationLevel
(IsolationLevel level) Scan.setIsolationLevel
(IsolationLevel level) ImmutableScan.setLimit
(int limit) Scan.setLimit
(int limit) Set the limit of rows for this scan.ImmutableScan.setLoadColumnFamiliesOnDemand
(boolean value) Scan.setLoadColumnFamiliesOnDemand
(boolean value) ImmutableScan.setMaxResultSize
(long maxResultSize) Scan.setMaxResultSize
(long maxResultSize) Set the maximum result size.ImmutableScan.setMaxResultsPerColumnFamily
(int limit) Scan.setMaxResultsPerColumnFamily
(int limit) Set the maximum number of values to return per row per Column FamilyScan.setMaxVersions()
Deprecated.since 2.0.0 and will be removed in 3.0.0.Scan.setMaxVersions
(int maxVersions) Deprecated.since 2.0.0 and will be removed in 3.0.0.(package private) Scan
ImmutableScan.setMvccReadPoint
(long mvccReadPoint) (package private) Scan
Scan.setMvccReadPoint
(long mvccReadPoint) Set the mvcc read point used to open a scanner.ImmutableScan.setNeedCursorResult
(boolean needCursorResult) Scan.setNeedCursorResult
(boolean needCursorResult) When the server is slow or we scan a table with many deleted data or we use a sparse filter, the server will response heartbeat to prevent timeout.ImmutableScan.setOneRowLimit()
Scan.setOneRowLimit()
Call this when you only want to get one row.ImmutableScan.setPriority
(int priority) Scan.setPriority
(int priority) ImmutableScan.setRaw
(boolean raw) Scan.setRaw
(boolean raw) Enable/disable "raw" mode for this scan.ImmutableScan.setReadType
(Scan.ReadType readType) Scan.setReadType
(Scan.ReadType readType) Set the read type for this scan.ImmutableScan.setReplicaId
(int id) Scan.setReplicaId
(int Id) ImmutableScan.setReversed
(boolean reversed) Scan.setReversed
(boolean reversed) Set whether this scan is a reversed oneImmutableScan.setRowOffsetPerColumnFamily
(int offset) Scan.setRowOffsetPerColumnFamily
(int offset) Set offset for the row per Column Family.Scan.setRowPrefixFilter
(byte[] rowPrefix) Deprecated.since 2.5.0, will be removed in 4.0.0.ImmutableScan.setScanMetricsEnabled
(boolean enabled) Scan.setScanMetricsEnabled
(boolean enabled) Enable collection ofScanMetrics
.ImmutableScan.setSmall
(boolean small) Deprecated.Scan.setSmall
(boolean small) Deprecated.since 2.0.0 and will be removed in 3.0.0.Scan.setStartRow
(byte[] startRow) Deprecated.since 2.0.0 and will be removed in 3.0.0.ImmutableScan.setStartStopRowForPrefixScan
(byte[] rowPrefix) Scan.setStartStopRowForPrefixScan
(byte[] rowPrefix) Set a filter (using stopRow and startRow) so the result set only contains rows where the rowKey starts with the specified prefix.Scan.setStopRow
(byte[] stopRow) Deprecated.since 2.0.0 and will be removed in 3.0.0.ImmutableScan.setTimeRange
(long minStamp, long maxStamp) Scan.setTimeRange
(long minStamp, long maxStamp) Get versions of columns only within the specified timestamp range, [minStamp, maxStamp).ImmutableScan.setTimestamp
(long timestamp) Scan.setTimestamp
(long timestamp) Get versions of columns with the specified timestamp.ImmutableScan.setTimeStamp
(long timestamp) Deprecated.Scan.setTimeStamp
(long timestamp) Deprecated.As of release 2.0.0, this will be removed in HBase 3.0.0.ImmutableScan.withStartRow
(byte[] startRow) ImmutableScan.withStartRow
(byte[] startRow, boolean inclusive) Scan.withStartRow
(byte[] startRow) Set the start row of the scan.Scan.withStartRow
(byte[] startRow, boolean inclusive) Set the start row of the scan.ImmutableScan.withStopRow
(byte[] stopRow) ImmutableScan.withStopRow
(byte[] stopRow, boolean inclusive) Scan.withStopRow
(byte[] stopRow) Set the stop row of the scan.Scan.withStopRow
(byte[] stopRow, boolean inclusive) Set the stop row of the scan.Methods in org.apache.hadoop.hbase.client that return types with arguments of type ScanModifier and TypeMethodDescriptionOnlineLogRecord.getScan()
If "hbase.slowlog.scan.payload.enabled" is enabled then this value may be present and should represent the Scan that produced the givenOnlineLogRecord
Methods in org.apache.hadoop.hbase.client with parameters of type ScanModifier and TypeMethodDescriptionstatic ScanResultCache
ConnectionUtils.createScanResultCache
(Scan scan) (package private) static RegionLocateType
ConnectionUtils.getLocateType
(Scan scan) static long
PackagePrivateFieldAccessor.getMvccReadPoint
(Scan scan) AsyncTable.getScanner
(Scan scan) Returns a scanner on the current table as specified by theScan
object.AsyncTableImpl.getScanner
(Scan scan) HTable.getScanner
(Scan scan) The underlyingHTable
must not be closed.RawAsyncTableImpl.getScanner
(Scan scan) default ResultScanner
Table.getScanner
(Scan scan) Returns a scanner on the current table as specified by theScan
object.private ResultScanner
HTable.getScannerInternal
(Scan scan, Scan scanForMetrics) protected void
AbstractClientScanner.initScanMetrics
(Scan scan) Check and initialize if application wants to collect scan metrics(package private) static boolean
ConnectionUtils.noMoreResultsForReverseScan
(Scan scan, RegionInfo info) (package private) static boolean
ConnectionUtils.noMoreResultsForScan
(Scan scan, RegionInfo info) void
The scan API uses the observer pattern.void
AsyncTableImpl.scan
(Scan scan, ScanResultConsumer consumer) void
RawAsyncTableImpl.scan
(Scan scan, AdvancedScanResultConsumer consumer) private void
AsyncTableImpl.scan0
(Scan scan, ScanResultConsumer consumer) Return all the results that match the given scan object.private Scan
RawAsyncTableImpl.setDefaultScanConfig
(Scan scan) static void
PackagePrivateFieldAccessor.setMvccReadPoint
(Scan scan, long mvccReadPoint) Constructors in org.apache.hadoop.hbase.client with parameters of type ScanModifierConstructorDescriptionAsyncClientScanner
(Scan scan, AdvancedScanResultConsumer consumer, TableName tableName, AsyncConnectionImpl conn, org.apache.hbase.thirdparty.io.netty.util.Timer retryTimer, long pauseNs, long pauseNsForServerOverloaded, int maxAttempts, long scanTimeoutNs, long rpcTimeoutNs, int startLogErrorsCnt, Map<String, byte[]> requestAttributes) AsyncScanSingleRegionRpcRetryingCaller
(org.apache.hbase.thirdparty.io.netty.util.Timer retryTimer, AsyncConnectionImpl conn, Scan scan, ScanMetrics scanMetrics, long scannerId, ScanResultCache resultCache, AdvancedScanResultConsumer consumer, org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ClientService.Interface stub, HRegionLocation loc, boolean isRegionServerRemote, int priority, long scannerLeaseTimeoutPeriodNs, long pauseNs, long pauseNsForServerOverloaded, int maxAttempts, long scanTimeoutNs, long rpcTimeoutNs, int startLogErrorsCnt, Map<String, byte[]> requestAttributes) AsyncTableResultScanner
(TableName tableName, Scan scan, long maxCacheSize) ClientAsyncPrefetchScanner
(org.apache.hadoop.conf.Configuration configuration, Scan scan, Scan scanForMetrics, TableName name, ClusterConnection connection, RpcRetryingCallerFactory rpcCallerFactory, RpcControllerFactory rpcControllerFactory, ExecutorService pool, int scanReadRpcTimeout, int scannerTimeout, int replicaCallTimeoutMicroSecondScan, ConnectionConfiguration connectionConfiguration, Map<String, byte[]> requestAttributes) ClientScanner
(org.apache.hadoop.conf.Configuration conf, Scan scan, Scan scanForMetrics, TableName tableName, ClusterConnection connection, RpcRetryingCallerFactory rpcFactory, RpcControllerFactory controllerFactory, ExecutorService pool, int scanReadRpcTimeout, int scannerTimeout, int primaryOperationTimeout, ConnectionConfiguration connectionConfiguration, Map<String, byte[]> requestAttributes) Create a new ClientScanner for the specified table Note that the passedScan
's start row maybe changed changed.ClientSideRegionScanner
(org.apache.hadoop.conf.Configuration conf, org.apache.hadoop.fs.FileSystem fs, org.apache.hadoop.fs.Path rootDir, TableDescriptor htd, RegionInfo hri, Scan scan, ScanMetrics scanMetrics) ClientSimpleScanner
(org.apache.hadoop.conf.Configuration configuration, Scan scan, Scan scanForMetrics, TableName name, ClusterConnection connection, RpcRetryingCallerFactory rpcCallerFactory, RpcControllerFactory rpcControllerFactory, ExecutorService pool, int scanReadRpcTimeout, int scannerTimeout, int replicaCallTimeoutMicroSecondScan, ConnectionConfiguration connectionConfiguration, Map<String, byte[]> requestAttributes) ImmutableScan
(Scan scan) Create Immutable instance of Scan from given Scan object(package private)
OnlineLogRecord
(long startTime, int processingTime, int queueTime, long responseSize, long blockBytesScanned, long fsReadTime, String clientAddress, String serverClass, String methodName, String callDetails, String param, String regionName, String userName, int multiGetsCount, int multiMutationsCount, int multiServiceCalls, Scan scan, Map<String, byte[]> requestAttributes, Map<String, byte[]> connectionAttributes) ReversedClientScanner
(org.apache.hadoop.conf.Configuration conf, Scan scan, Scan scanForMetrics, TableName tableName, ClusterConnection connection, RpcRetryingCallerFactory rpcFactory, RpcControllerFactory controllerFactory, ExecutorService pool, int scanReadRpcTimeout, int scannerTimeout, int primaryOperationTimeout, ConnectionConfiguration connectionConfiguration, Map<String, byte[]> requestAttributes) Create a new ReversibleClientScanner for the specified table Note that the passedScan
's start row maybe changed.ReversedScannerCallable
(ClusterConnection connection, TableName tableName, Scan scan, ScanMetrics scanMetrics, RpcControllerFactory rpcFactory, int replicaId, Map<String, byte[]> requestAttributes) Creates a new instance of this class while copying all values.ScannerCallable
(ClusterConnection connection, TableName tableName, Scan scan, ScanMetrics scanMetrics, RpcControllerFactory rpcControllerFactory, int id, Map<String, byte[]> requestAttributes) ScannerCallableWithReplicas
(TableName tableName, ClusterConnection cConnection, ScannerCallable baseCallable, ExecutorService pool, int timeBeforeReplicas, Scan scan, int retries, int readRpcTimeout, int scannerTimeout, boolean useScannerTimeoutForNextCalls, int caching, org.apache.hadoop.conf.Configuration conf, RpcRetryingCaller<Result[]> caller) TableSnapshotScanner
(org.apache.hadoop.conf.Configuration conf, org.apache.hadoop.fs.Path restoreDir, String snapshotName, Scan scan) Creates a TableSnapshotScanner.TableSnapshotScanner
(org.apache.hadoop.conf.Configuration conf, org.apache.hadoop.fs.Path rootDir, org.apache.hadoop.fs.Path restoreDir, String snapshotName, Scan scan) TableSnapshotScanner
(org.apache.hadoop.conf.Configuration conf, org.apache.hadoop.fs.Path rootDir, org.apache.hadoop.fs.Path restoreDir, String snapshotName, Scan scan, boolean snapshotAlreadyRestored) Creates a TableSnapshotScanner. -
Uses of Scan in org.apache.hadoop.hbase.client.trace
Methods in org.apache.hadoop.hbase.client.trace with parameters of type ScanModifier and TypeMethodDescriptionTableOperationSpanBuilder.setOperation
(Scan scan) private static HBaseSemanticAttributes.Operation
-
Uses of Scan in org.apache.hadoop.hbase.coprocessor
Methods in org.apache.hadoop.hbase.coprocessor with parameters of type ScanModifier and TypeMethodDescriptiondefault RegionScanner
RegionObserver.postScannerOpen
(ObserverContext<RegionCoprocessorEnvironment> c, Scan scan, RegionScanner s) Called after the client opens a new scanner.default void
RegionObserver.preScannerOpen
(ObserverContext<RegionCoprocessorEnvironment> c, Scan scan) Called before the client opens a new scanner. -
Uses of Scan in org.apache.hadoop.hbase.coprocessor.example
Methods in org.apache.hadoop.hbase.coprocessor.example with parameters of type ScanModifier and TypeMethodDescriptionvoid
ScanModifyingObserver.preScannerOpen
(ObserverContext<RegionCoprocessorEnvironment> c, Scan scan) -
Uses of Scan in org.apache.hadoop.hbase.io
Methods in org.apache.hadoop.hbase.io with parameters of type Scan -
Uses of Scan in org.apache.hadoop.hbase.mapred
Method parameters in org.apache.hadoop.hbase.mapred with type arguments of type ScanModifier and TypeMethodDescriptionstatic void
TableMapReduceUtil.initMultiTableSnapshotMapperJob
(Map<String, Collection<Scan>> snapshotScans, Class<? extends TableMap> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapred.JobConf job, boolean addDependencyJars, org.apache.hadoop.fs.Path tmpRestoreDir) Sets up the job for reading from one or more multiple table snapshots, with one or more scans per snapshot.static void
MultiTableSnapshotInputFormat.setInput
(org.apache.hadoop.conf.Configuration conf, Map<String, Collection<Scan>> snapshotScans, org.apache.hadoop.fs.Path restoreDir) Constructors in org.apache.hadoop.hbase.mapred with parameters of type ScanModifierConstructorDescriptionTableSnapshotRegionSplit
(HTableDescriptor htd, HRegionInfo regionInfo, List<String> locations, Scan scan, org.apache.hadoop.fs.Path restoreDir) -
Uses of Scan in org.apache.hadoop.hbase.mapreduce
Fields in org.apache.hadoop.hbase.mapreduce declared as ScanModifier and TypeFieldDescriptionprivate Scan
TableRecordReaderImpl.currentScan
private Scan
TableInputFormatBase.scan
Holds the details for the internal scanner.private Scan
TableRecordReaderImpl.scan
private Scan
TableSnapshotInputFormatImpl.RecordReader.scan
Fields in org.apache.hadoop.hbase.mapreduce with type parameters of type ScanModifier and TypeFieldDescriptionMultiTableInputFormatBase.scans
Holds the set of scans used to define the input.Methods in org.apache.hadoop.hbase.mapreduce that return ScanModifier and TypeMethodDescriptionstatic Scan
TableMapReduceUtil.convertStringToScan
(String base64) Converts the given Base64 string back into a Scan instance.static Scan
TableInputFormat.createScanFromConfiguration
(org.apache.hadoop.conf.Configuration conf) Sets up aScan
instance, applying settings from the configuration property constants defined inTableInputFormat
.static Scan
TableSnapshotInputFormatImpl.extractScanFromConf
(org.apache.hadoop.conf.Configuration conf) private static Scan
CellCounter.getConfiguredScanForJob
(org.apache.hadoop.conf.Configuration conf, String[] args) TableInputFormatBase.getScan()
Gets the scan defining the actual details like columns etc.TableSplit.getScan()
Returns a Scan object from the stored string representation.(package private) static Scan
ExportUtils.getScanFromCommandLine
(org.apache.hadoop.conf.Configuration conf, String[] args) (package private) Scan
HashTable.TableHash.initScan()
Methods in org.apache.hadoop.hbase.mapreduce that return types with arguments of type ScanModifier and TypeMethodDescriptionExportUtils.getArgumentsFromCommandLine
(org.apache.hadoop.conf.Configuration conf, String[] args) MultiTableInputFormatBase.getScans()
Allows subclasses to get the list ofScan
objects.MultiTableSnapshotInputFormatImpl.getSnapshotsToScans
(org.apache.hadoop.conf.Configuration conf) Retrieve the snapshot name -> list<scan> mapping pushed to configuration byMultiTableSnapshotInputFormatImpl.setSnapshotToScans(Configuration, Map)
Methods in org.apache.hadoop.hbase.mapreduce with parameters of type ScanModifier and TypeMethodDescriptionprivate static void
Parses a combined family and qualifier and adds either both or just the family in case there is no qualifier.static void
TableInputFormat.addColumns
(Scan scan, byte[][] columns) Adds an array of columns specified using old format, family:qualifier.private static void
TableInputFormat.addColumns
(Scan scan, String columns) Convenience method to parse a string representation of an array of column specifiers.static String
TableMapReduceUtil.convertScanToString
(Scan scan) Writes the given scan into a Base64 encoded string.TableSnapshotInputFormatImpl.getSplits
(Scan scan, SnapshotManifest manifest, List<HRegionInfo> regionManifests, org.apache.hadoop.fs.Path restoreDir, org.apache.hadoop.conf.Configuration conf) TableSnapshotInputFormatImpl.getSplits
(Scan scan, SnapshotManifest manifest, List<HRegionInfo> regionManifests, org.apache.hadoop.fs.Path restoreDir, org.apache.hadoop.conf.Configuration conf, RegionSplitter.SplitAlgorithm sa, int numSplits) private void
CopyTable.initCopyTableMapperReducerJob
(org.apache.hadoop.mapreduce.Job job, Scan scan) static void
GroupingTableMapper.initJob
(String table, Scan scan, String groupColumns, Class<? extends TableMapper> mapper, org.apache.hadoop.mapreduce.Job job) Use this before submitting a TableMap job.static void
IdentityTableMapper.initJob
(String table, Scan scan, Class<? extends TableMapper> mapper, org.apache.hadoop.mapreduce.Job job) Use this before submitting a TableMap job.static void
TableMapReduceUtil.initTableMapperJob
(byte[] table, Scan scan, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job) Use this before submitting a TableMap job.static void
TableMapReduceUtil.initTableMapperJob
(byte[] table, Scan scan, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars) Use this before submitting a TableMap job.static void
TableMapReduceUtil.initTableMapperJob
(byte[] table, Scan scan, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars, Class<? extends org.apache.hadoop.mapreduce.InputFormat> inputFormatClass) Use this before submitting a TableMap job.static void
TableMapReduceUtil.initTableMapperJob
(String table, Scan scan, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job) Use this before submitting a TableMap job.static void
TableMapReduceUtil.initTableMapperJob
(String table, Scan scan, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars) Use this before submitting a TableMap job.static void
TableMapReduceUtil.initTableMapperJob
(String table, Scan scan, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars, boolean initCredentials, Class<? extends org.apache.hadoop.mapreduce.InputFormat> inputFormatClass) Use this before submitting a TableMap job.static void
TableMapReduceUtil.initTableMapperJob
(String table, Scan scan, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars, Class<? extends org.apache.hadoop.mapreduce.InputFormat> inputFormatClass) Use this before submitting a TableMap job.static void
TableMapReduceUtil.initTableMapperJob
(TableName table, Scan scan, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job) Use this before submitting a TableMap job.static void
TableMapReduceUtil.initTableSnapshotMapperJob
(String snapshotName, Scan scan, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars, org.apache.hadoop.fs.Path tmpRestoreDir) Sets up the job for reading from a table snapshot.static void
TableMapReduceUtil.initTableSnapshotMapperJob
(String snapshotName, Scan scan, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars, org.apache.hadoop.fs.Path tmpRestoreDir, RegionSplitter.SplitAlgorithm splitAlgo, int numSplitsPerRegion) Sets up the job for reading from a table snapshot.void
Sets the scan defining the actual details like columns etc.void
Sets the scan defining the actual details like columns etc.void
Sets the scan defining the actual details like columns etc.private static void
RowCounter.setScanFilter
(Scan scan, List<MultiRowRangeFilter.RowRange> rowRangeList) Sets filterFilterBase
to theScan
instance.Method parameters in org.apache.hadoop.hbase.mapreduce with type arguments of type ScanModifier and TypeMethodDescriptionstatic void
TableMapReduceUtil.initMultiTableSnapshotMapperJob
(Map<String, Collection<Scan>> snapshotScans, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars, org.apache.hadoop.fs.Path tmpRestoreDir) Sets up the job for reading from one or more table snapshots, with one or more scans per snapshot.static void
TableMapReduceUtil.initTableMapperJob
(List<Scan> scans, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job) Use this before submitting a Multi TableMap job.static void
TableMapReduceUtil.initTableMapperJob
(List<Scan> scans, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars) Use this before submitting a Multi TableMap job.static void
TableMapReduceUtil.initTableMapperJob
(List<Scan> scans, Class<? extends TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars, boolean initCredentials) Use this before submitting a Multi TableMap job.static void
MultiTableSnapshotInputFormat.setInput
(org.apache.hadoop.conf.Configuration configuration, Map<String, Collection<Scan>> snapshotScans, org.apache.hadoop.fs.Path tmpRestoreDir) void
MultiTableSnapshotInputFormatImpl.setInput
(org.apache.hadoop.conf.Configuration conf, Map<String, Collection<Scan>> snapshotScans, org.apache.hadoop.fs.Path restoreDir) Configure conf to read from snapshotScans, with snapshots restored to a subdirectory of restoreDir.protected void
Allows subclasses to set the list ofScan
objects.void
MultiTableSnapshotInputFormatImpl.setSnapshotToScans
(org.apache.hadoop.conf.Configuration conf, Map<String, Collection<Scan>> snapshotScans) Push snapshotScans to conf (under the keyMultiTableSnapshotInputFormatImpl.SNAPSHOT_TO_SCANS_KEY
)Constructors in org.apache.hadoop.hbase.mapreduce with parameters of type ScanModifierConstructorDescriptionInputSplit
(TableDescriptor htd, HRegionInfo regionInfo, List<String> locations, Scan scan, org.apache.hadoop.fs.Path restoreDir) TableSnapshotRegionSplit
(HTableDescriptor htd, HRegionInfo regionInfo, List<String> locations, Scan scan, org.apache.hadoop.fs.Path restoreDir) TableSplit
(TableName tableName, Scan scan, byte[] startRow, byte[] endRow, String location) Creates a new instance while assigning all variables.TableSplit
(TableName tableName, Scan scan, byte[] startRow, byte[] endRow, String location, long length) Creates a new instance while assigning all variables.TableSplit
(TableName tableName, Scan scan, byte[] startRow, byte[] endRow, String location, String encodedRegionName, long length) Creates a new instance while assigning all variables. -
Uses of Scan in org.apache.hadoop.hbase.mapreduce.replication
Fields in org.apache.hadoop.hbase.mapreduce.replication declared as ScanModifier and TypeFieldDescriptionprivate Scan
VerifyReplication.Verifier.tableScan
private final Scan
VerifyReplicationRecompareRunnable.tableScan
Methods in org.apache.hadoop.hbase.mapreduce.replication with parameters of type ScanModifier and TypeMethodDescriptionprivate static void
VerifyReplication.setRowPrefixFilter
(Scan scan, String rowPrefixes) private static void
VerifyReplication.setStartAndStopRows
(Scan scan, byte[] startPrefixRow, byte[] lastPrefixRow) Constructors in org.apache.hadoop.hbase.mapreduce.replication with parameters of type ScanModifierConstructorDescriptionVerifyReplicationRecompareRunnable
(org.apache.hadoop.mapreduce.Mapper.Context context, Result sourceResult, Result replicatedResult, VerifyReplication.Verifier.Counters originalCounter, String delimiter, Scan tableScan, Table sourceTable, Table replicatedTable, int reCompareTries, int sleepMsBeforeReCompare, int reCompareBackoffExponent, boolean verbose) -
Uses of Scan in org.apache.hadoop.hbase.master.assignment
Methods in org.apache.hadoop.hbase.master.assignment that return ScanModifier and TypeMethodDescriptionprivate Scan
RegionStateStore.getScanForUpdateRegionReplicas
(TableName tableName) -
Uses of Scan in org.apache.hadoop.hbase.master.http
Methods in org.apache.hadoop.hbase.master.http that return Scan -
Uses of Scan in org.apache.hadoop.hbase.master.region
Methods in org.apache.hadoop.hbase.master.region with parameters of type ScanModifier and TypeMethodDescriptionMasterRegion.getRegionScanner
(Scan scan) MasterRegion.getScanner
(Scan scan) -
Uses of Scan in org.apache.hadoop.hbase.mob
Methods in org.apache.hadoop.hbase.mob with parameters of type ScanModifier and TypeMethodDescriptionstatic boolean
MobUtils.isCacheMobBlocks
(Scan scan) Indicates whether the scan contains the information of caching blocks.static boolean
MobUtils.isRawMobScan
(Scan scan) Indicates whether it's a raw scan.static boolean
MobUtils.isReadEmptyValueOnMobCellMiss
(Scan scan) Indicates whether return null value when the mob file is missing or corrupt.static boolean
MobUtils.isRefOnlyScan
(Scan scan) Indicates whether it's a reference only scan.static void
MobUtils.setCacheMobBlocks
(Scan scan, boolean cacheBlocks) Sets the attribute of caching blocks in the scan. -
Uses of Scan in org.apache.hadoop.hbase.quotas
Methods in org.apache.hadoop.hbase.quotas that return ScanModifier and TypeMethodDescription(package private) static Scan
QuotaTableUtil.createScanForNamespaceSnapshotSizes()
Returns a scanner for all existing namespace snapshot entries.(package private) static Scan
QuotaTableUtil.createScanForNamespaceSnapshotSizes
(String namespace) Returns a scanner for all namespace snapshot entries of the given namespace(package private) static Scan
QuotaTableUtil.createScanForSpaceSnapshotSizes()
(package private) static Scan
QuotaTableUtil.createScanForSpaceSnapshotSizes
(TableName table) static Scan
QuotaTableUtil.makeQuotaSnapshotScan()
Creates aScan
which returns only quota snapshots from the quota table.static Scan
QuotaTableUtil.makeQuotaSnapshotScanForTable
(TableName tn) Creates aScan
which returns onlySpaceQuotaSnapshot
from the quota table for a specific table.static Scan
QuotaTableUtil.makeScan
(QuotaFilter filter) Methods in org.apache.hadoop.hbase.quotas with parameters of type ScanModifier and TypeMethodDescriptionQuotaTableUtil.createDeletesForExistingSnapshotsFromScan
(Connection connection, Scan scan) Returns a list ofDelete
to remove all entries returned by the passed scanner.(package private) void
(package private) void
QuotaRetriever.init
(Connection conn, Scan scan) -
Uses of Scan in org.apache.hadoop.hbase.regionserver
Subclasses of Scan in org.apache.hadoop.hbase.regionserverModifier and TypeClassDescriptionclass
Special scanner, currently used for increment operations to allow additional server-side arguments for Scan operations.Fields in org.apache.hadoop.hbase.regionserver declared as ScanModifier and TypeFieldDescriptionprivate final Scan
CustomizedScanInfoBuilder.scan
private final Scan
StoreScanner.scan
private static final Scan
StoreScanner.SCAN_FOR_COMPACTION
Methods in org.apache.hadoop.hbase.regionserver that return ScanModifier and TypeMethodDescriptionCustomizedScanInfoBuilder.getScan()
ScanOptions.getScan()
Returns a copy of the Scan object.Methods in org.apache.hadoop.hbase.regionserver with parameters of type ScanModifier and TypeMethodDescriptionprotected KeyValueScanner
HMobStore.createScanner
(Scan scan, ScanInfo scanInfo, NavigableSet<byte[]> targetCols, long readPt) Gets the MobStoreScanner or MobReversedStoreScanner.protected KeyValueScanner
HStore.createScanner
(Scan scan, ScanInfo scanInfo, NavigableSet<byte[]> targetCols, long readPt) HRegion.getScanner
(Scan scan) HRegion.getScanner
(Scan scan, List<KeyValueScanner> additionalScanners) private RegionScannerImpl
HRegion.getScanner
(Scan scan, List<KeyValueScanner> additionalScanners, long nonceGroup, long nonce) HStore.getScanner
(Scan scan, NavigableSet<byte[]> targetCols, long readPt) Return a scanner for both the memstore and the HStore files.Region.getScanner
(Scan scan) Return an iterator that scans over the HRegion, returning the indicated columns and rows specified by theScan
.Region.getScanner
(Scan scan, List<KeyValueScanner> additionalScanners) Return an iterator that scans over the HRegion, returning the indicated columns and rows specified by theScan
.private void
RegionScannerImpl.initializeScanners
(Scan scan, List<KeyValueScanner> additionalScanners) protected RegionScannerImpl
HRegion.instantiateRegionScanner
(Scan scan, List<KeyValueScanner> additionalScanners, long nonceGroup, long nonce) private boolean
RSRpcServices.isFullRegionScan
(Scan scan, HRegion region) (package private) boolean
StoreFileReader.passesBloomFilter
(Scan scan, SortedSet<byte[]> columns) Checks whether the given scan passes the Bloom filter (if present).private boolean
StoreFileReader.passesGeneralRowPrefixBloomFilter
(Scan scan) A method for checking Bloom filters.boolean
StoreFileReader.passesKeyRangeFilter
(Scan scan) Checks whether the given scan rowkey range overlaps with the current storefile'sRegionCoprocessorHost.postScannerOpen
(Scan scan, RegionScanner s) void
RegionCoprocessorHost.preScannerOpen
(Scan scan) RegionCoprocessorHost.preStoreScannerOpen
(HStore store, Scan scan) Called before open store scanner for user scan.boolean
KeyValueScanner.shouldUseScanner
(Scan scan, HStore store, long oldestUnexpiredTS) Allows to filter out scanners (both StoreFile and memstore) that we don't want to use based on criteria such as Bloom filters and timestamp ranges.boolean
NonLazyKeyValueScanner.shouldUseScanner
(Scan scan, HStore store, long oldestUnexpiredTS) boolean
SegmentScanner.shouldUseScanner
(Scan scan, HStore store, long oldestUnexpiredTS) This functionality should be resolved in the higher level which is MemStoreScanner, currently returns true as default.boolean
StoreFileScanner.shouldUseScanner
(Scan scan, HStore store, long oldestUnexpiredTS) Constructors in org.apache.hadoop.hbase.regionserver with parameters of type ScanModifierConstructorDescriptionCustomizedScanInfoBuilder
(ScanInfo scanInfo, Scan scan) InternalScan
(Scan scan) MobStoreScanner
(HStore store, ScanInfo scanInfo, Scan scan, NavigableSet<byte[]> columns, long readPt) (package private)
RegionScannerImpl
(Scan scan, List<KeyValueScanner> additionalScanners, HRegion region, long nonceGroup, long nonce) (package private)
ReversedMobStoreScanner
(HStore store, ScanInfo scanInfo, Scan scan, NavigableSet<byte[]> columns, long readPt) (package private)
ReversedRegionScannerImpl
(Scan scan, List<KeyValueScanner> additionalScanners, HRegion region, long nonceGroup, long nonce) ReversedStoreScanner
(Scan scan, ScanInfo scanInfo, NavigableSet<byte[]> columns, List<? extends KeyValueScanner> scanners) Constructor for testing.ReversedStoreScanner
(HStore store, ScanInfo scanInfo, Scan scan, NavigableSet<byte[]> columns, long readPt) Opens a scanner across memstore, snapshot, and all StoreFiles.(package private)
StoreScanner
(Scan scan, ScanInfo scanInfo, NavigableSet<byte[]> columns, List<? extends KeyValueScanner> scanners) (package private)
StoreScanner
(Scan scan, ScanInfo scanInfo, NavigableSet<byte[]> columns, List<? extends KeyValueScanner> scanners, ScanType scanType) private
StoreScanner
(HStore store, Scan scan, ScanInfo scanInfo, int numColumns, long readPt, boolean cacheBlocks, ScanType scanType) An internal constructor.StoreScanner
(HStore store, ScanInfo scanInfo, Scan scan, NavigableSet<byte[]> columns, long readPt) Opens a scanner across memstore, snapshot, and all StoreFiles. -
Uses of Scan in org.apache.hadoop.hbase.regionserver.querymatcher
Methods in org.apache.hadoop.hbase.regionserver.querymatcher with parameters of type ScanModifier and TypeMethodDescriptionstatic NormalUserScanQueryMatcher
NormalUserScanQueryMatcher.create
(Scan scan, ScanInfo scanInfo, ColumnTracker columns, DeleteTracker deletes, boolean hasNullColumn, long oldestUnexpiredTS, long now) static RawScanQueryMatcher
RawScanQueryMatcher.create
(Scan scan, ScanInfo scanInfo, ColumnTracker columns, boolean hasNullColumn, long oldestUnexpiredTS, long now) static UserScanQueryMatcher
UserScanQueryMatcher.create
(Scan scan, ScanInfo scanInfo, NavigableSet<byte[]> columns, long oldestUnexpiredTS, long now, RegionCoprocessorHost regionCoprocessorHost) private static Cell
UserScanQueryMatcher.createStartKey
(Scan scan, ScanInfo scanInfo) protected static Pair<DeleteTracker,
ColumnTracker> ScanQueryMatcher.getTrackers
(RegionCoprocessorHost host, NavigableSet<byte[]> columns, ScanInfo scanInfo, long oldestUnexpiredTS, Scan userScan) Constructors in org.apache.hadoop.hbase.regionserver.querymatcher with parameters of type ScanModifierConstructorDescriptionprotected
NormalUserScanQueryMatcher
(Scan scan, ScanInfo scanInfo, ColumnTracker columns, boolean hasNullColumn, DeleteTracker deletes, long oldestUnexpiredTS, long now) protected
RawScanQueryMatcher
(Scan scan, ScanInfo scanInfo, ColumnTracker columns, boolean hasNullColumn, long oldestUnexpiredTS, long now) protected
UserScanQueryMatcher
(Scan scan, ScanInfo scanInfo, ColumnTracker columns, boolean hasNullColumn, long oldestUnexpiredTS, long now) -
Uses of Scan in org.apache.hadoop.hbase.rest.model
Methods in org.apache.hadoop.hbase.rest.model with parameters of type Scan -
Uses of Scan in org.apache.hadoop.hbase.security.access
Methods in org.apache.hadoop.hbase.security.access with parameters of type ScanModifier and TypeMethodDescriptionAccessController.postScannerOpen
(ObserverContext<RegionCoprocessorEnvironment> c, Scan scan, RegionScanner s) void
AccessController.preScannerOpen
(ObserverContext<RegionCoprocessorEnvironment> c, Scan scan) -
Uses of Scan in org.apache.hadoop.hbase.security.visibility
Methods in org.apache.hadoop.hbase.security.visibility with parameters of type ScanModifier and TypeMethodDescriptionVisibilityController.postScannerOpen
(ObserverContext<RegionCoprocessorEnvironment> c, Scan scan, RegionScanner s) void
VisibilityController.preScannerOpen
(ObserverContext<RegionCoprocessorEnvironment> e, Scan scan) -
Uses of Scan in org.apache.hadoop.hbase.thrift2
Methods in org.apache.hadoop.hbase.thrift2 that return ScanModifier and TypeMethodDescriptionstatic Scan
ThriftUtilities.scanFromThrift
(org.apache.hadoop.hbase.thrift2.generated.TScan in) Methods in org.apache.hadoop.hbase.thrift2 with parameters of type ScanModifier and TypeMethodDescriptionstatic org.apache.hadoop.hbase.thrift2.generated.TScan
ThriftUtilities.scanFromHBase
(Scan in) -
Uses of Scan in org.apache.hadoop.hbase.thrift2.client
Methods in org.apache.hadoop.hbase.thrift2.client with parameters of type ScanConstructors in org.apache.hadoop.hbase.thrift2.client with parameters of type Scan