Uses of Class
org.apache.hadoop.hbase.client.Scan
Packages that use org.apache.hadoop.hbase.client.Scan
Package
Description
Provides HBase Client
Table of Contents
Provides HBase MapReduce
Input/OutputFormats, a table indexing MapReduce job, and utility methods.
Provides HBase MapReduce
Input/OutputFormats, a table indexing MapReduce job, and utility methods.
-
Uses of org.apache.hadoop.hbase.client.Scan in org.apache.hadoop.hbase
Methods in org.apache.hadoop.hbase that return org.apache.hadoop.hbase.client.Scan in inModifier and TypeMethodDescriptionstatic org.apache.hadoop.hbase.client.ScanMetaTableAccessor.getScanForTableName(org.apache.hadoop.conf.Configuration conf, org.apache.hadoop.hbase.TableName tableName) This method creates a Scan object that will only scan catalog rows that belong to the specified table. -
Uses of org.apache.hadoop.hbase.client.Scan in org.apache.hadoop.hbase.client
Subclasses of org.apache.hadoop.hbase.client.Scan in in org.apache.hadoop.hbase.clientModifier and TypeClassDescriptionfinal classorg.apache.hadoop.hbase.client.ImmutableScanImmutable version of ScanFields in org.apache.hadoop.hbase.client declared as org.apache.hadoop.hbase.client.Scan in inModifier and TypeFieldDescriptionprotected final org.apache.hadoop.hbase.client.ScanClientScanner.scanprotected final org.apache.hadoop.hbase.client.ScanScannerCallable.scanMethods in org.apache.hadoop.hbase.client that return org.apache.hadoop.hbase.client.Scan in inModifier and TypeMethodDescriptionorg.apache.hadoop.hbase.client.ScanImmutableScan.addColumn(byte[] family, byte[] qualifier) org.apache.hadoop.hbase.client.ScanScan.addColumn(byte[] family, byte[] qualifier) Get the column from the specified family with the specified qualifier.org.apache.hadoop.hbase.client.ScanImmutableScan.addFamily(byte[] family) org.apache.hadoop.hbase.client.ScanScan.addFamily(byte[] family) Get all columns from the specified family.static org.apache.hadoop.hbase.client.ScanScan.createScanFromCursor(org.apache.hadoop.hbase.client.Cursor cursor) Create a new Scan with a cursor.protected org.apache.hadoop.hbase.client.ScanClientScanner.getScan()protected org.apache.hadoop.hbase.client.ScanScannerCallable.getScan()org.apache.hadoop.hbase.client.ScanImmutableScan.readAllVersions()org.apache.hadoop.hbase.client.ScanScan.readAllVersions()Get all available versions.org.apache.hadoop.hbase.client.ScanImmutableScan.readVersions(int versions) org.apache.hadoop.hbase.client.ScanScan.readVersions(int versions) Get up to the specified number of versions of each column.org.apache.hadoop.hbase.client.ScanImmutableScan.setACL(String user, org.apache.hadoop.hbase.security.access.Permission perms) org.apache.hadoop.hbase.client.Scanorg.apache.hadoop.hbase.client.ScanScan.setACL(String user, org.apache.hadoop.hbase.security.access.Permission perms) org.apache.hadoop.hbase.client.Scanorg.apache.hadoop.hbase.client.ScanImmutableScan.setAllowPartialResults(boolean allowPartialResults) org.apache.hadoop.hbase.client.ScanScan.setAllowPartialResults(boolean allowPartialResults) Setting whether the caller wants to see the partial results when server returns less-than-expected cells.org.apache.hadoop.hbase.client.ScanImmutableScan.setAsyncPrefetch(boolean asyncPrefetch) Deprecated.org.apache.hadoop.hbase.client.ScanScan.setAsyncPrefetch(boolean asyncPrefetch) org.apache.hadoop.hbase.client.ScanImmutableScan.setAttribute(String name, byte[] value) org.apache.hadoop.hbase.client.ScanScan.setAttribute(String name, byte[] value) org.apache.hadoop.hbase.client.ScanImmutableScan.setAuthorizations(org.apache.hadoop.hbase.security.visibility.Authorizations authorizations) org.apache.hadoop.hbase.client.ScanScan.setAuthorizations(org.apache.hadoop.hbase.security.visibility.Authorizations authorizations) org.apache.hadoop.hbase.client.ScanImmutableScan.setBatch(int batch) org.apache.hadoop.hbase.client.ScanScan.setBatch(int batch) Set the maximum number of cells to return for each call to next().org.apache.hadoop.hbase.client.ScanImmutableScan.setCacheBlocks(boolean cacheBlocks) org.apache.hadoop.hbase.client.ScanScan.setCacheBlocks(boolean cacheBlocks) Set whether blocks should be cached for this Scan.org.apache.hadoop.hbase.client.ScanImmutableScan.setCaching(int caching) org.apache.hadoop.hbase.client.ScanScan.setCaching(int caching) Set the number of rows for caching that will be passed to scanners.org.apache.hadoop.hbase.client.ScanImmutableScan.setColumnFamilyTimeRange(byte[] cf, long minStamp, long maxStamp) org.apache.hadoop.hbase.client.ScanScan.setColumnFamilyTimeRange(byte[] cf, long minStamp, long maxStamp) org.apache.hadoop.hbase.client.ScanImmutableScan.setConsistency(org.apache.hadoop.hbase.client.Consistency consistency) org.apache.hadoop.hbase.client.ScanScan.setConsistency(org.apache.hadoop.hbase.client.Consistency consistency) org.apache.hadoop.hbase.client.ScanImmutableScan.setFamilyMap(Map<byte[], NavigableSet<byte[]>> familyMap) org.apache.hadoop.hbase.client.ScanScan.setFamilyMap(Map<byte[], NavigableSet<byte[]>> familyMap) Setting the familyMaporg.apache.hadoop.hbase.client.ScanImmutableScan.setFilter(org.apache.hadoop.hbase.filter.Filter filter) org.apache.hadoop.hbase.client.ScanScan.setFilter(org.apache.hadoop.hbase.filter.Filter filter) org.apache.hadoop.hbase.client.ScanImmutableScan.setId(String id) org.apache.hadoop.hbase.client.ScanScan.setId(String id) org.apache.hadoop.hbase.client.ScanImmutableScan.setIsolationLevel(org.apache.hadoop.hbase.client.IsolationLevel level) org.apache.hadoop.hbase.client.ScanScan.setIsolationLevel(org.apache.hadoop.hbase.client.IsolationLevel level) org.apache.hadoop.hbase.client.ScanImmutableScan.setLimit(int limit) org.apache.hadoop.hbase.client.ScanScan.setLimit(int limit) Set the limit of rows for this scan.org.apache.hadoop.hbase.client.ScanImmutableScan.setLoadColumnFamiliesOnDemand(boolean value) org.apache.hadoop.hbase.client.ScanScan.setLoadColumnFamiliesOnDemand(boolean value) org.apache.hadoop.hbase.client.ScanImmutableScan.setMaxResultSize(long maxResultSize) org.apache.hadoop.hbase.client.ScanScan.setMaxResultSize(long maxResultSize) Set the maximum result size.org.apache.hadoop.hbase.client.ScanImmutableScan.setMaxResultsPerColumnFamily(int limit) org.apache.hadoop.hbase.client.ScanScan.setMaxResultsPerColumnFamily(int limit) Set the maximum number of values to return per row per Column Familyorg.apache.hadoop.hbase.client.ScanScan.setMaxVersions()Deprecated.since 2.0.0 and will be removed in 3.0.0.org.apache.hadoop.hbase.client.ScanScan.setMaxVersions(int maxVersions) Deprecated.since 2.0.0 and will be removed in 3.0.0.org.apache.hadoop.hbase.client.ScanImmutableScan.setNeedCursorResult(boolean needCursorResult) org.apache.hadoop.hbase.client.ScanScan.setNeedCursorResult(boolean needCursorResult) When the server is slow or we scan a table with many deleted data or we use a sparse filter, the server will response heartbeat to prevent timeout.org.apache.hadoop.hbase.client.ScanImmutableScan.setOneRowLimit()org.apache.hadoop.hbase.client.ScanScan.setOneRowLimit()Call this when you only want to get one row.org.apache.hadoop.hbase.client.ScanImmutableScan.setPriority(int priority) org.apache.hadoop.hbase.client.ScanScan.setPriority(int priority) org.apache.hadoop.hbase.client.ScanImmutableScan.setRaw(boolean raw) org.apache.hadoop.hbase.client.ScanScan.setRaw(boolean raw) Enable/disable "raw" mode for this scan.org.apache.hadoop.hbase.client.ScanImmutableScan.setReadType(org.apache.hadoop.hbase.client.Scan.ReadType readType) org.apache.hadoop.hbase.client.ScanScan.setReadType(org.apache.hadoop.hbase.client.Scan.ReadType readType) Set the read type for this scan.org.apache.hadoop.hbase.client.ScanImmutableScan.setReplicaId(int id) org.apache.hadoop.hbase.client.ScanScan.setReplicaId(int Id) org.apache.hadoop.hbase.client.ScanImmutableScan.setReversed(boolean reversed) org.apache.hadoop.hbase.client.ScanScan.setReversed(boolean reversed) Set whether this scan is a reversed oneorg.apache.hadoop.hbase.client.ScanImmutableScan.setRowOffsetPerColumnFamily(int offset) org.apache.hadoop.hbase.client.ScanScan.setRowOffsetPerColumnFamily(int offset) Set offset for the row per Column Family.org.apache.hadoop.hbase.client.ScanScan.setRowPrefixFilter(byte[] rowPrefix) Deprecated.since 2.5.0, will be removed in 4.0.0.org.apache.hadoop.hbase.client.ScanImmutableScan.setScanMetricsEnabled(boolean enabled) org.apache.hadoop.hbase.client.ScanScan.setScanMetricsEnabled(boolean enabled) Enable collection ofScanMetrics.org.apache.hadoop.hbase.client.ScanImmutableScan.setSmall(boolean small) Deprecated.org.apache.hadoop.hbase.client.ScanScan.setSmall(boolean small) Deprecated.since 2.0.0 and will be removed in 3.0.0.org.apache.hadoop.hbase.client.ScanScan.setStartRow(byte[] startRow) Deprecated.since 2.0.0 and will be removed in 3.0.0.org.apache.hadoop.hbase.client.ScanImmutableScan.setStartStopRowForPrefixScan(byte[] rowPrefix) org.apache.hadoop.hbase.client.ScanScan.setStartStopRowForPrefixScan(byte[] rowPrefix) Set a filter (using stopRow and startRow) so the result set only contains rows where the rowKey starts with the specified prefix.org.apache.hadoop.hbase.client.ScanScan.setStopRow(byte[] stopRow) Deprecated.since 2.0.0 and will be removed in 3.0.0.org.apache.hadoop.hbase.client.ScanImmutableScan.setTimeRange(long minStamp, long maxStamp) org.apache.hadoop.hbase.client.ScanScan.setTimeRange(long minStamp, long maxStamp) Get versions of columns only within the specified timestamp range, [minStamp, maxStamp).org.apache.hadoop.hbase.client.ScanImmutableScan.setTimestamp(long timestamp) org.apache.hadoop.hbase.client.ScanScan.setTimestamp(long timestamp) Get versions of columns with the specified timestamp.org.apache.hadoop.hbase.client.ScanImmutableScan.setTimeStamp(long timestamp) Deprecated.org.apache.hadoop.hbase.client.ScanScan.setTimeStamp(long timestamp) Deprecated.As of release 2.0.0, this will be removed in HBase 3.0.0.org.apache.hadoop.hbase.client.ScanImmutableScan.withStartRow(byte[] startRow) org.apache.hadoop.hbase.client.ScanImmutableScan.withStartRow(byte[] startRow, boolean inclusive) org.apache.hadoop.hbase.client.ScanScan.withStartRow(byte[] startRow) Set the start row of the scan.org.apache.hadoop.hbase.client.ScanScan.withStartRow(byte[] startRow, boolean inclusive) Set the start row of the scan.org.apache.hadoop.hbase.client.ScanImmutableScan.withStopRow(byte[] stopRow) org.apache.hadoop.hbase.client.ScanImmutableScan.withStopRow(byte[] stopRow, boolean inclusive) org.apache.hadoop.hbase.client.ScanScan.withStopRow(byte[] stopRow) Set the stop row of the scan.org.apache.hadoop.hbase.client.ScanScan.withStopRow(byte[] stopRow, boolean inclusive) Set the stop row of the scan.Methods in org.apache.hadoop.hbase.client that return types with arguments of type org.apache.hadoop.hbase.client.Scan in inModifier and TypeMethodDescriptionOptional<org.apache.hadoop.hbase.client.Scan>OnlineLogRecord.getScan()If "hbase.slowlog.scan.payload.enabled" is enabled then this value may be present and should represent the Scan that produced the givenOnlineLogRecordMethods in org.apache.hadoop.hbase.client with parameters of type org.apache.hadoop.hbase.client.Scan in inModifier and TypeMethodDescriptionstatic org.apache.hadoop.hbase.client.ScanResultCacheConnectionUtils.createScanResultCache(org.apache.hadoop.hbase.client.Scan scan) static longPackagePrivateFieldAccessor.getMvccReadPoint(org.apache.hadoop.hbase.client.Scan scan) org.apache.hadoop.hbase.client.ResultScannerAsyncTable.getScanner(org.apache.hadoop.hbase.client.Scan scan) Returns a scanner on the current table as specified by theScanobject.org.apache.hadoop.hbase.client.ResultScannerHTable.getScanner(org.apache.hadoop.hbase.client.Scan scan) The underlyingHTablemust not be closed.default org.apache.hadoop.hbase.client.ResultScannerTable.getScanner(org.apache.hadoop.hbase.client.Scan scan) Returns a scanner on the current table as specified by theScanobject.protected voidAbstractClientScanner.initScanMetrics(org.apache.hadoop.hbase.client.Scan scan) Check and initialize if application wants to collect scan metricsvoidAsyncTable.scan(org.apache.hadoop.hbase.client.Scan scan, C consumer) The scan API uses the observer pattern.CompletableFuture<List<org.apache.hadoop.hbase.client.Result>>AsyncTable.scanAll(org.apache.hadoop.hbase.client.Scan scan) Return all the results that match the given scan object.static voidPackagePrivateFieldAccessor.setMvccReadPoint(org.apache.hadoop.hbase.client.Scan scan, long mvccReadPoint) org.apache.hadoop.hbase.client.OnlineLogRecord.OnlineLogRecordBuilderOnlineLogRecord.OnlineLogRecordBuilder.setScan(org.apache.hadoop.hbase.client.Scan scan) Constructors in org.apache.hadoop.hbase.client with parameters of type org.apache.hadoop.hbase.client.Scan in inModifierConstructorDescriptionClientAsyncPrefetchScanner(org.apache.hadoop.conf.Configuration configuration, org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.client.Scan scanForMetrics, org.apache.hadoop.hbase.TableName name, org.apache.hadoop.hbase.client.ClusterConnection connection, org.apache.hadoop.hbase.client.RpcRetryingCallerFactory rpcCallerFactory, org.apache.hadoop.hbase.ipc.RpcControllerFactory rpcControllerFactory, ExecutorService pool, int scanReadRpcTimeout, int scannerTimeout, int replicaCallTimeoutMicroSecondScan, org.apache.hadoop.hbase.client.ConnectionConfiguration connectionConfiguration, Map<String, byte[]> requestAttributes) ClientScanner(org.apache.hadoop.conf.Configuration conf, org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.client.Scan scanForMetrics, org.apache.hadoop.hbase.TableName tableName, org.apache.hadoop.hbase.client.ClusterConnection connection, org.apache.hadoop.hbase.client.RpcRetryingCallerFactory rpcFactory, org.apache.hadoop.hbase.ipc.RpcControllerFactory controllerFactory, ExecutorService pool, int scanReadRpcTimeout, int scannerTimeout, int primaryOperationTimeout, org.apache.hadoop.hbase.client.ConnectionConfiguration connectionConfiguration, Map<String, byte[]> requestAttributes) Create a new ClientScanner for the specified table Note that the passedScan's start row maybe changed changed.ClientSideRegionScanner(org.apache.hadoop.conf.Configuration conf, org.apache.hadoop.fs.FileSystem fs, org.apache.hadoop.fs.Path rootDir, org.apache.hadoop.hbase.client.TableDescriptor htd, org.apache.hadoop.hbase.client.RegionInfo hri, org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.client.metrics.ScanMetrics scanMetrics) ClientSimpleScanner(org.apache.hadoop.conf.Configuration configuration, org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.client.Scan scanForMetrics, org.apache.hadoop.hbase.TableName name, org.apache.hadoop.hbase.client.ClusterConnection connection, org.apache.hadoop.hbase.client.RpcRetryingCallerFactory rpcCallerFactory, org.apache.hadoop.hbase.ipc.RpcControllerFactory rpcControllerFactory, ExecutorService pool, int scanReadRpcTimeout, int scannerTimeout, int replicaCallTimeoutMicroSecondScan, org.apache.hadoop.hbase.client.ConnectionConfiguration connectionConfiguration, Map<String, byte[]> requestAttributes) ImmutableScan(org.apache.hadoop.hbase.client.Scan scan) Create Immutable instance of Scan from given Scan objectReversedClientScanner(org.apache.hadoop.conf.Configuration conf, org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.client.Scan scanForMetrics, org.apache.hadoop.hbase.TableName tableName, org.apache.hadoop.hbase.client.ClusterConnection connection, org.apache.hadoop.hbase.client.RpcRetryingCallerFactory rpcFactory, org.apache.hadoop.hbase.ipc.RpcControllerFactory controllerFactory, ExecutorService pool, int scanReadRpcTimeout, int scannerTimeout, int primaryOperationTimeout, org.apache.hadoop.hbase.client.ConnectionConfiguration connectionConfiguration, Map<String, byte[]> requestAttributes) Create a new ReversibleClientScanner for the specified table Note that the passedScan's start row maybe changed.ReversedScannerCallable(org.apache.hadoop.hbase.client.ClusterConnection connection, org.apache.hadoop.hbase.TableName tableName, org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.client.metrics.ScanMetrics scanMetrics, org.apache.hadoop.hbase.ipc.RpcControllerFactory rpcFactory, int replicaId, Map<String, byte[]> requestAttributes) Scan(org.apache.hadoop.hbase.client.Scan scan) Creates a new instance of this class while copying all values.ScannerCallable(org.apache.hadoop.hbase.client.ClusterConnection connection, org.apache.hadoop.hbase.TableName tableName, org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.client.metrics.ScanMetrics scanMetrics, org.apache.hadoop.hbase.ipc.RpcControllerFactory rpcControllerFactory, int id, Map<String, byte[]> requestAttributes) TableSnapshotScanner(org.apache.hadoop.conf.Configuration conf, org.apache.hadoop.fs.Path restoreDir, String snapshotName, org.apache.hadoop.hbase.client.Scan scan) Creates a TableSnapshotScanner.TableSnapshotScanner(org.apache.hadoop.conf.Configuration conf, org.apache.hadoop.fs.Path rootDir, org.apache.hadoop.fs.Path restoreDir, String snapshotName, org.apache.hadoop.hbase.client.Scan scan) TableSnapshotScanner(org.apache.hadoop.conf.Configuration conf, org.apache.hadoop.fs.Path rootDir, org.apache.hadoop.fs.Path restoreDir, String snapshotName, org.apache.hadoop.hbase.client.Scan scan, boolean snapshotAlreadyRestored) Creates a TableSnapshotScanner. -
Uses of org.apache.hadoop.hbase.client.Scan in org.apache.hadoop.hbase.client.trace
Methods in org.apache.hadoop.hbase.client.trace with parameters of type org.apache.hadoop.hbase.client.Scan in inModifier and TypeMethodDescriptionorg.apache.hadoop.hbase.client.trace.TableOperationSpanBuilderTableOperationSpanBuilder.setOperation(org.apache.hadoop.hbase.client.Scan scan) -
Uses of org.apache.hadoop.hbase.client.Scan in org.apache.hadoop.hbase.coprocessor
Methods in org.apache.hadoop.hbase.coprocessor with parameters of type org.apache.hadoop.hbase.client.Scan in inModifier and TypeMethodDescriptiondefault org.apache.hadoop.hbase.regionserver.RegionScannerRegionObserver.postScannerOpen(org.apache.hadoop.hbase.coprocessor.ObserverContext<org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment> c, org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.regionserver.RegionScanner s) Called after the client opens a new scanner.default voidRegionObserver.preScannerOpen(org.apache.hadoop.hbase.coprocessor.ObserverContext<org.apache.hadoop.hbase.coprocessor.RegionCoprocessorEnvironment> c, org.apache.hadoop.hbase.client.Scan scan) Called before the client opens a new scanner. -
Uses of org.apache.hadoop.hbase.client.Scan in org.apache.hadoop.hbase.io
Methods in org.apache.hadoop.hbase.io with parameters of type org.apache.hadoop.hbase.client.Scan in inModifier and TypeMethodDescriptionbooleanHalfStoreFileReader.passesKeyRangeFilter(org.apache.hadoop.hbase.client.Scan scan) -
Uses of org.apache.hadoop.hbase.client.Scan in org.apache.hadoop.hbase.mapred
Method parameters in org.apache.hadoop.hbase.mapred with type arguments of type org.apache.hadoop.hbase.client.Scan in inModifier and TypeMethodDescriptionstatic voidTableMapReduceUtil.initMultiTableSnapshotMapperJob(Map<String, Collection<org.apache.hadoop.hbase.client.Scan>> snapshotScans, Class<? extends org.apache.hadoop.hbase.mapred.TableMap> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapred.JobConf job, boolean addDependencyJars, org.apache.hadoop.fs.Path tmpRestoreDir) Sets up the job for reading from one or more multiple table snapshots, with one or more scans per snapshot.static voidMultiTableSnapshotInputFormat.setInput(org.apache.hadoop.conf.Configuration conf, Map<String, Collection<org.apache.hadoop.hbase.client.Scan>> snapshotScans, org.apache.hadoop.fs.Path restoreDir) Constructors in org.apache.hadoop.hbase.mapred with parameters of type org.apache.hadoop.hbase.client.Scan in in -
Uses of org.apache.hadoop.hbase.client.Scan in org.apache.hadoop.hbase.mapreduce
Methods in org.apache.hadoop.hbase.mapreduce that return org.apache.hadoop.hbase.client.Scan in inModifier and TypeMethodDescriptionstatic org.apache.hadoop.hbase.client.ScanTableMapReduceUtil.convertStringToScan(String base64) Converts the given Base64 string back into a Scan instance.static org.apache.hadoop.hbase.client.ScanTableInputFormat.createScanFromConfiguration(org.apache.hadoop.conf.Configuration conf) Sets up aScaninstance, applying settings from the configuration property constants defined inTableInputFormat.static org.apache.hadoop.hbase.client.ScanTableSnapshotInputFormatImpl.extractScanFromConf(org.apache.hadoop.conf.Configuration conf) org.apache.hadoop.hbase.client.ScanTableInputFormatBase.getScan()Gets the scan defining the actual details like columns etc.org.apache.hadoop.hbase.client.ScanTableSplit.getScan()Returns a Scan object from the stored string representation.Methods in org.apache.hadoop.hbase.mapreduce that return types with arguments of type org.apache.hadoop.hbase.client.Scan in inModifier and TypeMethodDescriptionstatic org.apache.hadoop.hbase.util.Triple<org.apache.hadoop.hbase.TableName,org.apache.hadoop.hbase.client.Scan, org.apache.hadoop.fs.Path> ExportUtils.getArgumentsFromCommandLine(org.apache.hadoop.conf.Configuration conf, String[] args) protected List<org.apache.hadoop.hbase.client.Scan>MultiTableInputFormatBase.getScans()Allows subclasses to get the list ofScanobjects.Map<String,Collection<org.apache.hadoop.hbase.client.Scan>> MultiTableSnapshotInputFormatImpl.getSnapshotsToScans(org.apache.hadoop.conf.Configuration conf) Retrieve the snapshot name -> list<scan> mapping pushed to configuration byMultiTableSnapshotInputFormatImpl.setSnapshotToScans(Configuration, Map)Methods in org.apache.hadoop.hbase.mapreduce with parameters of type org.apache.hadoop.hbase.client.Scan in inModifier and TypeMethodDescriptionstatic voidTableInputFormat.addColumns(org.apache.hadoop.hbase.client.Scan scan, byte[][] columns) Adds an array of columns specified using old format, family:qualifier.static StringTableMapReduceUtil.convertScanToString(org.apache.hadoop.hbase.client.Scan scan) Writes the given scan into a Base64 encoded string.static List<org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl.InputSplit>TableSnapshotInputFormatImpl.getSplits(org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.snapshot.SnapshotManifest manifest, List<org.apache.hadoop.hbase.HRegionInfo> regionManifests, org.apache.hadoop.fs.Path restoreDir, org.apache.hadoop.conf.Configuration conf) static List<org.apache.hadoop.hbase.mapreduce.TableSnapshotInputFormatImpl.InputSplit>TableSnapshotInputFormatImpl.getSplits(org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.snapshot.SnapshotManifest manifest, List<org.apache.hadoop.hbase.HRegionInfo> regionManifests, org.apache.hadoop.fs.Path restoreDir, org.apache.hadoop.conf.Configuration conf, org.apache.hadoop.hbase.util.RegionSplitter.SplitAlgorithm sa, int numSplits) static voidGroupingTableMapper.initJob(String table, org.apache.hadoop.hbase.client.Scan scan, String groupColumns, Class<? extends org.apache.hadoop.hbase.mapreduce.TableMapper> mapper, org.apache.hadoop.mapreduce.Job job) Use this before submitting a TableMap job.static voidIdentityTableMapper.initJob(String table, org.apache.hadoop.hbase.client.Scan scan, Class<? extends org.apache.hadoop.hbase.mapreduce.TableMapper> mapper, org.apache.hadoop.mapreduce.Job job) Use this before submitting a TableMap job.static voidTableMapReduceUtil.initTableMapperJob(byte[] table, org.apache.hadoop.hbase.client.Scan scan, Class<? extends org.apache.hadoop.hbase.mapreduce.TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job) Use this before submitting a TableMap job.static voidTableMapReduceUtil.initTableMapperJob(byte[] table, org.apache.hadoop.hbase.client.Scan scan, Class<? extends org.apache.hadoop.hbase.mapreduce.TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars) Use this before submitting a TableMap job.static voidTableMapReduceUtil.initTableMapperJob(byte[] table, org.apache.hadoop.hbase.client.Scan scan, Class<? extends org.apache.hadoop.hbase.mapreduce.TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars, Class<? extends org.apache.hadoop.mapreduce.InputFormat> inputFormatClass) Use this before submitting a TableMap job.static voidTableMapReduceUtil.initTableMapperJob(String table, org.apache.hadoop.hbase.client.Scan scan, Class<? extends org.apache.hadoop.hbase.mapreduce.TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job) Use this before submitting a TableMap job.static voidTableMapReduceUtil.initTableMapperJob(String table, org.apache.hadoop.hbase.client.Scan scan, Class<? extends org.apache.hadoop.hbase.mapreduce.TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars) Use this before submitting a TableMap job.static voidTableMapReduceUtil.initTableMapperJob(String table, org.apache.hadoop.hbase.client.Scan scan, Class<? extends org.apache.hadoop.hbase.mapreduce.TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars, boolean initCredentials, Class<? extends org.apache.hadoop.mapreduce.InputFormat> inputFormatClass) Use this before submitting a TableMap job.static voidTableMapReduceUtil.initTableMapperJob(String table, org.apache.hadoop.hbase.client.Scan scan, Class<? extends org.apache.hadoop.hbase.mapreduce.TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars, Class<? extends org.apache.hadoop.mapreduce.InputFormat> inputFormatClass) Use this before submitting a TableMap job.static voidTableMapReduceUtil.initTableMapperJob(org.apache.hadoop.hbase.TableName table, org.apache.hadoop.hbase.client.Scan scan, Class<? extends org.apache.hadoop.hbase.mapreduce.TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job) Use this before submitting a TableMap job.static voidTableMapReduceUtil.initTableSnapshotMapperJob(String snapshotName, org.apache.hadoop.hbase.client.Scan scan, Class<? extends org.apache.hadoop.hbase.mapreduce.TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars, org.apache.hadoop.fs.Path tmpRestoreDir) Sets up the job for reading from a table snapshot.static voidTableMapReduceUtil.initTableSnapshotMapperJob(String snapshotName, org.apache.hadoop.hbase.client.Scan scan, Class<? extends org.apache.hadoop.hbase.mapreduce.TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars, org.apache.hadoop.fs.Path tmpRestoreDir, org.apache.hadoop.hbase.util.RegionSplitter.SplitAlgorithm splitAlgo, int numSplitsPerRegion) Sets up the job for reading from a table snapshot.voidTableInputFormatBase.setScan(org.apache.hadoop.hbase.client.Scan scan) Sets the scan defining the actual details like columns etc.voidTableRecordReader.setScan(org.apache.hadoop.hbase.client.Scan scan) Sets the scan defining the actual details like columns etc.voidTableRecordReaderImpl.setScan(org.apache.hadoop.hbase.client.Scan scan) Sets the scan defining the actual details like columns etc.Method parameters in org.apache.hadoop.hbase.mapreduce with type arguments of type org.apache.hadoop.hbase.client.Scan in inModifier and TypeMethodDescriptionstatic voidTableMapReduceUtil.initMultiTableSnapshotMapperJob(Map<String, Collection<org.apache.hadoop.hbase.client.Scan>> snapshotScans, Class<? extends org.apache.hadoop.hbase.mapreduce.TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars, org.apache.hadoop.fs.Path tmpRestoreDir) Sets up the job for reading from one or more table snapshots, with one or more scans per snapshot.static voidTableMapReduceUtil.initTableMapperJob(List<org.apache.hadoop.hbase.client.Scan> scans, Class<? extends org.apache.hadoop.hbase.mapreduce.TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job) Use this before submitting a Multi TableMap job.static voidTableMapReduceUtil.initTableMapperJob(List<org.apache.hadoop.hbase.client.Scan> scans, Class<? extends org.apache.hadoop.hbase.mapreduce.TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars) Use this before submitting a Multi TableMap job.static voidTableMapReduceUtil.initTableMapperJob(List<org.apache.hadoop.hbase.client.Scan> scans, Class<? extends org.apache.hadoop.hbase.mapreduce.TableMapper> mapper, Class<?> outputKeyClass, Class<?> outputValueClass, org.apache.hadoop.mapreduce.Job job, boolean addDependencyJars, boolean initCredentials) Use this before submitting a Multi TableMap job.static voidMultiTableSnapshotInputFormat.setInput(org.apache.hadoop.conf.Configuration configuration, Map<String, Collection<org.apache.hadoop.hbase.client.Scan>> snapshotScans, org.apache.hadoop.fs.Path tmpRestoreDir) voidMultiTableSnapshotInputFormatImpl.setInput(org.apache.hadoop.conf.Configuration conf, Map<String, Collection<org.apache.hadoop.hbase.client.Scan>> snapshotScans, org.apache.hadoop.fs.Path restoreDir) Configure conf to read from snapshotScans, with snapshots restored to a subdirectory of restoreDir.protected voidMultiTableInputFormatBase.setScans(List<org.apache.hadoop.hbase.client.Scan> scans) Allows subclasses to set the list ofScanobjects.voidMultiTableSnapshotInputFormatImpl.setSnapshotToScans(org.apache.hadoop.conf.Configuration conf, Map<String, Collection<org.apache.hadoop.hbase.client.Scan>> snapshotScans) Push snapshotScans to conf (under the keyMultiTableSnapshotInputFormatImpl.SNAPSHOT_TO_SCANS_KEY)Constructors in org.apache.hadoop.hbase.mapreduce with parameters of type org.apache.hadoop.hbase.client.Scan in inModifierConstructorDescriptionInputSplit(org.apache.hadoop.hbase.client.TableDescriptor htd, org.apache.hadoop.hbase.HRegionInfo regionInfo, List<String> locations, org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.fs.Path restoreDir) TableSnapshotRegionSplit(org.apache.hadoop.hbase.HTableDescriptor htd, org.apache.hadoop.hbase.HRegionInfo regionInfo, List<String> locations, org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.fs.Path restoreDir) TableSplit(org.apache.hadoop.hbase.TableName tableName, org.apache.hadoop.hbase.client.Scan scan, byte[] startRow, byte[] endRow, String location) Creates a new instance while assigning all variables.TableSplit(org.apache.hadoop.hbase.TableName tableName, org.apache.hadoop.hbase.client.Scan scan, byte[] startRow, byte[] endRow, String location, long length) Creates a new instance while assigning all variables.TableSplit(org.apache.hadoop.hbase.TableName tableName, org.apache.hadoop.hbase.client.Scan scan, byte[] startRow, byte[] endRow, String location, String encodedRegionName, long length) Creates a new instance while assigning all variables. -
Uses of org.apache.hadoop.hbase.client.Scan in org.apache.hadoop.hbase.mob
Methods in org.apache.hadoop.hbase.mob with parameters of type org.apache.hadoop.hbase.client.Scan in inModifier and TypeMethodDescriptionstatic booleanMobUtils.isCacheMobBlocks(org.apache.hadoop.hbase.client.Scan scan) Indicates whether the scan contains the information of caching blocks.static booleanMobUtils.isRawMobScan(org.apache.hadoop.hbase.client.Scan scan) Indicates whether it's a raw scan.static booleanMobUtils.isReadEmptyValueOnMobCellMiss(org.apache.hadoop.hbase.client.Scan scan) Indicates whether return null value when the mob file is missing or corrupt.static booleanMobUtils.isRefOnlyScan(org.apache.hadoop.hbase.client.Scan scan) Indicates whether it's a reference only scan.static voidMobUtils.setCacheMobBlocks(org.apache.hadoop.hbase.client.Scan scan, boolean cacheBlocks) Sets the attribute of caching blocks in the scan. -
Uses of org.apache.hadoop.hbase.client.Scan in org.apache.hadoop.hbase.quotas
Methods in org.apache.hadoop.hbase.quotas that return org.apache.hadoop.hbase.client.Scan in inModifier and TypeMethodDescriptionstatic org.apache.hadoop.hbase.client.ScanQuotaTableUtil.makeQuotaSnapshotScan()Creates aScanwhich returns only quota snapshots from the quota table.static org.apache.hadoop.hbase.client.ScanQuotaTableUtil.makeQuotaSnapshotScanForTable(org.apache.hadoop.hbase.TableName tn) Creates aScanwhich returns onlySpaceQuotaSnapshotfrom the quota table for a specific table.static org.apache.hadoop.hbase.client.ScanQuotaTableUtil.makeScan(org.apache.hadoop.hbase.quotas.QuotaFilter filter) -
Uses of org.apache.hadoop.hbase.client.Scan in org.apache.hadoop.hbase.regionserver
Subclasses of org.apache.hadoop.hbase.client.Scan in in org.apache.hadoop.hbase.regionserverModifier and TypeClassDescriptionclassorg.apache.hadoop.hbase.regionserver.InternalScanSpecial scanner, currently used for increment operations to allow additional server-side arguments for Scan operations.Methods in org.apache.hadoop.hbase.regionserver that return org.apache.hadoop.hbase.client.Scan in inModifier and TypeMethodDescriptionorg.apache.hadoop.hbase.client.ScanCustomizedScanInfoBuilder.getScan()org.apache.hadoop.hbase.client.ScanScanOptions.getScan()Returns a copy of the Scan object.Methods in org.apache.hadoop.hbase.regionserver with parameters of type org.apache.hadoop.hbase.client.Scan in inModifier and TypeMethodDescriptionprotected org.apache.hadoop.hbase.regionserver.KeyValueScannerHMobStore.createScanner(org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.regionserver.ScanInfo scanInfo, NavigableSet<byte[]> targetCols, long readPt) Gets the MobStoreScanner or MobReversedStoreScanner.protected org.apache.hadoop.hbase.regionserver.KeyValueScannerHStore.createScanner(org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.regionserver.ScanInfo scanInfo, NavigableSet<byte[]> targetCols, long readPt) org.apache.hadoop.hbase.regionserver.RegionScannerImplHRegion.getScanner(org.apache.hadoop.hbase.client.Scan scan) org.apache.hadoop.hbase.regionserver.RegionScannerImplHRegion.getScanner(org.apache.hadoop.hbase.client.Scan scan, List<org.apache.hadoop.hbase.regionserver.KeyValueScanner> additionalScanners) org.apache.hadoop.hbase.regionserver.KeyValueScannerHStore.getScanner(org.apache.hadoop.hbase.client.Scan scan, NavigableSet<byte[]> targetCols, long readPt) Return a scanner for both the memstore and the HStore files.org.apache.hadoop.hbase.regionserver.RegionScannerRegion.getScanner(org.apache.hadoop.hbase.client.Scan scan) Return an iterator that scans over the HRegion, returning the indicated columns and rows specified by theScan.org.apache.hadoop.hbase.regionserver.RegionScannerRegion.getScanner(org.apache.hadoop.hbase.client.Scan scan, List<org.apache.hadoop.hbase.regionserver.KeyValueScanner> additionalScanners) Return an iterator that scans over the HRegion, returning the indicated columns and rows specified by theScan.protected org.apache.hadoop.hbase.regionserver.RegionScannerImplHRegion.instantiateRegionScanner(org.apache.hadoop.hbase.client.Scan scan, List<org.apache.hadoop.hbase.regionserver.KeyValueScanner> additionalScanners, long nonceGroup, long nonce) booleanStoreFileReader.passesKeyRangeFilter(org.apache.hadoop.hbase.client.Scan scan) Checks whether the given scan rowkey range overlaps with the current storefile'sorg.apache.hadoop.hbase.regionserver.RegionScannerRegionCoprocessorHost.postScannerOpen(org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.regionserver.RegionScanner s) voidRegionCoprocessorHost.preScannerOpen(org.apache.hadoop.hbase.client.Scan scan) org.apache.hadoop.hbase.regionserver.ScanInfoRegionCoprocessorHost.preStoreScannerOpen(org.apache.hadoop.hbase.regionserver.HStore store, org.apache.hadoop.hbase.client.Scan scan) Called before open store scanner for user scan.booleanKeyValueScanner.shouldUseScanner(org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.regionserver.HStore store, long oldestUnexpiredTS) Allows to filter out scanners (both StoreFile and memstore) that we don't want to use based on criteria such as Bloom filters and timestamp ranges.booleanNonLazyKeyValueScanner.shouldUseScanner(org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.regionserver.HStore store, long oldestUnexpiredTS) booleanSegmentScanner.shouldUseScanner(org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.regionserver.HStore store, long oldestUnexpiredTS) This functionality should be resolved in the higher level which is MemStoreScanner, currently returns true as default.booleanStoreFileScanner.shouldUseScanner(org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.regionserver.HStore store, long oldestUnexpiredTS) Constructors in org.apache.hadoop.hbase.regionserver with parameters of type org.apache.hadoop.hbase.client.Scan in inModifierConstructorDescriptionCustomizedScanInfoBuilder(org.apache.hadoop.hbase.regionserver.ScanInfo scanInfo, org.apache.hadoop.hbase.client.Scan scan) InternalScan(org.apache.hadoop.hbase.client.Scan scan) MobStoreScanner(org.apache.hadoop.hbase.regionserver.HStore store, org.apache.hadoop.hbase.regionserver.ScanInfo scanInfo, org.apache.hadoop.hbase.client.Scan scan, NavigableSet<byte[]> columns, long readPt) ReversedStoreScanner(org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.regionserver.ScanInfo scanInfo, NavigableSet<byte[]> columns, List<? extends org.apache.hadoop.hbase.regionserver.KeyValueScanner> scanners) Constructor for testing.ReversedStoreScanner(org.apache.hadoop.hbase.regionserver.HStore store, org.apache.hadoop.hbase.regionserver.ScanInfo scanInfo, org.apache.hadoop.hbase.client.Scan scan, NavigableSet<byte[]> columns, long readPt) Opens a scanner across memstore, snapshot, and all StoreFiles.StoreScanner(org.apache.hadoop.hbase.regionserver.HStore store, org.apache.hadoop.hbase.regionserver.ScanInfo scanInfo, org.apache.hadoop.hbase.client.Scan scan, NavigableSet<byte[]> columns, long readPt) Opens a scanner across memstore, snapshot, and all StoreFiles. -
Uses of org.apache.hadoop.hbase.client.Scan in org.apache.hadoop.hbase.regionserver.querymatcher
Methods in org.apache.hadoop.hbase.regionserver.querymatcher with parameters of type org.apache.hadoop.hbase.client.Scan in inModifier and TypeMethodDescriptionstatic org.apache.hadoop.hbase.regionserver.querymatcher.NormalUserScanQueryMatcherNormalUserScanQueryMatcher.create(org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.regionserver.ScanInfo scanInfo, org.apache.hadoop.hbase.regionserver.querymatcher.ColumnTracker columns, org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker deletes, boolean hasNullColumn, long oldestUnexpiredTS, long now) static org.apache.hadoop.hbase.regionserver.querymatcher.RawScanQueryMatcherRawScanQueryMatcher.create(org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.regionserver.ScanInfo scanInfo, org.apache.hadoop.hbase.regionserver.querymatcher.ColumnTracker columns, boolean hasNullColumn, long oldestUnexpiredTS, long now) static org.apache.hadoop.hbase.regionserver.querymatcher.UserScanQueryMatcherUserScanQueryMatcher.create(org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.regionserver.ScanInfo scanInfo, NavigableSet<byte[]> columns, long oldestUnexpiredTS, long now, org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost regionCoprocessorHost) protected static org.apache.hadoop.hbase.util.Pair<org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker,org.apache.hadoop.hbase.regionserver.querymatcher.ColumnTracker> ScanQueryMatcher.getTrackers(org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost host, NavigableSet<byte[]> columns, org.apache.hadoop.hbase.regionserver.ScanInfo scanInfo, long oldestUnexpiredTS, org.apache.hadoop.hbase.client.Scan userScan) Constructors in org.apache.hadoop.hbase.regionserver.querymatcher with parameters of type org.apache.hadoop.hbase.client.Scan in inModifierConstructorDescriptionprotectedNormalUserScanQueryMatcher(org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.regionserver.ScanInfo scanInfo, org.apache.hadoop.hbase.regionserver.querymatcher.ColumnTracker columns, boolean hasNullColumn, org.apache.hadoop.hbase.regionserver.querymatcher.DeleteTracker deletes, long oldestUnexpiredTS, long now) protectedRawScanQueryMatcher(org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.regionserver.ScanInfo scanInfo, org.apache.hadoop.hbase.regionserver.querymatcher.ColumnTracker columns, boolean hasNullColumn, long oldestUnexpiredTS, long now) protectedUserScanQueryMatcher(org.apache.hadoop.hbase.client.Scan scan, org.apache.hadoop.hbase.regionserver.ScanInfo scanInfo, org.apache.hadoop.hbase.regionserver.querymatcher.ColumnTracker columns, boolean hasNullColumn, long oldestUnexpiredTS, long now) -
Uses of org.apache.hadoop.hbase.client.Scan in org.apache.hadoop.hbase.shaded.protobuf
Methods in org.apache.hadoop.hbase.shaded.protobuf that return org.apache.hadoop.hbase.client.Scan in inModifier and TypeMethodDescriptionstatic org.apache.hadoop.hbase.client.ScanProtobufUtil.toScan(org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.Scan proto) Convert a protocol buffer Scan to a client ScanMethods in org.apache.hadoop.hbase.shaded.protobuf with parameters of type org.apache.hadoop.hbase.client.Scan in inModifier and TypeMethodDescriptionstatic org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ScanRequestRequestConverter.buildScanRequest(byte[] regionName, org.apache.hadoop.hbase.client.Scan scan, int numberOfRows, boolean closeScanner) Create a protocol buffer ScanRequest for a client Scanstatic org.apache.hadoop.hbase.shaded.protobuf.generated.ClientProtos.ScanProtobufUtil.toScan(org.apache.hadoop.hbase.client.Scan scan) Convert a client Scan to a protocol buffer Scan