Modifier and Type | Method and Description |
---|---|
protected ScanInfo |
ZooKeeperScanPolicyObserver.getScanInfo(Store store,
RegionCoprocessorEnvironment e) |
InternalScanner |
ZooKeeperScanPolicyObserver.preCompactScannerOpen(ObserverContext<RegionCoprocessorEnvironment> c,
Store store,
List<? extends KeyValueScanner> scanners,
ScanType scanType,
long earliestPutTs,
InternalScanner s) |
InternalScanner |
ZooKeeperScanPolicyObserver.preFlushScannerOpen(ObserverContext<RegionCoprocessorEnvironment> c,
Store store,
KeyValueScanner memstoreScanner,
InternalScanner s) |
KeyValueScanner |
ZooKeeperScanPolicyObserver.preStoreScannerOpen(ObserverContext<RegionCoprocessorEnvironment> c,
Store store,
Scan scan,
NavigableSet<byte[]> targetCols,
KeyValueScanner s) |
Modifier and Type | Class and Description |
---|---|
class |
HStore
A Store holds a column family in a Region.
|
Modifier and Type | Field and Description |
---|---|
protected Store |
StoreScanner.store |
Modifier and Type | Field and Description |
---|---|
protected Map<byte[],Store> |
HRegion.stores |
Modifier and Type | Method and Description |
---|---|
Store |
HRegion.getStore(byte[] column) |
Store |
Region.getStore(byte[] family)
Return the Store for the given family
|
Modifier and Type | Method and Description |
---|---|
List<Store> |
HRegion.getStores() |
List<Store> |
Region.getStores()
Return the list of Stores managed by this region
|
Collection<Store> |
FlushLargeStoresPolicy.selectStoresToFlush() |
abstract Collection<Store> |
FlushPolicy.selectStoresToFlush() |
Collection<Store> |
FlushAllStoresPolicy.selectStoresToFlush() |
Modifier and Type | Method and Description |
---|---|
boolean |
HRegion.compact(CompactionContext compaction,
Store store,
ThroughputController throughputController) |
boolean |
HRegion.compact(CompactionContext compaction,
Store store,
ThroughputController throughputController,
User user) |
static StoreEngine<?,?,?,?> |
StoreEngine.create(Store store,
org.apache.hadoop.conf.Configuration conf,
KeyValue.KVComparator kvComparator)
Create the StoreEngine configured for the given Store.
|
protected void |
DateTieredStoreEngine.createComponents(org.apache.hadoop.conf.Configuration conf,
Store store,
KeyValue.KVComparator kvComparator) |
protected void |
DefaultStoreEngine.createComponents(org.apache.hadoop.conf.Configuration conf,
Store store,
KeyValue.KVComparator kvComparator) |
protected abstract void |
StoreEngine.createComponents(org.apache.hadoop.conf.Configuration conf,
Store store,
KeyValue.KVComparator kvComparator)
Create the StoreEngine's components.
|
protected void |
StripeStoreEngine.createComponents(org.apache.hadoop.conf.Configuration conf,
Store store,
KeyValue.KVComparator comparator) |
void |
RegionCoprocessorHost.postCompact(Store store,
StoreFile resultFile,
CompactionRequest request)
Called after the store compaction has completed.
|
void |
RegionCoprocessorHost.postCompactSelection(Store store,
com.google.common.collect.ImmutableList<StoreFile> selected,
CompactionRequest request)
Called after the
StoreFile s to be compacted have been selected from the available
candidates. |
void |
RegionCoprocessorHost.postFlush(Store store,
StoreFile storeFile)
Invoked after a memstore flush
|
InternalScanner |
RegionCoprocessorHost.preCompact(Store store,
InternalScanner scanner,
ScanType scanType,
CompactionRequest request)
Called prior to rewriting the store files selected for compaction
|
InternalScanner |
RegionCoprocessorHost.preCompactScannerOpen(Store store,
List<StoreFileScanner> scanners,
ScanType scanType,
long earliestPutTs,
CompactionRequest request,
long readPoint)
|
boolean |
RegionCoprocessorHost.preCompactSelection(Store store,
List<StoreFile> candidates,
CompactionRequest request)
Called prior to selecting the
StoreFile s for compaction from the list of currently
available candidates. |
InternalScanner |
RegionCoprocessorHost.preFlush(Store store,
InternalScanner scanner)
Invoked before a memstore flush
|
InternalScanner |
RegionCoprocessorHost.preFlushScannerOpen(Store store,
KeyValueScanner memstoreScanner,
long readPoint)
|
KeyValueScanner |
RegionCoprocessorHost.preStoreScannerOpen(Store store,
Scan scan,
NavigableSet<byte[]> targetCols)
|
CompactionRequest |
CompactionRequestor.requestCompaction(Region r,
Store s,
String why,
CompactionRequest request) |
CompactionRequest |
CompactSplitThread.requestCompaction(Region r,
Store s,
String why,
CompactionRequest request) |
CompactionRequest |
CompactionRequestor.requestCompaction(Region r,
Store s,
String why,
int pri,
CompactionRequest request,
User user) |
CompactionRequest |
CompactSplitThread.requestCompaction(Region r,
Store s,
String why,
int priority,
CompactionRequest request,
User user) |
void |
CompactSplitThread.requestSystemCompaction(Region r,
Store s,
String why) |
protected boolean |
HRegion.restoreEdit(Store s,
Cell cell)
Used by tests
|
boolean |
DefaultMemStore.shouldSeek(Scan scan,
Store store,
long oldestUnexpiredTS)
Check if this memstore may contain the required keys
|
boolean |
StoreFileScanner.shouldUseScanner(Scan scan,
Store store,
long oldestUnexpiredTS) |
boolean |
KeyValueScanner.shouldUseScanner(Scan scan,
Store store,
long oldestUnexpiredTS)
Allows to filter out scanners (both StoreFile and memstore) that we don't
want to use based on criteria such as Bloom filters and timestamp ranges.
|
boolean |
DefaultMemStore.MemStoreScanner.shouldUseScanner(Scan scan,
Store store,
long oldestUnexpiredTS) |
boolean |
NonLazyKeyValueScanner.shouldUseScanner(Scan scan,
Store store,
long oldestUnexpiredTS) |
Modifier and Type | Method and Description |
---|---|
protected Region.FlushResult |
HRegion.internalFlushcache(WAL wal,
long myseqid,
Collection<Store> storesToFlush,
MonitoredTask status,
boolean writeFlushWalMarker)
Flush the memstore.
|
protected Region.FlushResult |
HRegion.internalFlushCacheAndCommit(WAL wal,
MonitoredTask status,
org.apache.hadoop.hbase.regionserver.HRegion.PrepareFlushResult prepareResult,
Collection<Store> storesToFlush) |
protected org.apache.hadoop.hbase.regionserver.HRegion.PrepareFlushResult |
HRegion.internalPrepareFlushCache(WAL wal,
long myseqid,
Collection<Store> storesToFlush,
MonitoredTask status,
boolean writeFlushWalMarker) |
List<CompactionRequest> |
CompactionRequestor.requestCompaction(Region r,
String why,
int pri,
List<Pair<CompactionRequest,Store>> requests,
User user) |
List<CompactionRequest> |
CompactSplitThread.requestCompaction(Region r,
String why,
int p,
List<Pair<CompactionRequest,Store>> requests,
User user) |
List<CompactionRequest> |
CompactionRequestor.requestCompaction(Region r,
String why,
List<Pair<CompactionRequest,Store>> requests) |
List<CompactionRequest> |
CompactSplitThread.requestCompaction(Region r,
String why,
List<Pair<CompactionRequest,Store>> requests) |
Constructor and Description |
---|
DefaultStoreFlusher(org.apache.hadoop.conf.Configuration conf,
Store store) |
StoreScanner(Store store,
ScanInfo scanInfo,
Scan scan,
List<? extends KeyValueScanner> scanners,
long smallestReadPoint,
long earliestPutTs,
byte[] dropDeletesFromRow,
byte[] dropDeletesToRow)
Used for compactions that drop deletes from a limited range of rows.
|
StoreScanner(Store store,
ScanInfo scanInfo,
Scan scan,
List<? extends KeyValueScanner> scanners,
ScanType scanType,
long smallestReadPoint,
long earliestPutTs)
Used for compactions.
|
StoreScanner(Store store,
ScanInfo scanInfo,
Scan scan,
NavigableSet<byte[]> columns,
long readPt)
Opens a scanner across memstore, snapshot, and all StoreFiles.
|
StoreScanner(Store store,
Scan scan,
ScanInfo scanInfo,
NavigableSet<byte[]> columns,
long readPt,
boolean cacheBlocks)
An internal constructor.
|
StripeStoreFlusher(org.apache.hadoop.conf.Configuration conf,
Store store,
StripeCompactionPolicy policy,
StripeStoreFileManager stripes) |
Modifier and Type | Field and Description |
---|---|
protected Store |
Compactor.store |
Modifier and Type | Method and Description |
---|---|
protected InternalScanner |
Compactor.createScanner(Store store,
List<StoreFileScanner> scanners,
long smallestReadPoint,
long earliestPutTs,
byte[] dropDeletesFromRow,
byte[] dropDeletesToRow) |
protected InternalScanner |
Compactor.createScanner(Store store,
List<StoreFileScanner> scanners,
ScanType scanType,
long smallestReadPoint,
long earliestPutTs) |
Constructor and Description |
---|
AbstractMultiOutputCompactor(org.apache.hadoop.conf.Configuration conf,
Store store) |
DateTieredCompactor(org.apache.hadoop.conf.Configuration conf,
Store store) |
DefaultCompactor(org.apache.hadoop.conf.Configuration conf,
Store store) |
StripeCompactor(org.apache.hadoop.conf.Configuration conf,
Store store) |
Modifier and Type | Method and Description |
---|---|
static String |
ThroughputControlUtil.getNameForThrottling(Store store,
String opName)
Generate a name for throttling, to prevent name conflict when multiple IO operation running
parallel on the same store.
|
Modifier and Type | Method and Description |
---|---|
InternalScanner |
AccessController.preCompact(ObserverContext<RegionCoprocessorEnvironment> e,
Store store,
InternalScanner scanner,
ScanType scanType) |
Copyright © 2007–2019 The Apache Software Foundation. All rights reserved.