Modifier and Type | Method and Description |
---|---|
List<org.apache.hadoop.fs.Path> |
DefaultMobStoreCompactor.compact(CompactionRequestImpl request,
ThroughputController throughputController,
User user) |
List<org.apache.hadoop.fs.Path> |
DefaultMobStoreFlusher.flushSnapshot(MemStoreSnapshot snapshot,
long cacheFlushId,
MonitoredTask status,
ThroughputController throughputController,
FlushLifeCycleTracker tracker)
Flushes the snapshot of the MemStore.
|
protected boolean |
DefaultMobStoreCompactor.performCompaction(Compactor.FileDetails fd,
InternalScanner scanner,
CellSink writer,
long smallestReadPoint,
boolean cleanSeqId,
ThroughputController throughputController,
boolean major,
int numofFilesToCompact)
Performs compaction on a column family with the mob flag enabled.
|
protected void |
DefaultMobStoreFlusher.performMobFlush(MemStoreSnapshot snapshot,
long cacheFlushId,
InternalScanner scanner,
StoreFileWriter writer,
MonitoredTask status,
ThroughputController throughputController)
Flushes the cells in the mob store.
|
Modifier and Type | Field and Description |
---|---|
private ThroughputController |
CompactSplit.compactionThroughputController |
private ThroughputController |
HRegionServer.flushThroughputController |
Modifier and Type | Method and Description |
---|---|
ThroughputController |
CompactSplit.getCompactionThroughputController() |
ThroughputController |
RegionServerServices.getFlushThroughputController() |
ThroughputController |
HRegionServer.getFlushThroughputController() |
Modifier and Type | Method and Description |
---|---|
boolean |
HRegion.compact(CompactionContext compaction,
HStore store,
ThroughputController throughputController)
Called by compaction thread and after region is opened to compact the
HStores if necessary.
|
boolean |
HRegion.compact(CompactionContext compaction,
HStore store,
ThroughputController throughputController,
User user) |
List<HStoreFile> |
HStore.compact(CompactionContext compaction,
ThroughputController throughputController,
User user)
Compact the StoreFiles.
|
List<org.apache.hadoop.fs.Path> |
StripeStoreEngine.StripeCompaction.compact(ThroughputController throughputController,
User user) |
List<org.apache.hadoop.fs.Path> |
DefaultStoreEngine.DefaultCompactionContext.compact(ThroughputController throughputController,
User user) |
List<org.apache.hadoop.fs.Path> |
DateTieredStoreEngine.DateTieredCompactionContext.compact(ThroughputController throughputController,
User user) |
(package private) void |
HRegion.compactStore(byte[] family,
ThroughputController throughputController)
This is a helper function that compact the given store.
|
protected List<org.apache.hadoop.fs.Path> |
HStore.flushCache(long logCacheFlushId,
MemStoreSnapshot snapshot,
MonitoredTask status,
ThroughputController throughputController,
FlushLifeCycleTracker tracker)
Write out current snapshot.
|
List<org.apache.hadoop.fs.Path> |
DefaultStoreFlusher.flushSnapshot(MemStoreSnapshot snapshot,
long cacheFlushId,
MonitoredTask status,
ThroughputController throughputController,
FlushLifeCycleTracker tracker) |
abstract List<org.apache.hadoop.fs.Path> |
StoreFlusher.flushSnapshot(MemStoreSnapshot snapshot,
long cacheFlushSeqNum,
MonitoredTask status,
ThroughputController throughputController,
FlushLifeCycleTracker tracker)
Turns a snapshot of memstore into a set of store files.
|
List<org.apache.hadoop.fs.Path> |
StripeStoreFlusher.flushSnapshot(MemStoreSnapshot snapshot,
long cacheFlushSeqNum,
MonitoredTask status,
ThroughputController throughputController,
FlushLifeCycleTracker tracker) |
protected void |
StoreFlusher.performFlush(InternalScanner scanner,
CellSink sink,
long smallestReadPoint,
ThroughputController throughputController)
Performs memstore flush, writing data from scanner into sink.
|
Modifier and Type | Method and Description |
---|---|
protected List<org.apache.hadoop.fs.Path> |
Compactor.compact(CompactionRequestImpl request,
Compactor.InternalScannerFactory scannerFactory,
Compactor.CellSinkFactory<T> sinkFactory,
ThroughputController throughputController,
User user) |
List<org.apache.hadoop.fs.Path> |
StripeCompactor.compact(CompactionRequestImpl request,
int targetCount,
long targetSize,
byte[] left,
byte[] right,
byte[] majorRangeFromRow,
byte[] majorRangeToRow,
ThroughputController throughputController,
User user) |
List<org.apache.hadoop.fs.Path> |
StripeCompactor.compact(CompactionRequestImpl request,
List<byte[]> targetBoundaries,
byte[] majorRangeFromRow,
byte[] majorRangeToRow,
ThroughputController throughputController,
User user) |
List<org.apache.hadoop.fs.Path> |
DateTieredCompactor.compact(CompactionRequestImpl request,
List<Long> lowerBoundaries,
ThroughputController throughputController,
User user) |
List<org.apache.hadoop.fs.Path> |
DefaultCompactor.compact(CompactionRequestImpl request,
ThroughputController throughputController,
User user)
Do a minor/major compaction on an explicit set of storefiles from a Store.
|
abstract List<org.apache.hadoop.fs.Path> |
CompactionContext.compact(ThroughputController throughputController,
User user) |
List<org.apache.hadoop.fs.Path> |
StripeCompactionPolicy.StripeCompactionRequest.execute(StripeCompactor compactor,
ThroughputController throughputController) |
abstract List<org.apache.hadoop.fs.Path> |
StripeCompactionPolicy.StripeCompactionRequest.execute(StripeCompactor compactor,
ThroughputController throughputController,
User user)
Executes the request against compactor (essentially, just calls correct overload of
compact method), to simulate more dynamic dispatch.
|
List<org.apache.hadoop.fs.Path> |
StripeCompactionPolicy.BoundaryStripeCompactionRequest.execute(StripeCompactor compactor,
ThroughputController throughputController,
User user) |
List<org.apache.hadoop.fs.Path> |
StripeCompactionPolicy.SplitStripeCompactionRequest.execute(StripeCompactor compactor,
ThroughputController throughputController,
User user) |
protected boolean |
Compactor.performCompaction(Compactor.FileDetails fd,
InternalScanner scanner,
CellSink writer,
long smallestReadPoint,
boolean cleanSeqId,
ThroughputController throughputController,
boolean major,
int numofFilesToCompact)
Performs the compaction.
|
Modifier and Type | Class and Description |
---|---|
class |
NoLimitThroughputController |
class |
PressureAwareCompactionThroughputController
A throughput controller which uses the follow schema to limit throughput
If compaction pressure is greater than 1.0, no limitation.
In off peak hours, use a fixed throughput limitation
"hbase.hstore.compaction.throughput.offpeak"
In normal hours, the max throughput is tuned between
"hbase.hstore.compaction.throughput.lower.bound" and
"hbase.hstore.compaction.throughput.higher.bound", using the formula "lower +
(higer - lower) * compactionPressure", where compactionPressure is in range [0.0, 1.0]
|
class |
PressureAwareFlushThroughputController
A throughput controller which uses the follow schema to limit throughput
If flush pressure is greater than or equal to 1.0, no limitation.
In normal case, the max throughput is tuned between
"hbase.hstore.flush.throughput.lower.bound" and
"hbase.hstore.flush.throughput.upper.bound", using the formula "lower +
(upper - lower) * flushPressure", where flushPressure is in range [0.0, 1.0)
|
class |
PressureAwareThroughputController |
Modifier and Type | Field and Description |
---|---|
private static Class<? extends ThroughputController> |
FlushThroughputControllerFactory.DEFAULT_FLUSH_THROUGHPUT_CONTROLLER_CLASS |
private static Class<? extends ThroughputController> |
CompactionThroughputControllerFactory.DEFAULT_THROUGHPUT_CONTROLLER_CLASS |
Modifier and Type | Method and Description |
---|---|
static ThroughputController |
FlushThroughputControllerFactory.create(RegionServerServices server,
org.apache.hadoop.conf.Configuration conf) |
static ThroughputController |
CompactionThroughputControllerFactory.create(RegionServerServices server,
org.apache.hadoop.conf.Configuration conf) |
Modifier and Type | Method and Description |
---|---|
static Class<? extends ThroughputController> |
FlushThroughputControllerFactory.getThroughputControllerClass(org.apache.hadoop.conf.Configuration conf) |
static Class<? extends ThroughputController> |
CompactionThroughputControllerFactory.getThroughputControllerClass(org.apache.hadoop.conf.Configuration conf) |
Copyright © 2007–2020 The Apache Software Foundation. All rights reserved.