Package | Description |
---|---|
org.apache.hadoop.hbase.mapreduce |
Provides HBase MapReduce
Input/OutputFormats, a table indexing MapReduce job, and utility methods.
|
org.apache.hadoop.hbase.regionserver.wal | |
org.apache.hadoop.hbase.replication |
Multi Cluster Replication
|
org.apache.hadoop.hbase.replication.regionserver | |
org.apache.hadoop.hbase.wal |
Modifier and Type | Field and Description |
---|---|
(package private) WAL.Entry |
WALInputFormat.WALRecordReader.currentEntry |
Modifier and Type | Class and Description |
---|---|
(package private) class |
FSWALEntry
A WAL Entry for
AbstractFSWAL implementation. |
Modifier and Type | Method and Description |
---|---|
WAL.Entry |
ReaderBase.next() |
WAL.Entry |
ReaderBase.next(WAL.Entry reuse) |
Modifier and Type | Method and Description |
---|---|
void |
AsyncProtobufLogWriter.append(WAL.Entry entry) |
void |
ProtobufLogWriter.append(WAL.Entry entry) |
WAL.Entry |
ReaderBase.next(WAL.Entry reuse) |
private long |
AbstractFSWAL.postAppend(WAL.Entry e,
long elapsedTime) |
protected abstract boolean |
ReaderBase.readNext(WAL.Entry e)
Read next entry.
|
protected boolean |
ProtobufLogReader.readNext(WAL.Entry entry) |
Modifier and Type | Field and Description |
---|---|
(package private) List<WAL.Entry> |
ReplicationEndpoint.ReplicateContext.entries |
Modifier and Type | Method and Description |
---|---|
WAL.Entry |
WALEntryFilter.filter(WAL.Entry entry)
Applies the filter, possibly returning a different Entry instance.
|
WAL.Entry |
ClusterMarkingEntryFilter.filter(WAL.Entry entry) |
WAL.Entry |
ScopeWALEntryFilter.filter(WAL.Entry entry) |
WAL.Entry |
SystemTableWALEntryFilter.filter(WAL.Entry entry) |
WAL.Entry |
ChainWALEntryFilter.filter(WAL.Entry entry) |
WAL.Entry |
NamespaceTableCfWALEntryFilter.filter(WAL.Entry entry) |
Modifier and Type | Method and Description |
---|---|
List<WAL.Entry> |
ReplicationEndpoint.ReplicateContext.getEntries() |
Modifier and Type | Method and Description |
---|---|
WAL.Entry |
WALEntryFilter.filter(WAL.Entry entry)
Applies the filter, possibly returning a different Entry instance.
|
WAL.Entry |
ClusterMarkingEntryFilter.filter(WAL.Entry entry) |
WAL.Entry |
ScopeWALEntryFilter.filter(WAL.Entry entry) |
WAL.Entry |
SystemTableWALEntryFilter.filter(WAL.Entry entry) |
WAL.Entry |
ChainWALEntryFilter.filter(WAL.Entry entry) |
WAL.Entry |
NamespaceTableCfWALEntryFilter.filter(WAL.Entry entry) |
Cell |
ScopeWALEntryFilter.filterCell(WAL.Entry entry,
Cell cell) |
Cell |
WALCellFilter.filterCell(WAL.Entry entry,
Cell cell)
Applies the filter, possibly returning a different Cell instance.
|
Cell |
NamespaceTableCfWALEntryFilter.filterCell(WAL.Entry entry,
Cell cell) |
private void |
ChainWALEntryFilter.filterCells(WAL.Entry entry) |
Modifier and Type | Method and Description |
---|---|
ReplicationEndpoint.ReplicateContext |
ReplicationEndpoint.ReplicateContext.setEntries(List<WAL.Entry> entries) |
Modifier and Type | Field and Description |
---|---|
private WAL.Entry |
WALEntryStream.currentEntry |
Modifier and Type | Field and Description |
---|---|
private List<WAL.Entry> |
RegionReplicaReplicationEndpoint.RegionReplicaReplayCallable.entries |
private List<WAL.Entry> |
WALEntryBatch.walEntries |
Modifier and Type | Method and Description |
---|---|
protected WAL.Entry |
ReplicationSourceWALReader.filterEntry(WAL.Entry entry) |
WAL.Entry |
WALEntryStream.next()
Returns the next WAL entry in this stream and advance the stream.
|
WAL.Entry |
WALEntryStream.peek()
Returns the next WAL entry in this stream but does not advance.
|
Modifier and Type | Method and Description |
---|---|
private List<List<WAL.Entry>> |
HBaseInterClusterReplicationEndpoint.createBatches(List<WAL.Entry> entries)
Divide the entries into multiple batches, so that we can replicate each batch in a thread pool
concurrently.
|
private List<List<WAL.Entry>> |
HBaseInterClusterReplicationEndpoint.createParallelBatches(List<WAL.Entry> entries) |
private List<List<WAL.Entry>> |
HBaseInterClusterReplicationEndpoint.createSerialBatches(List<WAL.Entry> entries) |
private List<List<WAL.Entry>> |
HBaseInterClusterReplicationEndpoint.filterBatches(List<List<WAL.Entry>> oldEntryList,
TableName table) |
List<WAL.Entry> |
WALEntryBatch.getWalEntries() |
Modifier and Type | Method and Description |
---|---|
void |
WALEntryBatch.addEntry(WAL.Entry entry) |
protected boolean |
ReplicationSourceWALReader.addEntryToBatch(WALEntryBatch batch,
WAL.Entry entry) |
private boolean |
SerialReplicationChecker.canPush(WAL.Entry entry,
byte[] row) |
boolean |
SerialReplicationChecker.canPush(WAL.Entry entry,
Cell firstCellInEdit) |
protected WAL.Entry |
ReplicationSourceWALReader.filterEntry(WAL.Entry entry) |
static long |
ReplicationSourceWALReader.getEntrySizeExcludeBulkLoad(WAL.Entry entry) |
private long |
ReplicationSourceWALReader.getEntrySizeIncludeBulkLoad(WAL.Entry entry) |
private int |
HBaseInterClusterReplicationEndpoint.getEstimatedEntrySize(WAL.Entry e) |
boolean |
RegionReplicaReplicationEndpoint.RegionReplicaOutputSink.keepRegionEvent(WAL.Entry entry) |
private void |
ReplicationSourceWALReader.updateBatchStats(WALEntryBatch batch,
WAL.Entry entry,
long entrySize) |
void |
SerialReplicationChecker.waitUntilCanPush(WAL.Entry entry,
Cell firstCellInEdit) |
Modifier and Type | Method and Description |
---|---|
void |
RegionReplicaReplicationEndpoint.RegionReplicaSinkWriter.append(TableName tableName,
byte[] encodedRegionName,
byte[] row,
List<WAL.Entry> entries) |
private List<List<WAL.Entry>> |
HBaseInterClusterReplicationEndpoint.createBatches(List<WAL.Entry> entries)
Divide the entries into multiple batches, so that we can replicate each batch in a thread pool
concurrently.
|
private List<List<WAL.Entry>> |
HBaseInterClusterReplicationEndpoint.createParallelBatches(List<WAL.Entry> entries) |
protected Callable<Integer> |
HBaseInterClusterReplicationEndpoint.createReplicator(List<WAL.Entry> entries,
int batchIndex,
int timeout) |
private List<List<WAL.Entry>> |
HBaseInterClusterReplicationEndpoint.createSerialBatches(List<WAL.Entry> entries) |
private List<List<WAL.Entry>> |
HBaseInterClusterReplicationEndpoint.filterBatches(List<List<WAL.Entry>> oldEntryList,
TableName table) |
private long |
HBaseInterClusterReplicationEndpoint.parallelReplicate(CompletionService<Integer> pool,
ReplicationEndpoint.ReplicateContext replicateContext,
List<List<WAL.Entry>> batches) |
void |
ReplicationSourceInterface.postShipEdits(List<WAL.Entry> entries,
int batchSize)
Call this after the shipper thread ship some entries to peer cluster.
|
void |
ReplicationSource.postShipEdits(List<WAL.Entry> entries,
int batchSize) |
protected int |
HBaseInterClusterReplicationEndpoint.replicateEntries(List<WAL.Entry> entries,
int batchIndex,
int timeout) |
private boolean |
RegionReplicaReplicationEndpoint.RegionReplicaOutputSink.requiresReplication(TableName tableName,
List<WAL.Entry> entries)
returns true if the specified entry must be replicated.
|
private int |
HBaseInterClusterReplicationEndpoint.serialReplicateRegionEntries(List<WAL.Entry> entries,
int batchIndex,
int timeout) |
Constructor and Description |
---|
RegionReplicaReplayCallable(ClusterConnection connection,
RpcControllerFactory rpcControllerFactory,
TableName tableName,
HRegionLocation location,
RegionInfo regionInfo,
byte[] row,
List<WAL.Entry> entries,
AtomicLong skippedEntries) |
Modifier and Type | Field and Description |
---|---|
(package private) List<WAL.Entry> |
WALSplitter.RegionEntryBuffer.entryBuffer |
Modifier and Type | Method and Description |
---|---|
private static WAL.Entry |
WALSplitter.getNextLogLine(WAL.Reader in,
org.apache.hadoop.fs.Path path,
boolean skipErrors) |
WAL.Entry |
WAL.Reader.next() |
WAL.Entry |
WAL.Reader.next(WAL.Entry reuse) |
Modifier and Type | Method and Description |
---|---|
List<WAL.Entry> |
WALSplitter.RegionEntryBuffer.getEntryBuffer() |
Modifier and Type | Method and Description |
---|---|
void |
WALProvider.Writer.append(WAL.Entry entry) |
void |
WALProvider.AsyncWriter.append(WAL.Entry entry) |
void |
EntryBuffers.appendEntry(WAL.Entry entry)
Append a log entry into the corresponding region buffer.
|
(package private) long |
WALSplitter.RegionEntryBuffer.appendEntry(WAL.Entry entry) |
(package private) WALSplitter.WriterAndPath |
LogRecoveredEditsOutputSink.createWAP(byte[] region,
WAL.Entry entry) |
(package private) void |
LogRecoveredEditsOutputSink.filterCellByStore(WAL.Entry logEntry) |
(package private) static org.apache.hadoop.fs.Path |
WALSplitUtil.getRegionSplitEditsPath(WAL.Entry walEntry,
String fileNameBeingSplit,
String tmpDirName,
org.apache.hadoop.conf.Configuration conf)
Path to a file under RECOVERED_EDITS_DIR directory of the region found in
logEntry
named for the sequenceid in the passed logEntry : e.g. |
(package private) WALSplitter.WriterAndPath |
LogRecoveredEditsOutputSink.getWriterAndPath(WAL.Entry entry,
boolean reusable)
Get a writer and path for a log starting at the given entry.
|
private void |
WALSplitter.RegionEntryBuffer.internify(WAL.Entry entry) |
abstract boolean |
OutputSink.keepRegionEvent(WAL.Entry entry)
Some WALEdit's contain only KV's for account on what happened to a region.
|
boolean |
LogRecoveredEditsOutputSink.keepRegionEvent(WAL.Entry entry) |
WAL.Entry |
WAL.Reader.next(WAL.Entry reuse) |
(package private) void |
OutputSink.updateRegionMaximumEditLogSeqNum(WAL.Entry entry)
Update region's maximum edit log SeqNum.
|
Copyright © 2007–2020 The Apache Software Foundation. All rights reserved.