Package | Description |
---|---|
org.apache.hadoop.hbase.client |
Provides HBase Client
|
org.apache.hadoop.hbase.mapred |
Provides HBase MapReduce
Input/OutputFormats, a table indexing MapReduce job, and utility methods.
|
org.apache.hadoop.hbase.mapreduce |
Provides HBase MapReduce
Input/OutputFormats, a table indexing MapReduce job, and utility methods.
|
Modifier and Type | Field and Description |
---|---|
static Result |
Result.EMPTY_RESULT |
Modifier and Type | Method and Description |
---|---|
default Result |
Table.append(Append append)
Appends values to one or more columns within a single row.
|
static Result |
Result.create(Cell[] cells)
Instantiate a Result with the specified array of KeyValues.
|
static Result |
Result.create(Cell[] cells,
Boolean exists,
boolean stale) |
static Result |
Result.create(Cell[] cells,
Boolean exists,
boolean stale,
boolean mayHaveMoreCellsInRow) |
static Result |
Result.create(List<Cell> cells)
Instantiate a Result with the specified List of KeyValues.
|
static Result |
Result.create(List<Cell> cells,
Boolean exists) |
static Result |
Result.create(List<Cell> cells,
Boolean exists,
boolean stale) |
static Result |
Result.create(List<Cell> cells,
Boolean exists,
boolean stale,
boolean mayHaveMoreCellsInRow) |
static Result |
Result.createCompleteResult(Iterable<Result> partialResults)
Forms a single result from the partial results in the partialResults list.
|
static Result |
Result.createCursorResult(Cursor cursor) |
default Result |
Table.get(Get get)
Extracts certain cells from a given row.
|
default Result[] |
Table.get(List<Get> gets)
Extracts specified cells from the given rows, as a batch.
|
default Result |
Table.increment(Increment increment)
Increments one or more columns within a single row.
|
Result |
ResultScanner.next()
Grab the next row's worth of values.
|
default Result[] |
ResultScanner.next(int nbRows)
Get nbRows rows.
|
Modifier and Type | Method and Description |
---|---|
CompletableFuture<Result> |
AsyncTable.append(Append append)
Appends values to one or more columns within a single row.
|
CompletableFuture<Result> |
AsyncTable.get(Get get)
Extracts certain cells from a given row.
|
List<CompletableFuture<Result>> |
AsyncTable.get(List<Get> gets)
Extracts certain cells from the given rows, in batch.
|
default CompletableFuture<List<Result>> |
AsyncTable.getAll(List<Get> gets)
A simple version for batch get.
|
CompletableFuture<Result> |
AsyncTable.increment(Increment increment)
Increments one or more columns within a single row.
|
default Iterator<Result> |
ResultScanner.iterator() |
CompletableFuture<List<Result>> |
AsyncTable.scanAll(Scan scan)
Return all the results that match the given scan object.
|
Modifier and Type | Method and Description |
---|---|
static void |
Result.compareResults(Result res1,
Result res2)
Does a deep comparison of two Results, down to the byte arrays.
|
void |
Result.copyFrom(Result other)
Copy another Result into this one.
|
static long |
Result.getTotalSizeOfCells(Result result)
Get total size of raw cells
|
boolean |
ScanResultConsumer.onNext(Result result) |
void |
AdvancedScanResultConsumer.onNext(Result[] results,
AdvancedScanResultConsumer.ScanController controller)
Indicate that we have receive some data.
|
Modifier and Type | Method and Description |
---|---|
static Result |
Result.createCompleteResult(Iterable<Result> partialResults)
Forms a single result from the partial results in the partialResults list.
|
Modifier and Type | Method and Description |
---|---|
Result |
TableRecordReaderImpl.createValue() |
Result |
TableRecordReader.createValue() |
Modifier and Type | Method and Description |
---|---|
org.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,Result> |
TableSnapshotInputFormat.getRecordReader(org.apache.hadoop.mapred.InputSplit split,
org.apache.hadoop.mapred.JobConf job,
org.apache.hadoop.mapred.Reporter reporter) |
org.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,Result> |
TableInputFormatBase.getRecordReader(org.apache.hadoop.mapred.InputSplit split,
org.apache.hadoop.mapred.JobConf job,
org.apache.hadoop.mapred.Reporter reporter)
Builds a TableRecordReader.
|
org.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,Result> |
MultiTableSnapshotInputFormat.getRecordReader(org.apache.hadoop.mapred.InputSplit split,
org.apache.hadoop.mapred.JobConf job,
org.apache.hadoop.mapred.Reporter reporter) |
Modifier and Type | Method and Description |
---|---|
protected byte[][] |
GroupingTableMap.extractKeyValues(Result r)
Extract columns values from the current record.
|
void |
GroupingTableMap.map(ImmutableBytesWritable key,
Result value,
org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output,
org.apache.hadoop.mapred.Reporter reporter)
Extract the grouping columns from value to construct a new key.
|
void |
IdentityTableMap.map(ImmutableBytesWritable key,
Result value,
org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output,
org.apache.hadoop.mapred.Reporter reporter)
Pass the key, value to reduce
|
boolean |
TableRecordReaderImpl.next(ImmutableBytesWritable key,
Result value) |
boolean |
TableRecordReader.next(ImmutableBytesWritable key,
Result value) |
Modifier and Type | Method and Description |
---|---|
void |
GroupingTableMap.map(ImmutableBytesWritable key,
Result value,
org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output,
org.apache.hadoop.mapred.Reporter reporter)
Extract the grouping columns from value to construct a new key.
|
void |
IdentityTableMap.map(ImmutableBytesWritable key,
Result value,
org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output,
org.apache.hadoop.mapred.Reporter reporter)
Pass the key, value to reduce
|
Modifier and Type | Method and Description |
---|---|
Result |
TableRecordReaderImpl.getCurrentValue()
Returns the current value.
|
Result |
TableRecordReader.getCurrentValue()
Returns the current value.
|
Modifier and Type | Method and Description |
---|---|
org.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,Result> |
TableSnapshotInputFormat.createRecordReader(org.apache.hadoop.mapreduce.InputSplit split,
org.apache.hadoop.mapreduce.TaskAttemptContext context) |
org.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,Result> |
TableInputFormatBase.createRecordReader(org.apache.hadoop.mapreduce.InputSplit split,
org.apache.hadoop.mapreduce.TaskAttemptContext context)
Builds a
TableRecordReader . |
org.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,Result> |
MultiTableInputFormatBase.createRecordReader(org.apache.hadoop.mapreduce.InputSplit split,
org.apache.hadoop.mapreduce.TaskAttemptContext context)
Builds a TableRecordReader.
|
org.apache.hadoop.io.serializer.Deserializer<Result> |
ResultSerialization.getDeserializer(Class<Result> c) |
org.apache.hadoop.io.serializer.Serializer<Result> |
ResultSerialization.getSerializer(Class<Result> c) |
Modifier and Type | Method and Description |
---|---|
protected byte[][] |
GroupingTableMapper.extractKeyValues(Result r)
Extract columns values from the current record.
|
void |
GroupingTableMapper.map(ImmutableBytesWritable key,
Result value,
org.apache.hadoop.mapreduce.Mapper.Context context)
Extract the grouping columns from value to construct a new key.
|
void |
IdentityTableMapper.map(ImmutableBytesWritable key,
Result value,
org.apache.hadoop.mapreduce.Mapper.Context context)
Pass the key, value to reduce.
|
Modifier and Type | Method and Description |
---|---|
org.apache.hadoop.io.serializer.Deserializer<Result> |
ResultSerialization.getDeserializer(Class<Result> c) |
org.apache.hadoop.io.serializer.Serializer<Result> |
ResultSerialization.getSerializer(Class<Result> c) |
Copyright © 2007–2020 The Apache Software Foundation. All rights reserved.