Uses of Class
org.apache.hadoop.hbase.io.ImmutableBytesWritable
Package
Description
Provides HBase MapReduce
Input/OutputFormats, a table indexing MapReduce job, and utility methods.
Provides HBase MapReduce
Input/OutputFormats, a table indexing MapReduce job, and utility methods.
-
Uses of ImmutableBytesWritable in org.apache.hadoop.hbase
Modifier and TypeMethodDescription(package private) static ImmutableBytesWritable
HFilePerformanceEvaluation.format
(int i, ImmutableBytesWritable w) Modifier and TypeMethodDescription(package private) static ImmutableBytesWritable
HFilePerformanceEvaluation.format
(int i, ImmutableBytesWritable w) protected void
ScanPerformanceEvaluation.MyMapper.map
(ImmutableBytesWritable key, Result value, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, KEYOUT, VALUEOUT>.org.apache.hadoop.mapreduce.Mapper.Context context) -
Uses of ImmutableBytesWritable in org.apache.hadoop.hbase.io
Modifier and TypeMethodDescriptionint
ImmutableBytesWritable.compareTo
(ImmutableBytesWritable that) Define the sort order of the BytesWritable.ModifierConstructorDescriptionSet the new ImmutableBytesWritable to the contents of the passedibw
. -
Uses of ImmutableBytesWritable in org.apache.hadoop.hbase.mapred
Modifier and TypeMethodDescriptionprotected ImmutableBytesWritable
GroupingTableMap.createGroupKey
(byte[][] vals) Create a key by concatenating multiple column values.TableRecordReader.createKey()
TableRecordReaderImpl.createKey()
TableSnapshotInputFormat.TableSnapshotRecordReader.createKey()
Modifier and TypeMethodDescriptionorg.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,
Result> MultiTableSnapshotInputFormat.getRecordReader
(org.apache.hadoop.mapred.InputSplit split, org.apache.hadoop.mapred.JobConf job, org.apache.hadoop.mapred.Reporter reporter) org.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,
Result> TableInputFormatBase.getRecordReader
(org.apache.hadoop.mapred.InputSplit split, org.apache.hadoop.mapred.JobConf job, org.apache.hadoop.mapred.Reporter reporter) Builds a TableRecordReader.org.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,
Result> TableSnapshotInputFormat.getRecordReader
(org.apache.hadoop.mapred.InputSplit split, org.apache.hadoop.mapred.JobConf job, org.apache.hadoop.mapred.Reporter reporter) Modifier and TypeMethodDescriptionint
HRegionPartitioner.getPartition
(ImmutableBytesWritable key, V2 value, int numPartitions) void
GroupingTableMap.map
(ImmutableBytesWritable key, Result value, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable, Result> output, org.apache.hadoop.mapred.Reporter reporter) Extract the grouping columns from value to construct a new key.void
IdentityTableMap.map
(ImmutableBytesWritable key, Result value, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable, Result> output, org.apache.hadoop.mapred.Reporter reporter) Pass the key, value to reducevoid
RowCounter.RowCounterMapper.map
(ImmutableBytesWritable row, Result values, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable, Result> output, org.apache.hadoop.mapred.Reporter reporter) boolean
TableRecordReader.next
(ImmutableBytesWritable key, Result value) boolean
TableRecordReaderImpl.next
(ImmutableBytesWritable key, Result value) boolean
TableSnapshotInputFormat.TableSnapshotRecordReader.next
(ImmutableBytesWritable key, Result value) void
IdentityTableReduce.reduce
(ImmutableBytesWritable key, Iterator<Put> values, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable, Put> output, org.apache.hadoop.mapred.Reporter reporter) No aggregation, output pairs of (key, record)void
TableOutputFormat.TableRecordWriter.write
(ImmutableBytesWritable key, Put value) Modifier and TypeMethodDescriptionvoid
GroupingTableMap.map
(ImmutableBytesWritable key, Result value, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable, Result> output, org.apache.hadoop.mapred.Reporter reporter) Extract the grouping columns from value to construct a new key.void
IdentityTableMap.map
(ImmutableBytesWritable key, Result value, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable, Result> output, org.apache.hadoop.mapred.Reporter reporter) Pass the key, value to reducevoid
RowCounter.RowCounterMapper.map
(ImmutableBytesWritable row, Result values, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable, Result> output, org.apache.hadoop.mapred.Reporter reporter) void
IdentityTableReduce.reduce
(ImmutableBytesWritable key, Iterator<Put> values, org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable, Put> output, org.apache.hadoop.mapred.Reporter reporter) No aggregation, output pairs of (key, record) -
Uses of ImmutableBytesWritable in org.apache.hadoop.hbase.mapreduce
Modifier and TypeFieldDescriptionprivate ImmutableBytesWritable
HashTable.ResultHasher.batchHash
private ImmutableBytesWritable
HashTable.ResultHasher.batchStartKey
private ImmutableBytesWritable
HashTable.HashMapper.currentRow
(package private) ImmutableBytesWritable
SyncTable.SyncMapper.currentSourceHash
private ImmutableBytesWritable
HashTable.TableHash.Reader.hash
private ImmutableBytesWritable
HashTable.TableHash.Reader.key
private ImmutableBytesWritable
MultithreadedTableMapper.SubMapRecordReader.key
private ImmutableBytesWritable
TableRecordReaderImpl.key
(package private) ImmutableBytesWritable
SyncTable.SyncMapper.nextSourceKey
private ImmutableBytesWritable
TableSnapshotInputFormatImpl.RecordReader.row
Modifier and TypeFieldDescriptionprivate TreeMap<byte[],
ImmutableBytesWritable> IndexBuilder.Map.indexes
private Class<? extends org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable,
Result, K2, V2>> MultithreadedTableMapper.mapClass
private org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable,
Result, K2, V2> MultithreadedTableMapper.MapRunner.mapper
(package private) Map<ImmutableBytesWritable,
BufferedMutator> MultiTableOutputFormat.MultiTableRecordWriter.mutatorMap
(package private) List<ImmutableBytesWritable>
HashTable.TableHash.partitions
Modifier and TypeMethodDescriptionprotected ImmutableBytesWritable
GroupingTableMapper.createGroupKey
(byte[][] vals) Create a key by concatenating multiple column values.HashTable.ResultHasher.getBatchHash()
HashTable.ResultHasher.getBatchStartKey()
HashTable.TableHash.Reader.getCurrentHash()
Get the current hashHashTable.TableHash.Reader.getCurrentKey()
Get the current keyMultithreadedTableMapper.SubMapRecordReader.getCurrentKey()
TableRecordReader.getCurrentKey()
Returns the current key.TableRecordReaderImpl.getCurrentKey()
Returns the current key.TableSnapshotInputFormat.TableSnapshotRegionRecordReader.getCurrentKey()
TableSnapshotInputFormatImpl.RecordReader.getCurrentKey()
Modifier and TypeMethodDescriptionorg.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,
Result> MultiTableInputFormatBase.createRecordReader
(org.apache.hadoop.mapreduce.InputSplit split, org.apache.hadoop.mapreduce.TaskAttemptContext context) Builds a TableRecordReader.org.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,
Result> TableInputFormatBase.createRecordReader
(org.apache.hadoop.mapreduce.InputSplit split, org.apache.hadoop.mapreduce.TaskAttemptContext context) Builds aTableRecordReader
.org.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,
Result> TableSnapshotInputFormat.createRecordReader
(org.apache.hadoop.mapreduce.InputSplit split, org.apache.hadoop.mapreduce.TaskAttemptContext context) (package private) static <V extends Cell>
org.apache.hadoop.mapreduce.RecordWriter<ImmutableBytesWritable,V> HFileOutputFormat2.createRecordWriter
(org.apache.hadoop.mapreduce.TaskAttemptContext context, org.apache.hadoop.mapreduce.OutputCommitter committer) static <K2,
V2> Class<org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, K2, V2>> MultithreadedTableMapper.getMapperClass
(org.apache.hadoop.mapreduce.JobContext job) Get the application's mapper class.org.apache.hadoop.mapreduce.RecordWriter<ImmutableBytesWritable,
Cell> HFileOutputFormat2.getRecordWriter
(org.apache.hadoop.mapreduce.TaskAttemptContext context) org.apache.hadoop.mapreduce.RecordWriter<ImmutableBytesWritable,
Mutation> MultiTableOutputFormat.getRecordWriter
(org.apache.hadoop.mapreduce.TaskAttemptContext context) private static List<ImmutableBytesWritable>
HFileOutputFormat2.getRegionStartKeys
(List<RegionLocator> regionLocators, boolean writeMultipleTables) Return the start keys of all of the regions in this table, as a list of ImmutableBytesWritable.Modifier and TypeMethodDescriptionstatic byte[]
MultiTableHFileOutputFormat.createCompositeKey
(byte[] tableName, ImmutableBytesWritable suffix) Alternate api which accepts an ImmutableBytesWritable for the suffixstatic byte[]
MultiTableHFileOutputFormat.createCompositeKey
(String tableName, ImmutableBytesWritable suffix) Alternate api which accepts a String for the tableName and ImmutableBytesWritable for the suffix(package private) BufferedMutator
MultiTableOutputFormat.MultiTableRecordWriter.getBufferedMutator
(ImmutableBytesWritable tableName) the name of the table, as a stringint
HRegionPartitioner.getPartition
(ImmutableBytesWritable key, VALUE value, int numPartitions) Gets the partition number for a given key (hence record) given the total number of partitions i.e.int
SimpleTotalOrderPartitioner.getPartition
(ImmutableBytesWritable key, VALUE value, int reduces) void
CellCounter.CellCounterMapper.map
(ImmutableBytesWritable row, Result values, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, org.apache.hadoop.io.Text, org.apache.hadoop.io.LongWritable>.org.apache.hadoop.mapreduce.Mapper.Context context) Maps the data.void
GroupingTableMapper.map
(ImmutableBytesWritable key, Result value, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Result>.org.apache.hadoop.mapreduce.Mapper.Context context) Extract the grouping columns from value to construct a new key.protected void
HashTable.HashMapper.map
(ImmutableBytesWritable key, Result value, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, ImmutableBytesWritable>.org.apache.hadoop.mapreduce.Mapper.Context context) void
IdentityTableMapper.map
(ImmutableBytesWritable key, Result value, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Result>.org.apache.hadoop.mapreduce.Mapper.Context context) Pass the key, value to reduce.void
Import.CellImporter.map
(ImmutableBytesWritable row, Result value, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Cell>.org.apache.hadoop.mapreduce.Mapper.Context context) void
Import.CellSortImporter.map
(ImmutableBytesWritable row, Result value, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, Import.CellWritableComparable, Cell>.org.apache.hadoop.mapreduce.Mapper.Context context) void
Import.Importer.map
(ImmutableBytesWritable row, Result value, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Mutation>.org.apache.hadoop.mapreduce.Mapper.Context context) protected void
IndexBuilder.Map.map
(ImmutableBytesWritable rowKey, Result result, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Put>.org.apache.hadoop.mapreduce.Mapper.Context context) void
RowCounter.RowCounterMapper.map
(ImmutableBytesWritable row, Result values, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Result>.org.apache.hadoop.mapreduce.Mapper.Context context) Maps the data.protected void
SyncTable.SyncMapper.map
(ImmutableBytesWritable key, Result value, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Mutation>.org.apache.hadoop.mapreduce.Mapper.Context context) HashTable.TableHash.newReader
(org.apache.hadoop.conf.Configuration conf, ImmutableBytesWritable startKey) Open a TableHash.Reader starting at the first hash at or after the given key.protected void
Import.Importer.processKV
(ImmutableBytesWritable key, Result result, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Mutation>.org.apache.hadoop.mapreduce.Mapper.Context context, Put put, Delete delete) protected void
CellSortReducer.reduce
(ImmutableBytesWritable row, Iterable<Cell> kvs, org.apache.hadoop.mapreduce.Reducer<ImmutableBytesWritable, Cell, ImmutableBytesWritable, Cell>.org.apache.hadoop.mapreduce.Reducer.Context context) protected void
PutSortReducer.reduce
(ImmutableBytesWritable row, Iterable<Put> puts, org.apache.hadoop.mapreduce.Reducer<ImmutableBytesWritable, Put, ImmutableBytesWritable, KeyValue>.org.apache.hadoop.mapreduce.Reducer.Context context) protected void
TextSortReducer.reduce
(ImmutableBytesWritable rowKey, Iterable<org.apache.hadoop.io.Text> lines, org.apache.hadoop.mapreduce.Reducer<ImmutableBytesWritable, org.apache.hadoop.io.Text, ImmutableBytesWritable, KeyValue>.org.apache.hadoop.mapreduce.Reducer.Context context) void
HashTable.ResultHasher.startBatch
(ImmutableBytesWritable row) private void
SyncTable.SyncMapper.syncRange
(org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Mutation>.org.apache.hadoop.mapreduce.Mapper.Context context, ImmutableBytesWritable startRow, ImmutableBytesWritable stopRow) Rescan the given range directly from the source and target tables.private static String
SyncTable.SyncMapper.toHex
(ImmutableBytesWritable bytes) void
MultiTableOutputFormat.MultiTableRecordWriter.write
(ImmutableBytesWritable tableName, Mutation action) Writes an action (Put or Delete) to the specified table.private void
Import.Importer.writeResult
(ImmutableBytesWritable key, Result result, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Mutation>.org.apache.hadoop.mapreduce.Mapper.Context context) Modifier and TypeMethodDescription(package private) static void
HFileOutputFormat2.configurePartitioner
(org.apache.hadoop.mapreduce.Job job, List<ImmutableBytesWritable> splitPoints, boolean writeMultipleTables) Configurejob
with a TotalOrderPartitioner, partitioning againstsplitPoints
.static <K2,
V2> void MultithreadedTableMapper.setMapperClass
(org.apache.hadoop.mapreduce.Job job, Class<? extends org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, K2, V2>> cls) Set the application's mapper class.private static void
HFileOutputFormat2.writePartitions
(org.apache.hadoop.conf.Configuration conf, org.apache.hadoop.fs.Path partitionsPath, List<ImmutableBytesWritable> startKeys, boolean writeMultipleTables) Write out aSequenceFile
that can be read byTotalOrderPartitioner
that contains the split points in startKeys.ModifierConstructorDescription(package private)
Reader
(org.apache.hadoop.conf.Configuration conf, ImmutableBytesWritable startKey) -
Uses of ImmutableBytesWritable in org.apache.hadoop.hbase.mapreduce.replication
Modifier and TypeMethodDescriptionvoid
VerifyReplication.Verifier.map
(ImmutableBytesWritable row, Result value, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, ImmutableBytesWritable, Put>.org.apache.hadoop.mapreduce.Mapper.Context context) Map method that compares every scanned row with the equivalent from a distant cluster. -
Uses of ImmutableBytesWritable in org.apache.hadoop.hbase.mob.mapreduce
Modifier and TypeMethodDescriptionvoid
MobRefReporter.MobRefMapper.map
(ImmutableBytesWritable r, Result columns, org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable, Result, org.apache.hadoop.io.Text, ImmutableBytesWritable>.org.apache.hadoop.mapreduce.Mapper.Context context) Modifier and TypeMethodDescriptionprivate org.apache.hadoop.io.Text
MobRefReporter.MobRefReducer.encodeRows
(org.apache.hadoop.mapreduce.Reducer<org.apache.hadoop.io.Text, ImmutableBytesWritable, org.apache.hadoop.io.Text, org.apache.hadoop.io.Text>.org.apache.hadoop.mapreduce.Reducer.Context context, org.apache.hadoop.io.Text key, Iterable<ImmutableBytesWritable> rows) reuses the passed Text key.void
MobRefReporter.MobRefReducer.reduce
(org.apache.hadoop.io.Text key, Iterable<ImmutableBytesWritable> rows, org.apache.hadoop.mapreduce.Reducer<org.apache.hadoop.io.Text, ImmutableBytesWritable, org.apache.hadoop.io.Text, org.apache.hadoop.io.Text>.org.apache.hadoop.mapreduce.Reducer.Context context)