Package org.apache.hadoop.hbase.mapred
Class TableSnapshotInputFormat
java.lang.Object
org.apache.hadoop.hbase.mapred.TableSnapshotInputFormat
- All Implemented Interfaces:
org.apache.hadoop.mapred.InputFormat<ImmutableBytesWritable,Result>
- Direct Known Subclasses:
MultiTableSnapshotInputFormat
@Public
public class TableSnapshotInputFormat
extends Object
implements org.apache.hadoop.mapred.InputFormat<ImmutableBytesWritable,Result>
TableSnapshotInputFormat allows a MapReduce job to run over a table snapshot. Further
documentation available on
TableSnapshotInputFormat.- See Also:
-
Nested Class Summary
Nested ClassesModifier and TypeClassDescription(package private) static classstatic class -
Constructor Summary
Constructors -
Method Summary
Modifier and TypeMethodDescriptionorg.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,Result> getRecordReader(org.apache.hadoop.mapred.InputSplit split, org.apache.hadoop.mapred.JobConf job, org.apache.hadoop.mapred.Reporter reporter) org.apache.hadoop.mapred.InputSplit[]getSplits(org.apache.hadoop.mapred.JobConf job, int numSplits) static voidsetInput(org.apache.hadoop.mapred.JobConf job, String snapshotName, org.apache.hadoop.fs.Path restoreDir) Configures the job to use TableSnapshotInputFormat to read from a snapshot.static voidsetInput(org.apache.hadoop.mapred.JobConf job, String snapshotName, org.apache.hadoop.fs.Path restoreDir, RegionSplitter.SplitAlgorithm splitAlgo, int numSplitsPerRegion) Configures the job to use TableSnapshotInputFormat to read from a snapshot.
-
Constructor Details
-
TableSnapshotInputFormat
public TableSnapshotInputFormat()
-
-
Method Details
-
getSplits
public org.apache.hadoop.mapred.InputSplit[] getSplits(org.apache.hadoop.mapred.JobConf job, int numSplits) throws IOException - Specified by:
getSplitsin interfaceorg.apache.hadoop.mapred.InputFormat<ImmutableBytesWritable,Result> - Throws:
IOException
-
getRecordReader
public org.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,Result> getRecordReader(org.apache.hadoop.mapred.InputSplit split, org.apache.hadoop.mapred.JobConf job, org.apache.hadoop.mapred.Reporter reporter) throws IOException - Specified by:
getRecordReaderin interfaceorg.apache.hadoop.mapred.InputFormat<ImmutableBytesWritable,Result> - Throws:
IOException
-
setInput
public static void setInput(org.apache.hadoop.mapred.JobConf job, String snapshotName, org.apache.hadoop.fs.Path restoreDir) throws IOException Configures the job to use TableSnapshotInputFormat to read from a snapshot.- Parameters:
job- the job to configuresnapshotName- the name of the snapshot to read fromrestoreDir- a temporary directory to restore the snapshot into. Current user should have write permissions to this directory, and this should not be a subdirectory of rootdir. After the job is finished, restoreDir can be deleted.- Throws:
IOException- if an error occurs
-
setInput
public static void setInput(org.apache.hadoop.mapred.JobConf job, String snapshotName, org.apache.hadoop.fs.Path restoreDir, RegionSplitter.SplitAlgorithm splitAlgo, int numSplitsPerRegion) throws IOException Configures the job to use TableSnapshotInputFormat to read from a snapshot.- Parameters:
job- the job to configuresnapshotName- the name of the snapshot to read fromrestoreDir- a temporary directory to restore the snapshot into. Current user should have write permissions to this directory, and this should not be a subdirectory of rootdir. After the job is finished, restoreDir can be deleted.splitAlgo- split algorithm to generate splits from regionnumSplitsPerRegion- how many input splits to generate per one region- Throws:
IOException- if an error occurs
-