Package | Description |
---|---|
org.apache.hadoop.hbase.io | |
org.apache.hadoop.hbase.mapred |
Provides HBase MapReduce
Input/OutputFormats, a table indexing MapReduce job, and utility methods.
|
org.apache.hadoop.hbase.mapreduce |
Provides HBase MapReduce
Input/OutputFormats, a table indexing MapReduce job, and utility methods.
|
org.apache.hadoop.hbase.mapreduce.replication |
Modifier and Type | Method and Description |
---|---|
int |
ImmutableBytesWritable.compareTo(ImmutableBytesWritable that)
Define the sort order of the BytesWritable.
|
Constructor and Description |
---|
ImmutableBytesWritable(ImmutableBytesWritable ibw)
Set the new ImmutableBytesWritable to the contents of the passed
ibw . |
Modifier and Type | Method and Description |
---|---|
protected ImmutableBytesWritable |
GroupingTableMap.createGroupKey(byte[][] vals)
Create a key by concatenating multiple column values.
|
ImmutableBytesWritable |
TableRecordReaderImpl.createKey() |
ImmutableBytesWritable |
TableRecordReader.createKey() |
Modifier and Type | Method and Description |
---|---|
org.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,Result> |
TableInputFormatBase.getRecordReader(org.apache.hadoop.mapred.InputSplit split,
org.apache.hadoop.mapred.JobConf job,
org.apache.hadoop.mapred.Reporter reporter)
Builds a TableRecordReader.
|
org.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,Result> |
TableSnapshotInputFormat.getRecordReader(org.apache.hadoop.mapred.InputSplit split,
org.apache.hadoop.mapred.JobConf job,
org.apache.hadoop.mapred.Reporter reporter) |
org.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,Result> |
MultiTableSnapshotInputFormat.getRecordReader(org.apache.hadoop.mapred.InputSplit split,
org.apache.hadoop.mapred.JobConf job,
org.apache.hadoop.mapred.Reporter reporter) |
Modifier and Type | Method and Description |
---|---|
int |
HRegionPartitioner.getPartition(ImmutableBytesWritable key,
V2 value,
int numPartitions) |
void |
IdentityTableMap.map(ImmutableBytesWritable key,
Result value,
org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output,
org.apache.hadoop.mapred.Reporter reporter)
Pass the key, value to reduce
|
void |
GroupingTableMap.map(ImmutableBytesWritable key,
Result value,
org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output,
org.apache.hadoop.mapred.Reporter reporter)
Extract the grouping columns from value to construct a new key.
|
boolean |
TableRecordReaderImpl.next(ImmutableBytesWritable key,
Result value) |
boolean |
TableRecordReader.next(ImmutableBytesWritable key,
Result value) |
void |
IdentityTableReduce.reduce(ImmutableBytesWritable key,
Iterator<Put> values,
org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Put> output,
org.apache.hadoop.mapred.Reporter reporter)
No aggregation, output pairs of (key, record)
|
void |
TableOutputFormat.TableRecordWriter.write(ImmutableBytesWritable key,
Put value) |
Modifier and Type | Method and Description |
---|---|
void |
IdentityTableMap.map(ImmutableBytesWritable key,
Result value,
org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output,
org.apache.hadoop.mapred.Reporter reporter)
Pass the key, value to reduce
|
void |
GroupingTableMap.map(ImmutableBytesWritable key,
Result value,
org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output,
org.apache.hadoop.mapred.Reporter reporter)
Extract the grouping columns from value to construct a new key.
|
void |
IdentityTableReduce.reduce(ImmutableBytesWritable key,
Iterator<Put> values,
org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Put> output,
org.apache.hadoop.mapred.Reporter reporter)
No aggregation, output pairs of (key, record)
|
Modifier and Type | Method and Description |
---|---|
protected ImmutableBytesWritable |
GroupingTableMapper.createGroupKey(byte[][] vals)
Create a key by concatenating multiple column values.
|
ImmutableBytesWritable |
HashTable.TableHash.Reader.getCurrentHash()
Get the current hash
|
ImmutableBytesWritable |
HashTable.TableHash.Reader.getCurrentKey()
Get the current key
|
ImmutableBytesWritable |
TableRecordReaderImpl.getCurrentKey()
Returns the current key.
|
ImmutableBytesWritable |
TableRecordReader.getCurrentKey()
Returns the current key.
|
ImmutableBytesWritable |
TableSnapshotInputFormatImpl.RecordReader.getCurrentKey() |
Modifier and Type | Method and Description |
---|---|
org.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,Result> |
TableSnapshotInputFormat.createRecordReader(org.apache.hadoop.mapreduce.InputSplit split,
org.apache.hadoop.mapreduce.TaskAttemptContext context) |
org.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,Result> |
TableInputFormatBase.createRecordReader(org.apache.hadoop.mapreduce.InputSplit split,
org.apache.hadoop.mapreduce.TaskAttemptContext context)
Builds a
TableRecordReader . |
org.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,Result> |
MultiTableInputFormatBase.createRecordReader(org.apache.hadoop.mapreduce.InputSplit split,
org.apache.hadoop.mapreduce.TaskAttemptContext context)
Builds a TableRecordReader.
|
static <K2,V2> Class<org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable,Result,K2,V2>> |
MultithreadedTableMapper.getMapperClass(org.apache.hadoop.mapreduce.JobContext job)
Get the application's mapper class.
|
org.apache.hadoop.mapreduce.RecordWriter<ImmutableBytesWritable,Mutation> |
MultiTableOutputFormat.getRecordWriter(org.apache.hadoop.mapreduce.TaskAttemptContext context) |
org.apache.hadoop.mapreduce.RecordWriter<ImmutableBytesWritable,Cell> |
HFileOutputFormat2.getRecordWriter(org.apache.hadoop.mapreduce.TaskAttemptContext context) |
Modifier and Type | Method and Description |
---|---|
static byte[] |
MultiTableHFileOutputFormat.createCompositeKey(byte[] tableName,
ImmutableBytesWritable suffix)
Alternate api which accepts an ImmutableBytesWritable for the suffix
|
static byte[] |
MultiTableHFileOutputFormat.createCompositeKey(String tableName,
ImmutableBytesWritable suffix)
Alternate api which accepts a String for the tableName and ImmutableBytesWritable for the
suffix
|
int |
HRegionPartitioner.getPartition(ImmutableBytesWritable key,
VALUE value,
int numPartitions)
Gets the partition number for a given key (hence record) given the total
number of partitions i.e.
|
int |
SimpleTotalOrderPartitioner.getPartition(ImmutableBytesWritable key,
VALUE value,
int reduces) |
void |
Import.CellImporter.map(ImmutableBytesWritable row,
Result value,
org.apache.hadoop.mapreduce.Mapper.Context context) |
protected void |
HashTable.HashMapper.map(ImmutableBytesWritable key,
Result value,
org.apache.hadoop.mapreduce.Mapper.Context context) |
void |
Import.KeyValueImporter.map(ImmutableBytesWritable row,
Result value,
org.apache.hadoop.mapreduce.Mapper.Context context)
Deprecated.
|
protected void |
SyncTable.SyncMapper.map(ImmutableBytesWritable key,
Result value,
org.apache.hadoop.mapreduce.Mapper.Context context) |
void |
Import.Importer.map(ImmutableBytesWritable row,
Result value,
org.apache.hadoop.mapreduce.Mapper.Context context) |
protected void |
IndexBuilder.Map.map(ImmutableBytesWritable rowKey,
Result result,
org.apache.hadoop.mapreduce.Mapper.Context context) |
void |
IdentityTableMapper.map(ImmutableBytesWritable key,
Result value,
org.apache.hadoop.mapreduce.Mapper.Context context)
Pass the key, value to reduce.
|
void |
GroupingTableMapper.map(ImmutableBytesWritable key,
Result value,
org.apache.hadoop.mapreduce.Mapper.Context context)
Extract the grouping columns from value to construct a new key.
|
void |
Import.CellSortImporter.map(ImmutableBytesWritable row,
Result value,
org.apache.hadoop.mapreduce.Mapper.Context context) |
void |
Import.KeyValueSortImporter.map(ImmutableBytesWritable row,
Result value,
org.apache.hadoop.mapreduce.Mapper.Context context)
Deprecated.
|
HashTable.TableHash.Reader |
HashTable.TableHash.newReader(org.apache.hadoop.conf.Configuration conf,
ImmutableBytesWritable startKey)
Open a TableHash.Reader starting at the first hash at or after the given key.
|
protected void |
Import.Importer.processKV(ImmutableBytesWritable key,
Result result,
org.apache.hadoop.mapreduce.Mapper.Context context,
Put put,
Delete delete) |
protected void |
CellSortReducer.reduce(ImmutableBytesWritable row,
Iterable<Cell> kvs,
org.apache.hadoop.mapreduce.Reducer.Context context) |
protected void |
KeyValueSortReducer.reduce(ImmutableBytesWritable row,
Iterable<KeyValue> kvs,
org.apache.hadoop.mapreduce.Reducer.Context context)
Deprecated.
|
protected void |
PutSortReducer.reduce(ImmutableBytesWritable row,
Iterable<Put> puts,
org.apache.hadoop.mapreduce.Reducer.Context context) |
protected void |
TextSortReducer.reduce(ImmutableBytesWritable rowKey,
Iterable<org.apache.hadoop.io.Text> lines,
org.apache.hadoop.mapreduce.Reducer.Context context) |
void |
MultiTableOutputFormat.MultiTableRecordWriter.write(ImmutableBytesWritable tableName,
Mutation action)
Writes an action (Put or Delete) to the specified table.
|
Modifier and Type | Method and Description |
---|---|
static <K2,V2> void |
MultithreadedTableMapper.setMapperClass(org.apache.hadoop.mapreduce.Job job,
Class<? extends org.apache.hadoop.mapreduce.Mapper<ImmutableBytesWritable,Result,K2,V2>> cls)
Set the application's mapper class.
|
Modifier and Type | Method and Description |
---|---|
void |
VerifyReplication.Verifier.map(ImmutableBytesWritable row,
Result value,
org.apache.hadoop.mapreduce.Mapper.Context context)
Map method that compares every scanned row with the equivalent from
a distant cluster.
|
Copyright © 2007–2019 Cloudera. All rights reserved.