Package | Description |
---|---|
org.apache.hadoop.hbase | |
org.apache.hadoop.hbase.client |
Provides HBase Client
|
org.apache.hadoop.hbase.client.coprocessor |
Provides client classes for invoking Coprocessor RPC protocols
Overview
Example Usage
|
org.apache.hadoop.hbase.client.example | |
org.apache.hadoop.hbase.mapred |
Provides HBase MapReduce
Input/OutputFormats, a table indexing MapReduce job, and utility methods.
|
org.apache.hadoop.hbase.mapreduce |
Provides HBase MapReduce
Input/OutputFormats, a table indexing MapReduce job, and utility methods.
|
org.apache.hadoop.hbase.rest.client | |
org.apache.hadoop.hbase.thrift |
Provides an HBase Thrift
service.
|
org.apache.hadoop.hbase.tool | |
org.apache.hadoop.hbase.util |
Modifier and Type | Method and Description |
---|---|
static Table |
MetaTableAccessor.getMetaHTable(Connection connection)
Callers should call close on the returned
Table instance. |
Modifier and Type | Class and Description |
---|---|
class |
HTable
An implementation of
Table . |
Modifier and Type | Method and Description |
---|---|
Table |
TableBuilder.build()
Create the
Table instance. |
default Table |
Connection.getTable(TableName tableName)
Retrieve a Table implementation for accessing a table.
|
default Table |
Connection.getTable(TableName tableName,
ExecutorService pool)
Retrieve a Table implementation for accessing a table.
|
Constructor and Description |
---|
SecureBulkLoadClient(org.apache.hadoop.conf.Configuration conf,
Table table) |
Modifier and Type | Method and Description |
---|---|
<R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message> |
AggregationClient.avg(Table table,
ColumnInterpreter<R,S,P,Q,T> ci,
Scan scan)
This is the client side interface/handle for calling the average method for
a given cf-cq combination.
|
<R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message> |
AggregationClient.max(Table table,
ColumnInterpreter<R,S,P,Q,T> ci,
Scan scan)
It gives the maximum value of a column for a given column family for the
given range.
|
<R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message> |
AggregationClient.median(Table table,
ColumnInterpreter<R,S,P,Q,T> ci,
Scan scan)
This is the client side interface/handler for calling the median method for a
given cf-cq combination.
|
<R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message> |
AggregationClient.min(Table table,
ColumnInterpreter<R,S,P,Q,T> ci,
Scan scan)
It gives the minimum value of a column for a given column family for the
given range.
|
<R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message> |
AggregationClient.rowCount(Table table,
ColumnInterpreter<R,S,P,Q,T> ci,
Scan scan)
It gives the row count, by summing up the individual results obtained from
regions.
|
<R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message> |
AggregationClient.std(Table table,
ColumnInterpreter<R,S,P,Q,T> ci,
Scan scan)
This is the client side interface/handle for calling the std method for a
given cf-cq combination.
|
<R,S,P extends com.google.protobuf.Message,Q extends com.google.protobuf.Message,T extends com.google.protobuf.Message> |
AggregationClient.sum(Table table,
ColumnInterpreter<R,S,P,Q,T> ci,
Scan scan)
It sums up the value returned from various regions.
|
Modifier and Type | Method and Description |
---|---|
void |
RefreshHFilesClient.refreshHFiles(Table table) |
Modifier and Type | Method and Description |
---|---|
protected Table |
TableInputFormatBase.getTable()
Allows subclasses to get the
Table . |
Modifier and Type | Method and Description |
---|---|
void |
TableRecordReaderImpl.setHTable(Table htable) |
void |
TableRecordReader.setHTable(Table htable) |
Modifier and Type | Method and Description |
---|---|
protected Table |
TableInputFormatBase.getTable()
Allows subclasses to get the
Table . |
Modifier and Type | Method and Description |
---|---|
static void |
HFileOutputFormat2.configureIncrementalLoad(org.apache.hadoop.mapreduce.Job job,
Table table,
RegionLocator regionLocator)
Configure a MapReduce Job to perform an incremental load into the given
table.
|
void |
TableRecordReaderImpl.setHTable(Table htable)
Sets the HBase table.
|
void |
TableRecordReader.setTable(Table table) |
Modifier and Type | Class and Description |
---|---|
class |
RemoteHTable
HTable interface to remote tables accessed via REST gateway
|
Modifier and Type | Method and Description |
---|---|
Table |
ThriftServerRunner.HBaseHandler.getTable(byte[] tableName)
Creates and returns a Table instance from a given table name.
|
Table |
ThriftServerRunner.HBaseHandler.getTable(ByteBuffer tableName) |
Modifier and Type | Method and Description |
---|---|
protected void |
LoadIncrementalHFiles.bulkLoadPhase(Table table,
Connection conn,
ExecutorService pool,
Deque<LoadIncrementalHFiles.LoadQueueItem> queue,
org.apache.hbase.thirdparty.com.google.common.collect.Multimap<ByteBuffer,LoadIncrementalHFiles.LoadQueueItem> regionGroups,
boolean copyFile,
Map<LoadIncrementalHFiles.LoadQueueItem,ByteBuffer> item2RegionMap)
This takes the LQI's grouped by likely regions and attempts to bulk load them.
|
Map<LoadIncrementalHFiles.LoadQueueItem,ByteBuffer> |
LoadIncrementalHFiles.doBulkLoad(Map<byte[],List<org.apache.hadoop.fs.Path>> map,
Admin admin,
Table table,
RegionLocator regionLocator,
boolean silence,
boolean copyFile)
Perform a bulk load of the given directory into the given pre-existing table.
|
Map<LoadIncrementalHFiles.LoadQueueItem,ByteBuffer> |
LoadIncrementalHFiles.doBulkLoad(org.apache.hadoop.fs.Path hfofDir,
Admin admin,
Table table,
RegionLocator regionLocator)
Perform a bulk load of the given directory into the given pre-existing table.
|
Map<LoadIncrementalHFiles.LoadQueueItem,ByteBuffer> |
LoadIncrementalHFiles.doBulkLoad(org.apache.hadoop.fs.Path hfofDir,
Admin admin,
Table table,
RegionLocator regionLocator,
boolean silence,
boolean copyFile)
Perform a bulk load of the given directory into the given pre-existing table.
|
protected Pair<List<LoadIncrementalHFiles.LoadQueueItem>,String> |
LoadIncrementalHFiles.groupOrSplit(org.apache.hbase.thirdparty.com.google.common.collect.Multimap<ByteBuffer,LoadIncrementalHFiles.LoadQueueItem> regionGroups,
LoadIncrementalHFiles.LoadQueueItem item,
Table table,
Pair<byte[][],byte[][]> startEndKeys)
Attempt to assign the given load queue item into its target region group.
|
void |
LoadIncrementalHFiles.loadHFileQueue(Table table,
Connection conn,
Deque<LoadIncrementalHFiles.LoadQueueItem> queue,
Pair<byte[][],byte[][]> startEndKeys)
Used by the replication sink to load the hfiles from the source cluster.
|
void |
LoadIncrementalHFiles.loadHFileQueue(Table table,
Connection conn,
Deque<LoadIncrementalHFiles.LoadQueueItem> queue,
Pair<byte[][],byte[][]> startEndKeys,
boolean copyFile)
Used by the replication sink to load the hfiles from the source cluster.
|
void |
LoadIncrementalHFiles.prepareHFileQueue(Map<byte[],List<org.apache.hadoop.fs.Path>> map,
Table table,
Deque<LoadIncrementalHFiles.LoadQueueItem> queue,
boolean silence)
Prepare a collection of
LoadIncrementalHFiles.LoadQueueItem from list of source hfiles contained in the
passed directory and validates whether the prepared queue has all the valid table column
families in it. |
void |
LoadIncrementalHFiles.prepareHFileQueue(org.apache.hadoop.fs.Path hfilesDir,
Table table,
Deque<LoadIncrementalHFiles.LoadQueueItem> queue,
boolean validateHFile)
Prepare a collection of
LoadIncrementalHFiles.LoadQueueItem from list of source hfiles contained in the
passed directory and validates whether the prepared queue has all the valid table column
families in it. |
void |
LoadIncrementalHFiles.prepareHFileQueue(org.apache.hadoop.fs.Path hfilesDir,
Table table,
Deque<LoadIncrementalHFiles.LoadQueueItem> queue,
boolean validateHFile,
boolean silence)
Prepare a collection of
LoadIncrementalHFiles.LoadQueueItem from list of source hfiles contained in the
passed directory and validates whether the prepared queue has all the valid table column
families in it. |
Modifier and Type | Method and Description |
---|---|
Table |
ConnectionCache.getTable(String tableName)
Caller closes the table afterwards.
|
Copyright © 2007–2019 Cloudera. All rights reserved.