Package | Description |
---|---|
org.apache.hadoop.hbase | |
org.apache.hadoop.hbase.client |
Provides HBase Client
|
org.apache.hadoop.hbase.mapred |
Provides HBase MapReduce
Input/OutputFormats, a table indexing MapReduce job, and utility
|
org.apache.hadoop.hbase.mapreduce |
Provides HBase MapReduce
Input/OutputFormats, a table indexing MapReduce job, and utility
|
org.apache.hadoop.hbase.rest.client |
Modifier and Type | Method and Description |
---|---|
static PairOfSameType<HRegionInfo> |
HRegionInfo.getDaughterRegions(Result data)
Returns the daughter regions by reading the corresponding columns of the catalog table
Result.
|
static HRegionInfo |
HRegionInfo.getHRegionInfo(Result data)
Returns HRegionInfo object from the column
HConstants.CATALOG_FAMILY:HConstants.REGIONINFO_QUALIFIER of the catalog
table Result.
|
static HRegionInfo |
HRegionInfo.getHRegionInfo(Result r,
byte[] qualifier)
Returns the HRegionInfo object from the column
HConstants.CATALOG_FAMILY and
qualifier of the catalog table result. |
static Pair<HRegionInfo,ServerName> |
HRegionInfo.getHRegionInfoAndServerName(Result r)
Extract a HRegionInfo and ServerName from catalog table
Result . |
static PairOfSameType<HRegionInfo> |
HRegionInfo.getMergeRegions(Result data)
Returns the merge regions by reading the corresponding columns of the catalog table
Result.
|
static long |
HRegionInfo.getSeqNumDuringOpen(Result r)
The latest seqnum that the server writing to meta observed when opening the region.
|
static ServerName |
HRegionInfo.getServerName(Result r)
Returns a
ServerName from catalog table Result . |
Modifier and Type | Field and Description |
---|---|
static Result |
Result.EMPTY_RESULT |
protected Result |
ClientScanner.lastResult |
Modifier and Type | Field and Description |
---|---|
protected LinkedList<Result> |
ClientScanner.cache |
Modifier and Type | Method and Description |
---|---|
Result |
HTable.append(Append append)
Appends values to one or more columns within a single row.
|
Result |
HTableInterface.append(Append append)
Appends values to one or more columns within a single row.
|
static Result |
Result.create(Cell[] cells)
Instantiate a Result with the specified array of KeyValues.
|
static Result |
Result.create(List<Cell> cells)
Instantiate a Result with the specified List of KeyValues.
|
static Result |
Result.create(List<Cell> cells,
Boolean exists) |
Result |
HTable.get(Get get)
Extracts certain cells from a given row.
|
Result |
HTableInterface.get(Get get)
Extracts certain cells from a given row.
|
Result[] |
HTable.get(List<Get> gets)
Extracts certain cells from the given rows, in batch.
|
Result[] |
HTableInterface.get(List<Get> gets)
Extracts certain cells from the given rows, in batch.
|
Result |
HTable.getRowOrBefore(byte[] row,
byte[] family)
Return the row that matches row exactly,
or the one that immediately precedes it.
|
Result |
HTableInterface.getRowOrBefore(byte[] row,
byte[] family)
Deprecated.
As of version 0.92 this method is deprecated without
replacement.
getRowOrBefore is used internally to find entries in hbase:meta and makes
various assumptions about the table (which are true for hbase:meta but not
in general) to be efficient.
|
Result |
HTable.increment(Increment increment)
Increments one or more columns within a single row.
|
Result |
HTableInterface.increment(Increment increment)
Increments one or more columns within a single row.
|
Result |
TableSnapshotScanner.next() |
Result |
ResultScanner.next()
Grab the next row's worth of values.
|
Result |
ClientSmallScanner.next() |
Result |
ClientScanner.next() |
Result |
ClientSmallReversedScanner.next() |
Result[] |
ResultScanner.next(int nbRows) |
Modifier and Type | Method and Description |
---|---|
static void |
Result.compareResults(Result res1,
Result res2)
Does a deep comparison of two Results, down to the byte arrays.
|
void |
Result.copyFrom(Result other)
Copy another Result into this one.
|
Modifier and Type | Method and Description |
---|---|
Result |
TableRecordReaderImpl.createValue()
Deprecated.
|
Result |
TableRecordReader.createValue()
Deprecated.
|
Modifier and Type | Method and Description |
---|---|
org.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,Result> |
TableInputFormatBase.getRecordReader(org.apache.hadoop.mapred.InputSplit split,
org.apache.hadoop.mapred.JobConf job,
org.apache.hadoop.mapred.Reporter reporter)
Deprecated.
Builds a TableRecordReader.
|
org.apache.hadoop.mapred.RecordReader<ImmutableBytesWritable,Result> |
TableSnapshotInputFormat.getRecordReader(org.apache.hadoop.mapred.InputSplit split,
org.apache.hadoop.mapred.JobConf job,
org.apache.hadoop.mapred.Reporter reporter) |
Modifier and Type | Method and Description |
---|---|
protected byte[][] |
GroupingTableMap.extractKeyValues(Result r)
Deprecated.
Extract columns values from the current record.
|
void |
IdentityTableMap.map(ImmutableBytesWritable key,
Result value,
org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output,
org.apache.hadoop.mapred.Reporter reporter)
Deprecated.
Pass the key, value to reduce
|
void |
GroupingTableMap.map(ImmutableBytesWritable key,
Result value,
org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output,
org.apache.hadoop.mapred.Reporter reporter)
Deprecated.
Extract the grouping columns from value to construct a new key.
|
boolean |
TableRecordReaderImpl.next(ImmutableBytesWritable key,
Result value)
Deprecated.
|
boolean |
TableRecordReader.next(ImmutableBytesWritable key,
Result value)
Deprecated.
|
Modifier and Type | Method and Description |
---|---|
void |
IdentityTableMap.map(ImmutableBytesWritable key,
Result value,
org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output,
org.apache.hadoop.mapred.Reporter reporter)
Deprecated.
Pass the key, value to reduce
|
void |
GroupingTableMap.map(ImmutableBytesWritable key,
Result value,
org.apache.hadoop.mapred.OutputCollector<ImmutableBytesWritable,Result> output,
org.apache.hadoop.mapred.Reporter reporter)
Deprecated.
Extract the grouping columns from value to construct a new key.
|
Modifier and Type | Method and Description |
---|---|
Result |
TableRecordReaderImpl.getCurrentValue()
Returns the current value.
|
Result |
TableRecordReader.getCurrentValue()
Returns the current value.
|
Modifier and Type | Method and Description |
---|---|
org.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,Result> |
TableInputFormatBase.createRecordReader(org.apache.hadoop.mapreduce.InputSplit split,
org.apache.hadoop.mapreduce.TaskAttemptContext context)
Builds a TableRecordReader.
|
org.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,Result> |
MultiTableInputFormatBase.createRecordReader(org.apache.hadoop.mapreduce.InputSplit split,
org.apache.hadoop.mapreduce.TaskAttemptContext context)
Builds a TableRecordReader.
|
org.apache.hadoop.mapreduce.RecordReader<ImmutableBytesWritable,Result> |
TableSnapshotInputFormat.createRecordReader(org.apache.hadoop.mapreduce.InputSplit split,
org.apache.hadoop.mapreduce.TaskAttemptContext context) |
Modifier and Type | Method and Description |
---|---|
protected byte[][] |
GroupingTableMapper.extractKeyValues(Result r)
Extract columns values from the current record.
|
void |
IdentityTableMapper.map(ImmutableBytesWritable key,
Result value,
org.apache.hadoop.mapreduce.Mapper.Context context)
Pass the key, value to reduce.
|
void |
GroupingTableMapper.map(ImmutableBytesWritable key,
Result value,
org.apache.hadoop.mapreduce.Mapper.Context context)
Extract the grouping columns from value to construct a new key.
|
Modifier and Type | Method and Description |
---|---|
Result |
RemoteHTable.append(Append append) |
protected Result[] |
RemoteHTable.buildResultFromModel(org.apache.hadoop.hbase.rest.model.CellSetModel model) |
Result |
RemoteHTable.get(Get get) |
Result[] |
RemoteHTable.get(List<Get> gets) |
Result |
RemoteHTable.getRowOrBefore(byte[] row,
byte[] family) |
Result |
RemoteHTable.increment(Increment increment) |
Copyright © 2014 The Apache Software Foundation. All rights reserved.