| Interface | Description |
|---|---|
| AcidInputFormat<KEY extends org.apache.hadoop.io.WritableComparable,VALUE> |
The interface required for input formats that what to support ACID
transactions.
|
| AcidInputFormat.AcidRecordReader<K,V> |
RecordReader returned by AcidInputFormat working in row-at-a-time mode should AcidRecordReader.
|
| AcidInputFormat.RawReader<V> | |
| AcidInputFormat.RowReader<V> | |
| AcidOutputFormat<K extends org.apache.hadoop.io.WritableComparable,V> |
An extension for OutputFormats that want to implement ACID transactions.
|
| AcidUtils.Directory | |
| CombineHiveInputFormat.AvoidSplitCombination |
This is a marker interface that is used to identify the formats where
combine split generation is not applicable
|
| ContentSummaryInputFormat |
ContentSummayInputFormat provides an interface to let the input format itself
figure the content summary for a give input path.
|
| FlatFileInputFormat.SerializationContext<S> |
An implementation of SerializationContext is responsible for looking up the
Serialization implementation for the given RecordReader.
|
| HiveOutputFormat<K,V> |
HiveOutputFormat describes the output-specification for Hive's
operators. |
| HivePartitioner<K2,V2> |
Partition keys by their
Object.hashCode(). |
| InputFormatChecker |
Check for validity of the input files.
|
| RecordUpdater |
API for supporting updating records.
|
| ReworkMapredInputFormat | |
| StatsProvidingRecordReader |
If a file format internally gathers statistics (like ORC) then it can expose
the statistics through this interface.
|
| StatsProvidingRecordWriter |
If a file format internally gathers statistics (like ORC) while writing then
it can expose the statistics through this record writer interface.
|
| StorageFormatDescriptor |
Subclasses represent a storage format for the
CREATE TABLE ...
|
| Class | Description |
|---|---|
| AbstractStorageFormatDescriptor | |
| AcidInputFormat.Options |
Options for controlling the record readers.
|
| AcidOutputFormat.Options |
Options to control how the files are written
|
| AcidUtils |
Utilities that are shared by all of the ACID input and output formats.
|
| AcidUtils.ParsedDelta | |
| AvroStorageFormatDescriptor | |
| BucketizedHiveInputFormat<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> |
BucketizedHiveInputFormat serves the similar function as hiveInputFormat but
its getSplits() always group splits from one input file into one wrapper
split.
|
| BucketizedHiveInputSplit |
HiveInputSplit encapsulates an InputSplit with its corresponding
inputFormatClass.
|
| BucketizedHiveRecordReader<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> |
BucketizedHiveRecordReader is a wrapper on a list of RecordReader.
|
| CodecPool |
A global compressor/decompressor pool used to save and reuse (possibly
native) compression/decompression codecs.
|
| CombineHiveInputFormat<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> |
CombineHiveInputFormat is a parameterized InputFormat which looks at the path
name and determine the correct InputFormat for that path name from
mapredPlan.pathToPartitionInfo().
|
| CombineHiveInputFormat.CombineHiveInputSplit |
CombineHiveInputSplit encapsulates an InputSplit with its corresponding
inputFormatClassName.
|
| CombineHiveRecordReader<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> |
CombineHiveRecordReader.
|
| DefaultHivePartitioner<K2,V2> |
Partition keys by their
Object.hashCode(). |
| FlatFileInputFormat<T> | Deprecated |
| FlatFileInputFormat.RowContainer<T> |
A work-around until HADOOP-1230 is fixed.
|
| FlatFileInputFormat.SerializationContextFromConf<S> |
An implementation of
FlatFileInputFormat.SerializationContext that reads the
Serialization class and specific subclass to be deserialized from the
JobConf. |
| HiveBinaryOutputFormat<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> |
HiveBinaryOutputFormat writes out the values consecutively without any
separators.
|
| HiveContextAwareRecordReader<K,V> |
This class prepares an IOContext, and provides the ability to perform a binary search on the
data.
|
| HiveFileFormatUtils |
An util class for various Hive file format tasks.
|
| HiveFileFormatUtils.NullOutputCommitter | |
| HiveIgnoreKeyTextOutputFormat<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> |
HiveIgnoreKeyTextOutputFormat replaces key with null before feeding the
|
| HiveIgnoreKeyTextOutputFormat.IgnoreKeyWriter<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> | |
| HiveInputFormat<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> |
HiveInputFormat is a parameterized InputFormat which looks at the path name
and determine the correct InputFormat for that path name from
mapredPlan.pathToPartitionInfo().
|
| HiveInputFormat.HiveInputSplit |
HiveInputSplit encapsulates an InputSplit with its corresponding
inputFormatClass.
|
| HiveKey |
HiveKey is a simple wrapper on Text which allows us to set the hashCode
easily.
|
| HiveKey.Comparator |
A Comparator optimized for HiveKey.
|
| HiveNullValueSequenceFileOutputFormat<K,V> |
A
HiveOutputFormat that writes SequenceFiles with the
content saved in the keys, and null in the values. |
| HiveOutputFormatImpl<K extends org.apache.hadoop.io.WritableComparable<K>,V extends org.apache.hadoop.io.Writable> |
Hive does not use OutputFormat's in a conventional way, but constructs and uses
the defined OutputFormat for each table from FileSinkOperator.
|
| HivePassThroughOutputFormat<K,V> |
This pass through class is used to wrap OutputFormat implementations such that new OutputFormats not derived from
HiveOutputFormat gets through the checker
|
| HivePassThroughRecordWriter<K extends org.apache.hadoop.io.WritableComparable<?>,V extends org.apache.hadoop.io.Writable> | |
| HiveRecordReader<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> |
HiveRecordReader is a simple wrapper on RecordReader.
|
| HiveSequenceFileOutputFormat<K,V> |
A
HiveOutputFormat that writes SequenceFiles. |
| IgnoreKeyTextOutputFormat<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> | Deprecated
use
HiveIgnoreKeyTextOutputFormat instead} |
| IgnoreKeyTextOutputFormat.IgnoreKeyWriter<K extends org.apache.hadoop.io.WritableComparable,V extends org.apache.hadoop.io.Writable> | |
| IOConstants | |
| IOContext |
IOContext basically contains the position information of the current
key/value.
|
| IOPrepareCache |
IOPrepareCache is used to cache pre-query io-related objects.
|
| NonSyncDataInputBuffer |
A thread-not-safe version of Hadoop's DataInputBuffer, which removes all
synchronized modifiers.
|
| NonSyncDataOutputBuffer |
A thread-not-safe version of Hadoop's DataOutputBuffer, which removes all
synchronized modifiers.
|
| NullRowsInputFormat |
NullRowsInputFormat outputs null rows, maximum 100.
|
| NullRowsInputFormat.DummyInputSplit | |
| NullRowsInputFormat.NullRowsRecordReader | |
| OneNullRowInputFormat |
OneNullRowInputFormat outputs one null row.
|
| OneNullRowInputFormat.OneNullRowRecordReader | |
| ORCFileStorageFormatDescriptor | |
| ParquetFileStorageFormatDescriptor | |
| RCFile |
RCFiles, short of Record Columnar File, are flat files
consisting of binary key/value pairs, which shares much similarity with
SequenceFile. |
| RCFile.KeyBuffer |
KeyBuffer is the key of each record in RCFile.
|
| RCFile.Reader |
Read KeyBuffer/ValueBuffer pairs from a RCFile.
|
| RCFile.ValueBuffer |
ValueBuffer is the value of each record in RCFile.
|
| RCFile.Writer |
Write KeyBuffer/ValueBuffer pairs to a RCFile.
|
| RCFileInputFormat<K extends org.apache.hadoop.io.LongWritable,V extends BytesRefArrayWritable> |
RCFileInputFormat.
|
| RCFileOutputFormat |
RCFileOutputFormat.
|
| RCFileRecordReader<K extends org.apache.hadoop.io.LongWritable,V extends BytesRefArrayWritable> |
RCFileRecordReader.
|
| RCFileStorageFormatDescriptor | |
| RecordIdentifier |
Gives the Record identifier information for the current record.
|
| RecordIdentifier.StructInfo |
RecordIdentifier is passed along the operator tree as a struct.
|
| SchemaAwareCompressionInputStream |
SchemaAwareCompressionInputStream adds the ability to inform the compression
stream what column is being read.
|
| SchemaAwareCompressionOutputStream |
SchemaAwareCompressionOutputStream adds the ability to inform the comression stream
the current column being compressed.
|
| SequenceFileInputFormatChecker |
SequenceFileInputFormatChecker.
|
| SequenceFileStorageFormatDescriptor | |
| StorageFormatFactory | |
| SymbolicInputFormat | |
| SymlinkTextInputFormat |
Symlink file is a text file which contains a list of filename / dirname.
|
| SymlinkTextInputFormat.SymlinkTextInputSplit |
This input split wraps the FileSplit generated from
TextInputFormat.getSplits(), while setting the original link file path
as job input path.
|
| TextFileStorageFormatDescriptor | |
| VectorizedRCFileInputFormat |
A MapReduce/Hive Vectorized input format for RC files.
|
| VectorizedRCFileRecordReader |
RCFileRecordReader.
|
| Enum | Description |
|---|---|
| AcidUtils.Operation | |
| IOContext.Comparison | |
| RecordIdentifier.Field |
This is in support of
VirtualColumn.ROWID
Contains metadata about each field in RecordIdentifier that needs to be part of ROWID
which is represented as a struct RecordIdentifier.StructInfo. |
Copyright © 2017 The Apache Software Foundation. All rights reserved.