public class SparkHashTableSinkOperator extends TerminalOperator<SparkHashTableSinkDesc> implements Serializable
Operator.OperatorFunc, Operator.State| Modifier and Type | Field and Description |
|---|---|
protected static org.apache.commons.logging.Log |
LOG |
alias, childOperators, childOperatorsArray, childOperatorsTag, colExprMap, conf, CONTEXT_NAME_KEY, done, groupKeyObject, HIVECOUNTERCREATEDFILES, HIVECOUNTERFATAL, id, inputObjInspectors, isLogDebugEnabled, isLogInfoEnabled, isLogTraceEnabled, operatorId, out, outputObjInspector, parentOperators, PLOG, reporter, state, statsMap| Constructor and Description |
|---|
SparkHashTableSinkOperator() |
| Modifier and Type | Method and Description |
|---|---|
void |
closeOp(boolean abort)
Operator specific close routine.
|
protected void |
flushToFile(MapJoinPersistableTableContainer tableContainer,
byte tag) |
String |
getName()
Implements the getName function for the Node Interface.
|
OperatorType |
getType()
Return the type of the specific operator among the
types in OperatorType.
|
protected Collection<Future<?>> |
initializeOp(org.apache.hadoop.conf.Configuration hconf)
Operator specific initialization.
|
void |
process(Object row,
int tag)
Process the row.
|
void |
setTag(byte tag) |
acceptLimitPushdown, allInitializedParentsAreClosed, areAllParentsInitialized, augmentPlan, cleanUpInputFileChanged, cleanUpInputFileChangedOp, clone, cloneOp, cloneRecursiveChildren, close, columnNamesRowResolvedCanBeObtained, completeInitializationOp, createDummy, defaultEndGroup, defaultStartGroup, dump, dump, endGroup, flush, forward, getAdditionalCounters, getChildOperators, getChildren, getColumnExprMap, getConf, getConfiguration, getDone, getExecContext, getGroupKeyObject, getIdentifier, getInputObjInspectors, getNextCntr, getNumChild, getNumParent, getOperatorId, getOperatorName, getOpTraits, getOutputObjInspector, getParentOperators, getSchema, getStatistics, getStats, initEvaluators, initEvaluators, initEvaluatorsAndReturnStruct, initialize, initialize, initializeChildren, initializeLocalWork, initOperatorId, isUseBucketizedHiveInputFormat, jobClose, jobCloseOp, logStats, opAllowedAfterMapJoin, opAllowedBeforeMapJoin, opAllowedBeforeSortMergeJoin, opAllowedConvertMapJoin, passExecContext, preorderMap, processGroup, removeChild, removeChildAndAdoptItsChildren, removeChildren, removeParent, replaceChild, replaceParent, reset, resetId, resetStats, setAlias, setChildOperators, setColumnExprMap, setConf, setDone, setExecContext, setGroupKeyObject, setId, setInputContext, setInputObjInspectors, setOperatorId, setOpTraits, setOutputCollector, setParentOperators, setReporter, setSchema, setStatistics, setUseBucketizedHiveInputFormat, startGroup, supportAutomaticSortMergeJoin, supportSkewJoinOptimization, supportUnionRemoveOptimization, toString, toStringprotected Collection<Future<?>> initializeOp(org.apache.hadoop.conf.Configuration hconf) throws HiveException
OperatorinitializeOp in class Operator<SparkHashTableSinkDesc>HiveExceptionpublic void process(Object row, int tag) throws HiveException
Operatorprocess in class Operator<SparkHashTableSinkDesc>row - The object representing the row.tag - The tag of the row usually means which parent this row comes from.
Rows with the same tag should have exactly the same rowInspector
all the time.HiveExceptionpublic void closeOp(boolean abort)
throws HiveException
OperatorcloseOp in class Operator<SparkHashTableSinkDesc>HiveExceptionprotected void flushToFile(MapJoinPersistableTableContainer tableContainer, byte tag) throws Exception
Exceptionpublic void setTag(byte tag)
public String getName()
getName in interface NodegetName in class Operator<SparkHashTableSinkDesc>public OperatorType getType()
OperatorgetType in class Operator<SparkHashTableSinkDesc>Copyright © 2017 The Apache Software Foundation. All rights reserved.