|
||||||||||
PREV NEXT | FRAMES NO FRAMES |
Uses of HiveException in org.apache.hadoop.hive.ql |
---|
Methods in org.apache.hadoop.hive.ql that throw HiveException | |
---|---|
boolean |
DriverContext.addToRunnable(Task<? extends Serializable> tsk)
|
Task<? extends Serializable> |
DriverContext.getRunnable(int maxthreads)
|
void |
DriverContext.launching(TaskRunner runner)
|
Uses of HiveException in org.apache.hadoop.hive.ql.exec |
---|
Subclasses of HiveException in org.apache.hadoop.hive.ql.exec | |
---|---|
class |
AmbiguousMethodException
Exception thrown by the UDF and UDAF method resolvers in case a unique method is not found. |
class |
NoMatchingMethodException
Exception thrown by the UDF and UDAF method resolvers in case no matching method is found. |
class |
UDFArgumentException
exception class, thrown when udf argument have something wrong. |
class |
UDFArgumentLengthException
exception class, thrown when udf arguments have wrong length. |
class |
UDFArgumentTypeException
exception class, thrown when udf arguments have wrong types. |
Methods in org.apache.hadoop.hive.ql.exec that throw HiveException | ||
---|---|---|
void |
FileSinkOperator.FSPaths.abortWriters(org.apache.hadoop.fs.FileSystem fs,
boolean abort,
boolean delete)
|
|
static void |
FunctionTask.addFunctionResources(List<org.apache.hadoop.hive.metastore.api.ResourceUri> resources)
|
|
static URI |
ArchiveUtils.addSlash(URI u)
Makes sure, that URI points to directory by adding slash to it. |
|
void |
PTFPartition.append(Object o)
|
|
void |
Operator.cleanUpInputFileChanged()
|
|
void |
MapJoinOperator.cleanUpInputFileChangedOp()
|
|
void |
MapOperator.cleanUpInputFileChangedOp()
|
|
void |
Operator.cleanUpInputFileChangedOp()
|
|
void |
SMBMapJoinOperator.cleanUpInputFileChangedOp()
|
|
void |
TableScanOperator.cleanUpInputFileChangedOp()
|
|
void |
FetchTask.clearFetch()
Clear the Fetch Operator. |
|
void |
FetchOperator.clearFetchContext()
Clear the context, if anything needs to be done. |
|
void |
Operator.close(boolean abort)
|
|
void |
ScriptOperator.close(boolean abort)
|
|
void |
SkewJoinHandler.close(boolean abort)
|
|
void |
AbstractMapJoinOperator.closeOp(boolean abort)
|
|
void |
CommonJoinOperator.closeOp(boolean abort)
All done. |
|
void |
FileSinkOperator.closeOp(boolean abort)
|
|
void |
GroupByOperator.closeOp(boolean abort)
We need to forward all the aggregations to children. |
|
void |
HashTableDummyOperator.closeOp(boolean abort)
|
|
void |
HashTableSinkOperator.closeOp(boolean abort)
|
|
void |
JoinOperator.closeOp(boolean abort)
All done. |
|
void |
LimitOperator.closeOp(boolean abort)
|
|
void |
MapJoinOperator.closeOp(boolean abort)
|
|
void |
MapOperator.closeOp(boolean abort)
close extra child operators that are initialized but are not executed. |
|
void |
SMBMapJoinOperator.closeOp(boolean abort)
|
|
void |
TableScanOperator.closeOp(boolean abort)
|
|
void |
FileSinkOperator.FSPaths.closeWriters(boolean abort)
|
|
Integer |
ExprNodeGenericFuncEvaluator.compare(Object row)
If the genericUDF is a base comparison, it returns an integer based on the result of comparing the two sides of the UDF, like the compareTo method in Comparable. |
|
static ArrayList<Object> |
JoinUtil.computeKeys(Object row,
List<ExprNodeEvaluator> keyFields,
List<ObjectInspector> keyFieldsOI)
Return the key as a standard object. |
|
static Object[] |
JoinUtil.computeMapJoinValues(Object row,
List<ExprNodeEvaluator> valueFields,
List<ObjectInspector> valueFieldsOI,
List<ExprNodeEvaluator> filters,
List<ObjectInspector> filtersOI,
int[] filterMap)
Return the value as a standard object. |
|
static List<Object> |
JoinUtil.computeValues(Object row,
List<ExprNodeEvaluator> valueFields,
List<ObjectInspector> valueFieldsOI,
boolean hasFilter)
Return the value as a standard object. |
|
static String |
ArchiveUtils.conflictingArchiveNameOrNull(Hive db,
Table tbl,
LinkedHashMap<String,String> partSpec)
Determines if one can insert into partition(s), or there's a conflict with archive. |
|
static void |
PTFOperator.connectLeadLagFunctionsToPartition(PTFDesc ptfDesc,
PTFPartition.PTFPartitionIterator<Object> pItr)
|
|
static PTFPartition |
PTFPartition.create(HiveConf cfg,
SerDe serDe,
StructObjectInspector inputOI,
StructObjectInspector outputOI)
|
|
static ArchiveUtils.PartSpecInfo |
ArchiveUtils.PartSpecInfo.create(Table tbl,
Map<String,String> partSpec)
Extract partial prefix specification from table and key-value map |
|
PTFPartition |
PTFOperator.createFirstPartitionForChain(ObjectInspector oi,
HiveConf hiveConf,
boolean isMapSide)
Create a new Partition. |
|
org.apache.hadoop.fs.Path |
ArchiveUtils.PartSpecInfo.createPath(Table tbl)
Creates path where partitions matching prefix should lie in filesystem |
|
void |
CommonJoinOperator.endGroup()
Forward a record of join results. |
|
void |
DemuxOperator.endGroup()
|
|
void |
GroupByOperator.endGroup()
|
|
void |
JoinOperator.endGroup()
Forward a record of join results. |
|
void |
MapJoinOperator.endGroup()
|
|
void |
MuxOperator.endGroup()
|
|
void |
Operator.endGroup()
|
|
Object |
ExprNodeEvaluator.evaluate(Object row)
|
|
void |
GroupByOperator.flush()
Forward all aggregations to children. |
|
void |
Operator.flush()
|
|
void |
TopNHash.flush()
Flushes all the rows cached in the heap. |
|
void |
DemuxOperator.forward(Object row,
ObjectInspector rowInspector)
|
|
void |
MuxOperator.forward(Object row,
ObjectInspector rowInspector)
|
|
void |
UDTFOperator.forwardUDTFOutput(Object o)
forwardUDTFOutput is typically called indirectly by the GenericUDTF when the GenericUDTF has generated output rows that should be passed on to the next operator(s) in the DAG. |
|
void |
MapJoinOperator.generateMapMetaData()
|
|
static ExprNodeEvaluator |
ExprNodeEvaluatorFactory.get(ExprNodeDesc desc)
|
|
static int |
ArchiveUtils.getArchivingLevel(Partition p)
Returns archiving level, which is how many fields were set in partial specification ARCHIVE was run for |
|
Object |
PTFPartition.getAt(int i)
|
|
static String[] |
Utilities.getDbTableName(String dbtable)
Extract db and table name from dbtable string, where db and table are separated by "." If there is no db name part, set the current sessions default db |
|
static List<LinkedHashMap<String,String>> |
Utilities.getFullDPSpecs(org.apache.hadoop.conf.Configuration conf,
DynamicPartitionCtx dpCtx)
Construct a list of full partition spec from Dynamic Partition Context and the directory names corresponding to these dynamic partitions. |
|
URI |
ArchiveUtils.HarPathHelper.getHarUri(URI original,
org.apache.hadoop.hive.shims.HadoopShims shim)
|
|
static Hive |
FunctionRegistry.getHive()
|
|
String |
ArchiveUtils.PartSpecInfo.getName()
Generates name for prefix partial partition specification. |
|
abstract void |
KeyWrapper.getNewKey(Object row,
ObjectInspector rowInspector)
|
|
static List<ObjectInspector>[] |
JoinUtil.getObjectInspectorsFromEvaluators(List<ExprNodeEvaluator>[] exprEntries,
ObjectInspector[] inputObjInspector,
int posBigTableAlias,
int tagLen)
|
|
ObjectInspector |
FetchOperator.getOutputObjectInspector()
returns output ObjectInspector, never null |
|
static String |
ArchiveUtils.getPartialName(Partition p,
int level)
Get a prefix of the given parition's string representation. |
|
static PartitionDesc |
Utilities.getPartitionDesc(Partition part)
|
|
static PartitionDesc |
Utilities.getPartitionDescFromTableDesc(TableDesc tblDesc,
Partition part)
|
|
static String[] |
FunctionUtils.getQualifiedFunctionNameParts(String name)
|
|
static RowContainer<List<Object>> |
JoinUtil.getRowContainer(org.apache.hadoop.conf.Configuration hconf,
List<ObjectInspector> structFieldObjectInspectors,
Byte alias,
int containerSize,
TableDesc[] spillTableDesc,
JoinDesc conf,
boolean noFilter,
org.apache.hadoop.mapred.Reporter reporter)
|
|
static
|
OperatorFactory.getVectorOperator(T conf,
VectorizationContext vContext)
|
|
void |
SkewJoinHandler.handleSkew(int tag)
|
|
void |
Operator.initialize(org.apache.hadoop.conf.Configuration hconf,
ObjectInspector[] inputOIs)
Initializes operators only if all parents have been initialized. |
|
void |
DefaultFetchFormatter.initialize(org.apache.hadoop.conf.Configuration hconf,
Properties props)
|
|
ObjectInspector |
ExprNodeColumnEvaluator.initialize(ObjectInspector rowInspector)
|
|
ObjectInspector |
ExprNodeConstantEvaluator.initialize(ObjectInspector rowInspector)
|
|
abstract ObjectInspector |
ExprNodeEvaluator.initialize(ObjectInspector rowInspector)
Initialize should be called once and only once. |
|
ObjectInspector |
ExprNodeEvaluatorHead.initialize(ObjectInspector rowInspector)
|
|
ObjectInspector |
ExprNodeEvaluatorRef.initialize(ObjectInspector rowInspector)
|
|
ObjectInspector |
ExprNodeFieldEvaluator.initialize(ObjectInspector rowInspector)
|
|
ObjectInspector |
ExprNodeGenericFuncEvaluator.initialize(ObjectInspector rowInspector)
|
|
ObjectInspector |
ExprNodeNullEvaluator.initialize(ObjectInspector rowInspector)
|
|
void |
MapOperator.initializeAsRoot(org.apache.hadoop.conf.Configuration hconf,
MapWork mapWork)
Initializes this map op as the root of the tree. |
|
void |
Operator.initializeLocalWork(org.apache.hadoop.conf.Configuration hconf)
|
|
void |
SMBMapJoinOperator.initializeLocalWork(org.apache.hadoop.conf.Configuration hconf)
|
|
void |
SMBMapJoinOperator.initializeMapredLocalWork(MapJoinDesc mjConf,
org.apache.hadoop.conf.Configuration hconf,
MapredLocalWork localWork,
org.apache.commons.logging.Log l4j)
|
|
void |
MapOperator.initializeOp(org.apache.hadoop.conf.Configuration hconf)
|
|
static Object |
FunctionRegistry.invoke(Method m,
Object thisObject,
Object... arguments)
|
|
PTFPartition.PTFPartitionIterator<Object> |
PTFPartition.iterator()
|
|
void |
Operator.jobClose(org.apache.hadoop.conf.Configuration conf,
boolean success)
Unlike other operator interfaces which are called from map or reduce task, jobClose is called from the jobclient side once the job has completed. |
|
void |
FileSinkOperator.jobCloseOp(org.apache.hadoop.conf.Configuration hconf,
boolean success)
|
|
void |
JoinOperator.jobCloseOp(org.apache.hadoop.conf.Configuration hconf,
boolean success)
|
|
void |
Operator.jobCloseOp(org.apache.hadoop.conf.Configuration conf,
boolean success)
|
|
T |
PTFPartition.PTFPartitionIterator.lag(int amt)
|
|
T |
PTFPartition.PTFPartitionIterator.lead(int amt)
|
|
void |
HashTableLoader.load(MapJoinTableContainer[] mapJoinTables,
MapJoinTableContainerSerDe[] mapJoinTableSerdes)
|
|
static void |
Utilities.mvFileToFinalPath(org.apache.hadoop.fs.Path specPath,
org.apache.hadoop.conf.Configuration hconf,
boolean success,
org.apache.commons.logging.Log log,
DynamicPartitionCtx dpCtx,
FileSinkDesc conf,
org.apache.hadoop.mapred.Reporter reporter)
|
|
static int |
JoinUtil.populateJoinKeyValue(List<ExprNodeEvaluator>[] outMap,
Map<Byte,List<ExprNodeDesc>> inputMap,
Byte[] order,
int posBigTableAlias)
|
|
static int |
JoinUtil.populateJoinKeyValue(List<ExprNodeEvaluator>[] outMap,
Map<Byte,List<ExprNodeDesc>> inputMap,
int posBigTableAlias)
|
|
void |
MapOperator.process(org.apache.hadoop.io.Writable value)
|
|
void |
MuxOperator.processGroup(int tag)
|
|
void |
Operator.processGroup(int tag)
|
|
void |
CollectOperator.processOp(Object row,
int tag)
|
|
void |
DemuxOperator.processOp(Object row,
int tag)
|
|
void |
DummyStoreOperator.processOp(Object row,
int tag)
|
|
void |
ExtractOperator.processOp(Object row,
int tag)
|
|
void |
FileSinkOperator.processOp(Object row,
int tag)
|
|
void |
FilterOperator.processOp(Object row,
int tag)
|
|
void |
ForwardOperator.processOp(Object row,
int tag)
|
|
void |
GroupByOperator.processOp(Object row,
int tag)
|
|
void |
HashTableDummyOperator.processOp(Object row,
int tag)
|
|
void |
HashTableSinkOperator.processOp(Object row,
int tag)
|
|
void |
JoinOperator.processOp(Object row,
int tag)
|
|
void |
LateralViewForwardOperator.processOp(Object row,
int tag)
|
|
void |
LateralViewJoinOperator.processOp(Object row,
int tag)
An important assumption for processOp() is that for a given row from the TS, the LVJ will first get the row from the left select operator, followed by all the corresponding rows from the UDTF operator. |
|
void |
LimitOperator.processOp(Object row,
int tag)
|
|
void |
ListSinkOperator.processOp(Object row,
int tag)
|
|
void |
MapJoinOperator.processOp(Object row,
int tag)
|
|
void |
MapOperator.processOp(Object row,
int tag)
|
|
void |
MuxOperator.processOp(Object row,
int tag)
|
|
abstract void |
Operator.processOp(Object row,
int tag)
Process the row. |
|
void |
PTFOperator.processOp(Object row,
int tag)
|
|
void |
ReduceSinkOperator.processOp(Object row,
int tag)
|
|
void |
ScriptOperator.processOp(Object row,
int tag)
|
|
void |
SelectOperator.processOp(Object row,
int tag)
|
|
void |
SMBMapJoinOperator.processOp(Object row,
int tag)
|
|
void |
TableScanOperator.processOp(Object row,
int tag)
Other than gathering statistics for the ANALYZE command, the table scan operator does not do anything special other than just forwarding the row. |
|
void |
UDTFOperator.processOp(Object row,
int tag)
|
|
void |
UnionOperator.processOp(Object row,
int tag)
|
|
boolean |
FetchOperator.pushRow()
Get the next row and push down it to operator tree. |
|
static void |
Utilities.rename(org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path src,
org.apache.hadoop.fs.Path dst)
Rename src to dst, or in the case dst already exists, move files in src to dst. |
|
static void |
Utilities.renameOrMoveFiles(org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path src,
org.apache.hadoop.fs.Path dst)
Rename src to dst, or in the case dst already exists, move files in src to dst. |
|
void |
PTFPartition.reset()
|
|
void |
PTFPartition.PTFPartitionIterator.reset()
|
|
Object |
PTFPartition.PTFPartitionIterator.resetToIndex(int idx)
|
|
void |
MapOperator.setChildren(org.apache.hadoop.conf.Configuration hconf)
|
|
int |
DDLTask.showColumns(Hive db,
ShowColumnsDesc showCols)
|
|
static String[] |
FunctionUtils.splitQualifiedFunctionName(String functionName)
Splits a qualified function name into an array containing the database name and function name. |
|
void |
CommonJoinOperator.startGroup()
|
|
void |
DemuxOperator.startGroup()
|
|
void |
FileSinkOperator.startGroup()
|
|
void |
GroupByOperator.startGroup()
|
|
void |
MapJoinOperator.startGroup()
|
|
void |
MuxOperator.startGroup()
|
|
void |
Operator.startGroup()
|
|
int |
TopNHash.startVectorizedBatch(int size)
Perform basic checks and initialize TopNHash for the new vectorized row batch. |
|
int |
TopNHash.tryStoreKey(HiveKey key)
Try store the non-vectorized key. |
|
void |
TopNHash.tryStoreVectorizedKey(HiveKey key,
int batchIndex)
Try to put the key from the current vectorized batch into the heap. |
|
static void |
FunctionRegistry.unregisterTemporaryUDF(String functionName)
|
Constructors in org.apache.hadoop.hive.ql.exec that throw HiveException | |
---|---|
ArchiveUtils.HarPathHelper(HiveConf hconf,
URI archive,
URI originalBase)
Creates helper for archive. |
|
ExprNodeFieldEvaluator(ExprNodeFieldDesc desc)
|
|
ExprNodeGenericFuncEvaluator(ExprNodeGenericFuncDesc expr)
|
|
SecureCmdDoAs(HiveConf conf)
|
Uses of HiveException in org.apache.hadoop.hive.ql.exec.mapjoin |
---|
Subclasses of HiveException in org.apache.hadoop.hive.ql.exec.mapjoin | |
---|---|
class |
MapJoinMemoryExhaustionException
|
Uses of HiveException in org.apache.hadoop.hive.ql.exec.mr |
---|
Methods in org.apache.hadoop.hive.ql.exec.mr that throw HiveException | |
---|---|
void |
HashTableLoader.load(MapJoinTableContainer[] mapJoinTables,
MapJoinTableContainerSerDe[] mapJoinTableSerdes)
|
static void |
ExecDriver.main(String[] args)
|
Constructors in org.apache.hadoop.hive.ql.exec.mr that throw HiveException | |
---|---|
ExecDriver(MapredWork plan,
org.apache.hadoop.mapred.JobConf job,
boolean isSilent)
Constructor/Initialization for invocation as independent utility. |
|
MapredLocalTask(MapredLocalWork plan,
org.apache.hadoop.mapred.JobConf job,
boolean isSilent)
|
Uses of HiveException in org.apache.hadoop.hive.ql.exec.persistence |
---|
Methods in org.apache.hadoop.hive.ql.exec.persistence that throw HiveException | |
---|---|
void |
LazyFlatRowContainer.add(MapJoinObjectSerDeContext context,
org.apache.hadoop.io.BytesWritable value,
boolean allowLazy)
Called when loading the hashtable. |
void |
LazyFlatRowContainer.addRow(List<Object> t)
|
void |
LazyFlatRowContainer.addRow(Object[] value)
|
void |
MapJoinRowContainer.addRow(Object[] value)
|
void |
PTFRowContainer.addRow(Row t)
|
void |
AbstractRowContainer.addRow(ROW t)
add a row into the RowContainer |
void |
RowContainer.addRow(ROW t)
|
void |
AbstractRowContainer.clearRows()
Remove all elements in the RowContainer. |
void |
PTFRowContainer.clearRows()
|
void |
RowContainer.clearRows()
Remove all elements in the RowContainer. |
void |
PTFRowContainer.close()
|
MapJoinRowContainer |
LazyFlatRowContainer.copy()
|
MapJoinRowContainer |
MapJoinRowContainer.copy()
|
void |
RowContainer.copyToDFSDirecory(org.apache.hadoop.fs.FileSystem destFs,
org.apache.hadoop.fs.Path destPath)
|
ROW |
AbstractRowContainer.RowIterator.first()
|
List<Object> |
LazyFlatRowContainer.first()
|
Row |
PTFRowContainer.first()
|
ROW |
RowContainer.first()
|
byte |
LazyFlatRowContainer.getAliasFilter()
|
byte |
MapJoinRowContainer.getAliasFilter()
|
Row |
PTFRowContainer.getAt(int rowIdx)
|
MapJoinTableContainer |
MapJoinTableContainerSerDe.load(ObjectInputStream in)
|
ROW |
AbstractRowContainer.RowIterator.next()
|
Row |
PTFRowContainer.next()
|
ROW |
RowContainer.next()
|
void |
MapJoinTableContainerSerDe.persist(ObjectOutputStream out,
MapJoinTableContainer tableContainer)
|
static MapJoinKey |
MapJoinKey.readFromRow(ByteStream.Output output,
MapJoinKey key,
Object row,
List<ExprNodeEvaluator> fields,
List<ObjectInspector> keyFieldsOI,
boolean mayReuseKey)
|
static MapJoinKey |
MapJoinKey.readFromVector(ByteStream.Output output,
MapJoinKey key,
VectorHashKeyWrapper kw,
VectorExpressionWriter[] keyOutputWriters,
VectorHashKeyWrapperBatch keyWrapperBatch,
boolean mayReuseKey)
|
void |
MapJoinKeyObject.readFromVector(VectorHashKeyWrapper kw,
VectorExpressionWriter[] keyOutputWriters,
VectorHashKeyWrapperBatch keyWrapperBatch)
|
int |
AbstractRowContainer.rowCount()
|
int |
LazyFlatRowContainer.rowCount()
|
AbstractRowContainer.RowIterator<ROW> |
AbstractRowContainer.rowIter()
|
AbstractRowContainer.RowIterator<List<Object>> |
LazyFlatRowContainer.rowIter()
|
Constructors in org.apache.hadoop.hive.ql.exec.persistence that throw HiveException | |
---|---|
PTFRowContainer(int bs,
org.apache.hadoop.conf.Configuration jc,
org.apache.hadoop.mapred.Reporter reporter)
|
|
RowContainer(org.apache.hadoop.conf.Configuration jc,
org.apache.hadoop.mapred.Reporter reporter)
|
|
RowContainer(int bs,
org.apache.hadoop.conf.Configuration jc,
org.apache.hadoop.mapred.Reporter reporter)
|
Uses of HiveException in org.apache.hadoop.hive.ql.exec.tez |
---|
Methods in org.apache.hadoop.hive.ql.exec.tez that throw HiveException | |
---|---|
void |
HashTableLoader.load(MapJoinTableContainer[] mapJoinTables,
MapJoinTableContainerSerDe[] mapJoinTableSerdes)
|
Uses of HiveException in org.apache.hadoop.hive.ql.exec.vector |
---|
Methods in org.apache.hadoop.hive.ql.exec.vector that throw HiveException | |
---|---|
void |
VectorizedRowBatchCtx.addPartitionColsToBatch(VectorizedRowBatch batch)
Add the partition values to the batch |
void |
VectorizedRowBatchCtx.addRowToBatch(int rowIndex,
org.apache.hadoop.io.Writable rowBlob,
VectorizedRowBatch batch,
org.apache.hadoop.io.DataOutputBuffer buffer)
Adds the row to the batch after deserializing the row |
static void |
VectorizedBatchUtil.addRowToBatch(Object row,
StructObjectInspector oi,
int rowIndex,
VectorizedRowBatch batch,
org.apache.hadoop.io.DataOutputBuffer buffer)
Iterates thru all the columns in a given row and populates the batch |
void |
VectorizationContext.addToColumnMap(String columnName,
int outputColumn)
|
T |
VectorUtilBatchObjectPool.IAllocator.alloc()
|
void |
VectorColumnAssign.assignObjectValue(Object val,
int destIndex)
|
void |
VectorColumnAssign.assignVectorValue(VectorizedRowBatch inBatch,
int batchIndex,
int valueColumn,
int destIndex)
|
static VectorColumnAssign[] |
VectorColumnAssignFactory.buildAssigners(VectorizedRowBatch outputBatch)
|
static VectorColumnAssign[] |
VectorColumnAssignFactory.buildAssigners(VectorizedRowBatch outputBatch,
ObjectInspector outputOI,
Map<String,Integer> columnMap,
List<String> outputColumnNames)
Builds the assigners from an object inspector and from a list of columns. |
static VectorColumnAssign |
VectorColumnAssignFactory.buildObjectAssign(VectorizedRowBatch outputBatch,
int outColIndex,
ObjectInspector objInspector)
|
void |
VectorGroupByOperator.closeOp(boolean aborted)
|
void |
VectorMapJoinOperator.closeOp(boolean aborted)
|
void |
VectorSMBMapJoinOperator.closeOp(boolean aborted)
|
static VectorHashKeyWrapperBatch |
VectorHashKeyWrapperBatch.compileKeyWrapperBatch(VectorExpression[] keyExpressions)
Prepares a VectorHashKeyWrapperBatch to work for a specific set of keys. |
VectorizedRowBatch |
VectorizedRowBatchCtx.createVectorizedRowBatch()
Creates a Vectorized row batch and the column vectors. |
void |
VectorHashKeyWrapperBatch.evaluateBatch(VectorizedRowBatch batch)
Processes a batch: Evaluates each key vector expression. Copies out each key's primitive values into the key wrappers computes the hashcode of the key wrappers |
VectorAggregateExpression |
VectorizationContext.getAggregatorExpression(AggregationDesc desc)
|
T |
VectorUtilBatchObjectPool.getFromPool()
|
void |
VectorHashKeyWrapper.getNewKey(Object row,
ObjectInspector rowInspector)
|
VectorExpression |
VectorizationContext.getVectorExpression(ExprNodeDesc exprDesc)
|
VectorExpression |
VectorizationContext.getVectorExpression(ExprNodeDesc exprDesc,
VectorExpressionDescriptor.Mode mode)
Returns a vector expression for a given expression description. |
Class<?> |
VectorExpressionDescriptor.getVectorExpressionClass(Class<?> udf,
VectorExpressionDescriptor.Descriptor descriptor)
|
VectorExpression[] |
VectorizationContext.getVectorExpressions(List<ExprNodeDesc> exprNodes)
|
VectorExpression[] |
VectorizationContext.getVectorExpressions(List<ExprNodeDesc> exprNodes,
VectorExpressionDescriptor.Mode mode)
|
Object |
VectorHashKeyWrapperBatch.getWritableKeyValue(VectorHashKeyWrapper kw,
int i,
VectorExpressionWriter keyOutputWriter)
Get the row-mode writable object value of a key from a key wrapper |
void |
VectorizedRowBatchCtx.init(org.apache.hadoop.conf.Configuration hiveConf,
org.apache.hadoop.mapred.FileSplit split)
Initializes VectorizedRowBatch context based on the split and Hive configuration (Job conf with hive Plan). |
void |
VectorMapJoinOperator.initializeOp(org.apache.hadoop.conf.Configuration hconf)
|
void |
VectorMapOperator.process(org.apache.hadoop.io.Writable value)
|
void |
VectorFileSinkOperator.processOp(Object data,
int tag)
|
void |
VectorFilterOperator.processOp(Object row,
int tag)
|
void |
VectorGroupByOperator.processOp(Object row,
int tag)
|
void |
VectorLimitOperator.processOp(Object row,
int tag)
|
void |
VectorMapJoinOperator.processOp(Object row,
int tag)
|
void |
VectorReduceSinkOperator.processOp(Object row,
int tag)
|
void |
VectorSelectOperator.processOp(Object row,
int tag)
|
void |
VectorSMBMapJoinOperator.processOp(Object row,
int tag)
|
Constructors in org.apache.hadoop.hive.ql.exec.vector that throw HiveException | |
---|---|
VectorFilterOperator(VectorizationContext vContext,
OperatorDesc conf)
|
|
VectorGroupByOperator(VectorizationContext vContext,
OperatorDesc conf)
|
|
VectorMapJoinOperator(VectorizationContext vContext,
OperatorDesc conf)
|
|
VectorReduceSinkOperator(VectorizationContext vContext,
OperatorDesc conf)
|
|
VectorSelectOperator(VectorizationContext vContext,
OperatorDesc conf)
|
|
VectorSMBMapJoinOperator(VectorizationContext vContext,
OperatorDesc conf)
|
Uses of HiveException in org.apache.hadoop.hive.ql.exec.vector.expressions |
---|
Methods in org.apache.hadoop.hive.ql.exec.vector.expressions that throw HiveException | |
---|---|
static VectorExpressionWriter |
VectorExpressionWriterFactory.genVectorExpressionWritable(ExprNodeDesc nodeDesc)
Compiles the appropriate vector expression writer based on an expression info (ExprNodeDesc) |
static VectorExpressionWriter |
VectorExpressionWriterFactory.genVectorExpressionWritable(ObjectInspector fieldObjInspector)
Compiles the appropriate vector expression writer based on an expression info (ExprNodeDesc) |
static VectorExpressionWriter[] |
VectorExpressionWriterFactory.getExpressionWriters(List<ExprNodeDesc> nodesDesc)
Helper function to create an array of writers from a list of expression descriptors. |
static VectorExpressionWriter[] |
VectorExpressionWriterFactory.getExpressionWriters(StructObjectInspector objInspector)
Returns VectorExpressionWriter objects for the fields in the given
object inspector. |
static VectorExpressionWriter[] |
VectorExpressionWriterFactory.getSettableExpressionWriters(SettableStructObjectInspector objInspector)
|
Object |
VectorExpressionWriter.initValue(Object ost)
|
static void |
VectorExpressionWriterFactory.processVectorExpressions(List<ExprNodeDesc> nodesDesc,
List<String> columnNames,
VectorExpressionWriterFactory.SingleOIDClosure closure)
Creates the value writers for a column vector expression list. |
static void |
VectorExpressionWriterFactory.processVectorExpressions(List<ExprNodeDesc> nodesDesc,
VectorExpressionWriterFactory.ListOIDClosure closure)
Creates the value writers for a column vector expression list. |
Object |
VectorExpressionWriter.setValue(Object row,
ColumnVector column,
int columnRow)
|
Object |
VectorExpressionWriter.writeValue(byte[] value,
int start,
int length)
|
Object |
VectorExpressionWriter.writeValue(ColumnVector column,
int row)
|
Object |
VectorExpressionWriter.writeValue(Decimal128 value)
|
Object |
VectorExpressionWriter.writeValue(double value)
|
Object |
VectorExpressionWriter.writeValue(long value)
|
Constructors in org.apache.hadoop.hive.ql.exec.vector.expressions that throw HiveException | |
---|---|
FilterStringColLikeStringScalar(int colNum,
byte[] likePattern)
|
|
FilterStringColRegExpStringScalar(int colNum,
byte[] regExpPattern)
|
Uses of HiveException in org.apache.hadoop.hive.ql.exec.vector.expressions.aggregates |
---|
Uses of HiveException in org.apache.hadoop.hive.ql.exec.vector.udf |
---|
Methods in org.apache.hadoop.hive.ql.exec.vector.udf that throw HiveException | |
---|---|
void |
VectorUDFAdaptor.init()
|
Constructors in org.apache.hadoop.hive.ql.exec.vector.udf that throw HiveException | |
---|---|
VectorUDFAdaptor(ExprNodeGenericFuncDesc expr,
int outputColumn,
String resultType,
VectorUDFArgDesc[] argDescs)
|
Uses of HiveException in org.apache.hadoop.hive.ql.index |
---|
Methods in org.apache.hadoop.hive.ql.index that throw HiveException | |
---|---|
void |
AggregateIndexHandler.analyzeIndexDefinition(org.apache.hadoop.hive.metastore.api.Table baseTable,
org.apache.hadoop.hive.metastore.api.Index idx,
org.apache.hadoop.hive.metastore.api.Table indexTable)
|
void |
HiveIndexHandler.analyzeIndexDefinition(org.apache.hadoop.hive.metastore.api.Table baseTable,
org.apache.hadoop.hive.metastore.api.Index index,
org.apache.hadoop.hive.metastore.api.Table indexTable)
Requests that the handler validate an index definition and fill in additional information about its stored representation. |
boolean |
HiveIndexResult.contains(org.apache.hadoop.mapred.FileSplit split)
|
List<Task<?>> |
HiveIndexHandler.generateIndexBuildTaskList(Table baseTbl,
org.apache.hadoop.hive.metastore.api.Index index,
List<Partition> indexTblPartitions,
List<Partition> baseTblPartitions,
Table indexTbl,
Set<ReadEntity> inputs,
Set<WriteEntity> outputs)
Requests that the handler generate a plan for building the index; the plan should read the base table and write out the index representation. |
List<Task<?>> |
TableBasedIndexHandler.generateIndexBuildTaskList(Table baseTbl,
org.apache.hadoop.hive.metastore.api.Index index,
List<Partition> indexTblPartitions,
List<Partition> baseTblPartitions,
Table indexTbl,
Set<ReadEntity> inputs,
Set<WriteEntity> outputs)
|
Constructors in org.apache.hadoop.hive.ql.index that throw HiveException | |
---|---|
HiveIndexResult(List<String> indexFiles,
org.apache.hadoop.mapred.JobConf conf)
|
Uses of HiveException in org.apache.hadoop.hive.ql.index.bitmap |
---|
Methods in org.apache.hadoop.hive.ql.index.bitmap that throw HiveException | |
---|---|
void |
BitmapIndexHandler.analyzeIndexDefinition(org.apache.hadoop.hive.metastore.api.Table baseTable,
org.apache.hadoop.hive.metastore.api.Index index,
org.apache.hadoop.hive.metastore.api.Table indexTable)
|
Uses of HiveException in org.apache.hadoop.hive.ql.index.compact |
---|
Methods in org.apache.hadoop.hive.ql.index.compact that throw HiveException | |
---|---|
void |
CompactIndexHandler.analyzeIndexDefinition(org.apache.hadoop.hive.metastore.api.Table baseTable,
org.apache.hadoop.hive.metastore.api.Index index,
org.apache.hadoop.hive.metastore.api.Table indexTable)
|
Uses of HiveException in org.apache.hadoop.hive.ql.io |
---|
Methods in org.apache.hadoop.hive.ql.io that throw HiveException | |
---|---|
static boolean |
HiveFileFormatUtils.checkInputFormat(org.apache.hadoop.fs.FileSystem fs,
HiveConf conf,
Class<? extends org.apache.hadoop.mapred.InputFormat> inputFormatCls,
ArrayList<org.apache.hadoop.fs.FileStatus> files)
checks if files are in same format as the given input format. |
static FileSinkOperator.RecordWriter |
HiveFileFormatUtils.getHiveRecordWriter(org.apache.hadoop.mapred.JobConf jc,
TableDesc tableInfo,
Class<? extends org.apache.hadoop.io.Writable> outputClass,
FileSinkDesc conf,
org.apache.hadoop.fs.Path outPath,
org.apache.hadoop.mapred.Reporter reporter)
|
static FileSinkOperator.RecordWriter |
HiveFileFormatUtils.getRecordWriter(org.apache.hadoop.mapred.JobConf jc,
HiveOutputFormat<?,?> hiveOutputFormat,
Class<? extends org.apache.hadoop.io.Writable> valueClass,
boolean isCompressed,
Properties tableProp,
org.apache.hadoop.fs.Path outPath,
org.apache.hadoop.mapred.Reporter reporter)
|
Uses of HiveException in org.apache.hadoop.hive.ql.io.rcfile.merge |
---|
Methods in org.apache.hadoop.hive.ql.io.rcfile.merge that throw HiveException | |
---|---|
static org.apache.hadoop.fs.Path |
RCFileMergeMapper.backupOutputPath(org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path outpath,
org.apache.hadoop.mapred.JobConf job)
|
static void |
RCFileMergeMapper.jobClose(org.apache.hadoop.fs.Path outputPath,
boolean success,
org.apache.hadoop.mapred.JobConf job,
SessionState.LogHelper console,
DynamicPartitionCtx dynPartCtx,
org.apache.hadoop.mapred.Reporter reporter)
|
Uses of HiveException in org.apache.hadoop.hive.ql.io.rcfile.truncate |
---|
Methods in org.apache.hadoop.hive.ql.io.rcfile.truncate that throw HiveException | |
---|---|
static org.apache.hadoop.fs.Path |
ColumnTruncateMapper.backupOutputPath(org.apache.hadoop.fs.FileSystem fs,
org.apache.hadoop.fs.Path outpath,
org.apache.hadoop.mapred.JobConf job)
|
static void |
ColumnTruncateMapper.jobClose(org.apache.hadoop.fs.Path outputPath,
boolean success,
org.apache.hadoop.mapred.JobConf job,
SessionState.LogHelper console,
DynamicPartitionCtx dynPartCtx,
org.apache.hadoop.mapred.Reporter reporter)
|
Uses of HiveException in org.apache.hadoop.hive.ql.lockmgr |
---|
Subclasses of HiveException in org.apache.hadoop.hive.ql.lockmgr | |
---|---|
class |
LockException
Exception from lock manager. |
Uses of HiveException in org.apache.hadoop.hive.ql.metadata |
---|
Subclasses of HiveException in org.apache.hadoop.hive.ql.metadata | |
---|---|
class |
HiveFatalException
|
class |
InvalidTableException
Generic exception class for Hive. |
Methods in org.apache.hadoop.hive.ql.metadata that throw HiveException | |
---|---|
void |
Hive.alterDatabase(String dbName,
org.apache.hadoop.hive.metastore.api.Database db)
|
void |
Hive.alterFunction(String dbName,
String funcName,
org.apache.hadoop.hive.metastore.api.Function newFunction)
|
void |
Hive.alterIndex(String dbName,
String baseTblName,
String idxName,
org.apache.hadoop.hive.metastore.api.Index newIdx)
Updates the existing index metadata with the new metadata. |
void |
Hive.alterPartition(String tblName,
Partition newPart)
Updates the existing partition metadata with the new metadata. |
void |
Hive.alterPartition(String dbName,
String tblName,
Partition newPart)
Updates the existing partition metadata with the new metadata. |
void |
Hive.alterPartitions(String tblName,
List<Partition> newParts)
Updates the existing table metadata with the new metadata. |
void |
Hive.alterTable(String tblName,
Table newTbl)
Updates the existing table metadata with the new metadata. |
void |
Hive.cancelDelegationToken(String tokenStrForm)
|
void |
HiveMetaStoreChecker.checkMetastore(String dbName,
String tableName,
List<? extends Map<String,String>> partitions,
CheckResult result)
Check the metastore for inconsistencies, data missing in either the metastore or on the dfs. |
void |
Table.checkValidity()
|
static org.apache.hadoop.hive.metastore.api.StorageDescriptor |
Partition.cloneSd(Table tbl)
We already have methods that clone stuff using XML or Kryo. |
void |
Hive.compact(String dbname,
String tableName,
String partName,
String compactType)
Enqueue a compaction request. |
Table |
Table.copy()
|
void |
Hive.createDatabase(org.apache.hadoop.hive.metastore.api.Database db)
Create a Database. |
void |
Hive.createDatabase(org.apache.hadoop.hive.metastore.api.Database db,
boolean ifNotExist)
Create a database |
void |
Hive.createFunction(org.apache.hadoop.hive.metastore.api.Function func)
|
void |
Hive.createIndex(String tableName,
String indexName,
String indexHandlerClass,
List<String> indexedCols,
String indexTblName,
boolean deferredRebuild,
String inputFormat,
String outputFormat,
String serde,
String storageHandler,
String location,
Map<String,String> idxProps,
Map<String,String> tblProps,
Map<String,String> serdeProps,
String collItemDelim,
String fieldDelim,
String fieldEscape,
String lineDelim,
String mapKeyDelim,
String indexComment)
|
static org.apache.hadoop.hive.metastore.api.Partition |
Partition.createMetaPartitionObject(Table tbl,
Map<String,String> partSpec,
org.apache.hadoop.fs.Path location)
|
Partition |
Hive.createPartition(Table tbl,
Map<String,String> partSpec)
Creates a partition. |
List<Partition> |
Hive.createPartitions(AddPartitionDesc addPartitionDesc)
|
void |
Hive.createRole(String roleName,
String ownerName)
|
void |
Hive.createTable(String tableName,
List<String> columns,
List<String> partCols,
Class<? extends org.apache.hadoop.mapred.InputFormat> fileInputFormat,
Class<?> fileOutputFormat)
Creates a table metdata and the directory for the table data |
void |
Hive.createTable(String tableName,
List<String> columns,
List<String> partCols,
Class<? extends org.apache.hadoop.mapred.InputFormat> fileInputFormat,
Class<?> fileOutputFormat,
int bucketCount,
List<String> bucketCols)
Creates a table metdata and the directory for the table data |
void |
Hive.createTable(Table tbl)
Creates the table with the give objects |
void |
Hive.createTable(Table tbl,
boolean ifNotExists)
Creates the table with the give objects |
boolean |
Hive.databaseExists(String dbName)
Query metadata to see if a database with the given name already exists. |
boolean |
Hive.deletePartitionColumnStatistics(String dbName,
String tableName,
String partName,
String colName)
|
boolean |
Hive.deleteTableColumnStatistics(String dbName,
String tableName,
String colName)
|
void |
Hive.dropDatabase(String name)
Drop a database. |
void |
Hive.dropDatabase(String name,
boolean deleteData,
boolean ignoreUnknownDb)
Drop a database |
void |
Hive.dropDatabase(String name,
boolean deleteData,
boolean ignoreUnknownDb,
boolean cascade)
Drop a database |
void |
Hive.dropFunction(String dbName,
String funcName)
|
boolean |
Hive.dropIndex(String db_name,
String tbl_name,
String index_name,
boolean deleteData)
|
boolean |
Hive.dropPartition(String tblName,
List<String> part_vals,
boolean deleteData)
|
boolean |
Hive.dropPartition(String db_name,
String tbl_name,
List<String> part_vals,
boolean deleteData)
|
List<Partition> |
Hive.dropPartitions(String tblName,
List<DropTableDesc.PartSpec> partSpecs,
boolean deleteData,
boolean ignoreProtection,
boolean ifExists)
|
List<Partition> |
Hive.dropPartitions(String dbName,
String tblName,
List<DropTableDesc.PartSpec> partSpecs,
boolean deleteData,
boolean ignoreProtection,
boolean ifExists)
|
void |
Hive.dropRole(String roleName)
|
void |
Hive.dropTable(String tableName)
Drops table along with the data in it. |
void |
Hive.dropTable(String dbName,
String tableName)
Drops table along with the data in it. |
void |
Hive.dropTable(String dbName,
String tableName,
boolean deleteData,
boolean ignoreUnknownTab)
Drops the table. |
InputEstimator.Estimation |
InputEstimator.estimate(org.apache.hadoop.mapred.JobConf job,
TableScanOperator ts,
long remaining)
Estimate input size based on filter and projection on table scan operator |
void |
Hive.exchangeTablePartitions(Map<String,String> partitionSpecs,
String sourceDb,
String sourceTable,
String destDb,
String destinationTableName)
|
org.apache.hadoop.hive.metastore.api.PrincipalPrivilegeSet |
Hive.get_privilege_set(org.apache.hadoop.hive.metastore.api.HiveObjectType objectType,
String db_name,
String table_name,
List<String> part_values,
String column_name,
String user_name,
List<String> group_names)
|
static Hive |
Hive.get()
|
static Hive |
Hive.get(HiveConf c)
Gets hive object for the current thread. |
static Hive |
Hive.get(HiveConf c,
boolean needsRefresh)
get a connection to metastore. |
List<String> |
Hive.getAllDatabases()
Get all existing database names. |
List<org.apache.hadoop.hive.metastore.api.Index> |
Table.getAllIndexes(short max)
|
Set<Partition> |
Hive.getAllPartitionsOf(Table tbl)
Get all the partitions; unlike Hive.getPartitions(Table) , does not include auth. |
List<String> |
Hive.getAllRoleNames()
Get all existing role names. |
List<String> |
Hive.getAllTables()
Get all table names for the current database. |
List<String> |
Hive.getAllTables(String dbName)
Get all table names for the specified database. |
static HiveAuthenticationProvider |
HiveUtils.getAuthenticator(org.apache.hadoop.conf.Configuration conf,
HiveConf.ConfVars authenticatorConfKey)
|
HiveAuthorizationProvider |
DefaultStorageHandler.getAuthorizationProvider()
|
HiveAuthorizationProvider |
HiveStorageHandler.getAuthorizationProvider()
Returns the implementation specific authorization provider |
static HiveAuthorizationProvider |
HiveUtils.getAuthorizeProviderManager(org.apache.hadoop.conf.Configuration conf,
HiveConf.ConfVars authorizationProviderConfKey,
HiveAuthenticationProvider authenticator)
|
static HiveAuthorizationProvider |
HiveUtils.getAuthorizeProviderManager(org.apache.hadoop.conf.Configuration conf,
HiveConf.ConfVars authorizationProviderConfKey,
HiveAuthenticationProvider authenticator,
boolean nullIfOtherClass)
Create a new instance of HiveAuthorizationProvider |
static HiveAuthorizerFactory |
HiveUtils.getAuthorizerFactory(org.apache.hadoop.conf.Configuration conf,
HiveConf.ConfVars authorizationProviderConfKey)
Return HiveAuthorizerFactory used by new authorization plugin interface. |
org.apache.hadoop.hive.metastore.api.Database |
Hive.getDatabase(String dbName)
Get the database by name. |
org.apache.hadoop.hive.metastore.api.Database |
Hive.getDatabaseCurrent()
Get the Database object for current database |
List<String> |
Hive.getDatabasesByPattern(String databasePattern)
Get all existing databases that match the given pattern. |
String |
Hive.getDelegationToken(String owner,
String renewer)
|
static List<org.apache.hadoop.hive.metastore.api.FieldSchema> |
Hive.getFieldsFromDeserializer(String name,
Deserializer serde)
|
org.apache.hadoop.hive.metastore.api.Function |
Hive.getFunction(String dbName,
String funcName)
|
List<String> |
Hive.getFunctions(String dbName,
String pattern)
|
org.apache.hadoop.hive.metastore.api.Index |
Hive.getIndex(String qualifiedIndexName)
|
org.apache.hadoop.hive.metastore.api.Index |
Hive.getIndex(String baseTableName,
String indexName)
|
org.apache.hadoop.hive.metastore.api.Index |
Hive.getIndex(String dbName,
String baseTableName,
String indexName)
|
List<org.apache.hadoop.hive.metastore.api.Index> |
Hive.getIndexes(String dbName,
String tblName,
short max)
|
static HiveIndexHandler |
HiveUtils.getIndexHandler(HiveConf conf,
String indexHandlerClass)
|
Class<? extends org.apache.hadoop.mapred.InputFormat> |
Partition.getInputFormatClass()
|
Class<? extends HiveOutputFormat> |
Partition.getOutputFormatClass()
|
Partition |
Hive.getPartition(Table tbl,
Map<String,String> partSpec,
boolean forceCreate)
|
Partition |
Hive.getPartition(Table tbl,
Map<String,String> partSpec,
boolean forceCreate,
String partPath,
boolean inheritTableSpecs)
Returns partition metadata |
Map<String,List<org.apache.hadoop.hive.metastore.api.ColumnStatisticsObj>> |
Hive.getPartitionColumnStatistics(String dbName,
String tableName,
List<String> partNames,
List<String> colNames)
|
List<String> |
Hive.getPartitionNames(String tblName,
short max)
|
List<String> |
Hive.getPartitionNames(String dbName,
String tblName,
Map<String,String> partSpec,
short max)
|
List<String> |
Hive.getPartitionNames(String dbName,
String tblName,
short max)
|
List<Partition> |
Hive.getPartitions(Table tbl)
get all the partitions that the table has |
List<Partition> |
Hive.getPartitions(Table tbl,
Map<String,String> partialPartSpec)
get all the partitions of the table that matches the given partial specification. |
List<Partition> |
Hive.getPartitions(Table tbl,
Map<String,String> partialPartSpec,
short limit)
get all the partitions of the table that matches the given partial specification. |
boolean |
Hive.getPartitionsByExpr(Table tbl,
ExprNodeGenericFuncDesc expr,
HiveConf conf,
List<Partition> result)
Get a list of Partitions by expr. |
List<Partition> |
Hive.getPartitionsByFilter(Table tbl,
String filter)
Get a list of Partitions by filter. |
List<Partition> |
Hive.getPartitionsByNames(Table tbl,
List<String> partNames)
Get all partitions of the table that matches the list of given partition names. |
List<Partition> |
Hive.getPartitionsByNames(Table tbl,
Map<String,String> partialPartSpec)
get all the partitions of the table that matches the given partial specification. |
org.apache.hadoop.fs.Path[] |
Partition.getPath(Sample s)
|
List<org.apache.hadoop.hive.metastore.api.RolePrincipalGrant> |
Hive.getRoleGrantInfoForPrincipal(String principalName,
org.apache.hadoop.hive.metastore.api.PrincipalType principalType)
|
static HiveStorageHandler |
HiveUtils.getStorageHandler(org.apache.hadoop.conf.Configuration conf,
String className)
|
Table |
Hive.getTable(String tableName)
Returns metadata for the table named tableName |
Table |
Hive.getTable(String tableName,
boolean throwException)
Returns metadata for the table named tableName |
Table |
Hive.getTable(String dbName,
String tableName)
Returns metadata of the table |
Table |
Hive.getTable(String dbName,
String tableName,
boolean throwException)
Returns metadata of the table |
List<org.apache.hadoop.hive.metastore.api.ColumnStatisticsObj> |
Hive.getTableColumnStatistics(String dbName,
String tableName,
List<String> colNames)
|
List<String> |
Hive.getTablesByPattern(String tablePattern)
Returns all existing tables from default database which match the given pattern. |
List<String> |
Hive.getTablesByPattern(String dbName,
String tablePattern)
Returns all existing tables from the specified database which match the given pattern. |
List<String> |
Hive.getTablesForDb(String database,
String tablePattern)
Returns all existing tables from the given database which match the given pattern. |
boolean |
Hive.grantPrivileges(org.apache.hadoop.hive.metastore.api.PrivilegeBag privileges)
|
boolean |
Hive.grantRole(String roleName,
String userName,
org.apache.hadoop.hive.metastore.api.PrincipalType principalType,
String grantor,
org.apache.hadoop.hive.metastore.api.PrincipalType grantorType,
boolean grantOption)
|
List<org.apache.hadoop.hive.metastore.api.Role> |
Hive.listRoles(String userName,
org.apache.hadoop.hive.metastore.api.PrincipalType principalType)
|
ArrayList<LinkedHashMap<String,String>> |
Hive.loadDynamicPartitions(org.apache.hadoop.fs.Path loadPath,
String tableName,
Map<String,String> partSpec,
boolean replace,
int numDP,
boolean holdDDLTime,
boolean listBucketingEnabled)
Given a source directory name of the load path, load all dynamically generated partitions into the specified table and return a list of strings that represent the dynamic partition paths. |
void |
Hive.loadPartition(org.apache.hadoop.fs.Path loadPath,
String tableName,
Map<String,String> partSpec,
boolean replace,
boolean holdDDLTime,
boolean inheritTableSpecs,
boolean isSkewedStoreAsSubdir)
Load a directory into a Hive Table Partition - Alters existing content of the partition with the contents of loadPath. |
void |
Hive.loadTable(org.apache.hadoop.fs.Path loadPath,
String tableName,
boolean replace,
boolean holdDDLTime)
Load a directory into a Hive Table. |
Table |
Hive.newTable(String tableName)
|
void |
Hive.renamePartition(Table tbl,
Map<String,String> oldPartSpec,
Partition newPart)
Rename a old partition to new partition |
boolean |
Hive.revokePrivileges(org.apache.hadoop.hive.metastore.api.PrivilegeBag privileges)
|
boolean |
Hive.revokeRole(String roleName,
String userName,
org.apache.hadoop.hive.metastore.api.PrincipalType principalType)
|
void |
Table.setBucketCols(List<String> bucketCols)
|
void |
Table.setInputFormatClass(String name)
|
void |
Table.setOutputFormatClass(String name)
|
void |
Table.setSkewedColNames(List<String> skewedColNames)
|
void |
Table.setSkewedColValues(List<List<String>> skewedValues)
|
void |
Table.setSkewedInfo(org.apache.hadoop.hive.metastore.api.SkewedInfo skewedInfo)
|
void |
Partition.setSkewedValueLocationMap(List<String> valList,
String dirName)
|
void |
Table.setSkewedValueLocationMap(List<String> valList,
String dirName)
|
void |
Table.setSortCols(List<org.apache.hadoop.hive.metastore.api.Order> sortOrder)
|
void |
Table.setStoredAsSubDirectories(boolean storedAsSubDirectories)
|
void |
Partition.setValues(Map<String,String> partSpec)
Set Partition's values |
org.apache.hadoop.hive.metastore.api.ShowCompactResponse |
Hive.showCompactions()
|
List<org.apache.hadoop.hive.metastore.api.HiveObjectPrivilege> |
Hive.showPrivilegeGrant(org.apache.hadoop.hive.metastore.api.HiveObjectType objectType,
String principalName,
org.apache.hadoop.hive.metastore.api.PrincipalType principalType,
String dbName,
String tableName,
List<String> partValues,
String columnName)
|
org.apache.hadoop.hive.metastore.api.GetOpenTxnsInfoResponse |
Hive.showTransactions()
|
boolean |
Hive.updatePartitionColumnStatistics(org.apache.hadoop.hive.metastore.api.ColumnStatistics statsObj)
|
boolean |
Hive.updateTableColumnStatistics(org.apache.hadoop.hive.metastore.api.ColumnStatistics statsObj)
|
void |
Hive.validatePartitionNameCharacters(List<String> partVals)
|
Constructors in org.apache.hadoop.hive.ql.metadata that throw HiveException | |
---|---|
DummyPartition(Table tbl,
String name)
|
|
DummyPartition(Table tbl,
String name,
Map<String,String> partSpec)
|
|
Partition(Table tbl)
create an empty partition. |
|
Partition(Table tbl,
Map<String,String> partSpec,
org.apache.hadoop.fs.Path location)
Create partition object with the given info. |
|
Partition(Table tbl,
org.apache.hadoop.hive.metastore.api.Partition tp)
|
|
Sample(int num,
int fraction,
Dimension d)
|
Uses of HiveException in org.apache.hadoop.hive.ql.metadata.formatting |
---|
Methods in org.apache.hadoop.hive.ql.metadata.formatting that throw HiveException | |
---|---|
void |
JsonMetaDataFormatter.describeTable(DataOutputStream out,
String colPath,
String tableName,
Table tbl,
Partition part,
List<org.apache.hadoop.hive.metastore.api.FieldSchema> cols,
boolean isFormatted,
boolean isExt,
boolean isPretty,
boolean isOutputPadded)
Describe table. |
void |
MetaDataFormatter.describeTable(DataOutputStream out,
String colPath,
String tableName,
Table tbl,
Partition part,
List<org.apache.hadoop.hive.metastore.api.FieldSchema> cols,
boolean isFormatted,
boolean isExt,
boolean isPretty,
boolean isOutputPadded)
Describe table. |
void |
JsonMetaDataFormatter.error(OutputStream out,
String msg,
int errorCode,
String sqlState)
Write an error message. |
void |
MetaDataFormatter.error(OutputStream out,
String msg,
int errorCode,
String sqlState)
Write an error message. |
void |
JsonMetaDataFormatter.error(OutputStream out,
String errorMessage,
int errorCode,
String sqlState,
String errorDetail)
|
void |
MetaDataFormatter.error(OutputStream out,
String errorMessage,
int errorCode,
String sqlState,
String errorDetail)
|
void |
JsonMetaDataFormatter.showDatabaseDescription(DataOutputStream out,
String database,
String comment,
String location,
String ownerName,
String ownerType,
Map<String,String> params)
Show the description of a database |
void |
MetaDataFormatter.showDatabaseDescription(DataOutputStream out,
String database,
String comment,
String location,
String ownerName,
String ownerType,
Map<String,String> params)
Describe a database. |
void |
JsonMetaDataFormatter.showDatabases(DataOutputStream out,
List<String> databases)
Show a list of databases |
void |
MetaDataFormatter.showDatabases(DataOutputStream out,
List<String> databases)
Show the databases |
void |
JsonMetaDataFormatter.showTablePartitons(DataOutputStream out,
List<String> parts)
Show the table partitions. |
void |
MetaDataFormatter.showTablePartitons(DataOutputStream out,
List<String> parts)
Show the table partitions. |
void |
JsonMetaDataFormatter.showTables(DataOutputStream out,
Set<String> tables)
Show a list of tables. |
void |
MetaDataFormatter.showTables(DataOutputStream out,
Set<String> tables)
Show a list of tables. |
void |
JsonMetaDataFormatter.showTableStatus(DataOutputStream out,
Hive db,
HiveConf conf,
List<Table> tbls,
Map<String,String> part,
Partition par)
|
void |
MetaDataFormatter.showTableStatus(DataOutputStream out,
Hive db,
HiveConf conf,
List<Table> tbls,
Map<String,String> part,
Partition par)
Show the table status. |
Uses of HiveException in org.apache.hadoop.hive.ql.optimizer |
---|
Methods in org.apache.hadoop.hive.ql.optimizer that throw HiveException | |
---|---|
static Set<Partition> |
IndexUtils.checkPartitionsCoveredByIndex(TableScanOperator tableScan,
ParseContext pctx,
Map<Table,List<org.apache.hadoop.hive.metastore.api.Index>> indexes)
Check the partitions used by the table scan to make sure they also exist in the index table. |
Uses of HiveException in org.apache.hadoop.hive.ql.optimizer.ppr |
---|
Methods in org.apache.hadoop.hive.ql.optimizer.ppr that throw HiveException | |
---|---|
static Object |
PartExprEvalUtils.evalExprWithPart(ExprNodeDesc expr,
Partition p,
List<VirtualColumn> vcs,
StructObjectInspector rowObjectInspector)
Evaluate expression with partition columns |
static Object |
PartExprEvalUtils.evaluateExprOnPart(ObjectPair<PrimitiveObjectInspector,ExprNodeEvaluator> pair,
Object partColValues)
|
static ObjectPair<PrimitiveObjectInspector,ExprNodeEvaluator> |
PartExprEvalUtils.prepareExpr(ExprNodeGenericFuncDesc expr,
List<String> partNames)
|
static PrunedPartitionList |
PartitionPruner.prune(TableScanOperator ts,
ParseContext parseCtx,
String alias)
Get the partition list for the TS operator that satisfies the partition pruner condition. |
static boolean |
PartitionPruner.prunePartitionNames(List<String> columnNames,
ExprNodeGenericFuncDesc prunerExpr,
String defaultPartitionName,
List<String> partNames)
Prunes partition names to see if they match the prune expression. |
Uses of HiveException in org.apache.hadoop.hive.ql.parse |
---|
Subclasses of HiveException in org.apache.hadoop.hive.ql.parse | |
---|---|
class |
SemanticException
Exception from SemanticAnalyzer. |
Methods in org.apache.hadoop.hive.ql.parse that throw HiveException | |
---|---|
PTFExpressionDef |
PTFTranslator.buildExpressionDef(ShapeDetails inpShape,
ASTNode arg)
|
List<Task<? extends Serializable>> |
IndexUpdater.generateUpdateTasks()
|
static ExprNodeEvaluator |
WindowingExprNodeEvaluatorFactory.get(LeadLagInfo llInfo,
ExprNodeDesc desc)
|
Hive |
HiveSemanticAnalyzerHookContext.getHive()
|
Hive |
HiveSemanticAnalyzerHookContextImpl.getHive()
|
PrunedPartitionList |
ParseContext.getPrunedPartitions(String alias,
TableScanOperator ts)
|
boolean |
BaseSemanticAnalyzer.isValidPrefixSpec(Table tTable,
Map<String,String> spec)
Checks if given specification is proper specification for prefix of partition cols, for table partitioned by ds, hr, min valid ones are (ds='2008-04-08'), (ds='2008-04-08', hr='12'), (ds='2008-04-08', hr='12', min='30') invalid one is for example (ds='2008-04-08', min='30') |
void |
WindowingExprNodeEvaluatorFactory.FindLeadLagFuncExprs.visit(ExprNodeGenericFuncDesc fnExpr)
|
Uses of HiveException in org.apache.hadoop.hive.ql.plan |
---|
Methods in org.apache.hadoop.hive.ql.plan that throw HiveException | |
---|---|
static void |
PlanUtils.addInputsForView(ParseContext parseCtx)
|
void |
PTFDeserializer.initializePTFChain(PartitionedTableFunctionDef tblFnDef)
|
void |
PTFDeserializer.initializeWindowing(WindowTableFunctionDef def)
|
Constructors in org.apache.hadoop.hive.ql.plan that throw HiveException | |
---|---|
PartitionDesc(Partition part)
|
|
PartitionDesc(Partition part,
TableDesc tblDesc)
|
Uses of HiveException in org.apache.hadoop.hive.ql.security |
---|
Methods in org.apache.hadoop.hive.ql.security that throw HiveException | |
---|---|
void |
HadoopDefaultAuthenticator.destroy()
|
void |
HiveAuthenticationProvider.destroy()
|
void |
SessionStateConfigUserAuthenticator.destroy()
|
void |
SessionStateUserAuthenticator.destroy()
|
Uses of HiveException in org.apache.hadoop.hive.ql.security.authorization |
---|
Methods in org.apache.hadoop.hive.ql.security.authorization that throw HiveException | |
---|---|
void |
BitSetCheckedAuthorizationProvider.authorize(org.apache.hadoop.hive.metastore.api.Database db,
Privilege[] inputRequiredPriv,
Privilege[] outputRequiredPriv)
|
void |
HiveAuthorizationProvider.authorize(org.apache.hadoop.hive.metastore.api.Database db,
Privilege[] readRequiredPriv,
Privilege[] writeRequiredPriv)
Authorization privileges against a database object. |
void |
StorageBasedAuthorizationProvider.authorize(org.apache.hadoop.hive.metastore.api.Database db,
Privilege[] readRequiredPriv,
Privilege[] writeRequiredPriv)
|
void |
BitSetCheckedAuthorizationProvider.authorize(Partition part,
Privilege[] inputRequiredPriv,
Privilege[] outputRequiredPriv)
|
void |
HiveAuthorizationProvider.authorize(Partition part,
Privilege[] readRequiredPriv,
Privilege[] writeRequiredPriv)
Authorization privileges against a hive partition object. |
void |
StorageBasedAuthorizationProvider.authorize(Partition part,
Privilege[] readRequiredPriv,
Privilege[] writeRequiredPriv)
|
void |
StorageBasedAuthorizationProvider.authorize(org.apache.hadoop.fs.Path path,
Privilege[] readRequiredPriv,
Privilege[] writeRequiredPriv)
Authorization privileges against a path. |
void |
BitSetCheckedAuthorizationProvider.authorize(Privilege[] inputRequiredPriv,
Privilege[] outputRequiredPriv)
|
void |
HiveAuthorizationProvider.authorize(Privilege[] readRequiredPriv,
Privilege[] writeRequiredPriv)
Authorization user level privileges. |
void |
StorageBasedAuthorizationProvider.authorize(Privilege[] readRequiredPriv,
Privilege[] writeRequiredPriv)
|
void |
BitSetCheckedAuthorizationProvider.authorize(Table table,
Partition part,
List<String> columns,
Privilege[] inputRequiredPriv,
Privilege[] outputRequiredPriv)
|
void |
HiveAuthorizationProvider.authorize(Table table,
Partition part,
List<String> columns,
Privilege[] readRequiredPriv,
Privilege[] writeRequiredPriv)
Authorization privileges against a list of columns. |
void |
StorageBasedAuthorizationProvider.authorize(Table table,
Partition part,
List<String> columns,
Privilege[] readRequiredPriv,
Privilege[] writeRequiredPriv)
|
void |
BitSetCheckedAuthorizationProvider.authorize(Table table,
Privilege[] inputRequiredPriv,
Privilege[] outputRequiredPriv)
|
void |
HiveAuthorizationProvider.authorize(Table table,
Privilege[] readRequiredPriv,
Privilege[] writeRequiredPriv)
Authorization privileges against a hive table object. |
void |
StorageBasedAuthorizationProvider.authorize(Table table,
Privilege[] readRequiredPriv,
Privilege[] writeRequiredPriv)
|
static HivePrincipal.HivePrincipalType |
AuthorizationUtils.getHivePrincipalType(org.apache.hadoop.hive.metastore.api.PrincipalType type)
Convert thrift principal type to authorization plugin principal type |
static org.apache.hadoop.hive.metastore.api.HiveObjectRef |
AuthorizationUtils.getThriftHiveObjectRef(HivePrivilegeObject privObj)
Convert thrift HiveObjectRef to plugin HivePrivilegeObject |
static org.apache.hadoop.hive.metastore.api.HiveObjectType |
AuthorizationUtils.getThriftHiveObjType(HivePrivilegeObject.HivePrivilegeObjectType type)
Convert plugin privilege object type to thrift type |
static org.apache.hadoop.hive.metastore.api.PrivilegeGrantInfo |
AuthorizationUtils.getThriftPrivilegeGrantInfo(HivePrivilege privilege,
HivePrincipal grantorPrincipal,
boolean grantOption,
int grantTime)
Get thrift privilege grant info |
void |
DefaultHiveAuthorizationProvider.init(org.apache.hadoop.conf.Configuration conf)
|
void |
DefaultHiveMetastoreAuthorizationProvider.init(org.apache.hadoop.conf.Configuration conf)
|
void |
HiveAuthorizationProvider.init(org.apache.hadoop.conf.Configuration conf)
|
void |
StorageBasedAuthorizationProvider.init(org.apache.hadoop.conf.Configuration conf)
|
Constructors in org.apache.hadoop.hive.ql.security.authorization that throw HiveException | |
---|---|
AuthorizationPreEventListener.PartitionWrapper(org.apache.hadoop.hive.metastore.api.Partition mapiPart,
PreEventContext context)
|
|
AuthorizationPreEventListener.PartitionWrapper(Table table,
org.apache.hadoop.hive.metastore.api.Partition mapiPart)
|
|
AuthorizationPreEventListener(org.apache.hadoop.conf.Configuration config)
|
Uses of HiveException in org.apache.hadoop.hive.ql.security.authorization.plugin |
---|
Subclasses of HiveException in org.apache.hadoop.hive.ql.security.authorization.plugin | |
---|---|
class |
HiveAccessControlException
Exception thrown by the Authorization plugin api (v2). |
class |
HiveAuthzPluginException
Exception thrown by the Authorization plugin api (v2). |
Uses of HiveException in org.apache.hadoop.hive.ql.session |
---|
Methods in org.apache.hadoop.hive.ql.session that throw HiveException | |
---|---|
void |
SessionState.applyAuthorizationPolicy()
If authorization mode is v2, then pass it through authorizer so that it can apply any security configuration changes. |
static CreateTableAutomaticGrant |
CreateTableAutomaticGrant.create(HiveConf conf)
|
Uses of HiveException in org.apache.hadoop.hive.ql.udf.generic |
---|
Methods in org.apache.hadoop.hive.ql.udf.generic that throw HiveException | |
---|---|
void |
NGramEstimator.add(ArrayList<String> ng)
Adds a new n-gram to the estimation. |
void |
GenericUDAFEvaluator.aggregate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
This function will be called by GroupByOperator when it sees a new input row. |
Object |
GenericUDFConcat.binaryEvaluate(GenericUDF.DeferredObject[] arguments)
|
abstract void |
GenericUDTF.close()
Called to notify the UDTF that there are no more rows to process. |
void |
GenericUDTFExplode.close()
|
void |
GenericUDTFInline.close()
|
void |
GenericUDTFJSONTuple.close()
|
void |
GenericUDTFParseUrlTuple.close()
|
void |
GenericUDTFPosExplode.close()
|
void |
GenericUDTFStack.close()
|
void |
Collector.collect(Object input)
Other classes will call collect() with the data that it has. |
void |
UDTFCollector.collect(Object input)
|
Integer |
GenericUDFBaseCompare.compare(GenericUDF.DeferredObject[] arguments)
|
void |
GenericUDAFAverage.GenericUDAFAverageEvaluatorDouble.doReset(org.apache.hadoop.hive.ql.udf.generic.GenericUDAFAverage.AverageAggregationBuffer<Double> aggregation)
|
void |
GenericUDAFAverage.GenericUDAFAverageEvaluatorDecimal.doReset(org.apache.hadoop.hive.ql.udf.generic.GenericUDAFAverage.AverageAggregationBuffer<HiveDecimal> aggregation)
|
Object |
GenericUDAFEvaluator.evaluate(GenericUDAFEvaluator.AggregationBuffer agg)
This function will be called by GroupByOperator when it sees a new input row. |
Object |
AbstractGenericUDFEWAHBitmapBop.evaluate(GenericUDF.DeferredObject[] arguments)
|
abstract Object |
GenericUDF.evaluate(GenericUDF.DeferredObject[] arguments)
Evaluate the GenericUDF with the arguments. |
Object |
GenericUDFAbs.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFArray.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFArrayContains.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFAssertTrue.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFBaseNumeric.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFBasePad.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFBaseTrim.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFBetween.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFBridge.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFCase.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFCoalesce.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFConcat.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFConcatWS.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFDate.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFDateAdd.evaluate(GenericUDF.DeferredObject[] arguments)
|
org.apache.hadoop.io.IntWritable |
GenericUDFDateDiff.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFDateSub.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFDecode.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFElt.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFEncode.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFEWAHBitmapEmpty.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFField.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFFloorCeilBase.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFFormatNumber.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFFromUtcTimestamp.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFHash.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFIf.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFIn.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFIndex.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFInFile.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFInstr.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFLeadLag.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFLocate.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFLower.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFMacro.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFMap.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFMapKeys.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFMapValues.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFNamedStruct.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFNvl.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFOPAnd.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFOPEqual.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFOPEqualNS.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFOPEqualOrGreaterThan.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFOPEqualOrLessThan.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFOPGreaterThan.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFOPLessThan.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFOPNegative.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFOPNot.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFOPNotEqual.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFOPNotNull.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFOPNull.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFOPOr.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFOPPositive.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFPower.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFPrintf.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFReflect.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFReflect2.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFRound.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFSentences.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFSize.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFSortArray.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFSplit.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFStringToMap.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFStruct.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFTimestamp.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFToBinary.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFToChar.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFToDate.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFToDecimal.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFToUnixTimeStamp.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFToVarchar.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFTranslate.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFUnion.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFUnixTimeStamp.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFUpper.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDFWhen.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
UDFCurrentDB.evaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDF.DeferredObject.get()
|
Object |
GenericUDF.DeferredJavaObject.get()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFAverage.GenericUDAFAverageEvaluatorDouble.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFAverage.GenericUDAFAverageEvaluatorDecimal.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFComputeStats.GenericUDAFBooleanStatsEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFComputeStats.GenericUDAFLongStatsEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFComputeStats.GenericUDAFDoubleStatsEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFComputeStats.GenericUDAFStringStatsEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFComputeStats.GenericUDAFBinaryStatsEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFComputeStats.GenericUDAFDecimalStatsEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFContextNGrams.GenericUDAFContextNGramEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFCorrelation.GenericUDAFCorrelationEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFCount.GenericUDAFCountEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFCovariance.GenericUDAFCovarianceEvaluator.getNewAggregationBuffer()
|
abstract GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFEvaluator.getNewAggregationBuffer()
Get a new aggregation object. |
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFEWAHBitmap.GenericUDAFEWAHBitmapEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFFirstValue.GenericUDAFFirstValueEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFHistogramNumeric.GenericUDAFHistogramNumericEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFLastValue.GenericUDAFLastValueEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFLeadLag.GenericUDAFLeadLagEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFMax.GenericUDAFMaxEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFMin.GenericUDAFMinEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFMkCollectionEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFnGrams.GenericUDAFnGramEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFNTile.GenericUDAFNTileEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFPercentileApprox.GenericUDAFPercentileApproxEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFRank.GenericUDAFRankEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFRowNumber.GenericUDAFRowNumberEvaluator.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFSum.GenericUDAFSumHiveDecimal.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFSum.GenericUDAFSumDouble.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFSum.GenericUDAFSumLong.getNewAggregationBuffer()
|
GenericUDAFEvaluator.AggregationBuffer |
GenericUDAFVariance.GenericUDAFVarianceEvaluator.getNewAggregationBuffer()
|
ArrayList<Object[]> |
NGramEstimator.getNGrams()
Returns the final top-k n-grams in a format suitable for returning to Hive. |
ObjectInspector |
GenericUDAFAverage.AbstractGenericUDAFAverageEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFBridge.GenericUDAFBridgeEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFComputeStats.GenericUDAFBooleanStatsEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFComputeStats.GenericUDAFLongStatsEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFComputeStats.GenericUDAFDoubleStatsEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFComputeStats.GenericUDAFStringStatsEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFComputeStats.GenericUDAFBinaryStatsEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFComputeStats.GenericUDAFDecimalStatsEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFContextNGrams.GenericUDAFContextNGramEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFCorrelation.GenericUDAFCorrelationEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFCount.GenericUDAFCountEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFCovariance.GenericUDAFCovarianceEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFCumeDist.GenericUDAFCumeDistEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
Initialize the evaluator. |
ObjectInspector |
GenericUDAFEWAHBitmap.GenericUDAFEWAHBitmapEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFFirstValue.GenericUDAFFirstValueEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFHistogramNumeric.GenericUDAFHistogramNumericEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFLastValue.GenericUDAFLastValueEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFLeadLag.GenericUDAFLeadLagEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFMax.GenericUDAFMaxEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFMin.GenericUDAFMinEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFMkCollectionEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFnGrams.GenericUDAFnGramEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFNTile.GenericUDAFNTileEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFPercentileApprox.GenericUDAFSinglePercentileApproxEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFPercentileApprox.GenericUDAFMultiplePercentileApproxEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFPercentRank.GenericUDAFPercentRankEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFRank.GenericUDAFRankEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFRowNumber.GenericUDAFRowNumberEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFSum.GenericUDAFSumHiveDecimal.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFSum.GenericUDAFSumDouble.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFSum.GenericUDAFSumLong.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
ObjectInspector |
GenericUDAFVariance.GenericUDAFVarianceEvaluator.init(GenericUDAFEvaluator.Mode m,
ObjectInspector[] parameters)
|
void |
NGramEstimator.initialize(int pk,
int ppf,
int pn)
Sets the 'k' and 'pf' parameters. |
void |
GenericUDAFAverage.AbstractGenericUDAFAverageEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer aggregation,
Object[] parameters)
|
void |
GenericUDAFBridge.GenericUDAFBridgeEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFComputeStats.GenericUDAFBooleanStatsEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFComputeStats.GenericUDAFLongStatsEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFComputeStats.GenericUDAFDoubleStatsEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFComputeStats.GenericUDAFStringStatsEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFComputeStats.GenericUDAFBinaryStatsEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFComputeStats.GenericUDAFDecimalStatsEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFContextNGrams.GenericUDAFContextNGramEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFCorrelation.GenericUDAFCorrelationEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFCount.GenericUDAFCountEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFCovariance.GenericUDAFCovarianceEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
abstract void |
GenericUDAFEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
Iterate through original data. |
void |
GenericUDAFEWAHBitmap.GenericUDAFEWAHBitmapEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFFirstValue.GenericUDAFFirstValueEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFHistogramNumeric.GenericUDAFHistogramNumericEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFLastValue.GenericUDAFLastValueEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFLeadLag.GenericUDAFLeadLagEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFMax.GenericUDAFMaxEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFMin.GenericUDAFMinEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFMkCollectionEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFnGrams.GenericUDAFnGramEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFNTile.GenericUDAFNTileEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFPercentileApprox.GenericUDAFPercentileApproxEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFRank.GenericUDAFRankEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFRowNumber.GenericUDAFRowNumberEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFSum.GenericUDAFSumHiveDecimal.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFSum.GenericUDAFSumDouble.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFSum.GenericUDAFSumLong.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDAFVariance.GenericUDAFVarianceEvaluator.iterate(GenericUDAFEvaluator.AggregationBuffer agg,
Object[] parameters)
|
void |
GenericUDFInFile.load(InputStream is)
Load the file from an InputStream. |
void |
GenericUDAFAverage.AbstractGenericUDAFAverageEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer aggregation,
Object partial)
|
void |
GenericUDAFBridge.GenericUDAFBridgeEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFComputeStats.GenericUDAFBooleanStatsEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFComputeStats.GenericUDAFLongStatsEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFComputeStats.GenericUDAFDoubleStatsEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFComputeStats.GenericUDAFStringStatsEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFComputeStats.GenericUDAFBinaryStatsEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFComputeStats.GenericUDAFDecimalStatsEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFContextNGrams.GenericUDAFContextNGramEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object obj)
|
void |
GenericUDAFCorrelation.GenericUDAFCorrelationEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFCount.GenericUDAFCountEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFCovariance.GenericUDAFCovarianceEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
abstract void |
GenericUDAFEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
Merge with partial aggregation result. |
void |
GenericUDAFEWAHBitmap.GenericUDAFEWAHBitmapEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFFirstValue.GenericUDAFFirstValueEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFHistogramNumeric.GenericUDAFHistogramNumericEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFLastValue.GenericUDAFLastValueEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFLeadLag.GenericUDAFLeadLagEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFMax.GenericUDAFMaxEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFMin.GenericUDAFMinEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFMkCollectionEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFnGrams.GenericUDAFnGramEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFNTile.GenericUDAFNTileEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFPercentileApprox.GenericUDAFPercentileApproxEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFRank.GenericUDAFRankEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFRowNumber.GenericUDAFRowNumberEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFSum.GenericUDAFSumHiveDecimal.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFSum.GenericUDAFSumDouble.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFSum.GenericUDAFSumLong.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
GenericUDAFVariance.GenericUDAFVarianceEvaluator.merge(GenericUDAFEvaluator.AggregationBuffer agg,
Object partial)
|
void |
NGramEstimator.merge(List<org.apache.hadoop.io.Text> other)
Takes a serialized n-gram estimator object created by the serialize() method and merges it with the current n-gram object. |
void |
GenericUDF.DeferredObject.prepare(int version)
|
void |
GenericUDF.DeferredJavaObject.prepare(int version)
|
abstract void |
GenericUDTF.process(Object[] args)
Give a set of arguments for the UDTF to process. |
void |
GenericUDTFExplode.process(Object[] o)
|
void |
GenericUDTFInline.process(Object[] os)
|
void |
GenericUDTFJSONTuple.process(Object[] o)
|
void |
GenericUDTFParseUrlTuple.process(Object[] o)
|
void |
GenericUDTFPosExplode.process(Object[] o)
|
void |
GenericUDTFStack.process(Object[] args)
|
void |
GenericUDAFAverage.AbstractGenericUDAFAverageEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer aggregation)
|
void |
GenericUDAFBridge.GenericUDAFBridgeEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFComputeStats.GenericUDAFBooleanStatsEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFComputeStats.GenericUDAFLongStatsEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFComputeStats.GenericUDAFDoubleStatsEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFComputeStats.GenericUDAFStringStatsEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFComputeStats.GenericUDAFBinaryStatsEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFComputeStats.GenericUDAFDecimalStatsEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFContextNGrams.GenericUDAFContextNGramEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFCorrelation.GenericUDAFCorrelationEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFCount.GenericUDAFCountEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFCovariance.GenericUDAFCovarianceEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
abstract void |
GenericUDAFEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
Reset the aggregation. |
void |
GenericUDAFEWAHBitmap.GenericUDAFEWAHBitmapEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFFirstValue.GenericUDAFFirstValueEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFHistogramNumeric.GenericUDAFHistogramNumericEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFLastValue.GenericUDAFLastValueEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFLeadLag.GenericUDAFLeadLagEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFMax.GenericUDAFMaxEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFMin.GenericUDAFMinEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFMkCollectionEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFnGrams.GenericUDAFnGramEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFNTile.GenericUDAFNTileEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFPercentileApprox.GenericUDAFPercentileApproxEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFRank.GenericUDAFRankEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFRowNumber.GenericUDAFRowNumberEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFSum.GenericUDAFSumHiveDecimal.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFSum.GenericUDAFSumDouble.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFSum.GenericUDAFSumLong.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
void |
GenericUDAFVariance.GenericUDAFVarianceEvaluator.reset(GenericUDAFEvaluator.AggregationBuffer agg)
|
ArrayList<org.apache.hadoop.io.Text> |
NGramEstimator.serialize()
In preparation for a Hive merge() call, serializes the current n-gram estimator object into an ArrayList of Text objects. |
String |
GenericUDFConcat.stringEvaluate(GenericUDF.DeferredObject[] arguments)
|
Object |
GenericUDAFAverage.AbstractGenericUDAFAverageEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer aggregation)
|
Object |
GenericUDAFBridge.GenericUDAFBridgeEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFComputeStats.GenericUDAFBooleanStatsEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFComputeStats.GenericUDAFLongStatsEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFComputeStats.GenericUDAFDoubleStatsEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFComputeStats.GenericUDAFStringStatsEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFComputeStats.GenericUDAFBinaryStatsEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFComputeStats.GenericUDAFDecimalStatsEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFContextNGrams.GenericUDAFContextNGramEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFCorrelation.GenericUDAFCorrelationEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFCount.GenericUDAFCountEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFCovariance.GenericUDAFCovarianceEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFCovarianceSample.GenericUDAFCovarianceSampleEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFCumeDist.GenericUDAFCumeDistEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
abstract Object |
GenericUDAFEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
Get final aggregation result. |
Object |
GenericUDAFEWAHBitmap.GenericUDAFEWAHBitmapEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFFirstValue.GenericUDAFFirstValueEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFHistogramNumeric.GenericUDAFHistogramNumericEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFLastValue.GenericUDAFLastValueEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFLeadLag.GenericUDAFLeadLagEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFMax.GenericUDAFMaxEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFMin.GenericUDAFMinEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFMkCollectionEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFnGrams.GenericUDAFnGramEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFNTile.GenericUDAFNTileEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFPercentileApprox.GenericUDAFSinglePercentileApproxEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFPercentileApprox.GenericUDAFMultiplePercentileApproxEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFPercentRank.GenericUDAFPercentRankEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFRank.GenericUDAFRankEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFRowNumber.GenericUDAFRowNumberEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFStd.GenericUDAFStdEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFStdSample.GenericUDAFStdSampleEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFSum.GenericUDAFSumHiveDecimal.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFSum.GenericUDAFSumDouble.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFSum.GenericUDAFSumLong.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFVariance.GenericUDAFVarianceEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFVarianceSample.GenericUDAFVarianceSampleEvaluator.terminate(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFAverage.AbstractGenericUDAFAverageEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer aggregation)
|
Object |
GenericUDAFBridge.GenericUDAFBridgeEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFComputeStats.GenericUDAFBooleanStatsEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFComputeStats.GenericUDAFLongStatsEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFComputeStats.GenericUDAFDoubleStatsEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFComputeStats.GenericUDAFStringStatsEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFComputeStats.GenericUDAFBinaryStatsEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFComputeStats.GenericUDAFDecimalStatsEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFContextNGrams.GenericUDAFContextNGramEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFCorrelation.GenericUDAFCorrelationEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFCount.GenericUDAFCountEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFCovariance.GenericUDAFCovarianceEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
abstract Object |
GenericUDAFEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
Get partial aggregation result. |
Object |
GenericUDAFEWAHBitmap.GenericUDAFEWAHBitmapEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFFirstValue.GenericUDAFFirstValueEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFHistogramNumeric.GenericUDAFHistogramNumericEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFLastValue.GenericUDAFLastValueEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFLeadLag.GenericUDAFLeadLagEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFMax.GenericUDAFMaxEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFMin.GenericUDAFMinEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFMkCollectionEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFnGrams.GenericUDAFnGramEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFNTile.GenericUDAFNTileEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFPercentileApprox.GenericUDAFPercentileApproxEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFRank.GenericUDAFRankEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFRowNumber.GenericUDAFRowNumberEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFSum.GenericUDAFSumHiveDecimal.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFSum.GenericUDAFSumDouble.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFSum.GenericUDAFSumLong.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Object |
GenericUDAFVariance.GenericUDAFVarianceEvaluator.terminatePartial(GenericUDAFEvaluator.AggregationBuffer agg)
|
Uses of HiveException in org.apache.hadoop.hive.ql.udf.ptf |
---|
Methods in org.apache.hadoop.hive.ql.udf.ptf that throw HiveException | |
---|---|
void |
WindowingTableFunction.execute(PTFPartition.PTFPartitionIterator<Object> pItr,
PTFPartition outP)
|
PTFPartition |
Noop.execute(PTFPartition iPart)
|
PTFPartition |
NoopWithMap.execute(PTFPartition iPart)
|
PTFPartition |
TableFunctionEvaluator.execute(PTFPartition iPart)
|
void |
TableFunctionResolver.initialize(PTFDesc ptfDesc,
PartitionedTableFunctionDef tDef,
TableFunctionEvaluator evaluator)
|
void |
Noop.NoopResolver.initializeOutputOI()
|
void |
NoopWithMap.NoopWithMapResolver.initializeOutputOI()
|
abstract void |
TableFunctionResolver.initializeOutputOI()
This method is invoked during runtime(during deserialization of theQueryDef). |
void |
WindowingTableFunction.WindowingTableFunctionResolver.initializeOutputOI()
|
void |
NoopWithMap.NoopWithMapResolver.initializeRawInputOI()
|
void |
TableFunctionResolver.initializeRawInputOI()
|
PTFPartition |
TableFunctionEvaluator.transformRawInput(PTFPartition iPart)
|
Uses of HiveException in org.apache.hadoop.hive.ql.udf.xml |
---|
Methods in org.apache.hadoop.hive.ql.udf.xml that throw HiveException | |
---|---|
Object |
GenericUDFXPath.evaluate(GenericUDF.DeferredObject[] arguments)
|
|
||||||||||
PREV NEXT | FRAMES NO FRAMES |