本文整理汇总了Java中org.apache.hadoop.hive.ql.exec.TableScanOperator类的典型用法代码示例。如果您正苦于以下问题:Java TableScanOperator类的具体用法?Java TableScanOperator怎么用?Java TableScanOperator使用的例子?那么恭喜您, 这里精选的类代码示例或许可以为您提供帮助。
TableScanOperator类属于org.apache.hadoop.hive.ql.exec包,在下文中一共展示了TableScanOperator类的6个代码示例,这些例子默认根据受欢迎程度排序。您可以为喜欢或者感觉有用的代码点赞,您的评价将有助于我们的系统推荐出更棒的Java代码示例。
示例1: estimate
import org.apache.hadoop.hive.ql.exec.TableScanOperator; //导入依赖的package包/类
@Override
public Estimation estimate(JobConf job, TableScanOperator ts, long remaining) throws HiveException {
String hiveTableName = ts.getConf().getTableMetadata().getTableName();
int reducerCount = job.getInt(hiveTableName + PhoenixStorageHandlerConstants.PHOENIX_REDUCER_NUMBER, 1);
if (LOG.isDebugEnabled()) {
LOG.debug("<<<<<<<<<< tableName : " + hiveTableName + " >>>>>>>>>>" );
LOG.debug("<<<<<<<<<< reducer count : " + reducerCount + " >>>>>>>>>>");
LOG.debug("<<<<<<<<<< remaining : " + remaining + " >>>>>>>>>>");
}
long bytesPerReducer = job.getLong(HiveConf.ConfVars.BYTESPERREDUCER.varname, Long.parseLong(HiveConf.ConfVars.BYTESPERREDUCER.getDefaultValue()));
long totalLength = reducerCount * bytesPerReducer;
return new Estimation(0, totalLength);
}
开发者ID:mini666,项目名称:hive-phoenix-handler,代码行数:17,代码来源:PhoenixStorageHandler.java
示例2: addSplitsForGroup
import org.apache.hadoop.hive.ql.exec.TableScanOperator; //导入依赖的package包/类
private void addSplitsForGroup(List<Path> dirs, TableScanOperator tableScan, JobConf conf,
InputFormat inputFormat, Class<? extends InputFormat> inputFormatClass, int splits,
TableDesc table, List<InputSplit> result) throws IOException {
Utilities.copyTablePropertiesToConf(table, conf);
if (tableScan != null) {
pushFilters(conf, tableScan);
}
FileInputFormat.setInputPaths(conf, dirs.toArray(new Path[dirs.size()]));
conf.setInputFormat(inputFormat.getClass());
int headerCount = 0;
int footerCount = 0;
if (table != null) {
headerCount = Utilities.getHeaderCount(table);
footerCount = Utilities.getFooterCount(table, conf);
if (headerCount != 0 || footerCount != 0) {
// Input file has header or footer, cannot be splitted.
conf.setLong(
ShimLoader.getHadoopShims().getHadoopConfNames().get("MAPREDMINSPLITSIZE"),
Long.MAX_VALUE);
}
}
InputSplit[] iss = inputFormat.getSplits(conf, splits);
for (InputSplit is : iss) {
result.add(new HiveInputSplit(is, inputFormatClass.getName()));
}
}
开发者ID:mini666,项目名称:hive-phoenix-handler,代码行数:32,代码来源:HiveInputFormat.java
示例3: pushFilters
import org.apache.hadoop.hive.ql.exec.TableScanOperator; //导入依赖的package包/类
private void pushFilters(final JobConf jobConf, final TableScanOperator tableScan) {
final TableScanDesc scanDesc = tableScan.getConf();
if (scanDesc == null) {
LOG.debug("Not pushing filters because TableScanDesc is null");
return;
}
// construct column name list for reference by filter push down
Utilities.setColumnNameList(jobConf, tableScan);
// push down filters
final ExprNodeDesc filterExpr = scanDesc.getFilterExpr();
if (filterExpr == null) {
LOG.debug("Not pushing filters because FilterExpr is null");
return;
}
final String filterText = filterExpr.getExprString();
final String filterExprSerialized = Utilities.serializeExpression(filterExpr);
jobConf.set(
TableScanDesc.FILTER_TEXT_CONF_STR,
filterText);
jobConf.set(
TableScanDesc.FILTER_EXPR_CONF_STR,
filterExprSerialized);
}
开发者ID:apache,项目名称:parquet-mr,代码行数:28,代码来源:Hive012Binding.java
示例4: pushFilters
import org.apache.hadoop.hive.ql.exec.TableScanOperator; //导入依赖的package包/类
public static void pushFilters(JobConf jobConf, TableScanOperator tableScan) {
// ensure filters are not set from previous pushFilters
jobConf.unset(TableScanDesc.FILTER_TEXT_CONF_STR);
jobConf.unset(TableScanDesc.FILTER_EXPR_CONF_STR);
TableScanDesc scanDesc = tableScan.getConf();
if (scanDesc == null) {
return;
}
// 2015-10-27 Added by JeongMin Ju
//////////////////////////////////////////////////////////////////////
setReadColumns(jobConf, tableScan);
//////////////////////////////////////////////////////////////////////
// construct column name list and types for reference by filter push down
Utilities.setColumnNameList(jobConf, tableScan);
Utilities.setColumnTypeList(jobConf, tableScan);
// push down filters
ExprNodeGenericFuncDesc filterExpr = (ExprNodeGenericFuncDesc)scanDesc.getFilterExpr();
if (filterExpr == null) {
return;
}
Serializable filterObject = scanDesc.getFilterObject();
if (filterObject != null) {
jobConf.set(TableScanDesc.FILTER_OBJECT_CONF_STR, Utilities.serializeObject(filterObject));
}
String filterText = filterExpr.getExprString();
String filterExprSerialized = Utilities.serializeExpression(filterExpr);
if (LOG.isDebugEnabled()) {
LOG.debug("Filter text = " + filterText);
LOG.debug("Filter expression = " + filterExprSerialized);
}
jobConf.set(
TableScanDesc.FILTER_TEXT_CONF_STR,
filterText);
jobConf.set(
TableScanDesc.FILTER_EXPR_CONF_STR,
filterExprSerialized);
}
开发者ID:mini666,项目名称:hive-phoenix-handler,代码行数:42,代码来源:HiveInputFormat.java
示例5: pushProjectionsAndFilters
import org.apache.hadoop.hive.ql.exec.TableScanOperator; //导入依赖的package包/类
protected void pushProjectionsAndFilters(JobConf jobConf, Class inputFormatClass,
String splitPath, String splitPathWithNoSchema, boolean nonNative) {
if (this.mrwork == null) {
init(job);
}
if(this.mrwork.getPathToAliases() == null) {
return;
}
ArrayList<String> aliases = new ArrayList<String>();
Iterator<Entry<String, ArrayList<String>>> iterator = this.mrwork
.getPathToAliases().entrySet().iterator();
while (iterator.hasNext()) {
Entry<String, ArrayList<String>> entry = iterator.next();
String key = entry.getKey();
boolean match;
if (nonNative) {
// For non-native tables, we need to do an exact match to avoid
// HIVE-1903. (The table location contains no files, and the string
// representation of its path does not have a trailing slash.)
match =
splitPath.equals(key) || splitPathWithNoSchema.equals(key);
} else {
// But for native tables, we need to do a prefix match for
// subdirectories. (Unlike non-native tables, prefix mixups don't seem
// to be a potential problem here since we are always dealing with the
// path to something deeper than the table location.)
match =
splitPath.startsWith(key) || splitPathWithNoSchema.startsWith(key);
}
if (match) {
ArrayList<String> list = entry.getValue();
for (String val : list) {
aliases.add(val);
}
}
}
for (String alias : aliases) {
Operator<? extends OperatorDesc> op = this.mrwork.getAliasToWork().get(
alias);
if (op instanceof TableScanOperator) {
TableScanOperator ts = (TableScanOperator) op;
// push down projections.
ColumnProjectionUtils.appendReadColumns(
jobConf, ts.getNeededColumnIDs(), ts.getNeededColumns());
// push down filters
pushFilters(jobConf, ts);
}
}
}
开发者ID:mini666,项目名称:hive-phoenix-handler,代码行数:54,代码来源:HiveInputFormat.java
示例6: setReadColumns
import org.apache.hadoop.hive.ql.exec.TableScanOperator; //导入依赖的package包/类
/**
* Set read columns.
* In case of select query(select t1.c1, t1.c2, ... from t1, t2 where t1.c1 = t2.c1), it can not read target columns at InputFormat class because columns of t1 table are not set.
* therefore, set to force read columns.
*
* @author JeongMin Ju
* @param jobConf
* @param tableScan
* @create-date 2016-01-15
*/
private static void setReadColumns(JobConf jobConf, TableScanOperator tableScan) {
List<String> readColumnList = tableScan.getNeededColumns();
if (readColumnList != null) {
jobConf.set(ColumnProjectionUtils.READ_COLUMN_NAMES_CONF_STR, Joiner.on(",").join(readColumnList));
}
}
开发者ID:mini666,项目名称:hive-phoenix-handler,代码行数:17,代码来源:HiveInputFormat.java
注:本文中的org.apache.hadoop.hive.ql.exec.TableScanOperator类示例整理自Github/MSDocs等源码及文档管理平台,相关代码片段筛选自各路编程大神贡献的开源项目,源码版权归原作者所有,传播和使用请参考对应项目的License;未经允许,请勿转载。 |
请发表评论