public class ParquetTableScan
extends org.apache.spark.sql.execution.SparkPlan
implements org.apache.spark.sql.execution.LeafNode, scala.Product, scala.Serializable
ParquetRelation
as a
RDD[Row]
.Constructor and Description |
---|
ParquetTableScan(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> attributes,
ParquetRelation relation,
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> columnPruningPred) |
Modifier and Type | Method and Description |
---|---|
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> |
attributes() |
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> |
columnPruningPred() |
RDD<org.apache.spark.sql.Row> |
execute() |
scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> |
output() |
ParquetTableScan |
pruneColumns(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> prunedAttributes)
Applies a (candidate) projection.
|
ParquetRelation |
relation() |
scala.Tuple2<Object,Object>[] |
requestedPartitionOrdinals() |
codegenEnabled, executeCollect, executeTake, isTraceEnabled, log, logDebug, logDebug, logError, logError, logInfo, logInfo, logName, logTrace, logTrace, logWarning, logWarning, makeCopy, newMutableProjection, newOrdering, newPredicate, newProjection, org$apache$spark$Logging$$log__$eq, org$apache$spark$Logging$$log_, outputPartitioning, requiredChildDistribution, sparkContext, sqlContext
expressions, inputSet, missingInput, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionDown$1, org$apache$spark$sql$catalyst$plans$QueryPlan$$transformExpressionUp$1, outputSet, printSchema, references, schema, schemaString, simpleString, statePrefix, transformAllExpressions, transformExpressions, transformExpressionsDown, transformExpressionsUp
apply, argString, asCode, children, collect, fastEquals, flatMap, foreach, foreachUp, generateTreeString, getNodeNumbered, map, mapChildren, nodeName, numberedTreeString, origin, otherCopyArgs, stringArgs, toString, transform, transformChildrenDown, transformChildrenUp, transformDown, transformUp, treeString, withNewChildren
productArity, productElement, productIterator, productPrefix
initializeIfNecessary, initializeLogging, log_
public ParquetTableScan(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> attributes, ParquetRelation relation, scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> columnPruningPred)
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> attributes()
public ParquetRelation relation()
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Expression> columnPruningPred()
public scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> output()
output
in class org.apache.spark.sql.catalyst.plans.QueryPlan<org.apache.spark.sql.execution.SparkPlan>
public scala.Tuple2<Object,Object>[] requestedPartitionOrdinals()
public RDD<org.apache.spark.sql.Row> execute()
execute
in class org.apache.spark.sql.execution.SparkPlan
public ParquetTableScan pruneColumns(scala.collection.Seq<org.apache.spark.sql.catalyst.expressions.Attribute> prunedAttributes)
prunedAttributes
- The list of attributes to be used in the projection.