WindowExec Unary Physical Operator
WindowExec is a unary physical operator (i.e. with one child physical operator) for window aggregation execution that represents Window unary logical operator at execution.
WindowExec is created exclusively when BasicOperators execution planning strategy resolves a Window unary logical operator.
|
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 |
// arguably the most trivial example // just a dataset of 3 rows per group // to demo how partitions and frames work // note the rows per groups are not consecutive (in the middle) val metrics = Seq( (0, 0, 0), (1, 0, 1), (2, 5, 2), (3, 0, 3), (4, 0, 1), (5, 5, 3), (6, 5, 0) ).toDF("id", "device", "level") scala> metrics.show +---+------+-----+ | id|device|level| +---+------+-----+ | 0| 0| 0| | 1| 0| 1| | 2| 5| 2| // <-- this row for device 5 is among the rows of device 0 | 3| 0| 3| // <-- as above but for device 0 | 4| 0| 1| // <-- almost as above but there is a group of two rows for device 0 | 5| 5| 3| | 6| 5| 0| +---+------+-----+ // create windows of rows to use window aggregate function over every window import org.apache.spark.sql.expressions.Window val rangeWithTwoDevicesById = Window. partitionBy('device). orderBy('id). rangeBetween(start = -1, end = Window.currentRow) // <-- use rangeBetween first val sumOverRange = metrics.withColumn("sum", sum('level) over rangeWithTwoDevicesById) // Logical plan with Window unary logical operator val optimizedPlan = sumOverRange.queryExecution.optimizedPlan scala> println(optimizedPlan) Window [sum(cast(level#9 as bigint)) windowspecdefinition(device#8, id#7 ASC NULLS FIRST, RANGE BETWEEN 1 PRECEDING AND CURRENT ROW) AS sum#15L], [device#8], [id#7 ASC NULLS FIRST] +- LocalRelation [id#7, device#8, level#9] // Physical plan with WindowExec unary physical operator (shown as Window) scala> sumOverRange.explain == Physical Plan == Window [sum(cast(level#9 as bigint)) windowspecdefinition(device#8, id#7 ASC NULLS FIRST, RANGE BETWEEN 1 PRECEDING AND CURRENT ROW) AS sum#15L], [device#8], [id#7 ASC NULLS FIRST] +- *Sort [device#8 ASC NULLS FIRST, id#7 ASC NULLS FIRST], false, 0 +- Exchange hashpartitioning(device#8, 200) +- LocalTableScan [id#7, device#8, level#9] // Going fairly low-level...you've been warned val plan = sumOverRange.queryExecution.executedPlan import org.apache.spark.sql.execution.window.WindowExec val we = plan.asInstanceOf[WindowExec] val windowRDD = we.execute() scala> :type windowRDD org.apache.spark.rdd.RDD[org.apache.spark.sql.catalyst.InternalRow] scala> windowRDD.toDebugString res0: String = (200) MapPartitionsRDD[5] at execute at <console>:35 [] | MapPartitionsRDD[4] at execute at <console>:35 [] | ShuffledRowRDD[3] at execute at <console>:35 [] +-(7) MapPartitionsRDD[2] at execute at <console>:35 [] | MapPartitionsRDD[1] at execute at <console>:35 [] | ParallelCollectionRDD[0] at execute at <console>:35 [] // no computation on the source dataset has really occurred // Let's trigger a RDD action scala> windowRDD.first res0: org.apache.spark.sql.catalyst.InternalRow = [0,2,5,2,2] scala> windowRDD.foreach(println) [0,2,5,2,2] [0,0,0,0,0] [0,5,5,3,3] [0,6,5,0,3] [0,1,0,1,1] [0,3,0,3,3] [0,4,0,1,4] scala> sumOverRange.show +---+------+-----+---+ | id|device|level|sum| +---+------+-----+---+ | 2| 5| 2| 2| | 5| 5| 3| 3| | 6| 5| 0| 3| | 0| 0| 0| 0| | 1| 0| 1| 1| | 3| 0| 3| 3| | 4| 0| 1| 4| +---+------+-----+---+ // use rowsBetween val rowsWithTwoDevicesById = Window. partitionBy('device). orderBy('id). rowsBetween(start = -1, end = Window.currentRow) val sumOverRows = metrics.withColumn("sum", sum('level) over rowsWithTwoDevicesById) // let's see the result first to have them close // and compare row- vs range-based windows scala> sumOverRows.show +---+------+-----+---+ | id|device|level|sum| +---+------+-----+---+ | 2| 5| 2| 2| | 5| 5| 3| 5| <-- a difference | 6| 5| 0| 3| | 0| 0| 0| 0| | 1| 0| 1| 1| | 3| 0| 3| 4| <-- another difference | 4| 0| 1| 4| +---+------+-----+---+ val rowsOptimizedPlan = sumOverRows.queryExecution.optimizedPlan scala> println(rowsOptimizedPlan) Window [sum(cast(level#901 as bigint)) windowspecdefinition(device#900, id#899 ASC NULLS FIRST, ROWS BETWEEN 1 PRECEDING AND CURRENT ROW) AS sum#1458L], [device#900], [id#899 ASC NULLS FIRST] +- LocalRelation [id#899, device#900, level#901] scala> sumOverRows.explain == Physical Plan == Window [sum(cast(level#901 as bigint)) windowspecdefinition(device#900, id#899 ASC NULLS FIRST, ROWS BETWEEN 1 PRECEDING AND CURRENT ROW) AS sum#1458L], [device#900], [id#899 ASC NULLS FIRST] +- *Sort [device#900 ASC NULLS FIRST, id#899 ASC NULLS FIRST], false, 0 +- Exchange hashpartitioning(device#900, 200) +- LocalTableScan [id#899, device#900, level#901] |
|
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 |
// a more involved example val dataset = spark.range(start = 0, end = 13, step = 1, numPartitions = 4) import org.apache.spark.sql.expressions.Window val groupsOrderById = Window.partitionBy('group).rangeBetween(-2, Window.currentRow).orderBy('id) val query = dataset. withColumn("group", 'id % 4). select('*, sum('id) over groupsOrderById as "sum") scala> query.explain == Physical Plan == Window [sum(id#25L) windowspecdefinition(group#244L, id#25L ASC NULLS FIRST, RANGE BETWEEN 2 PRECEDING AND CURRENT ROW) AS sum#249L], [group#244L], [id#25L ASC NULLS FIRST] +- *Sort [group#244L ASC NULLS FIRST, id#25L ASC NULLS FIRST], false, 0 +- Exchange hashpartitioning(group#244L, 200) +- *Project [id#25L, (id#25L % 4) AS group#244L] +- *Range (0, 13, step=1, splits=4) val plan = query.queryExecution.executedPlan import org.apache.spark.sql.execution.window.WindowExec val we = plan.asInstanceOf[WindowExec] |
The output schema of WindowExec are the attributes of child physical operator and window expressions.
|
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 |
val schema = query.queryExecution.executedPlan.output.toStructType scala> println(schema.treeString) root |-- id: long (nullable = false) |-- group: long (nullable = true) |-- sum: long (nullable = true) // we is WindowExec created earlier // child's output scala> println(we.child.output.toStructType.treeString) root |-- id: long (nullable = false) |-- group: long (nullable = true) // window expressions' output scala> println(we.windowExpression.map(_.toAttribute).toStructType.treeString) root |-- sum: long (nullable = true) |
| Single Child |
|---|
|
ClusteredDistribution (per window partition specifications expressions) |
If no window partition specification is specified, WindowExec prints out the following WARN message to the logs (and the child’s distribution requirement is AllTuples):
|
1 2 3 4 5 |
WARN WindowExec: No Partition Defined for Window operation! Moving all data to a single partition, this can cause serious performance degradation. |
|
Tip
|
Enable Add the following line to
Refer to Logging. |
Executing Physical Operator (Generating RDD[InternalRow]) — doExecute Method
|
1 2 3 4 5 |
doExecute(): RDD[InternalRow] |
|
Note
|
doExecute is part of SparkPlan Contract to generate the runtime representation of a structured query as a distributed computation over internal binary rows on Apache Spark (i.e. RDD[InternalRow]).
|
doExecute executes the single child physical operator and maps over partitions using a custom Iterator[InternalRow].
|
Note
|
When executed, doExecute creates a MapPartitionsRDD with the child physical operator’s RDD[InternalRow].
|
|
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 |
scala> :type we org.apache.spark.sql.execution.window.WindowExec val windowRDD = we.execute scala> :type windowRDD org.apache.spark.rdd.RDD[org.apache.spark.sql.catalyst.InternalRow] scala> println(windowRDD.toDebugString) (200) MapPartitionsRDD[5] at execute at <console>:35 [] | MapPartitionsRDD[4] at execute at <console>:35 [] | ShuffledRowRDD[3] at execute at <console>:35 [] +-(7) MapPartitionsRDD[2] at execute at <console>:35 [] | MapPartitionsRDD[1] at execute at <console>:35 [] | ParallelCollectionRDD[0] at execute at <console>:35 [] |
Internally, doExecute first takes WindowExpressions and their WindowFunctionFrame factory functions (from window frame factories) followed by executing the single child physical operator and mapping over partitions (using RDD.mapPartitions operator).
doExecute creates an Iterator[InternalRow] (of UnsafeRow exactly).
Mapping Over UnsafeRows per Partition — Iterator[InternalRow]
When created, Iterator[InternalRow] first creates two UnsafeProjection conversion functions (to convert InternalRows to UnsafeRows) as result and grouping.
|
Note
|
grouping conversion function is created for window partition specifications expressions and used exclusively to create nextGroup when Iterator[InternalRow] is requested next row.
|
|
Tip
|
Enable Add the following line to
Refer to Logging. |
Iterator[InternalRow] then fetches the first row from the upstream RDD and initializes nextRow and nextGroup UnsafeRows.
|
Note
|
nextGroup is the result of converting nextRow using grouping conversion function.
|
doExecute creates a ExternalAppendOnlyUnsafeRowArray buffer using spark.sql.windowExec.buffer.spill.threshold property (default: 4096) as the threshold for the number of rows buffered.
doExecute creates a SpecificInternalRow for the window function result (as windowFunctionResult).
|
Note
|
SpecificInternalRow is also used in the generated code for the UnsafeProjection for the result.
|
doExecute takes the window frame factories and generates WindowFunctionFrame per factory (using the SpecificInternalRow created earlier).
|
Caution
|
FIXME |
|
Note
|
ExternalAppendOnlyUnsafeRowArray is used to collect UnsafeRow objects from the child’s partitions (one partition per buffer and up to spark.sql.windowExec.buffer.spill.threshold).
|
next Method
|
1 2 3 4 5 |
override final def next(): InternalRow |
|
Note
|
next is part of Scala’s scala.collection.Iterator interface that returns the next element and discards it from the iterator.
|
next method of the final Iterator is…FIXME
next first fetches a new partition, but only when…FIXME
|
Note
|
next loads all the rows in nextGroup.
|
|
Caution
|
FIXME What’s nextGroup?
|
next takes one UnsafeRow from bufferIterator.
|
Caution
|
FIXME bufferIterator seems important for the iteration.
|
next then requests every WindowFunctionFrame to write the current rowIndex and UnsafeRow.
|
Caution
|
FIXME rowIndex?
|
next joins the current UnsafeRow and windowFunctionResult (i.e. takes two InternalRows and makes them appear as a single concatenated InternalRow).
next increments rowIndex.
In the end, next uses the UnsafeProjection function (that was created using createResultProjection) and projects the joined InternalRow to the result UnsafeRow.
Fetching All Rows In Partition — fetchNextPartition Internal Method
|
1 2 3 4 5 |
fetchNextPartition(): Unit |
fetchNextPartition first copies the current nextGroup UnsafeRow (that was created using grouping projection function) and clears the internal buffer.
fetchNextPartition then collects all UnsafeRows for the current nextGroup in buffer.
With the buffer filled in (with UnsafeRows per partition), fetchNextPartition prepares every WindowFunctionFrame function in frames one by one (and passing buffer).
In the end, fetchNextPartition resets rowIndex to 0 and requests buffer to generate an iterator (available as bufferIterator).
|
Note
|
fetchNextPartition is used internally when doExecute‘s Iterator is requested for the next UnsafeRow (when bufferIterator is uninitialized or was drained, i.e. holds no elements, but there are still rows in the upstream operator’s partition).
|
fetchNextRow Internal Method
|
1 2 3 4 5 |
fetchNextRow(): Unit |
fetchNextRow checks whether there is the next row available (using the upstream Iterator.hasNext) and sets nextRowAvailable mutable internal flag.
If there is a row available, fetchNextRow sets nextRow internal variable to the next UnsafeRow from the upstream’s RDD.
fetchNextRow also sets nextGroup internal variable as an UnsafeRow for nextRow using grouping function.
|
Note
|
|
If no row is available, fetchNextRow nullifies nextRow and nextGroup internal variables.
|
Note
|
fetchNextRow is used internally when doExecute‘s Iterator is created and fetchNextPartition is called.
|
createResultProjection Internal Method
|
1 2 3 4 5 |
createResultProjection(expressions: Seq[Expression]): UnsafeProjection |
createResultProjection creates a UnsafeProjection function for expressions window function Catalyst expressions so that the window expressions are on the right side of child’s output.
|
Note
|
UnsafeProjection is a Scala function that produces UnsafeRow for an InternalRow. |
Internally, createResultProjection first creates a translation table with a BoundReference per expression (in the input expressions).
|
Note
|
BoundReference is a Catalyst expression that is a reference to a value in internal binary row at a specified position and of specified data type.
|
createResultProjection then creates a window function bound references for window expressions so unbound expressions are transformed to the BoundReferences.
In the end, createResultProjection creates a UnsafeProjection with:
|
Note
|
createResultProjection is used exclusively when WindowExec is executed.
|
Creating WindowExec Instance
WindowExec takes the following when created:
-
Window named expressions
-
Window partition specification expressions
-
Window order specification (as a collection of
SortOrderexpressions) -
Child physical operator
Lookup Table for WindowExpressions and Factory Functions for WindowFunctionFrame — windowFrameExpressionFactoryPairs Lazy Value
|
1 2 3 4 5 6 |
windowFrameExpressionFactoryPairs: Seq[(mutable.Buffer[WindowExpression], InternalRow => WindowFunctionFrame)] |
windowFrameExpressionFactoryPairs is a lookup table with window expressions and factory functions for WindowFunctionFrame (per key-value pair in framedFunctions lookup table).
A factory function is a function that takes an InternalRow and produces a WindowFunctionFrame (described in the table below)
Internally, windowFrameExpressionFactoryPairs first builds framedFunctions lookup table with 4-element tuple keys and 2-element expression list values (described in the table below).
windowFrameExpressionFactoryPairs finds WindowExpression expressions in the input windowExpression and for every WindowExpression takes the window frame specification (of type SpecifiedWindowFrame that is used to find frame type and start and end frame positions).
| Element | Description |
|---|---|
|
Name of the kind of function |
|
|
|
|
|
Window frame’s start position |
|
|
Window frame’s end position |
|
| Element | Description |
|---|---|
|
Collection of window expressions |
|
|
Collection of window functions |
|
windowFrameExpressionFactoryPairs creates a AggregateProcessor for AGGREGATE frame keys in framedFunctions lookup table.
| Frame Name | FrameKey | WindowFunctionFrame |
|---|---|---|
|
Offset Frame |
|
|
|
Growing Frame |
|
|
|
Shrinking Frame |
|
|
|
Moving Frame |
|
|
|
|
|
Note
|
lazy val in Scala is computed when first accessed and once only (for the entire lifetime of the owning object instance).
|
|
Note
|
windowFrameExpressionFactoryPairs is used exclusively when WindowExec is executed.
|
createBoundOrdering Internal Method
|
1 2 3 4 5 |
createBoundOrdering(frame: FrameType, bound: Expression, timeZone: String): BoundOrdering |
createBoundOrdering…FIXME
|
Note
|
createBoundOrdering is used exclusively when WindowExec physical operator is requested for the window frame factories.
|
spark技术分享