-
Notifications
You must be signed in to change notification settings - Fork 29k
[SPARK-26459][SQL] replace UpdateNullabilityInAttributeReferences with FixNullability #23390
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from 1 commit
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
- Loading branch information
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -197,8 +197,8 @@ class Analyzer( | |
| PullOutNondeterministic), | ||
| Batch("UDF", Once, | ||
| HandleNullInputsForUDF), | ||
| Batch("FixNullability", Once, | ||
| FixNullability), | ||
| Batch("UpdateNullability", Once, | ||
| UpdateNullability), | ||
| Batch("Subquery", Once, | ||
| UpdateOuterReferences), | ||
| Batch("Cleanup", fixedPoint, | ||
|
|
@@ -1821,40 +1821,6 @@ class Analyzer( | |
| } | ||
| } | ||
|
|
||
| /** | ||
| * Fixes nullability of Attributes in a resolved LogicalPlan by using the nullability of | ||
| * corresponding Attributes of its children output Attributes. This step is needed because | ||
| * users can use a resolved AttributeReference in the Dataset API and outer joins | ||
| * can change the nullability of an AttribtueReference. Without the fix, a nullable column's | ||
| * nullable field can be actually set as non-nullable, which cause illegal optimization | ||
| * (e.g., NULL propagation) and wrong answers. | ||
| * See SPARK-13484 and SPARK-13801 for the concrete queries of this case. | ||
| */ | ||
| object FixNullability extends Rule[LogicalPlan] { | ||
|
|
||
| def apply(plan: LogicalPlan): LogicalPlan = plan resolveOperatorsUp { | ||
| case p if !p.resolved => p // Skip unresolved nodes. | ||
| case p: LogicalPlan if p.resolved => | ||
| val childrenOutput = p.children.flatMap(c => c.output).groupBy(_.exprId).flatMap { | ||
| case (exprId, attributes) => | ||
| // If there are multiple Attributes having the same ExprId, we need to resolve | ||
| // the conflict of nullable field. We do not really expect this happen. | ||
| val nullable = attributes.exists(_.nullable) | ||
| attributes.map(attr => attr.withNullability(nullable)) | ||
| }.toSeq | ||
| // At here, we create an AttributeMap that only compare the exprId for the lookup | ||
| // operation. So, we can find the corresponding input attribute's nullability. | ||
| val attributeMap = AttributeMap[Attribute](childrenOutput.map(attr => attr -> attr)) | ||
| // For an Attribute used by the current LogicalPlan, if it is from its children, | ||
| // we fix the nullable field by using the nullability setting of the corresponding | ||
| // output Attribute from the children. | ||
| p.transformExpressions { | ||
| case attr: Attribute if attributeMap.contains(attr) => | ||
| attr.withNullability(attributeMap(attr).nullable) | ||
| } | ||
| } | ||
| } | ||
|
|
||
| /** | ||
| * Extracts [[WindowExpression]]s from the projectList of a [[Project]] operator and | ||
| * aggregateExpressions of an [[Aggregate]] operator and creates individual [[Window]] | ||
|
|
@@ -2848,3 +2814,43 @@ object UpdateOuterReferences extends Rule[LogicalPlan] { | |
| } | ||
| } | ||
| } | ||
|
|
||
| /** | ||
| * Updates nullability of Attributes in a resolved LogicalPlan by using the nullability of | ||
| * corresponding Attributes of its children output Attributes. This step is needed because | ||
| * users can use a resolved AttributeReference in the Dataset API and outer joins | ||
| * can change the nullability of an AttribtueReference. Without this rule, a nullable column's | ||
| * nullable field can be actually set as non-nullable, which cause illegal optimization | ||
| * (e.g., NULL propagation) and wrong answers. | ||
| * See SPARK-13484 and SPARK-13801 for the concrete queries of this case. | ||
| * | ||
| * This rule should be executed again at the end of optimization phase, as optimizer may change | ||
| * some expressions and their nullabilities as well. See SPARK-21351 for more details. | ||
| */ | ||
| object UpdateNullability extends Rule[LogicalPlan] { | ||
|
||
|
|
||
| def apply(plan: LogicalPlan): LogicalPlan = plan resolveOperatorsUp { | ||
|
||
| // Skip unresolved nodes. | ||
| case p if !p.resolved => p | ||
| // Skip leaf node, as it has no child and no need to update nullability. | ||
| case p: LeafNode => p | ||
|
||
| case p: LogicalPlan => | ||
| val childrenOutput = p.children.flatMap(c => c.output).groupBy(_.exprId).flatMap { | ||
| case (exprId, attributes) => | ||
| // If there are multiple Attributes having the same ExprId, we need to resolve | ||
| // the conflict of nullable field. We do not really expect this happen. | ||
| val nullable = attributes.exists(_.nullable) | ||
| attributes.map(attr => attr.withNullability(nullable)) | ||
| }.toSeq | ||
| // At here, we create an AttributeMap that only compare the exprId for the lookup | ||
| // operation. So, we can find the corresponding input attribute's nullability. | ||
| val attributeMap = AttributeMap[Attribute](childrenOutput.map(attr => attr -> attr)) | ||
| // For an Attribute used by the current LogicalPlan, if it is from its children, | ||
| // we fix the nullable field by using the nullability setting of the corresponding | ||
| // output Attribute from the children. | ||
| p.transformExpressions { | ||
| case attr: Attribute if attributeMap.contains(attr) => | ||
| attr.withNullability(attributeMap(attr).nullable) | ||
| } | ||
| } | ||
| } | ||
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -17,6 +17,7 @@ | |
|
|
||
| package org.apache.spark.sql.catalyst.optimizer | ||
|
|
||
| import org.apache.spark.sql.catalyst.analysis.UpdateNullability | ||
| import org.apache.spark.sql.catalyst.dsl.expressions._ | ||
| import org.apache.spark.sql.catalyst.dsl.plans._ | ||
| import org.apache.spark.sql.catalyst.expressions.{CreateArray, GetArrayItem} | ||
|
|
@@ -25,7 +26,7 @@ import org.apache.spark.sql.catalyst.plans.logical.{LocalRelation, LogicalPlan} | |
| import org.apache.spark.sql.catalyst.rules.RuleExecutor | ||
|
|
||
|
|
||
| class UpdateNullabilityInAttributeReferencesSuite extends PlanTest { | ||
| class UpdateNullabilityInOptimizerSuite extends PlanTest { | ||
|
|
||
| object Optimizer extends RuleExecutor[LogicalPlan] { | ||
| val batches = | ||
|
|
@@ -36,8 +37,8 @@ class UpdateNullabilityInAttributeReferencesSuite extends PlanTest { | |
| SimplifyConditionals, | ||
| SimplifyBinaryComparison, | ||
| SimplifyExtractValueOps) :: | ||
| Batch("UpdateAttributeReferences", Once, | ||
| UpdateNullabilityInAttributeReferences) :: Nil | ||
| Batch("UpdateNullability", Once, | ||
| UpdateNullability) :: Nil | ||
| } | ||
|
|
||
| test("update nullability in AttributeReference") { | ||
|
||
|
|
||
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
move this rule out of the
Analyzer, so that it can be used in other places.There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
+1