Skip to content
Closed
Show file tree
Hide file tree
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Next Next commit
adjust code based on more commented rules
  • Loading branch information
Eric5553 committed Feb 12, 2020
commit 183eb4102784c17a91f30822c267524ecffa97b9
Original file line number Diff line number Diff line change
Expand Up @@ -797,7 +797,7 @@ object SubqueryExec {
/**
* A wrapper for reused [[BaseSubqueryExec]].
*/
case class ReusedSubqueryExec(override val child: BaseSubqueryExec)
case class ReusedSubqueryExec(child: BaseSubqueryExec)
extends BaseSubqueryExec with LeafExecNode {

override def name: String = child.name
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -37,9 +37,9 @@ import org.apache.spark.sql.util.CaseInsensitiveStringMap
* Rows in the output data set are appended.
*/
case class AppendDataExecV1(
override val table: SupportsWrite,
override val writeOptions: CaseInsensitiveStringMap,
override val plan: LogicalPlan) extends V1FallbackWriters {
table: SupportsWrite,
writeOptions: CaseInsensitiveStringMap,
plan: LogicalPlan) extends V1FallbackWriters {

override protected def doExecute(): RDD[InternalRow] = {
writeWithV1(newWriteBuilder().buildForV1Write())
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -119,8 +119,8 @@ case class AtomicCreateTableAsSelectExec(
catalog: StagingTableCatalog,
ident: Identifier,
partitioning: Seq[Transform],
override val plan: LogicalPlan,
override val query: SparkPlan,
plan: LogicalPlan,
query: SparkPlan,
properties: Map[String, String],
writeOptions: CaseInsensitiveStringMap,
ifNotExists: Boolean) extends AtomicTableWriteExec {
Expand Down Expand Up @@ -153,8 +153,8 @@ case class ReplaceTableAsSelectExec(
catalog: TableCatalog,
ident: Identifier,
partitioning: Seq[Transform],
override val plan: LogicalPlan,
override val query: SparkPlan,
plan: LogicalPlan,
query: SparkPlan,
properties: Map[String, String],
writeOptions: CaseInsensitiveStringMap,
orCreate: Boolean) extends V2TableWriteExec with SupportsV1Write {
Expand Down Expand Up @@ -219,8 +219,8 @@ case class AtomicReplaceTableAsSelectExec(
catalog: StagingTableCatalog,
ident: Identifier,
partitioning: Seq[Transform],
override val plan: LogicalPlan,
override val query: SparkPlan,
plan: LogicalPlan,
query: SparkPlan,
properties: Map[String, String],
writeOptions: CaseInsensitiveStringMap,
orCreate: Boolean) extends AtomicTableWriteExec {
Expand Down Expand Up @@ -253,7 +253,7 @@ case class AtomicReplaceTableAsSelectExec(
case class AppendDataExec(
table: SupportsWrite,
writeOptions: CaseInsensitiveStringMap,
override val query: SparkPlan) extends V2TableWriteExec with BatchWriteHelper {
query: SparkPlan) extends V2TableWriteExec with BatchWriteHelper {

override protected def doExecute(): RDD[InternalRow] = {
writeWithV2(newWriteBuilder().buildForBatch())
Expand All @@ -274,7 +274,7 @@ case class OverwriteByExpressionExec(
table: SupportsWrite,
deleteWhere: Array[Filter],
writeOptions: CaseInsensitiveStringMap,
override val query: SparkPlan) extends V2TableWriteExec with BatchWriteHelper {
query: SparkPlan) extends V2TableWriteExec with BatchWriteHelper {

private def isTruncate(filters: Array[Filter]): Boolean = {
filters.length == 1 && filters(0).isInstanceOf[AlwaysTrue]
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -71,7 +71,7 @@ case class CollectLimitExec(limit: Int, child: SparkPlan) extends LimitExec {
* This operator will be used when a logical `Tail` operation is the final operator in an
* logical plan, which happens when the user is collecting results back to the driver.
*/
case class CollectTailExec(override val limit: Int, child: SparkPlan) extends LimitExec {
case class CollectTailExec(limit: Int, child: SparkPlan) extends LimitExec {
override def output: Seq[Attribute] = child.output
override def outputPartitioning: Partitioning = SinglePartition
override def executeCollect(): Array[InternalRow] = child.executeTail(limit)
Expand Down Expand Up @@ -142,7 +142,7 @@ trait BaseLimitExec extends LimitExec with CodegenSupport {
/**
* Take the first `limit` elements of each child partition, but do not collect or shuffle them.
*/
case class LocalLimitExec(override val limit: Int, child: SparkPlan) extends BaseLimitExec {
case class LocalLimitExec(limit: Int, child: SparkPlan) extends BaseLimitExec {

override def outputOrdering: Seq[SortOrder] = child.outputOrdering

Expand All @@ -152,7 +152,7 @@ case class LocalLimitExec(override val limit: Int, child: SparkPlan) extends Bas
/**
* Take the first `limit` elements of the child's single output partition.
*/
case class GlobalLimitExec(override val limit: Int, child: SparkPlan) extends BaseLimitExec {
case class GlobalLimitExec(limit: Int, child: SparkPlan) extends BaseLimitExec {

override def requiredChildDistribution: List[Distribution] = AllTuples :: Nil

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -427,10 +427,10 @@ case class StateStoreSaveExec(

/** Physical operator for executing streaming Deduplicate. */
case class StreamingDeduplicateExec(
override val keyExpressions: Seq[Attribute],
keyExpressions: Seq[Attribute],
child: SparkPlan,
stateInfo: Option[StatefulOperatorStateInfo] = None,
override val eventTimeWatermark: Option[Long] = None)
eventTimeWatermark: Option[Long] = None)
extends UnaryExecNode with StateStoreWriter with WatermarkSupport {

/** Distribute by grouping attributes */
Expand Down