diff --git a/connector/connect/server/src/main/scala/org/apache/spark/sql/connect/planner/SparkConnectPlanner.scala b/connector/connect/server/src/main/scala/org/apache/spark/sql/connect/planner/SparkConnectPlanner.scala index 10528843e407..4868f5f66410 100644 --- a/connector/connect/server/src/main/scala/org/apache/spark/sql/connect/planner/SparkConnectPlanner.scala +++ b/connector/connect/server/src/main/scala/org/apache/spark/sql/connect/planner/SparkConnectPlanner.scala @@ -31,13 +31,9 @@ import org.apache.spark.connect.proto import org.apache.spark.connect.proto.{ExecutePlanResponse, SqlCommand, StreamingQueryCommand, StreamingQueryCommandResult, StreamingQueryInstanceId, WriteStreamOperationStart, WriteStreamOperationStartResult} import org.apache.spark.connect.proto.ExecutePlanResponse.SqlCommandResult import org.apache.spark.connect.proto.Parse.ParseFormat -import org.apache.spark.connect.proto.StreamingQueryCommand -import org.apache.spark.connect.proto.StreamingQueryCommandResult -import org.apache.spark.connect.proto.StreamingQueryInstanceId import org.apache.spark.connect.proto.StreamingQueryManagerCommand import org.apache.spark.connect.proto.StreamingQueryManagerCommandResult import org.apache.spark.connect.proto.StreamingQueryManagerCommandResult.StreamingQueryInstance -import org.apache.spark.connect.proto.WriteStreamOperationStart import org.apache.spark.connect.proto.WriteStreamOperationStart.TriggerCase import org.apache.spark.ml.{functions => MLFunctions} import org.apache.spark.sql.{Column, Dataset, Encoders, ForeachWriter, RelationalGroupedDataset, SparkSession} diff --git a/pom.xml b/pom.xml index 8e91e9f9cbe9..aa4376856d01 100644 --- a/pom.xml +++ b/pom.xml @@ -2855,9 +2855,9 @@ -explaintypes -target:jvm-1.8 -Wconf:cat=deprecation:wv,any:e + -Wunused:imports -Wconf:cat=scaladoc:wv -Wconf:cat=lint-multiarg-infix:wv @@ -2891,6 +2891,12 @@ -Wconf:cat=unchecked&msg=outer reference:s -Wconf:cat=unchecked&msg=eliminated by erasure:s -Wconf:msg=^(?=.*?a value of type)(?=.*?cannot also be).+$:s + + -Wconf:cat=unused-imports&src=org\/apache\/spark\/graphx\/impl\/VertexPartitionBase.scala:s + -Wconf:cat=unused-imports&src=org\/apache\/spark\/graphx\/impl\/VertexPartitionBaseOps.scala:s -Xss128m diff --git a/project/SparkBuild.scala b/project/SparkBuild.scala index 27cdca55ddee..7c9b7d9f1701 100644 --- a/project/SparkBuild.scala +++ b/project/SparkBuild.scala @@ -256,8 +256,7 @@ object SparkBuild extends PomBuild { // see `scalac -Wconf:help` for details "-Wconf:cat=deprecation:wv,any:e", // 2.13-specific warning hits to be muted (as narrowly as possible) and addressed separately - // TODO(SPARK-33499): Enable this option when Scala 2.12 is no longer supported. - // "-Wunused:imports", + "-Wunused:imports", "-Wconf:cat=lint-multiarg-infix:wv", "-Wconf:cat=other-nullary-override:wv", "-Wconf:cat=other-match-analysis&site=org.apache.spark.sql.catalyst.catalog.SessionCatalog.lookupFunction.catalogFunction:wv", @@ -283,7 +282,11 @@ object SparkBuild extends PomBuild { // 4. `fruitless type test: a value of TypeA cannot also be a TypeB` "-Wconf:cat=unchecked&msg=outer reference:s", "-Wconf:cat=unchecked&msg=eliminated by erasure:s", - "-Wconf:msg=^(?=.*?a value of type)(?=.*?cannot also be).+$:s" + "-Wconf:msg=^(?=.*?a value of type)(?=.*?cannot also be).+$:s", + // TODO(SPARK-43850): Remove the following suppression rules and remove `import scala.language.higherKinds` + // from the corresponding files when Scala 2.12 is no longer supported. + "-Wconf:cat=unused-imports&src=org\\/apache\\/spark\\/graphx\\/impl\\/VertexPartitionBase.scala:s", + "-Wconf:cat=unused-imports&src=org\\/apache\\/spark\\/graphx\\/impl\\/VertexPartitionBaseOps.scala:s" ) } } diff --git a/sql/catalyst/src/main/scala-2.13/org/apache/spark/sql/catalyst/expressions/ExpressionSet.scala b/sql/catalyst/src/main/scala-2.13/org/apache/spark/sql/catalyst/expressions/ExpressionSet.scala index a615223ef79a..ba18b7a2b86c 100644 --- a/sql/catalyst/src/main/scala-2.13/org/apache/spark/sql/catalyst/expressions/ExpressionSet.scala +++ b/sql/catalyst/src/main/scala-2.13/org/apache/spark/sql/catalyst/expressions/ExpressionSet.scala @@ -17,7 +17,7 @@ package org.apache.spark.sql.catalyst.expressions -import scala.collection.{mutable, IterableFactory, IterableOps} +import scala.collection.mutable import scala.collection.mutable.ArrayBuffer object ExpressionSet {