File tree Expand file tree Collapse file tree 2 files changed +3
-3
lines changed
core/src/main/scala/org/apache/spark/sql/execution
hive/src/test/scala/org/apache/spark/sql/sources Expand file tree Collapse file tree 2 files changed +3
-3
lines changed Original file line number Diff line number Diff line change @@ -219,7 +219,7 @@ private[sql] case class RowDataSourceScanExec(
219219 private [sql] override lazy val metrics =
220220 Map (" numOutputRows" -> SQLMetrics .createMetric(sparkContext, " number of output rows" ))
221221
222- val outputUnsafeRows = relation match {
222+ private val outputUnsafeRows = relation match {
223223 case r : HadoopFsRelation if r.fileFormat.isInstanceOf [ParquetSource ] =>
224224 ! SparkSession .getActiveSession.get.sessionState.conf.getConf(
225225 SQLConf .PARQUET_VECTORIZED_READER_ENABLED )
Original file line number Diff line number Diff line change @@ -862,8 +862,8 @@ abstract class HadoopFsRelationTest extends QueryTest with SQLTestUtils with Tes
862862 .load(path)
863863
864864 val Some (fileScanRDD) = df2.queryExecution.executedPlan.collectFirst {
865- case scan : DataSourceScanExec if scan.rdd .isInstanceOf [FileScanRDD ] =>
866- scan.rdd .asInstanceOf [FileScanRDD ]
865+ case scan : DataSourceScanExec if scan.inputRDDs().head .isInstanceOf [FileScanRDD ] =>
866+ scan.inputRDDs().head .asInstanceOf [FileScanRDD ]
867867 }
868868
869869 val partitions = fileScanRDD.partitions
You can’t perform that action at this time.
0 commit comments