Skip to content
Closed
Prev Previous commit
Next Next commit
SPARK-25004: Use SparkEnv instead of SparkSession.
  • Loading branch information
rdblue committed Aug 23, 2018
commit f11b3bbc3e39c250835b75fa9fa59e46282ab288
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,7 @@ import scala.collection.JavaConverters._

import net.razorvine.pickle.{Pickler, Unpickler}

import org.apache.spark.TaskContext
import org.apache.spark.{SparkEnv, TaskContext}
import org.apache.spark.api.python.{ChainedPythonFunctions, PythonEvalType}
import org.apache.spark.sql.catalyst.InternalRow
import org.apache.spark.sql.catalyst.expressions._
Expand Down Expand Up @@ -67,7 +67,7 @@ case class BatchEvalPythonExec(udfs: Seq[PythonUDF], output: Seq[Attribute], chi

// Output iterator for results from Python.
val outputIterator = new PythonUDFRunner(
funcs, PythonEvalType.SQL_BATCHED_UDF, argOffsets, sparkContext.conf)
funcs, PythonEvalType.SQL_BATCHED_UDF, argOffsets, SparkEnv.get.conf)
.compute(inputIterator, context.partitionId(), context)

val unpickle = new Unpickler
Expand Down