Skip to content
Closed
Show file tree
Hide file tree
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Next Next commit
create SparkHadoopWriterConfig directly in PairRDDFunctions.
  • Loading branch information
jiangxb1987 committed Nov 19, 2016
commit 7134e55c5802bbac9cf176833591491a3d4cac11
Original file line number Diff line number Diff line change
Expand Up @@ -82,27 +82,3 @@ abstract class SparkHadoopWriterConfig[K, V: ClassTag] extends Serializable {
def checkOutputSpecs(jobContext: JobContext): Unit

}

object SparkHadoopWriterConfig {

/**
* Instantiates a SparkHadoopWriterConfig using the given configuration.
*/
def instantiate[K, V](className: String, conf: Configuration)(
implicit ctorArgTag: ClassTag[(K, V)]): SparkHadoopWriterConfig[K, V] = {
val clazz = Utils.classForName(className).asInstanceOf[Class[SparkHadoopWriterConfig[K, V]]]

// First try the one with argument (conf: SerializableConfiguration).
// If that doesn't exist, try the one with (conf: SerializableJobConf).
try {
val ctor = clazz.getDeclaredConstructor(
classOf[SerializableConfiguration], classOf[ClassTag[(K, V)]])
ctor.newInstance(new SerializableConfiguration(conf), ctorArgTag)
} catch {
case _: NoSuchMethodException =>
val ctor = clazz.getDeclaredConstructor(
classOf[SerializableJobConf], classOf[ClassTag[(K, V)]])
ctor.newInstance(new SerializableJobConf(conf.asInstanceOf[JobConf]), ctorArgTag)
}
}
}
10 changes: 3 additions & 7 deletions core/src/main/scala/org/apache/spark/rdd/PairRDDFunctions.scala
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,7 @@ import org.apache.spark.internal.io._
import org.apache.spark.internal.Logging
import org.apache.spark.partial.{BoundedDouble, PartialResult}
import org.apache.spark.serializer.Serializer
import org.apache.spark.util.Utils
import org.apache.spark.util.{SerializableConfiguration, SerializableJobConf, Utils}
import org.apache.spark.util.collection.CompactBuffer
import org.apache.spark.util.random.StratifiedSamplingUtils

Expand Down Expand Up @@ -1051,9 +1051,7 @@ class PairRDDFunctions[K, V](self: RDD[(K, V)])
* configured for a Hadoop MapReduce job.
*/
def saveAsNewAPIHadoopDataset(conf: Configuration): Unit = self.withScope {
val config = SparkHadoopWriterConfig.instantiate[K, V](
className = classOf[SparkHadoopMapReduceWriterConfig[K, V]].getName,
conf = conf)
val config = new SparkHadoopMapReduceWriterConfig[K, V](new SerializableConfiguration(conf))
SparkHadoopWriter.write(
rdd = self,
config = config)
Expand All @@ -1066,9 +1064,7 @@ class PairRDDFunctions[K, V](self: RDD[(K, V)])
* MapReduce job.
*/
def saveAsHadoopDataset(conf: JobConf): Unit = self.withScope {
val config = SparkHadoopWriterConfig.instantiate[K, V](
className = classOf[SparkHadoopMapRedWriterConfig[K, V]].getName,
conf = conf)
val config = new SparkHadoopMapRedWriterConfig[K, V](new SerializableJobConf(conf))
SparkHadoopWriter.write(
rdd = self,
config = config)
Expand Down