Skip to content

Commit 17b6953

Browse files
committed
Resolve conflicts
Conflicts: sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/jdbc/JDBCRDD.scala
2 parents 9384a55 + ad5b7cf commit 17b6953

File tree

265 files changed

+655
-2500
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

265 files changed

+655
-2500
lines changed

core/src/main/scala/org/apache/spark/SparkConf.scala

Lines changed: 2 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -544,7 +544,8 @@ private[spark] object SparkConf extends Logging {
544544
DeprecatedConfig("spark.kryoserializer.buffer.mb", "1.4",
545545
"Please use spark.kryoserializer.buffer instead. The default value for " +
546546
"spark.kryoserializer.buffer.mb was previously specified as '0.064'. Fractional values " +
547-
"are no longer accepted. To specify the equivalent now, one may use '64k'.")
547+
"are no longer accepted. To specify the equivalent now, one may use '64k'."),
548+
DeprecatedConfig("spark.rpc", "2.0", "Not used any more.")
548549
)
549550

550551
Map(configs.map { cfg => (cfg.key -> cfg) } : _*)

core/src/main/scala/org/apache/spark/SparkEnv.scala

Lines changed: 3 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,6 @@ import org.apache.spark.memory.{MemoryManager, StaticMemoryManager, UnifiedMemor
3434
import org.apache.spark.network.BlockTransferService
3535
import org.apache.spark.network.netty.NettyBlockTransferService
3636
import org.apache.spark.rpc.{RpcEndpointRef, RpcEndpoint, RpcEnv}
37-
import org.apache.spark.rpc.akka.AkkaRpcEnv
3837
import org.apache.spark.scheduler.{OutputCommitCoordinator, LiveListenerBus}
3938
import org.apache.spark.scheduler.OutputCommitCoordinator.OutputCommitCoordinatorEndpoint
4039
import org.apache.spark.serializer.Serializer
@@ -97,9 +96,7 @@ class SparkEnv (
9796
blockManager.master.stop()
9897
metricsSystem.stop()
9998
outputCommitCoordinator.stop()
100-
if (!rpcEnv.isInstanceOf[AkkaRpcEnv]) {
101-
actorSystem.shutdown()
102-
}
99+
actorSystem.shutdown()
103100
rpcEnv.shutdown()
104101

105102
// Unfortunately Akka's awaitTermination doesn't actually wait for the Netty server to shut
@@ -248,14 +245,11 @@ object SparkEnv extends Logging {
248245

249246
val securityManager = new SecurityManager(conf)
250247

251-
// Create the ActorSystem for Akka and get the port it binds to.
252248
val actorSystemName = if (isDriver) driverActorSystemName else executorActorSystemName
249+
// Create the ActorSystem for Akka and get the port it binds to.
253250
val rpcEnv = RpcEnv.create(actorSystemName, hostname, port, conf, securityManager,
254251
clientMode = !isDriver)
255-
val actorSystem: ActorSystem =
256-
if (rpcEnv.isInstanceOf[AkkaRpcEnv]) {
257-
rpcEnv.asInstanceOf[AkkaRpcEnv].actorSystem
258-
} else {
252+
val actorSystem: ActorSystem = {
259253
val actorSystemPort =
260254
if (port == 0 || rpcEnv.address == null) {
261255
port

core/src/main/scala/org/apache/spark/broadcast/BroadcastFactory.scala

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -24,14 +24,12 @@ import org.apache.spark.SparkConf
2424
import org.apache.spark.annotation.DeveloperApi
2525

2626
/**
27-
* :: DeveloperApi ::
2827
* An interface for all the broadcast implementations in Spark (to allow
2928
* multiple broadcast implementations). SparkContext uses a user-specified
3029
* BroadcastFactory implementation to instantiate a particular broadcast for the
3130
* entire Spark job.
3231
*/
33-
@DeveloperApi
34-
trait BroadcastFactory {
32+
private[spark] trait BroadcastFactory {
3533

3634
def initialize(isDriver: Boolean, conf: SparkConf, securityMgr: SecurityManager): Unit
3735

core/src/main/scala/org/apache/spark/broadcast/BroadcastManager.scala

Lines changed: 3 additions & 10 deletions
Original file line numberDiff line numberDiff line change
@@ -21,8 +21,8 @@ import java.util.concurrent.atomic.AtomicLong
2121

2222
import scala.reflect.ClassTag
2323

24-
import org.apache.spark._
25-
import org.apache.spark.util.Utils
24+
import org.apache.spark.{Logging, SparkConf, SecurityManager}
25+
2626

2727
private[spark] class BroadcastManager(
2828
val isDriver: Boolean,
@@ -39,15 +39,8 @@ private[spark] class BroadcastManager(
3939
private def initialize() {
4040
synchronized {
4141
if (!initialized) {
42-
val broadcastFactoryClass =
43-
conf.get("spark.broadcast.factory", "org.apache.spark.broadcast.TorrentBroadcastFactory")
44-
45-
broadcastFactory =
46-
Utils.classForName(broadcastFactoryClass).newInstance.asInstanceOf[BroadcastFactory]
47-
48-
// Initialize appropriate BroadcastFactory and BroadcastObject
42+
broadcastFactory = new TorrentBroadcastFactory
4943
broadcastFactory.initialize(isDriver, conf, securityManager)
50-
5144
initialized = true
5245
}
5346
}

core/src/main/scala/org/apache/spark/broadcast/HttpBroadcast.scala

Lines changed: 0 additions & 269 deletions
This file was deleted.

0 commit comments

Comments
 (0)