Skip to content
Closed
Show file tree
Hide file tree
Changes from 1 commit
Commits
Show all changes
23 commits
Select commit Hold shift + click to select a range
51ca7bd
Improve building with maven docs
Mar 6, 2014
cda381f
SPARK-1184: Update the distribution tar.gz to include spark-assembly jar
markgrover Mar 6, 2014
3eb009f
SPARK-1156: allow user to login into a cluster without slaves
CodingCat Mar 6, 2014
3d3acef
SPARK-1187, Added missing Python APIs
Mar 6, 2014
40566e1
SPARK-942: Do not materialize partitions when DISK_ONLY storage level…
kellrott Mar 6, 2014
7edbea4
SPARK-1189: Add Security to Spark - Akka, Http, ConnectionManager, UI…
tgravescs Mar 7, 2014
328c73d
SPARK-1197. Change yarn-standalone to yarn-cluster and fix up running…
sryza Mar 7, 2014
9ae919c
Example for cassandra CQL read/write from spark
anitatailor Mar 7, 2014
33baf14
Small clean-up to flatmap tests
pwendell Mar 7, 2014
dabeb6f
SPARK-1136: Fix FaultToleranceTest for Docker 0.8.1
aarondav Mar 7, 2014
b7cd9e9
SPARK-1195: set map_input_file environment variable in PipedRDD
tgravescs Mar 7, 2014
6e730ed
Spark 1165 rdd.intersection in python and java
ScrapCodes Mar 8, 2014
a99fb37
SPARK-1193. Fix indentation in pom.xmls
sryza Mar 8, 2014
8ad486a
Allow sbt to use more than 1G of heap.
rxin Mar 8, 2014
0b7b7fd
[SPARK-1194] Fix the same-RDD rule for cache replacement
liancheng Mar 8, 2014
c2834ec
Update junitxml plugin to the latest version to avoid recompilation i…
rxin Mar 8, 2014
e59a3b6
SPARK-1190: Do not initialize log4j if slf4j log4j backend is not bei…
pwendell Mar 9, 2014
52834d7
SPARK-929: Fully deprecate usage of SPARK_MEM
aarondav Mar 9, 2014
f6f9d02
Add timeout for fetch file
guojc Mar 9, 2014
faf4cad
Fix markup errors introduced in #33 (SPARK-1189)
pwendell Mar 9, 2014
b9be160
SPARK-782 Clean up for ASM dependency.
pwendell Mar 9, 2014
5d98cfc
maintain arbitrary state data for each key
CrazyJvm Mar 10, 2014
32ad348
[SPARK-1186] : Enrich the Spark Shell to support additional arguments.
berngp Mar 10, 2014
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Next Next commit
Add timeout for fetch file
    Currently, when fetch a file, the connection's connect timeout
    and read timeout is based on the default jvm setting, in this change, I change it to
    use spark.worker.timeout. This can be usefull, when the
    connection status between worker is not perfect. And prevent
    prematurely remove task set.

Author: Jiacheng Guo <[email protected]>

Closes #98 from guojc/master and squashes the following commits:

abfe698 [Jiacheng Guo] add space according request
2a37c34 [Jiacheng Guo] Add timeout for fetch file     Currently, when fetch a file, the connection's connect timeout     and read timeout is based on the default jvm setting, in this change, I change it to     use spark.worker.timeout. This can be usefull, when the     connection status between worker is not perfect. And prevent     prematurely remove task set.
  • Loading branch information
guojc authored and pwendell committed Mar 9, 2014
commit f6f9d02e85d17da2f742ed0062f1648a9293e73c
4 changes: 4 additions & 0 deletions core/src/main/scala/org/apache/spark/util/Utils.scala
Original file line number Diff line number Diff line change
Expand Up @@ -278,6 +278,10 @@ private[spark] object Utils extends Logging {
uc = new URL(url).openConnection()
}

val timeout = conf.getInt("spark.files.fetchTimeout", 60) * 1000
uc.setConnectTimeout(timeout)
uc.setReadTimeout(timeout)
uc.connect()
val in = uc.getInputStream();
val out = new FileOutputStream(tempFile)
Utils.copyStream(in, out, true)
Expand Down
9 changes: 9 additions & 0 deletions docs/configuration.md
Original file line number Diff line number Diff line change
Expand Up @@ -523,6 +523,15 @@ Apart from these, the following properties are also available, and may be useful
<td>
Whether to overwrite files added through SparkContext.addFile() when the target file exists and its contents do not match those of the source.
</td>
</tr>
<tr>
<td>spark.files.fetchTimeout</td>
<td>false</td>
<td>
Communication timeout to use when fetching files added through SparkContext.addFile() from
the driver.
</td>
</tr>
<tr>
<td>spark.authenticate</td>
<td>false</td>
Expand Down