-
Notifications
You must be signed in to change notification settings - Fork 29k
[SPARK-5931][CORE] Use consistent naming for time properties #5236
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from 1 commit
59bf9e1
404f8c3
7db6d2a
4933fda
c9f5cad
21ef3dd
064ebd6
7320c87
272c215
3352d34
3f1cfc8
6d1518e
2fcc91c
5181597
c6a0095
cde9bff
42477aa
9a29d8d
34f87c2
8f741e1
9e2547c
499bdf0
5232a36
3a12dd8
68f4e93
70ac213
647b5ac
1c0c07c
8613631
bac9edf
1858197
3b126e1
39164f9
b2fc965
dd0a680
bf779b0
76cfa27
5193d5f
6387772
19c31af
ff40bfe
28187bf
1465390
cbf41db
d4efd26
4e48679
1a1122c
cbd2ca6
6f651a8
7d19cdd
dc7bd08
69fedcc
8927e66
642a06d
25d3f52
bc04e05
951ca2d
f5fafcd
de3bff9
4526c81
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
- Loading branch information
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -261,7 +261,7 @@ class ConnectionManagerSuite extends FunSuite { | |
| val clientConf = new SparkConf | ||
| clientConf.set("spark.authenticate", "false") | ||
| val ackTimeout = 30 | ||
|
Member
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. You could consider renaming this |
||
| clientConf.set("spark.core.connection.ack.wait.timeout", s"${ackTimeout}") | ||
| clientConf.set("spark.core.connection.ack.wait.timeout", s"${ackTimeout}s") | ||
|
|
||
| val clientSecurityManager = new SecurityManager(clientConf) | ||
| val manager = new ConnectionManager(0, clientConf, clientSecurityManager) | ||
|
|
||
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -120,7 +120,7 @@ of the most common options to set are: | |
| <td> | ||
| Amount of memory to use for the driver process, i.e. where SparkContext is initialized. | ||
| (e.g. <code>512m</code>, <code>2g</code>). | ||
|
|
||
| <br /><em>Note:</em> In client mode, this config must not be set through the <code>SparkConf</code> | ||
| directly in your application, because the driver JVM has already started at that point. | ||
| Instead, please set this through the <code>--driver-memory</code> command line option | ||
|
|
@@ -188,7 +188,7 @@ Apart from these, the following properties are also available, and may be useful | |
|
|
||
| <br /><em>Note:</em> In client mode, this config must not be set through the <code>SparkConf</code> | ||
| directly in your application, because the driver JVM has already started at that point. | ||
| Instead, please set this through the <code>--driver-class-path</code> command line option or in | ||
| Instead, please set this through the <code>--driver-class-path</code> command line option or in | ||
| your default properties file.</td> | ||
| </td> | ||
| </tr> | ||
|
|
@@ -197,10 +197,10 @@ Apart from these, the following properties are also available, and may be useful | |
| <td>(none)</td> | ||
| <td> | ||
| A string of extra JVM options to pass to the driver. For instance, GC settings or other logging. | ||
|
|
||
| <br /><em>Note:</em> In client mode, this config must not be set through the <code>SparkConf</code> | ||
| directly in your application, because the driver JVM has already started at that point. | ||
| Instead, please set this through the <code>--driver-java-options</code> command line option or in | ||
| Instead, please set this through the <code>--driver-java-options</code> command line option or in | ||
| your default properties file.</td> | ||
| </td> | ||
| </tr> | ||
|
|
@@ -209,10 +209,10 @@ Apart from these, the following properties are also available, and may be useful | |
| <td>(none)</td> | ||
| <td> | ||
| Set a special library path to use when launching the driver JVM. | ||
|
|
||
|
Member
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Not a big deal but I see a fair bit of white-space-only change in the diff. It makes it a little harder to see what's changed. I don't know how much it's possible or worth fixing by hand but might check your IDE settings and not strip trailing spaces.
Author
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. @srowen I gave this some thought. I think we can both support the new interface and maintain compatibility. Put simply, throughout the code I use This will allow us to introduce the new interface, update all usage of constants throughout the code to clarify units, and to support old usage. It will also cut down on the number of code changes since documentation can be left as is.
Author
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. Sean - to follow up on the white-space comment. The reason that there's white space changes is that there's a lot of trailing white space in these files which my IDE is stripping. I could disable white-space stripping but in this case, I think that this is a fix worth making since it shouldn't be there in the first place.
Member
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I would turn it off. The whitespace itself doesn't matter at all functionally in most contexts. But it can create more noise in the diff. More important it can create spurious merge conflicts. An instance here or there doesn't matter much but I actively try to not make whitespace changes for this reason. |
||
| <br /><em>Note:</em> In client mode, this config must not be set through the <code>SparkConf</code> | ||
| directly in your application, because the driver JVM has already started at that point. | ||
| Instead, please set this through the <code>--driver-library-path</code> command line option or in | ||
| Instead, please set this through the <code>--driver-library-path</code> command line option or in | ||
| your default properties file.</td> | ||
| </td> | ||
| </tr> | ||
|
|
@@ -223,26 +223,26 @@ Apart from these, the following properties are also available, and may be useful | |
| (Experimental) Whether to give user-added jars precedence over Spark's own jars when loading | ||
| classes in the the driver. This feature can be used to mitigate conflicts between Spark's | ||
| dependencies and user dependencies. It is currently an experimental feature. | ||
|
|
||
| This is used in cluster mode only. | ||
| </td> | ||
| </tr> | ||
| <tr> | ||
| <td><code>spark.executor.extraClassPath</code></td> | ||
| <td>(none)</td> | ||
| <td> | ||
| Extra classpath entries to append to the classpath of executors. This exists primarily for | ||
| backwards-compatibility with older versions of Spark. Users typically should not need to set | ||
| Extra classpath entries to append to the classpath of executors. This exists primarily for | ||
| backwards-compatibility with older versions of Spark. Users typically should not need to set | ||
| this option. | ||
| </td> | ||
| </tr> | ||
| <tr> | ||
| <td><code>spark.executor.extraJavaOptions</code></td> | ||
| <td>(none)</td> | ||
| <td> | ||
| A string of extra JVM options to pass to executors. For instance, GC settings or other logging. | ||
| Note that it is illegal to set Spark properties or heap size settings with this option. Spark | ||
| properties should be set using a SparkConf object or the spark-defaults.conf file used with the | ||
| A string of extra JVM options to pass to executors. For instance, GC settings or other logging. | ||
| Note that it is illegal to set Spark properties or heap size settings with this option. Spark | ||
| properties should be set using a SparkConf object or the spark-defaults.conf file used with the | ||
| spark-submit script. Heap size settings can be set with spark.executor.memory. | ||
| </td> | ||
| </tr> | ||
|
|
@@ -732,17 +732,17 @@ Apart from these, the following properties are also available, and may be useful | |
| </tr> | ||
| <tr> | ||
| <td><code>spark.executor.heartbeatInterval</code></td> | ||
| <td>10000</td> | ||
| <td>Interval (milliseconds) between each executor's heartbeats to the driver. Heartbeats let | ||
| <td>10s</td> | ||
| <td>Interval between each executor's heartbeats to the driver. Heartbeats let | ||
| the driver know that the executor is still alive and update it with metrics for in-progress | ||
| tasks.</td> | ||
| </tr> | ||
| <tr> | ||
| <td><code>spark.files.fetchTimeout</code></td> | ||
| <td>60</td> | ||
| <td>60s</td> | ||
| <td> | ||
| Communication timeout to use when fetching files added through SparkContext.addFile() from | ||
| the driver, in seconds. | ||
| the driver. | ||
| </td> | ||
| </tr> | ||
| <tr> | ||
|
|
@@ -855,14 +855,14 @@ Apart from these, the following properties are also available, and may be useful | |
| <td><code>spark.akka.heartbeat.interval</code></td> | ||
| <td>1000</td> | ||
| <td> | ||
| This is set to a larger value to disable the transport failure detector that comes built in to | ||
| Akka. It can be enabled again, if you plan to use this feature (Not recommended). A larger | ||
| interval value in seconds reduces network overhead and a smaller value ( ~ 1 s) might be more | ||
| informative for Akka's failure detector. Tune this in combination of `spark.akka.heartbeat.pauses` | ||
| if you need to. A likely positive use case for using failure detector would be: a sensistive | ||
| failure detector can help evict rogue executors quickly. However this is usually not the case | ||
| as GC pauses and network lags are expected in a real Spark cluster. Apart from that enabling | ||
| this leads to a lot of exchanges of heart beats between nodes leading to flooding the network | ||
| This is set to a larger value to disable the transport failure detector that comes built in to | ||
| Akka. It can be enabled again, if you plan to use this feature (Not recommended). A larger | ||
| interval value in seconds reduces network overhead and a smaller value ( ~ 1 s) might be more | ||
| informative for Akka's failure detector. Tune this in combination of `spark.akka.heartbeat.pauses` | ||
| if you need to. A likely positive use case for using failure detector would be: a sensistive | ||
| failure detector can help evict rogue executors quickly. However this is usually not the case | ||
| as GC pauses and network lags are expected in a real Spark cluster. Apart from that enabling | ||
| this leads to a lot of exchanges of heart beats between nodes leading to flooding the network | ||
| with those. | ||
| </td> | ||
| </tr> | ||
|
|
@@ -871,7 +871,7 @@ Apart from these, the following properties are also available, and may be useful | |
| <td>6000</td> | ||
| <td> | ||
| This is set to a larger value to disable the transport failure detector that comes built in to Akka. | ||
| It can be enabled again, if you plan to use this feature (Not recommended). Acceptable heart | ||
| It can be enabled again, if you plan to use this feature (Not recommended). Acceptable heart | ||
| beat pause in seconds for Akka. This can be used to control sensitivity to GC pauses. Tune | ||
| this along with `spark.akka.heartbeat.interval` if you need to. | ||
| </td> | ||
|
|
@@ -938,9 +938,9 @@ Apart from these, the following properties are also available, and may be useful | |
| </tr> | ||
| <tr> | ||
| <td><code>spark.network.timeout</code></td> | ||
| <td>120</td> | ||
| <td>120s</td> | ||
| <td> | ||
| Default timeout for all network interactions, in seconds. This config will be used in | ||
| Default timeout for all network interactions. This config will be used in | ||
| place of <code>spark.core.connection.ack.wait.timeout</code>, <code>spark.akka.timeout</code>, | ||
| <code>spark.storage.blockManagerSlaveTimeoutMs</code> or | ||
| <code>spark.shuffle.io.connectionTimeout</code>, if they are not configured. | ||
|
|
@@ -1215,9 +1215,9 @@ Apart from these, the following properties are also available, and may be useful | |
| </tr> | ||
| <tr> | ||
| <td><code>spark.core.connection.ack.wait.timeout</code></td> | ||
| <td>60</td> | ||
| <td>60s</td> | ||
| <td> | ||
| Number of seconds for the connection to wait for ack to occur before timing | ||
| How long for the connection to wait for ack to occur before timing | ||
| out and giving up. To avoid unwilling timeout caused by long pause like GC, | ||
| you can set larger value. | ||
| </td> | ||
|
|
||
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I don't think the "interval" here was referring to the variable, so could you please revert this :)