-
Notifications
You must be signed in to change notification settings - Fork 29k
SPARK-1126. spark-app preliminary #86
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Closed
Closed
Changes from 1 commit
Commits
Show all changes
6 commits
Select commit
Hold shift + click to select a range
04bc4e2
SPARK-1126. spark-submit script
sryza a94c627
Add newline at end of SparkSubmit
sryza 299ddca
Fix scalastyle
sryza 34de899
Change --more-jars to --jars and fix docs
sryza e7315c6
Fix failing tests
sryza d428d85
Commenting, doc, and import fixes from Patrick's comments
sryza File filter
Filter by extension
Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev
Previous commit
Commenting, doc, and import fixes from Patrick's comments
- Loading branch information
commit d428d857971866aa5c8f75970a9655071bf016fd
There are no files selected for viewing
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -56,37 +56,40 @@ The recommended way to launch a compiled Spark application is through the spark- | |
| bin directory), which takes care of setting up the classpath with Spark and its dependencies, as well as | ||
| provides a layer over the different cluster managers and deploy modes that Spark supports. It's usage is | ||
|
|
||
| spark-submit <jar> <options> | ||
| spark-submit `<jar>` `<options>` | ||
|
|
||
| Where options are any of: | ||
|
|
||
| - **--class** - The main class to run. | ||
| - **--master** - The URL of the cluster manager master, e.g. spark://host:port, mesos://host:port, yarn, | ||
| - **\--class** - The main class to run. | ||
| - **\--master** - The URL of the cluster manager master, e.g. spark://host:port, mesos://host:port, yarn, | ||
| or local. | ||
| - **--deploy-mode** - "client" to run the driver in the client process or "cluster" to run the driver in | ||
| - **\--deploy-mode** - "client" to run the driver in the client process or "cluster" to run the driver in | ||
| a process on the cluster. For Mesos, only "client" is supported. | ||
| - **--executor-memory** - Memory per executor (e.g. 1000M, 2G). | ||
| - **--executor-cores** - Number of cores per executor. | ||
| - **--driver-memory** - Memory for driver (e.g. 1000M, 2G) | ||
| - **--name** - Name of the application. | ||
| - **--arg** - Argument to be passed to the application's main class. This option can be specified | ||
| - **\--executor-memory** - Memory per executor (e.g. 1000M, 2G). | ||
| - **\--executor-cores** - Number of cores per executor. (Default: 2) | ||
| - **\--driver-memory** - Memory for driver (e.g. 1000M, 2G) | ||
| - **\--name** - Name of the application. | ||
| - **\--arg** - Argument to be passed to the application's main class. This option can be specified | ||
| multiple times to pass multiple arguments. | ||
| - **--jars** - A comma-separated list of local jars to include on the driver classpath and that | ||
| - **\--jars** - A comma-separated list of local jars to include on the driver classpath and that | ||
| SparkContext.addJar will work with. Doesn't work on standalone with 'cluster' deploy mode. | ||
|
|
||
| The following currently only work for Spark standalone with cluster deploy mode: | ||
| - **--driver-cores** - Cores for driver (Default: 1). | ||
| - **--supervise** - If given, restarts the driver on failure. | ||
|
|
||
| - **\--driver-cores** - Cores for driver (Default: 1). | ||
| - **\--supervise** - If given, restarts the driver on failure. | ||
|
|
||
| The following only works for Spark standalone and Mesos only: | ||
| - **--total-executor-cores** - Total cores for all executors. | ||
|
|
||
| - **\--total-executor-cores** - Total cores for all executors. | ||
|
|
||
| The following currently only work for YARN: | ||
|
|
||
| - **--queue** - The YARN queue to place the application in. | ||
| - **--files** - Comma separated list of files to be placed next to all executors | ||
| - **--archives** - Comma separated list of archives to be extracted next to all executors | ||
| - **--num-executors** - Number of executors to start. | ||
| - **\--queue** - The YARN queue to place the application in. | ||
| - **\--files** - Comma separated list of files to be placed in the working dir of each executor. | ||
| - **\--archives** - Comma separated list of archives to be extracted into the working dir of each | ||
| executor. | ||
| - **\--num-executors** - Number of executors (Default: 2). | ||
|
|
||
| The master and deploy mode can also be set with the MASTER and DEPLOY_MODE environment variables. | ||
|
Contributor
There was a problem hiding this comment. Choose a reason for hiding this commentThe reason will be displayed to describe this comment to others. Learn more. I think "deploy mode" is a new term that this PR introduces. Would you mind adding it to the glossary below? I think it's something like: |
||
| Values for these options passed via command line will override the environment variables. | ||
|
|
@@ -143,6 +146,12 @@ The following table summarizes terms you'll see used to refer to cluster concept | |
| <td>Cluster manager</td> | ||
| <td>An external service for acquiring resources on the cluster (e.g. standalone manager, Mesos, YARN)</td> | ||
| </tr> | ||
| <tr> | ||
| <td>Deploy mode</td> | ||
| <td>Distinguishes where the driver process runs. In "cluster" mode, the framework launches | ||
| the driver inside of the cluster. In "client" mode, the submitter launches the driver | ||
| outside of the cluster.</td> | ||
| <tr> | ||
| <tr> | ||
| <td>Worker node</td> | ||
| <td>Any node that can run application code in the cluster</td> | ||
|
|
||
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Do you mind adding a high level comment here? It can be very brief (1 line) - just something to make it clear to developers what this is if someone runs in to this file.