aboutsummaryrefslogtreecommitdiff
path: root/docs/running-on-yarn.md
diff options
context:
space:
mode:
authorMatei Zaharia <matei.zaharia@gmail.com>2013-08-28 12:44:46 -0700
committerMatei Zaharia <matei.zaharia@gmail.com>2013-08-28 12:44:46 -0700
commitbaa84e7e4c5e0afc8bc3b177379311d309c00cd2 (patch)
tree76aeeb15613a583c9472eefc6e82d3b9b582dd5c /docs/running-on-yarn.md
parentcd043cf922692aa493308cf1e6da6f7522d80b78 (diff)
parentaac1214ee48ef143b0164f740380cdb0a5a7383b (diff)
downloadspark-baa84e7e4c5e0afc8bc3b177379311d309c00cd2.tar.gz
spark-baa84e7e4c5e0afc8bc3b177379311d309c00cd2.tar.bz2
spark-baa84e7e4c5e0afc8bc3b177379311d309c00cd2.zip
Merge pull request #865 from tgravescs/fixtmpdir
Spark on Yarn should use yarn approved directories for spark.local.dir and tmp
Diffstat (limited to 'docs/running-on-yarn.md')
-rw-r--r--docs/running-on-yarn.md4
1 files changed, 1 insertions, 3 deletions
diff --git a/docs/running-on-yarn.md b/docs/running-on-yarn.md
index cac9c5e4b6..1a0afd19d4 100644
--- a/docs/running-on-yarn.md
+++ b/docs/running-on-yarn.md
@@ -63,7 +63,6 @@ The command to launch the YARN Client is as follows:
--master-memory <MEMORY_FOR_MASTER> \
--worker-memory <MEMORY_PER_WORKER> \
--worker-cores <CORES_PER_WORKER> \
- --user <hadoop_user> \
--queue <queue_name>
For example:
@@ -83,5 +82,4 @@ The above starts a YARN Client programs which periodically polls the Application
- When your application instantiates a Spark context it must use a special "yarn-standalone" master url. This starts the scheduler without forcing it to connect to a cluster. A good way to handle this is to pass "yarn-standalone" as an argument to your program, as shown in the example above.
- We do not requesting container resources based on the number of cores. Thus the numbers of cores given via command line arguments cannot be guaranteed.
-- Currently, we have not yet integrated with hadoop security. If --user is present, the hadoop_user specified will be used to run the tasks on the cluster. If unspecified, current user will be used (which should be valid in cluster).
- Once hadoop security support is added, and if hadoop cluster is enabled with security, additional restrictions would apply via delegation tokens passed.
+- The local directories used for spark will be the local directories configured for YARN (Hadoop Yarn config yarn.nodemanager.local-dirs). If the user specifies spark.local.dir, it will be ignored.