aboutsummaryrefslogtreecommitdiff
path: root/docs/index.md
diff options
context:
space:
mode:
authorAndrew Or <andrewor14@gmail.com>2014-05-16 22:36:23 -0700
committerPatrick Wendell <pwendell@gmail.com>2014-05-16 22:36:23 -0700
commitcf6cbe9f76c3b322a968c836d039fc5b70d4ce43 (patch)
tree7f1269166db1364d6f9393bd65d830a9948ce884 /docs/index.md
parent4b8ec6fcfd7a7ef0857d5b21917183c181301c95 (diff)
downloadspark-cf6cbe9f76c3b322a968c836d039fc5b70d4ce43.tar.gz
spark-cf6cbe9f76c3b322a968c836d039fc5b70d4ce43.tar.bz2
spark-cf6cbe9f76c3b322a968c836d039fc5b70d4ce43.zip
[SPARK-1824] Remove <master> from Python examples
A recent PR (#552) fixed this for all Scala / Java examples. We need to do it for python too. Note that this blocks on #799, which makes `bin/pyspark` go through Spark submit. With only the changes in this PR, the only way to run these examples is through Spark submit. Once #799 goes in, you can use `bin/pyspark` to run them too. For example, ``` bin/pyspark examples/src/main/python/pi.py 100 --master local-cluster[4,1,512] ``` Author: Andrew Or <andrewor14@gmail.com> Closes #802 from andrewor14/python-examples and squashes the following commits: cf50b9f [Andrew Or] De-indent python comments (minor) 50f80b1 [Andrew Or] Remove pyFiles from SparkContext construction c362f69 [Andrew Or] Update docs to use spark-submit for python applications 7072c6a [Andrew Or] Merge branch 'master' of github.com:apache/spark into python-examples 427a5f0 [Andrew Or] Update docs d32072c [Andrew Or] Remove <master> from examples + update usages
Diffstat (limited to 'docs/index.md')
-rw-r--r--docs/index.md11
1 files changed, 7 insertions, 4 deletions
diff --git a/docs/index.md b/docs/index.md
index 48182a27d2..c9b10376cc 100644
--- a/docs/index.md
+++ b/docs/index.md
@@ -43,12 +43,15 @@ The `--master` option specifies the
locally with one thread, or `local[N]` to run locally with N threads. You should start by using
`local` for testing. For a full list of options, run Spark shell with the `--help` option.
-Spark also provides a Python interface. To run an example Spark application written in Python, use
-`bin/pyspark <program> [params]`. For example,
+Spark also provides a Python interface. To run Spark interactively in a Python interpreter, use
+`bin/pyspark`. As in Spark shell, you can also pass in the `--master` option to configure your
+master URL.
- ./bin/pyspark examples/src/main/python/pi.py local[2] 10
+ ./bin/pyspark --master local[2]
-or simply `bin/pyspark` without any arguments to run Spark interactively in a python interpreter.
+Example applications are also provided in Python. For example,
+
+ ./bin/spark-submit examples/src/main/python/pi.py 10
# Launching on a Cluster