aboutsummaryrefslogtreecommitdiff
path: root/docs/python-programming-guide.md
diff options
context:
space:
mode:
authorAaron Davidson <aaron@databricks.com>2014-04-07 13:06:30 -0700
committerPatrick Wendell <pwendell@gmail.com>2014-04-07 13:06:30 -0700
commit0307db0f55b714930c7ea118d5451190ea8c1a94 (patch)
treeaef07717fd1658760a51d77d2b22445bbfe9921e /docs/python-programming-guide.md
parent2a2ca48be61ed0d72c4347e1c042a264b94db3e8 (diff)
downloadspark-0307db0f55b714930c7ea118d5451190ea8c1a94.tar.gz
spark-0307db0f55b714930c7ea118d5451190ea8c1a94.tar.bz2
spark-0307db0f55b714930c7ea118d5451190ea8c1a94.zip
SPARK-1099: Introduce local[*] mode to infer number of cores
This is the default mode for running spark-shell and pyspark, intended to allow users running spark for the first time to see the performance benefits of using multiple cores, while not breaking backwards compatibility for users who use "local" mode and expect exactly 1 core. Author: Aaron Davidson <aaron@databricks.com> Closes #182 from aarondav/110 and squashes the following commits: a88294c [Aaron Davidson] Rebased changes for new spark-shell a9f393e [Aaron Davidson] SPARK-1099: Introduce local[*] mode to infer number of cores
Diffstat (limited to 'docs/python-programming-guide.md')
-rw-r--r--docs/python-programming-guide.md7
1 files changed, 4 insertions, 3 deletions
diff --git a/docs/python-programming-guide.md b/docs/python-programming-guide.md
index c2e5327324..888631e702 100644
--- a/docs/python-programming-guide.md
+++ b/docs/python-programming-guide.md
@@ -82,15 +82,16 @@ The Python shell can be used explore data interactively and is a simple way to l
>>> help(pyspark) # Show all pyspark functions
{% endhighlight %}
-By default, the `bin/pyspark` shell creates SparkContext that runs applications locally on a single core.
-To connect to a non-local cluster, or use multiple cores, set the `MASTER` environment variable.
+By default, the `bin/pyspark` shell creates SparkContext that runs applications locally on all of
+your machine's logical cores.
+To connect to a non-local cluster, or to specify a number of cores, set the `MASTER` environment variable.
For example, to use the `bin/pyspark` shell with a [standalone Spark cluster](spark-standalone.html):
{% highlight bash %}
$ MASTER=spark://IP:PORT ./bin/pyspark
{% endhighlight %}
-Or, to use four cores on the local machine:
+Or, to use exactly four cores on the local machine:
{% highlight bash %}
$ MASTER=local[4] ./bin/pyspark