diff options
author | Patrick Wendell <pwendell@gmail.com> | 2013-01-30 14:49:18 -0800 |
---|---|---|
committer | Patrick Wendell <pwendell@gmail.com> | 2013-01-30 15:04:02 -0800 |
commit | 58a7d320d7287f3773976f0efdf2bc2c1474f7f9 (patch) | |
tree | f8c13cf2900843209a2b8a191e0adac58cc8279e /docs | |
parent | 86057ec7c868262763d1e31b3f3c94bd43eeafb3 (diff) | |
download | spark-58a7d320d7287f3773976f0efdf2bc2c1474f7f9.tar.gz spark-58a7d320d7287f3773976f0efdf2bc2c1474f7f9.tar.bz2 spark-58a7d320d7287f3773976f0efdf2bc2c1474f7f9.zip |
Inclue packaging and launching pyspark in guide.
It's nicer if all the commands you need are made explicit.
Diffstat (limited to 'docs')
-rw-r--r-- | docs/python-programming-guide.md | 10 |
1 files changed, 8 insertions, 2 deletions
diff --git a/docs/python-programming-guide.md b/docs/python-programming-guide.md index a840b9b34b..94311bd102 100644 --- a/docs/python-programming-guide.md +++ b/docs/python-programming-guide.md @@ -67,8 +67,14 @@ The script automatically adds the `pyspark` package to the `PYTHONPATH`. # Interactive Use -The `pyspark` script launches a Python interpreter that is configured to run PySpark jobs. -When run without any input files, `pyspark` launches a shell that can be used explore data interactively, which is a simple way to learn the API: +The `pyspark` script launches a Python interpreter that is configured to run PySpark jobs. To use `pyspark` interactively, first build Spark, then launch it directly from the command line without any options: + +{% highlight bash %} +$ sbt/sbt package +$ ./pyspark +{% endhighlight %} + +The Python shell can be used explore data interactively and is a simple way to learn the API: {% highlight python %} >>> words = sc.textFile("/usr/share/dict/words") |