aboutsummaryrefslogtreecommitdiff
path: root/docs/quick-start.md
diff options
context:
space:
mode:
authorPrashant Sharma <prashant.s@imaginea.com>2014-01-02 18:37:40 +0530
committerPrashant Sharma <prashant.s@imaginea.com>2014-01-02 18:37:40 +0530
commitb810a85cdddb247e1a104f4daad905b97222ad85 (patch)
tree89744bc45d5e895aae5d32fd76e8ba35232f1139 /docs/quick-start.md
parent980afd280a331103ce7391adaf484dd497218741 (diff)
downloadspark-b810a85cdddb247e1a104f4daad905b97222ad85.tar.gz
spark-b810a85cdddb247e1a104f4daad905b97222ad85.tar.bz2
spark-b810a85cdddb247e1a104f4daad905b97222ad85.zip
spark-shell -> bin/spark-shell
Diffstat (limited to 'docs/quick-start.md')
-rw-r--r--docs/quick-start.md4
1 files changed, 2 insertions, 2 deletions
diff --git a/docs/quick-start.md b/docs/quick-start.md
index bdf127d404..912f332236 100644
--- a/docs/quick-start.md
+++ b/docs/quick-start.md
@@ -20,7 +20,7 @@ $ sbt/sbt assembly
## Basics
Spark's interactive shell provides a simple way to learn the API, as well as a powerful tool to analyze datasets interactively.
-Start the shell by running `./spark-shell` in the Spark directory.
+Start the shell by running `./bin/spark-shell` in the Spark directory.
Spark's primary abstraction is a distributed collection of items called a Resilient Distributed Dataset (RDD). RDDs can be created from Hadoop InputFormats (such as HDFS files) or by transforming other RDDs. Let's make a new RDD from the text of the README file in the Spark source directory:
@@ -99,7 +99,7 @@ scala> linesWithSpark.count()
res9: Long = 15
{% endhighlight %}
-It may seem silly to use Spark to explore and cache a 30-line text file. The interesting part is that these same functions can be used on very large data sets, even when they are striped across tens or hundreds of nodes. You can also do this interactively by connecting `spark-shell` to a cluster, as described in the [programming guide](scala-programming-guide.html#initializing-spark).
+It may seem silly to use Spark to explore and cache a 30-line text file. The interesting part is that these same functions can be used on very large data sets, even when they are striped across tens or hundreds of nodes. You can also do this interactively by connecting `bin/spark-shell` to a cluster, as described in the [programming guide](scala-programming-guide.html#initializing-spark).
# A Standalone App in Scala
Now say we wanted to write a standalone application using the Spark API. We will walk through a simple application in both Scala (with SBT), Java (with Maven), and Python. If you are using other build systems, consider using the Spark assembly JAR described in the developer guide.