aboutsummaryrefslogtreecommitdiff
path: root/docs/index.md
diff options
context:
space:
mode:
authorMatei Zaharia <matei@eecs.berkeley.edu>2013-09-02 13:35:28 -0700
committerMatei Zaharia <matei@eecs.berkeley.edu>2013-09-02 13:35:28 -0700
commit2615cad30b0dff32ffc0d070a4f763265c603514 (patch)
treed7cb96dbf267df9d6a3494204f1778e3d0be662e /docs/index.md
parent12b2f1f9c9044a91018e5d5644dee33a65d9da95 (diff)
downloadspark-2615cad30b0dff32ffc0d070a4f763265c603514.tar.gz
spark-2615cad30b0dff32ffc0d070a4f763265c603514.tar.bz2
spark-2615cad30b0dff32ffc0d070a4f763265c603514.zip
Some doc improvements
- List higher-level projects that run on Spark - Tweak CSS
Diffstat (limited to 'docs/index.md')
-rw-r--r--docs/index.md12
1 files changed, 11 insertions, 1 deletions
diff --git a/docs/index.md b/docs/index.md
index c7018d8846..7d73929940 100644
--- a/docs/index.md
+++ b/docs/index.md
@@ -5,7 +5,7 @@ title: Spark Overview
Apache Spark is a fast and general-purpose cluster computing system.
It provides high-level APIs in [Scala](scala-programming-guide.html), [Java](java-programming-guide.html), and [Python](python-programming-guide.html) that make parallel jobs easy to write, and an optimized engine that supports general computation graphs.
-Spark can run on the Apache Mesos cluster manager, Hadoop YARN, Amazon EC2, or without an independent resource manager ("standalone mode").
+It also supports a rich set of higher-level tools including [Shark](http://shark.cs.berkeley.edu) (Hive on Spark), [MLlib](mllib-guide.html) for machine learning, [Bagel](bagel-programming-guide.html) for graph processing, and [Spark Streaming](streaming-programming-guide.html).
# Downloading
@@ -37,6 +37,15 @@ or `local` to run locally with one thread, or `local[N]` to run locally with N t
Finally, Spark can be used interactively through modified versions of the Scala shell (`./spark-shell`) or
Python interpreter (`./pyspark`). These are a great way to learn Spark.
+# Running on a Cluster
+
+Spark supports several options for deployment:
+
+* [Amazon EC2](ec2-scripts.html): our scripts let you launch a cluster in about 5 minutes
+* [Standalone Deploy Mode](spark-standalone.html): simplest way to deploy Spark on a private cluster
+* [Apache Mesos](running-on-mesos.html)
+* [Hadoop YARN](running-on-yarn.html)
+
# A Note About Hadoop Versions
Spark uses the Hadoop-client library to talk to HDFS and other Hadoop-supported
@@ -94,6 +103,7 @@ In addition, if you wish to run Spark on [YARN](running-on-yarn.md), set
**External resources:**
* [Spark Homepage](http://spark.incubator.apache.org)
+* [Shark](http://shark.cs.berkeley.edu): Apache Hive over Spark
* [Mailing Lists](http://spark.incubator.apache.org/mailing-lists.html): ask questions about Spark here
* [AMP Camps](http://ampcamp.berkeley.edu/): a series of training camps at UC Berkeley that featured talks and
exercises about Spark, Shark, Mesos, and more. [Videos](http://ampcamp.berkeley.edu/agenda-2012),