aboutsummaryrefslogtreecommitdiff
path: root/README.md
diff options
context:
space:
mode:
authorReynold Xin <rxin@databricks.com>2015-05-31 23:55:45 -0700
committerReynold Xin <rxin@databricks.com>2015-05-31 23:55:45 -0700
commit3c0156899dc1ec1f7dfe6d7c8af47fa6dc7d00bf (patch)
tree445d5dfc4bd5c92a34c01335649febfbd3a75df3 /README.md
parenta0e46a0d2ad23ce6a64e6ebdf2ccc776208696b6 (diff)
downloadspark-3c0156899dc1ec1f7dfe6d7c8af47fa6dc7d00bf.tar.gz
spark-3c0156899dc1ec1f7dfe6d7c8af47fa6dc7d00bf.tar.bz2
spark-3c0156899dc1ec1f7dfe6d7c8af47fa6dc7d00bf.zip
Update README to include DataFrames and zinc.
Also cut trailing whitespaces. Author: Reynold Xin <rxin@databricks.com> Closes #6548 from rxin/readme and squashes the following commits: 630efc3 [Reynold Xin] Update README to include DataFrames and zinc.
Diffstat (limited to 'README.md')
-rw-r--r--README.md16
1 files changed, 8 insertions, 8 deletions
diff --git a/README.md b/README.md
index 9c09d40e2b..380422ca00 100644
--- a/README.md
+++ b/README.md
@@ -3,8 +3,8 @@
Spark is a fast and general cluster computing system for Big Data. It provides
high-level APIs in Scala, Java, and Python, and an optimized engine that
supports general computation graphs for data analysis. It also supports a
-rich set of higher-level tools including Spark SQL for SQL and structured
-data processing, MLlib for machine learning, GraphX for graph processing,
+rich set of higher-level tools including Spark SQL for SQL and DataFrames,
+MLlib for machine learning, GraphX for graph processing,
and Spark Streaming for stream processing.
<http://spark.apache.org/>
@@ -22,7 +22,7 @@ This README file only contains basic setup instructions.
Spark is built using [Apache Maven](http://maven.apache.org/).
To build Spark and its example programs, run:
- mvn -DskipTests clean package
+ build/mvn -DskipTests clean package
(You do not need to do this if you downloaded a pre-built package.)
More detailed documentation is available from the project site, at
@@ -43,7 +43,7 @@ Try the following command, which should return 1000:
Alternatively, if you prefer Python, you can use the Python shell:
./bin/pyspark
-
+
And run the following command, which should also return 1000:
>>> sc.parallelize(range(1000)).count()
@@ -58,9 +58,9 @@ To run one of them, use `./bin/run-example <class> [params]`. For example:
will run the Pi example locally.
You can set the MASTER environment variable when running examples to submit
-examples to a cluster. This can be a mesos:// or spark:// URL,
-"yarn-cluster" or "yarn-client" to run on YARN, and "local" to run
-locally with one thread, or "local[N]" to run locally with N threads. You
+examples to a cluster. This can be a mesos:// or spark:// URL,
+"yarn-cluster" or "yarn-client" to run on YARN, and "local" to run
+locally with one thread, or "local[N]" to run locally with N threads. You
can also use an abbreviated class name if the class is in the `examples`
package. For instance:
@@ -75,7 +75,7 @@ can be run using:
./dev/run-tests
-Please see the guidance on how to
+Please see the guidance on how to
[run tests for a module, or individual tests](https://cwiki.apache.org/confluence/display/SPARK/Useful+Developer+Tools).
## A Note About Hadoop Versions