diff options
author | Olivier Grisel <olivier.grisel@ensta.org> | 2011-06-23 02:24:04 +0200 |
---|---|---|
committer | Olivier Grisel <olivier.grisel@ensta.org> | 2011-06-23 02:24:04 +0200 |
commit | 236bcd0d9b41987b771500b0643e1b73a87a3ca4 (patch) | |
tree | 3625a76249a15037778bf1a0664f1aa86a6c6372 /README | |
parent | 214250016a789aa1a7da9a1ccaa4efc9c5587926 (diff) | |
download | spark-236bcd0d9b41987b771500b0643e1b73a87a3ca4.tar.gz spark-236bcd0d9b41987b771500b0643e1b73a87a3ca4.tar.bz2 spark-236bcd0d9b41987b771500b0643e1b73a87a3ca4.zip |
Markdown rendering for the toplevel README.md to improve readability on github
Diffstat (limited to 'README')
-rw-r--r-- | README | 48 |
1 files changed, 0 insertions, 48 deletions
diff --git a/README b/README deleted file mode 100644 index 0738780395..0000000000 --- a/README +++ /dev/null @@ -1,48 +0,0 @@ -ONLINE DOCUMENTATION - -You can find the latest Spark documentation, including a programming guide, -on the project wiki at http://github.com/mesos/spark/wiki. This file only -contains basic setup instructions. - - - -BUILDING - -Spark requires Scala 2.8. This version has been tested with 2.8.1.final. - -The project is built using Simple Build Tool (SBT), which is packaged with it. -To build Spark and its example programs, run sbt/sbt update compile. - -To run Spark, you will need to have Scala's bin in your $PATH, or you -will need to set the SCALA_HOME environment variable to point to where -you've installed Scala. Scala must be accessible through one of these -methods on Mesos slave nodes as well as on the master. - -To run one of the examples, use ./run <class> <params>. For example, -./run spark.examples.SparkLR will run the Logistic Regression example. -Each of the example programs prints usage help if no params are given. - -All of the Spark samples take a <host> parameter that is the Mesos master -to connect to. This can be a Mesos URL, or "local" to run locally with one -thread, or "local[N]" to run locally with N threads. - - - -CONFIGURATION - -Spark can be configured through two files: conf/java-opts and conf/spark-env.sh. - -In java-opts, you can add flags to be passed to the JVM when running Spark. - -In spark-env.sh, you can set any environment variables you wish to be available -when running Spark programs, such as PATH, SCALA_HOME, etc. There are also -several Spark-specific variables you can set: -- SPARK_CLASSPATH: Extra entries to be added to the classpath, separated by ":". -- SPARK_MEM: Memory for Spark to use, in the format used by java's -Xmx option - (for example, 200m meams 200 MB, 1g means 1 GB, etc). -- SPARK_LIBRARY_PATH: Extra entries to add to java.library.path for locating - shared libraries. -- SPARK_JAVA_OPTS: Extra options to pass to JVM. - -Note that spark-env.sh must be a shell script (it must be executable and start -with a #! header to specify the shell to use). |