diff options
author | Patrick Wendell <pwendell@gmail.com> | 2013-09-07 14:37:54 -0700 |
---|---|---|
committer | Patrick Wendell <pwendell@gmail.com> | 2013-09-07 14:38:54 -0700 |
commit | 22b982d2bc76197c85ed50558a0fc676dffcc5ef (patch) | |
tree | a249f7fefe78a37f22958328034df813b29fff88 | |
parent | 61c4762d454c7b62bdccd33a5aab5c1f7e8809c5 (diff) | |
download | spark-22b982d2bc76197c85ed50558a0fc676dffcc5ef.tar.gz spark-22b982d2bc76197c85ed50558a0fc676dffcc5ef.tar.bz2 spark-22b982d2bc76197c85ed50558a0fc676dffcc5ef.zip |
File rename
-rwxr-xr-x | docs/_layouts/global.html | 2 | ||||
-rw-r--r-- | docs/hadoop-third-party-distributions.md (renamed from docs/cdh-hdp.md) | 4 |
2 files changed, 2 insertions, 4 deletions
diff --git a/docs/_layouts/global.html b/docs/_layouts/global.html index 3a3b8dce37..2f6bdcabe8 100755 --- a/docs/_layouts/global.html +++ b/docs/_layouts/global.html @@ -98,7 +98,7 @@ <ul class="dropdown-menu"> <li><a href="configuration.html">Configuration</a></li> <li><a href="tuning.html">Tuning Guide</a></li> - <li><a href="cdh-hdp.html">Running with CDH/HDP</a></li> + <li><a href="hadoop-third-party-distributions.html">Running with CDH/HDP</a></li> <li><a href="hardware-provisioning.html">Hardware Provisioning</a></li> <li><a href="building-with-maven.html">Building Spark with Maven</a></li> <li><a href="contributing-to-spark.html">Contributing to Spark</a></li> diff --git a/docs/cdh-hdp.md b/docs/hadoop-third-party-distributions.md index 679fb1100d..9f4f354525 100644 --- a/docs/cdh-hdp.md +++ b/docs/hadoop-third-party-distributions.md @@ -54,9 +54,7 @@ Spark can run in a variety of deployment modes: cores dedicated to Spark on each node. * Run Spark alongside Hadoop using a cluster resource manager, such as YARN or Mesos. -These options are identical for those using CDH and HDP. Note that if you have a YARN cluster, -but still prefer to run Spark on a dedicated set of nodes rather than scheduling through YARN, -use `mr1` versions of HADOOP_HOME when compiling. +These options are identical for those using CDH and HDP. # Inheriting Cluster Configuration If you plan to read and write from HDFS using Spark, there are two Hadoop configuration files that |