summaryrefslogtreecommitdiff
path: root/site/releases/spark-release-0-8-0.html
diff options
context:
space:
mode:
authorSean Owen <sowen@cloudera.com>2016-07-26 22:34:05 +0100
committerSean Owen <sowen@cloudera.com>2016-07-29 12:20:53 -0700
commit7d3c47dd63fe3cb09a74505d860bcf5bc3967c6c (patch)
treefd0c128b8a76bc7c1ae77dfdd4b53d122fc2a1d0 /site/releases/spark-release-0-8-0.html
parent65ec30e24bc7f1f81b0767d11867c6611d28fc6f (diff)
downloadspark-website-7d3c47dd63fe3cb09a74505d860bcf5bc3967c6c.tar.gz
spark-website-7d3c47dd63fe3cb09a74505d860bcf5bc3967c6c.tar.bz2
spark-website-7d3c47dd63fe3cb09a74505d860bcf5bc3967c6c.zip
Sync markdown with generated HTML according to what latest jekyll outputs; fix some markdown problems along the way
Diffstat (limited to 'site/releases/spark-release-0-8-0.html')
-rw-r--r--site/releases/spark-release-0-8-0.html4
1 files changed, 2 insertions, 2 deletions
diff --git a/site/releases/spark-release-0-8-0.html b/site/releases/spark-release-0-8-0.html
index ae67c9b10..b33e71f80 100644
--- a/site/releases/spark-release-0-8-0.html
+++ b/site/releases/spark-release-0-8-0.html
@@ -210,13 +210,13 @@
<p>Spark’s internal job scheduler has been refactored and extended to include more sophisticated scheduling policies. In particular, a <a href="http://spark.incubator.apache.org/docs/0.8.0/job-scheduling.html#scheduling-within-an-application">fair scheduler</a> implementation now allows multiple users to share an instance of Spark, which helps users running shorter jobs to achieve good performance, even when longer-running jobs are running in parallel. Support for topology-aware scheduling has been extended, including the ability to take into account rack locality and support for multiple executors on a single machine.</p>
<h3 id="easier-deployment-and-linking">Easier Deployment and Linking</h3>
-<p>User programs can now link to Spark no matter which Hadoop version they need, without having to publish a version of <code>spark-core</code> specifically for that Hadoop version. An explanation of how to link against different Hadoop versions is provided <a href="http://spark.incubator.apache.org/docs/0.8.0/scala-programming-guide.html#linking-with-spark">here</a>. </p>
+<p>User programs can now link to Spark no matter which Hadoop version they need, without having to publish a version of <code>spark-core</code> specifically for that Hadoop version. An explanation of how to link against different Hadoop versions is provided <a href="http://spark.incubator.apache.org/docs/0.8.0/scala-programming-guide.html#linking-with-spark">here</a>.</p>
<h3 id="expanded-ec2-capabilities">Expanded EC2 Capabilities</h3>
<p>Spark’s EC2 scripts now support launching in any availability zone. Support has also been added for EC2 instance types which use the newer “HVM” architecture. This includes the cluster compute (cc1/cc2) family of instance types. We’ve also added support for running newer versions of HDFS alongside Spark. Finally, we’ve added the ability to launch clusters with maintenance releases of Spark in addition to launching the newest release.</p>
<h3 id="improved-documentation">Improved Documentation</h3>
-<p>This release adds documentation about cluster hardware provisioning and inter-operation with common Hadoop distributions. Docs are also included to cover the MLlib machine learning functions and new cluster monitoring features. Existing documentation has been updated to reflect changes in building and deploying Spark. </p>
+<p>This release adds documentation about cluster hardware provisioning and inter-operation with common Hadoop distributions. Docs are also included to cover the MLlib machine learning functions and new cluster monitoring features. Existing documentation has been updated to reflect changes in building and deploying Spark.</p>
<h3 id="other-improvements">Other Improvements</h3>
<ul>