summaryrefslogtreecommitdiff
path: root/site/releases/spark-release-0-8-0.html
diff options
context:
space:
mode:
authorSean Owen <sowen@cloudera.com>2016-08-31 12:38:33 +0100
committerSean Owen <sowen@cloudera.com>2016-08-31 12:38:33 +0100
commit0845f49def1dd1fc5fb439d2f1e22f03297944ed (patch)
tree587fddee53464850254e2b65a4b2aa0d7b2a2443 /site/releases/spark-release-0-8-0.html
parentfcd0bc3dd9812c263a17bd4df4e85ce3a89d2b8b (diff)
downloadspark-website-0845f49def1dd1fc5fb439d2f1e22f03297944ed.tar.gz
spark-website-0845f49def1dd1fc5fb439d2f1e22f03297944ed.tar.bz2
spark-website-0845f49def1dd1fc5fb439d2f1e22f03297944ed.zip
Re-sync Spark site HTML to output of latest jekyll
Diffstat (limited to 'site/releases/spark-release-0-8-0.html')
-rw-r--r--site/releases/spark-release-0-8-0.html4
1 files changed, 2 insertions, 2 deletions
diff --git a/site/releases/spark-release-0-8-0.html b/site/releases/spark-release-0-8-0.html
index 9815cfea9..78fcbcf4c 100644
--- a/site/releases/spark-release-0-8-0.html
+++ b/site/releases/spark-release-0-8-0.html
@@ -210,13 +210,13 @@
<p>Spark’s internal job scheduler has been refactored and extended to include more sophisticated scheduling policies. In particular, a <a href="http://spark.incubator.apache.org/docs/0.8.0/job-scheduling.html#scheduling-within-an-application">fair scheduler</a> implementation now allows multiple users to share an instance of Spark, which helps users running shorter jobs to achieve good performance, even when longer-running jobs are running in parallel. Support for topology-aware scheduling has been extended, including the ability to take into account rack locality and support for multiple executors on a single machine.</p>
<h3 id="easier-deployment-and-linking">Easier Deployment and Linking</h3>
-<p>User programs can now link to Spark no matter which Hadoop version they need, without having to publish a version of <code>spark-core</code> specifically for that Hadoop version. An explanation of how to link against different Hadoop versions is provided <a href="http://spark.incubator.apache.org/docs/0.8.0/scala-programming-guide.html#linking-with-spark">here</a>. </p>
+<p>User programs can now link to Spark no matter which Hadoop version they need, without having to publish a version of <code>spark-core</code> specifically for that Hadoop version. An explanation of how to link against different Hadoop versions is provided <a href="http://spark.incubator.apache.org/docs/0.8.0/scala-programming-guide.html#linking-with-spark">here</a>.</p>
<h3 id="expanded-ec2-capabilities">Expanded EC2 Capabilities</h3>
<p>Spark’s EC2 scripts now support launching in any availability zone. Support has also been added for EC2 instance types which use the newer “HVM” architecture. This includes the cluster compute (cc1/cc2) family of instance types. We’ve also added support for running newer versions of HDFS alongside Spark. Finally, we’ve added the ability to launch clusters with maintenance releases of Spark in addition to launching the newest release.</p>
<h3 id="improved-documentation">Improved Documentation</h3>
-<p>This release adds documentation about cluster hardware provisioning and inter-operation with common Hadoop distributions. Docs are also included to cover the MLlib machine learning functions and new cluster monitoring features. Existing documentation has been updated to reflect changes in building and deploying Spark. </p>
+<p>This release adds documentation about cluster hardware provisioning and inter-operation with common Hadoop distributions. Docs are also included to cover the MLlib machine learning functions and new cluster monitoring features. Existing documentation has been updated to reflect changes in building and deploying Spark.</p>
<h3 id="other-improvements">Other Improvements</h3>
<ul>