summaryrefslogtreecommitdiff
path: root/site/releases
diff options
context:
space:
mode:
authorPatrick Wendell <pwendell@apache.org>2013-12-19 23:48:38 +0000
committerPatrick Wendell <pwendell@apache.org>2013-12-19 23:48:38 +0000
commitf64cac396b1a479fe4412d728c1b6797ded7873a (patch)
tree4f62841800848e9cfbe6b758ae9429da261ec27b /site/releases
parent5641130a8ed2a2917e64298d6832ec3b201e49a0 (diff)
downloadspark-website-f64cac396b1a479fe4412d728c1b6797ded7873a.tar.gz
spark-website-f64cac396b1a479fe4412d728c1b6797ded7873a.tar.bz2
spark-website-f64cac396b1a479fe4412d728c1b6797ded7873a.zip
Adding 0.8.1 release
Diffstat (limited to 'site/releases')
-rw-r--r--site/releases/spark-release-0-3.html2
-rw-r--r--site/releases/spark-release-0-5-0.html8
-rw-r--r--site/releases/spark-release-0-5-1.html2
-rw-r--r--site/releases/spark-release-0-6-0.html6
-rw-r--r--site/releases/spark-release-0-7-0.html4
-rw-r--r--site/releases/spark-release-0-8-0.html138
6 files changed, 80 insertions, 80 deletions
diff --git a/site/releases/spark-release-0-3.html b/site/releases/spark-release-0-3.html
index 281bd958f..6464b9838 100644
--- a/site/releases/spark-release-0-3.html
+++ b/site/releases/spark-release-0-3.html
@@ -148,7 +148,7 @@
<h3>Native Types for SequenceFiles</h3>
-<p>In working with SequenceFiles, which store objects that implement Hadoop’s Writable interface, Spark will now let you use native types for certain common Writable types, like IntWritable and Text. For example:</p>
+<p>In working with SequenceFiles, which store objects that implement Hadoop&#8217;s Writable interface, Spark will now let you use native types for certain common Writable types, like IntWritable and Text. For example:</p>
<div class="code">
<span class="comment">// Will read a SequenceFile of (IntWritable, Text)</span><br />
diff --git a/site/releases/spark-release-0-5-0.html b/site/releases/spark-release-0-5-0.html
index c48dd5d49..d4acfded1 100644
--- a/site/releases/spark-release-0-5-0.html
+++ b/site/releases/spark-release-0-5-0.html
@@ -136,10 +136,10 @@
<h3>Mesos 0.9 Support</h3>
-<p>This release runs on <a href="http://www.mesosproject.org/">Apache Mesos 0.9</a>, the first Apache Incubator release of Mesos, which contains significant usability and stability improvements. Most notable are better memory accounting for applications with long-term memory use, easier access of old jobs’ traces and logs (by keeping a history of executed tasks on the web UI), and simpler installation.</p>
+<p>This release runs on <a href="http://www.mesosproject.org/">Apache Mesos 0.9</a>, the first Apache Incubator release of Mesos, which contains significant usability and stability improvements. Most notable are better memory accounting for applications with long-term memory use, easier access of old jobs&#8217; traces and logs (by keeping a history of executed tasks on the web UI), and simpler installation.</p>
<h3>Performance Improvements</h3>
-<p>Spark’s scheduling is more communication-efficient when sending out operations on RDDs with large lineage graphs. In addition, the cache replacement policy has been improved to more smartly replace data when an RDD does not fit in the cache, shuffles are more efficient, and the serializer used for shipping closures is now configurable, making it possible to use faster libraries than Java serialization there.</p>
+<p>Spark&#8217;s scheduling is more communication-efficient when sending out operations on RDDs with large lineage graphs. In addition, the cache replacement policy has been improved to more smartly replace data when an RDD does not fit in the cache, shuffles are more efficient, and the serializer used for shipping closures is now configurable, making it possible to use faster libraries than Java serialization there.</p>
<h3>Debug Improvements</h3>
@@ -151,11 +151,11 @@
<h3>EC2 Launch Script Improvements</h3>
-<p>Spark’s EC2 launch scripts are now included in the main package, and have the ability to discover and use the latest Spark AMI automatically instead of launching a hardcoded machine image ID.</p>
+<p>Spark&#8217;s EC2 launch scripts are now included in the main package, and have the ability to discover and use the latest Spark AMI automatically instead of launching a hardcoded machine image ID.</p>
<h3>New Hadoop API Support</h3>
-<p>You can now use Spark to read and write data to storage formats in the new <tt>org.apache.mapreduce</tt> packages (the “new Hadoop” API). In addition, this release fixes an issue caused by a HDFS initialization bug in some recent versions of HDFS.</p>
+<p>You can now use Spark to read and write data to storage formats in the new <tt>org.apache.mapreduce</tt> packages (the &#8220;new Hadoop&#8221; API). In addition, this release fixes an issue caused by a HDFS initialization bug in some recent versions of HDFS.</p>
</article><!-- #post -->
diff --git a/site/releases/spark-release-0-5-1.html b/site/releases/spark-release-0-5-1.html
index c89309323..8945d59b1 100644
--- a/site/releases/spark-release-0-5-1.html
+++ b/site/releases/spark-release-0-5-1.html
@@ -165,7 +165,7 @@
<h3>EC2 Improvements</h3>
-<p>Spark’s EC2 launch script now configures Spark’s memory limit automatically based on the machine’s available RAM.</p>
+<p>Spark&#8217;s EC2 launch script now configures Spark&#8217;s memory limit automatically based on the machine&#8217;s available RAM.</p>
</article><!-- #post -->
diff --git a/site/releases/spark-release-0-6-0.html b/site/releases/spark-release-0-6-0.html
index 5dbf063f4..9e75fcdf4 100644
--- a/site/releases/spark-release-0-6-0.html
+++ b/site/releases/spark-release-0-6-0.html
@@ -144,11 +144,11 @@
<h3>Java API</h3>
-<p>Java programmers can now use Spark through a new <a href="/docs/0.6.0/java-programming-guide.html">Java API layer</a>. This layer makes available all of Spark’s features, including parallel transformations, distributed datasets, broadcast variables, and accumulators, in a Java-friendly manner.</p>
+<p>Java programmers can now use Spark through a new <a href="/docs/0.6.0/java-programming-guide.html">Java API layer</a>. This layer makes available all of Spark&#8217;s features, including parallel transformations, distributed datasets, broadcast variables, and accumulators, in a Java-friendly manner.</p>
<h3>Expanded Documentation</h3>
-<p>Spark’s <a href="/docs/0.6.0/">documentation</a> has been expanded with a new <a href="/docs/0.6.0/quick-start.html">quick start guide</a>, additional deployment instructions, configuration guide, tuning guide, and improved <a href="/docs/0.6.0/api/core">Scaladoc</a> API documentation.</p>
+<p>Spark&#8217;s <a href="/docs/0.6.0/">documentation</a> has been expanded with a new <a href="/docs/0.6.0/quick-start.html">quick start guide</a>, additional deployment instructions, configuration guide, tuning guide, and improved <a href="/docs/0.6.0/api/core">Scaladoc</a> API documentation.</p>
<h3>Engine Changes</h3>
@@ -171,7 +171,7 @@
<h3>Enhanced Debugging</h3>
-<p>Spark’s log now prints which operation in your program each RDD and job described in your logs belongs to, making it easier to tie back to which parts of your code experience problems.</p>
+<p>Spark&#8217;s log now prints which operation in your program each RDD and job described in your logs belongs to, making it easier to tie back to which parts of your code experience problems.</p>
<h3>Maven Artifacts</h3>
diff --git a/site/releases/spark-release-0-7-0.html b/site/releases/spark-release-0-7-0.html
index 4f75f7b46..8a1dda5d3 100644
--- a/site/releases/spark-release-0-7-0.html
+++ b/site/releases/spark-release-0-7-0.html
@@ -158,7 +158,7 @@
<h3>New Operations</h3>
-<p>This release adds several RDD transformations, including <tt>keys</tt>, <tt>values</tt>, <tt>keyBy</tt>, <tt>subtract</tt>, <tt>coalesce</tt>, <tt>zip</tt>. It also adds <tt>SparkContext.hadoopConfiguration</tt> to allow programs to configure Hadoop input/output settings globally across operations. Finally, it adds the <tt>RDD.toDebugString()</tt> method, which can be used to print an RDD’s lineage graph for troubleshooting.</p>
+<p>This release adds several RDD transformations, including <tt>keys</tt>, <tt>values</tt>, <tt>keyBy</tt>, <tt>subtract</tt>, <tt>coalesce</tt>, <tt>zip</tt>. It also adds <tt>SparkContext.hadoopConfiguration</tt> to allow programs to configure Hadoop input/output settings globally across operations. Finally, it adds the <tt>RDD.toDebugString()</tt> method, which can be used to print an RDD&#8217;s lineage graph for troubleshooting.</p>
<h3>EC2 Improvements</h3>
@@ -195,7 +195,7 @@
<h3>Credits</h3>
-<p>Spark 0.7 was the work of many contributors from Berkeley and outside—in total, 31 different contributors, of which 20 were from outside Berkeley. Here are the people who contributed, along with areas they worked on:</p>
+<p>Spark 0.7 was the work of many contributors from Berkeley and outside&#8212;in total, 31 different contributors, of which 20 were from outside Berkeley. Here are the people who contributed, along with areas they worked on:</p>
<ul>
<li>Mikhail Bautin -- Maven build</li>
diff --git a/site/releases/spark-release-0-8-0.html b/site/releases/spark-release-0-8-0.html
index 393b0eb22..f5f049b18 100644
--- a/site/releases/spark-release-0-8-0.html
+++ b/site/releases/spark-release-0-8-0.html
@@ -176,13 +176,13 @@
<li>The examples build has been isolated from the core build, substantially reducing the potential for dependency conflicts.</li>
<li>The Spark Streaming Twitter API has been updated to use OAuth authentication instead of the deprecated username/password authentication in Spark 0.7.0.</li>
<li>Several new example jobs have been added, including PageRank implementations in Java, Scala and Python, examples for accessing HBase and Cassandra, and MLlib examples.</li>
- <li>Support for running on Mesos has been improved – now you can deploy a Spark assembly JAR as part of the Mesos job, instead of having Spark pre-installed on each machine. The default Mesos version has also been updated to 0.13.</li>
+ <li>Support for running on Mesos has been improved &#8211; now you can deploy a Spark assembly JAR as part of the Mesos job, instead of having Spark pre-installed on each machine. The default Mesos version has also been updated to 0.13.</li>
<li>This release includes various optimizations to PySpark and to the job scheduler.</li>
</ul>
<h3 id="compatibility">Compatibility</h3>
<ul>
- <li><strong>This release changes Spark’s package name to ‘org.apache.spark’</strong>, so those upgrading from Spark 0.7 will need to adjust their imports accordingly. In addition, we’ve moved the <code>RDD</code> class to the org.apache.spark.rdd package (it was previously in the top-level package). The Spark artifacts published through Maven have also changed to the new package name.</li>
+ <li><strong>This release changes Spark’s package name to &#8216;org.apache.spark&#8217;</strong>, so those upgrading from Spark 0.7 will need to adjust their imports accordingly. In addition, we’ve moved the <code>RDD</code> class to the org.apache.spark.rdd package (it was previously in the top-level package). The Spark artifacts published through Maven have also changed to the new package name.</li>
<li>In the Java API, use of Scala’s <code>Option</code> class has been replaced with <code>Optional</code> from the Guava library.</li>
<li>Linking against Spark for arbitrary Hadoop versions is now possible by specifying a dependency on <code>hadoop-client</code>, instead of rebuilding <code>spark-core</code> against your version of Hadoop. See the documentation <a href="http://spark.incubator.apache.org/docs/0.8.0/scala-programming-guide.html#linking-with-spark">here</a> for details.</li>
<li>If you are building Spark, you’ll now need to run <code>sbt/sbt assembly</code> instead of <code>package</code>.</li>
@@ -192,73 +192,73 @@
<p>Spark 0.8.0 was the result of the largest team of contributors yet. The following developers contributed to this release:</p>
<ul>
- <li>Andrew Ash – documentation, code cleanup and logging improvements</li>
- <li>Mikhail Bautin – bug fix</li>
- <li>Konstantin Boudnik – Maven build, bug fixes, and documentation</li>
- <li>Ian Buss – sbt configuration improvement</li>
- <li>Evan Chan – API improvement, bug fix, and documentation</li>
- <li>Lian Cheng – bug fix</li>
- <li>Tathagata Das – performance improvement in streaming receiver and streaming bug fix</li>
- <li>Aaron Davidson – Python improvements, bug fix, and unit tests</li>
- <li>Giovanni Delussu – coalesced RDD feature</li>
- <li>Joseph E. Gonzalez – improvement to zipPartitions</li>
- <li>Karen Feng – several improvements to web UI</li>
- <li>Andy Feng – HDFS metrics</li>
- <li>Ali Ghodsi – configuration improvements and locality-aware coalesce</li>
- <li>Christoph Grothaus – bug fix</li>
- <li>Thomas Graves – support for secure YARN cluster and various YARN-related improvements</li>
- <li>Stephen Haberman – bug fix, documentation, and code cleanup</li>
- <li>Mark Hamstra – bug fixes and Maven build</li>
- <li>Benjamin Hindman – Mesos compatibility and documentation</li>
- <li>Liang-Chi Hsieh – bug fixes in build and in YARN mode</li>
- <li>Shane Huang – shuffle improvements, bug fix</li>
- <li>Ethan Jewett – Spark/HBase example</li>
- <li>Holden Karau – bug fix and EC2 improvement</li>
- <li>Kody Koeniger – JDBV RDD implementation</li>
- <li>Andy Konwinski – documentation</li>
- <li>Jey Kottalam – PySpark optimizations, Hadoop agnostic build (lead), and bug fixes</li>
- <li>Andrey Kouznetsov – Bug fix</li>
- <li>S. Kumar – Spark Streaming example</li>
- <li>Ryan LeCompte – topK method optimization and serialization improvements</li>
- <li>Gavin Li – compression codecs and pipe support</li>
- <li>Harold Lim – fair scheduler</li>
- <li>Dmitriy Lyubimov – bug fix</li>
- <li>Chris Mattmann – Apache mentor</li>
- <li>David McCauley – JSON API improvement</li>
- <li>Sean McNamara – added <code>takeOrdered</code> function, bug fixes, and a build fix</li>
- <li>Mridul Muralidharan – YARN integration (lead) and scheduler improvements</li>
- <li>Marc Mercer – improvements to UI json output</li>
- <li>Christopher Nguyen – bug fixes</li>
- <li>Erik van Oosten – example fix</li>
- <li>Kay Ousterhout – fix for scheduler regression and bug fixes</li>
- <li>Xinghao Pan – MLLib contributions</li>
- <li>Hiral Patel – bug fix</li>
- <li>James Phillpotts – updated Twitter API for Spark streaming</li>
- <li>Nick Pentreath – scala pageRank example, bagel improvement, and several Java examples</li>
- <li>Alexander Pivovarov – logging improvement and Maven build</li>
- <li>Mike Potts – configuration improvement</li>
- <li>Rohit Rai – Spark/Cassandra example</li>
- <li>Imran Rashid – bug fixes and UI improvement</li>
- <li>Charles Reiss – bug fixes, code cleanup, performance improvements</li>
- <li>Josh Rosen – Python API improvements, Java API improvements, EC2 scripts and bug fixes</li>
- <li>Henry Saputra – Apache mentor</li>
- <li>Jerry Shao – bug fixes, metrics system</li>
- <li>Prashant Sharma – documentation</li>
- <li>Mingfei Shi – joblogger and bug fix</li>
- <li>Andre Schumacher – several PySpark features</li>
- <li>Ginger Smith – MLLib contribution</li>
- <li>Evan Sparks – contributions to MLLib</li>
- <li>Ram Sriharsha – bug fix and RDD removal feature</li>
- <li>Ameet Talwalkar – MLlib contributions</li>
- <li>Roman Tkalenko – code refactoring and cleanup</li>
- <li>Chu Tong – Java PageRank algorithm and bug fix in bash scripts</li>
- <li>Shivaram Venkataraman – bug fixes, contributions to MLLib, netty shuffle fixes, and Java API additions</li>
- <li>Patrick Wendell – release manager, bug fixes, documentation, metrics system, and web UI</li>
- <li>Andrew Xia – fair scheduler (lead), metrics system, and ui improvements</li>
- <li>Reynold Xin – shuffle improvements, bug fixes, code refactoring, usability improvements, MLLib contributions</li>
- <li>Matei Zaharia – MLLib contributions, documentation, examples, UI improvements, PySpark improvements, and bug fixes</li>
- <li>Wu Zeming – bug fix in scheduler</li>
- <li>Bill Zhao – log message improvement</li>
+ <li>Andrew Ash &#8211; documentation, code cleanup and logging improvements</li>
+ <li>Mikhail Bautin &#8211; bug fix</li>
+ <li>Konstantin Boudnik &#8211; Maven build, bug fixes, and documentation</li>
+ <li>Ian Buss &#8211; sbt configuration improvement</li>
+ <li>Evan Chan &#8211; API improvement, bug fix, and documentation</li>
+ <li>Lian Cheng &#8211; bug fix</li>
+ <li>Tathagata Das &#8211; performance improvement in streaming receiver and streaming bug fix</li>
+ <li>Aaron Davidson &#8211; Python improvements, bug fix, and unit tests</li>
+ <li>Giovanni Delussu &#8211; coalesced RDD feature</li>
+ <li>Joseph E. Gonzalez &#8211; improvement to zipPartitions</li>
+ <li>Karen Feng &#8211; several improvements to web UI</li>
+ <li>Andy Feng &#8211; HDFS metrics</li>
+ <li>Ali Ghodsi &#8211; configuration improvements and locality-aware coalesce</li>
+ <li>Christoph Grothaus &#8211; bug fix</li>
+ <li>Thomas Graves &#8211; support for secure YARN cluster and various YARN-related improvements</li>
+ <li>Stephen Haberman &#8211; bug fix, documentation, and code cleanup</li>
+ <li>Mark Hamstra &#8211; bug fixes and Maven build</li>
+ <li>Benjamin Hindman &#8211; Mesos compatibility and documentation</li>
+ <li>Liang-Chi Hsieh &#8211; bug fixes in build and in YARN mode</li>
+ <li>Shane Huang &#8211; shuffle improvements, bug fix</li>
+ <li>Ethan Jewett &#8211; Spark/HBase example</li>
+ <li>Holden Karau &#8211; bug fix and EC2 improvement</li>
+ <li>Kody Koeniger &#8211; JDBV RDD implementation</li>
+ <li>Andy Konwinski &#8211; documentation</li>
+ <li>Jey Kottalam &#8211; PySpark optimizations, Hadoop agnostic build (lead), and bug fixes</li>
+ <li>Andrey Kouznetsov &#8211; Bug fix</li>
+ <li>S. Kumar &#8211; Spark Streaming example</li>
+ <li>Ryan LeCompte &#8211; topK method optimization and serialization improvements</li>
+ <li>Gavin Li &#8211; compression codecs and pipe support</li>
+ <li>Harold Lim &#8211; fair scheduler</li>
+ <li>Dmitriy Lyubimov &#8211; bug fix</li>
+ <li>Chris Mattmann &#8211; Apache mentor</li>
+ <li>David McCauley &#8211; JSON API improvement</li>
+ <li>Sean McNamara &#8211; added <code>takeOrdered</code> function, bug fixes, and a build fix</li>
+ <li>Mridul Muralidharan &#8211; YARN integration (lead) and scheduler improvements</li>
+ <li>Marc Mercer &#8211; improvements to UI json output</li>
+ <li>Christopher Nguyen &#8211; bug fixes</li>
+ <li>Erik van Oosten &#8211; example fix</li>
+ <li>Kay Ousterhout &#8211; fix for scheduler regression and bug fixes</li>
+ <li>Xinghao Pan &#8211; MLLib contributions</li>
+ <li>Hiral Patel &#8211; bug fix</li>
+ <li>James Phillpotts &#8211; updated Twitter API for Spark streaming</li>
+ <li>Nick Pentreath &#8211; scala pageRank example, bagel improvement, and several Java examples</li>
+ <li>Alexander Pivovarov &#8211; logging improvement and Maven build</li>
+ <li>Mike Potts &#8211; configuration improvement</li>
+ <li>Rohit Rai &#8211; Spark/Cassandra example</li>
+ <li>Imran Rashid &#8211; bug fixes and UI improvement</li>
+ <li>Charles Reiss &#8211; bug fixes, code cleanup, performance improvements</li>
+ <li>Josh Rosen &#8211; Python API improvements, Java API improvements, EC2 scripts and bug fixes</li>
+ <li>Henry Saputra &#8211; Apache mentor</li>
+ <li>Jerry Shao &#8211; bug fixes, metrics system</li>
+ <li>Prashant Sharma &#8211; documentation</li>
+ <li>Mingfei Shi &#8211; joblogger and bug fix</li>
+ <li>Andre Schumacher &#8211; several PySpark features</li>
+ <li>Ginger Smith &#8211; MLLib contribution</li>
+ <li>Evan Sparks &#8211; contributions to MLLib</li>
+ <li>Ram Sriharsha &#8211; bug fix and RDD removal feature</li>
+ <li>Ameet Talwalkar &#8211; MLlib contributions</li>
+ <li>Roman Tkalenko &#8211; code refactoring and cleanup</li>
+ <li>Chu Tong &#8211; Java PageRank algorithm and bug fix in bash scripts</li>
+ <li>Shivaram Venkataraman &#8211; bug fixes, contributions to MLLib, netty shuffle fixes, and Java API additions</li>
+ <li>Patrick Wendell &#8211; release manager, bug fixes, documentation, metrics system, and web UI</li>
+ <li>Andrew Xia &#8211; fair scheduler (lead), metrics system, and ui improvements</li>
+ <li>Reynold Xin &#8211; shuffle improvements, bug fixes, code refactoring, usability improvements, MLLib contributions</li>
+ <li>Matei Zaharia &#8211; MLLib contributions, documentation, examples, UI improvements, PySpark improvements, and bug fixes</li>
+ <li>Wu Zeming &#8211; bug fix in scheduler</li>
+ <li>Bill Zhao &#8211; log message improvement</li>
</ul>
<p>Thanks to everyone who contributed!