summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorMatei Alexandru Zaharia <matei@apache.org>2014-05-30 21:53:40 +0000
committerMatei Alexandru Zaharia <matei@apache.org>2014-05-30 21:53:40 +0000
commit13ca2f192e1cffa1b0f5e25ca123190c33e6695b (patch)
treea72af908a720209572ed834bd69415f7011ee505
parent8af35f4705088cffa94bced8d0ef0e1886e14c84 (diff)
downloadspark-website-13ca2f192e1cffa1b0f5e25ca123190c33e6695b.tar.gz
spark-website-13ca2f192e1cffa1b0f5e25ca123190c33e6695b.tar.bz2
spark-website-13ca2f192e1cffa1b0f5e25ca123190c33e6695b.zip
tweaks
-rw-r--r--news/_posts/2014-05-30-spark-1-0-0-released.md4
-rw-r--r--site/downloads.html2
-rw-r--r--site/news/amp-camp-2013-registration-ope.html2
-rw-r--r--site/news/index.html40
-rw-r--r--site/news/run-spark-and-shark-on-amazon-emr.html2
-rw-r--r--site/news/spark-0-6-1-and-0-5-2-released.html2
-rw-r--r--site/news/spark-0-7-0-released.html2
-rw-r--r--site/news/spark-0-7-2-released.html2
-rw-r--r--site/news/spark-0-7-3-released.html2
-rw-r--r--site/news/spark-0-8-0-released.html2
-rw-r--r--site/news/spark-0-8-1-released.html2
-rw-r--r--site/news/spark-0-9-0-released.html4
-rw-r--r--site/news/spark-1-0-0-released.html4
-rw-r--r--site/news/spark-and-shark-in-the-news.html4
-rw-r--r--site/news/spark-becomes-tlp.html4
-rw-r--r--site/news/spark-meetups.html2
-rw-r--r--site/news/spark-user-survey-and-powered-by-page.html4
-rw-r--r--site/news/strata-exercises-now-available-online.html2
-rw-r--r--site/news/submit-talks-to-spark-summit-2014.html4
-rw-r--r--site/news/video-from-first-spark-development-meetup.html2
-rw-r--r--site/releases/spark-release-0-3.html2
-rw-r--r--site/releases/spark-release-0-5-0.html8
-rw-r--r--site/releases/spark-release-0-5-1.html2
-rw-r--r--site/releases/spark-release-0-6-0.html6
-rw-r--r--site/releases/spark-release-0-7-0.html4
-rw-r--r--site/releases/spark-release-0-8-0.html138
-rw-r--r--site/releases/spark-release-0-8-1.html86
-rw-r--r--site/releases/spark-release-0-9-0.html168
-rw-r--r--site/releases/spark-release-0-9-1.html8
-rw-r--r--site/releases/spark-release-1-0-0.html236
30 files changed, 375 insertions, 375 deletions
diff --git a/news/_posts/2014-05-30-spark-1-0-0-released.md b/news/_posts/2014-05-30-spark-1-0-0-released.md
index 04d389922..a7ea4e6a4 100644
--- a/news/_posts/2014-05-30-spark-1-0-0-released.md
+++ b/news/_posts/2014-05-30-spark-1-0-0-released.md
@@ -11,7 +11,7 @@ meta:
_edit_last: '4'
_wpas_done_all: '1'
---
-We are happy to announce the availability of <a href="{{site.url}}releases/spark-release-1-0-0.html" title="Spark Release 1.0.0">Spark 1.0.0</a>! Spark 1.0.0 is the first in the 1.0 line of releases, providing API stability for Spark's core interfaces. It is Spark's largest release ever, with contributions from 117 developers.
-This release expands Spark's standard libraries, introducing a new SQL package (SparkSQL) which lets users integrate SQL queries into existing Spark workflows. MLlib, Spark's machine learning library, is expanded with sparse vector support and several new algorithms. The GraphX and Streaming libraries also introduce new features and optimizations. Spark's core engine adds support for secured YARN clusters, a unified tool for submitting Spark applications, and several performance and stability improvements.
+We are happy to announce the availability of <a href="{{site.url}}releases/spark-release-1-0-0.html" title="Spark Release 1.0.0">Spark 1.0.0</a>! Spark 1.0.0 is the first in the 1.X line of releases, providing API stability for Spark's core interfaces. It is Spark's largest release ever, with contributions from 117 developers.
+This release expands Spark's standard libraries, introducing a new SQL package (Spark SQL) that lets users integrate SQL queries into existing Spark workflows. MLlib, Spark's machine learning library, is expanded with sparse vector support and several new algorithms. The GraphX and Streaming libraries also introduce new features and optimizations. Spark's core engine adds support for secured YARN clusters, a unified tool for submitting Spark applications, and several performance and stability improvements.
Visit the <a href="{{site.url}}releases/spark-release-1-0-0.html" title="Spark Release 1.0.0">release notes</a> to read about the new features, or <a href="{{site.url}}downloads.html">download</a> the release today.
diff --git a/site/downloads.html b/site/downloads.html
index 9ca3a79fa..d26ac938f 100644
--- a/site/downloads.html
+++ b/site/downloads.html
@@ -202,7 +202,7 @@ version: 1.0.0
<h3 id="development-version">Development Version</h3>
<p>If you are interested in working with the newest under-development code or contributing to Spark development, you can also check out the master branch from Git: <tt>git clone git://github.com/apache/spark.git</tt>.</p>
-<p>Once you&#8217;ve downloaded Spark, you can find instructions for installing and building it on the <a href="/documentation.html">documentation page</a>.</p>
+<p>Once you’ve downloaded Spark, you can find instructions for installing and building it on the <a href="/documentation.html">documentation page</a>.</p>
<h3 id="all-releases">All Releases</h3>
<ul>
diff --git a/site/news/amp-camp-2013-registration-ope.html b/site/news/amp-camp-2013-registration-ope.html
index 75042c9af..c1ea9ec5d 100644
--- a/site/news/amp-camp-2013-registration-ope.html
+++ b/site/news/amp-camp-2013-registration-ope.html
@@ -160,7 +160,7 @@
<h2>Registration open for AMP Camp training camp in Berkeley</h2>
-<p>Want to learn how to use Spark, Shark, GraphX, and related technologies in person? The AMP Lab is hosting a two-day training workshop for them on August 29th and 30th in Berkeley. The workshop will include tutorials, talks from users, and over four hours of hands-on exercises. <a href="http://ampcamp.berkeley.edu/amp-camp-three-berkeley-2013/">Registration is now open on the AMP Camp website</a>, for a price of $250 per person. We recommend signing up early because last year&#8217;s workshop was sold out.</p>
+<p>Want to learn how to use Spark, Shark, GraphX, and related technologies in person? The AMP Lab is hosting a two-day training workshop for them on August 29th and 30th in Berkeley. The workshop will include tutorials, talks from users, and over four hours of hands-on exercises. <a href="http://ampcamp.berkeley.edu/amp-camp-three-berkeley-2013/">Registration is now open on the AMP Camp website</a>, for a price of $250 per person. We recommend signing up early because last year’s workshop was sold out.</p>
<p>
diff --git a/site/news/index.html b/site/news/index.html
index f9d59de17..e6867945b 100644
--- a/site/news/index.html
+++ b/site/news/index.html
@@ -164,8 +164,8 @@
<h3 class="entry-title"><a href="/news/spark-1-0-0-released.html">Spark 1.0.0 released</a></h3>
<div class="entry-date">May 30, 2014</div>
</header>
- <div class="entry-content"><p>We are happy to announce the availability of <a href="/releases/spark-release-1-0-0.html" title="Spark Release 1.0.0">Spark 1.0.0</a>! Spark 1.0.0 is the first in the 1.0 line of releases, providing API stability for Spark&#8217;s core interfaces. It is Spark&#8217;s largest release ever, with contributions from 117 developers.
-This release expands Spark&#8217;s standard libraries, introducing a new SQL package (SparkSQL) which lets users integrate SQL queries into existing Spark workflows. MLlib, Spark&#8217;s machine learning library, is expanded with sparse vector support and several new algorithms. The GraphX and Streaming libraries also introduce new features and optimizations. Spark&#8217;s core engine adds support for secured YARN clusters, a unified tool for submitting Spark applications, and several performance and stability improvements.</p>
+ <div class="entry-content"><p>We are happy to announce the availability of <a href="/releases/spark-release-1-0-0.html" title="Spark Release 1.0.0">Spark 1.0.0</a>! Spark 1.0.0 is the first in the 1.X line of releases, providing API stability for Spark’s core interfaces. It is Spark’s largest release ever, with contributions from 117 developers.
+This release expands Spark’s standard libraries, introducing a new SQL package (Spark SQL) that lets users integrate SQL queries into existing Spark workflows. MLlib, Spark’s machine learning library, is expanded with sparse vector support and several new algorithms. The GraphX and Streaming libraries also introduce new features and optimizations. Spark’s core engine adds support for secured YARN clusters, a unified tool for submitting Spark applications, and several performance and stability improvements.</p>
</div>
</article>
@@ -202,7 +202,7 @@ Contributions to this release came from 37 developers. </p>
<h3 class="entry-title"><a href="/news/submit-talks-to-spark-summit-2014.html">Submissions and registration open for Spark Summit 2014</a></h3>
<div class="entry-date">March 20, 2014</div>
</header>
- <div class="entry-content"><p>After last year&#8217;s successful <a href="http://spark-summit.org/2013">first Spark Summit</a>, registrations
+ <div class="entry-content"><p>After last year’s successful <a href="http://spark-summit.org/2013">first Spark Summit</a>, registrations
and talk submissions are now open for <a href="http://spark-summit.org/2014">Spark Summit 2014</a>.
This will be a 3-day event in San Francisco organized by multiple companies in the Spark community.
The event will run <strong>June 30th to July 2nd</strong> in San Francisco, CA.</p>
@@ -215,7 +215,7 @@ The event will run <strong>June 30th to July 2nd</strong> in San Francisco, CA.<
<h3 class="entry-title"><a href="/news/spark-becomes-tlp.html">Spark becomes top-level Apache project</a></h3>
<div class="entry-date">February 27, 2014</div>
</header>
- <div class="entry-content"><p>The Apache Software Foundation <a href="https://blogs.apache.org/foundation/entry/the_apache_software_foundation_announces50">announced</a> today that Spark has graduated from the Apache Incubator to become a top-level Apache project, signifying that the project&#8217;s community and products have been well-governed under the ASF&#8217;s meritocratic process and principles. This is a major step for the community and we are very proud to share this news with users as we complete Spark&#8217;s move to Apache. Read more about Spark&#8217;s growth during the past year and from contributors and users in the ASF&#8217;s <a href="https://blogs.apache.org/foundation/entry/the_apache_software_foundation_announces50">press release</a>.</p>
+ <div class="entry-content"><p>The Apache Software Foundation <a href="https://blogs.apache.org/foundation/entry/the_apache_software_foundation_announces50">announced</a> today that Spark has graduated from the Apache Incubator to become a top-level Apache project, signifying that the project’s community and products have been well-governed under the ASF’s meritocratic process and principles. This is a major step for the community and we are very proud to share this news with users as we complete Spark’s move to Apache. Read more about Spark’s growth during the past year and from contributors and users in the ASF’s <a href="https://blogs.apache.org/foundation/entry/the_apache_software_foundation_announces50">press release</a>.</p>
</div>
</article>
@@ -226,8 +226,8 @@ The event will run <strong>June 30th to July 2nd</strong> in San Francisco, CA.<
<div class="entry-date">February 2, 2014</div>
</header>
<div class="entry-content"><p>We are happy to announce the availability of <a href="/releases/spark-release-0-9-0.html" title="Spark Release 0.9.0">
-Spark 0.9.0</a>! Spark 0.9.0 is a major release and Spark&#8217;s largest release ever, with contributions from 83 developers.
-This release expands Spark&#8217;s standard libraries, introducing a new graph computation package (GraphX) and adding several new features to the machine learning and stream-processing packages. It also makes major improvements to the core engine,
+Spark 0.9.0</a>! Spark 0.9.0 is a major release and Spark’s largest release ever, with contributions from 83 developers.
+This release expands Spark’s standard libraries, introducing a new graph computation package (GraphX) and adding several new features to the machine learning and stream-processing packages. It also makes major improvements to the core engine,
including external aggregations, a simplified H/A mode for long lived applications, and
hardened YARN support.</p>
@@ -239,7 +239,7 @@ hardened YARN support.</p>
<h3 class="entry-title"><a href="/news/spark-0-8-1-released.html">Spark 0.8.1 released</a></h3>
<div class="entry-date">December 19, 2013</div>
</header>
- <div class="entry-content"><p>We&#8217;ve just posted <a href="/releases/spark-release-0-8-1.html" title="Spark Release 0.8.1">Spark Release 0.8.1</a>, a maintenance and performance release for the Scala 2.9 version of Spark. 0.8.1 includes support for YARN 2.2, a high availability mode for the standalone scheduler, optimizations to the shuffle, and many other improvements. We recommend that all users update to this release. Visit the <a href="/releases/spark-release-0-8-1.html" title="Spark Release 0.8.1">release notes</a> to read about the new features, or <a href="/downloads.html">download</a> the release today.</p>
+ <div class="entry-content"><p>We’ve just posted <a href="/releases/spark-release-0-8-1.html" title="Spark Release 0.8.1">Spark Release 0.8.1</a>, a maintenance and performance release for the Scala 2.9 version of Spark. 0.8.1 includes support for YARN 2.2, a high availability mode for the standalone scheduler, optimizations to the shuffle, and many other improvements. We recommend that all users update to this release. Visit the <a href="/releases/spark-release-0-8-1.html" title="Spark Release 0.8.1">release notes</a> to read about the new features, or <a href="/downloads.html">download</a> the release today.</p>
</div>
</article>
@@ -270,7 +270,7 @@ Over 450 Spark developers and enthusiasts from 13 countries and more than 180 co
<h3 class="entry-title"><a href="/news/spark-0-8-0-released.html">Spark 0.8.0 released</a></h3>
<div class="entry-date">September 25, 2013</div>
</header>
- <div class="entry-content"><p>We&#8217;re proud to announce the release of <a href="/releases/spark-release-0-8-0.html" title="Spark Release 0.8.0">Apache Spark 0.8.0</a>. Spark 0.8.0 is a major release that includes many new capabilities and usability improvements. It’s also our first release under the Apache incubator. It is the largest Spark release yet, with contributions from 67 developers and 24 companies. Major new features include an expanded monitoring framework and UI, a machine learning library, and support for running Spark inside of YARN.</p>
+ <div class="entry-content"><p>We’re proud to announce the release of <a href="/releases/spark-release-0-8-0.html" title="Spark Release 0.8.0">Apache Spark 0.8.0</a>. Spark 0.8.0 is a major release that includes many new capabilities and usability improvements. It’s also our first release under the Apache incubator. It is the largest Spark release yet, with contributions from 67 developers and 24 companies. Major new features include an expanded monitoring framework and UI, a machine learning library, and support for running Spark inside of YARN.</p>
</div>
</article>
@@ -280,7 +280,7 @@ Over 450 Spark developers and enthusiasts from 13 countries and more than 180 co
<h3 class="entry-title"><a href="/news/spark-user-survey-and-powered-by-page.html">Spark user survey and "Powered By" page</a></h3>
<div class="entry-date">September 5, 2013</div>
</header>
- <div class="entry-content"><p>As we continue developing Spark, we would love to get feedback from users and hear what you&#8217;d like us to work on next. We&#8217;ve decided that a good way to do that is a survey &#8211; we hope to run this at regular intervals. If you have a few minutes to participate, <a href="https://docs.google.com/forms/d/1eMXp4GjcIXglxJe5vYYBzXKVm-6AiYt1KThJwhCjJiY/viewform">fill in the survey here</a>. Your time is greatly appreciated.</p>
+ <div class="entry-content"><p>As we continue developing Spark, we would love to get feedback from users and hear what you’d like us to work on next. We’ve decided that a good way to do that is a survey – we hope to run this at regular intervals. If you have a few minutes to participate, <a href="https://docs.google.com/forms/d/1eMXp4GjcIXglxJe5vYYBzXKVm-6AiYt1KThJwhCjJiY/viewform">fill in the survey here</a>. Your time is greatly appreciated.</p>
</div>
</article>
@@ -300,7 +300,7 @@ Over 450 Spark developers and enthusiasts from 13 countries and more than 180 co
<h3 class="entry-title"><a href="/news/amp-camp-2013-registration-ope.html">Registration open for AMP Camp training camp in Berkeley</a></h3>
<div class="entry-date">July 23, 2013</div>
</header>
- <div class="entry-content"><p>Want to learn how to use Spark, Shark, GraphX, and related technologies in person? The AMP Lab is hosting a two-day training workshop for them on August 29th and 30th in Berkeley. The workshop will include tutorials, talks from users, and over four hours of hands-on exercises. <a href="http://ampcamp.berkeley.edu/amp-camp-three-berkeley-2013/">Registration is now open on the AMP Camp website</a>, for a price of $250 per person. We recommend signing up early because last year&#8217;s workshop was sold out.</p>
+ <div class="entry-content"><p>Want to learn how to use Spark, Shark, GraphX, and related technologies in person? The AMP Lab is hosting a two-day training workshop for them on August 29th and 30th in Berkeley. The workshop will include tutorials, talks from users, and over four hours of hands-on exercises. <a href="http://ampcamp.berkeley.edu/amp-camp-three-berkeley-2013/">Registration is now open on the AMP Camp website</a>, for a price of $250 per person. We recommend signing up early because last year’s workshop was sold out.</p>
</div>
</article>
@@ -331,7 +331,7 @@ Over 450 Spark developers and enthusiasts from 13 countries and more than 180 co
<h3 class="entry-title"><a href="/news/spark-0-7-3-released.html">Spark 0.7.3 released</a></h3>
<div class="entry-date">July 16, 2013</div>
</header>
- <div class="entry-content"><p>We&#8217;ve just posted <a href="/releases/spark-release-0-7-3.html" title="Spark Release 0.7.3">Spark Release 0.7.3</a>, a maintenance release that contains several fixes, including streaming API updates and new functionality for adding JARs to a <code>spark-shell</code> session. We recommend that all users update to this release. Visit the <a href="/releases/spark-release-0-7-3.html" title="Spark Release 0.7.3">release notes</a> to read about the new features, or <a href="/downloads.html">download</a> the release today.</p>
+ <div class="entry-content"><p>We’ve just posted <a href="/releases/spark-release-0-7-3.html" title="Spark Release 0.7.3">Spark Release 0.7.3</a>, a maintenance release that contains several fixes, including streaming API updates and new functionality for adding JARs to a <code>spark-shell</code> session. We recommend that all users update to this release. Visit the <a href="/releases/spark-release-0-7-3.html" title="Spark Release 0.7.3">release notes</a> to read about the new features, or <a href="/downloads.html">download</a> the release today.</p>
</div>
</article>
@@ -361,7 +361,7 @@ Over 450 Spark developers and enthusiasts from 13 countries and more than 180 co
<h3 class="entry-title"><a href="/news/spark-0-7-2-released.html">Spark 0.7.2 released</a></h3>
<div class="entry-date">June 2, 2013</div>
</header>
- <div class="entry-content"><p>We&#8217;re happy to announce the release of <a href="/releases/spark-release-0-7-2.html" title="Spark Release 0.7.2">Spark 0.7.2</a>, a new maintenance release that includes several bug fixes and improvements, as well as new code examples and API features. We recommend that all users update to this release. Head over to the <a href="/releases/spark-release-0-7-2.html" title="Spark Release 0.7.2">release notes</a> to read about the new features, or <a href="/downloads.html">download</a> the release today.</p>
+ <div class="entry-content"><p>We’re happy to announce the release of <a href="/releases/spark-release-0-7-2.html" title="Spark Release 0.7.2">Spark 0.7.2</a>, a new maintenance release that includes several bug fixes and improvements, as well as new code examples and API features. We recommend that all users update to this release. Head over to the <a href="/releases/spark-release-0-7-2.html" title="Spark Release 0.7.2">release notes</a> to read about the new features, or <a href="/downloads.html">download</a> the release today.</p>
</div>
</article>
@@ -387,7 +387,7 @@ Over 450 Spark developers and enthusiasts from 13 countries and more than 180 co
<h3 class="entry-title"><a href="/news/strata-exercises-now-available-online.html">Strata exercises now available online</a></h3>
<div class="entry-date">March 17, 2013</div>
</header>
- <div class="entry-content"><p>At this year&#8217;s <a href="http://strataconf.com/strata2013">Strata</a> conference, the AMP Lab hosted a full day of tutorials on Spark, Shark, and Spark Streaming, including online exercises on Amazon EC2. Those exercises are now <a href="http://ampcamp.berkeley.edu/big-data-mini-course/">available online</a>, letting you learn Spark and Shark at your own pace on an EC2 cluster with real data. They are a great resource for learning the systems. You can also find <a href="http://ampcamp.berkeley.edu/amp-camp-two-strata-2013/">slides</a> from the Strata tutorials online, as well as <a href="http://ampcamp.berkeley.edu/amp-camp-one-berkeley-2012/">videos</a> from the AMP Camp workshop we held at Berkeley in August.</p>
+ <div class="entry-content"><p>At this year’s <a href="http://strataconf.com/strata2013">Strata</a> conference, the AMP Lab hosted a full day of tutorials on Spark, Shark, and Spark Streaming, including online exercises on Amazon EC2. Those exercises are now <a href="http://ampcamp.berkeley.edu/big-data-mini-course/">available online</a>, letting you learn Spark and Shark at your own pace on an EC2 cluster with real data. They are a great resource for learning the systems. You can also find <a href="http://ampcamp.berkeley.edu/amp-camp-two-strata-2013/">slides</a> from the Strata tutorials online, as well as <a href="http://ampcamp.berkeley.edu/amp-camp-one-berkeley-2012/">videos</a> from the AMP Camp workshop we held at Berkeley in August.</p>
</div>
</article>
@@ -397,7 +397,7 @@ Over 450 Spark developers and enthusiasts from 13 countries and more than 180 co
<h3 class="entry-title"><a href="/news/spark-0-7-0-released.html">Spark 0.7.0 released</a></h3>
<div class="entry-date">February 27, 2013</div>
</header>
- <div class="entry-content"><p>We&#8217;re proud to announce the release of <a href="/releases/spark-release-0-7-0.html" title="Spark Release 0.7.0">Spark 0.7.0</a>, a new major version of Spark that adds several key features, including a <a href="/docs/latest/python-programming-guide.html">Python API</a> for Spark and an <a href="/docs/latest/streaming-programming-guide.html">alpha of Spark Streaming</a>. This release is the result of the largest group of contributors yet behind a Spark release &#8211; 31 contributors from inside and outside Berkeley. Head over to the <a href="/releases/spark-release-0-7-0.html" title="Spark Release 0.7.0">release notes</a> to read more about the new features, or <a href="/downloads.html">download</a> the release today.</p>
+ <div class="entry-content"><p>We’re proud to announce the release of <a href="/releases/spark-release-0-7-0.html" title="Spark Release 0.7.0">Spark 0.7.0</a>, a new major version of Spark that adds several key features, including a <a href="/docs/latest/python-programming-guide.html">Python API</a> for Spark and an <a href="/docs/latest/streaming-programming-guide.html">alpha of Spark Streaming</a>. This release is the result of the largest group of contributors yet behind a Spark release – 31 contributors from inside and outside Berkeley. Head over to the <a href="/releases/spark-release-0-7-0.html" title="Spark Release 0.7.0">release notes</a> to read more about the new features, or <a href="/downloads.html">download</a> the release today.</p>
</div>
</article>
@@ -407,7 +407,7 @@ Over 450 Spark developers and enthusiasts from 13 countries and more than 180 co
<h3 class="entry-title"><a href="/news/run-spark-and-shark-on-amazon-emr.html">Spark/Shark Tutorial for Amazon EMR</a></h3>
<div class="entry-date">February 24, 2013</div>
</header>
- <div class="entry-content"><p>This weekend, Amazon posted an <a href="http://aws.amazon.com/articles/Elastic-MapReduce/4926593393724923">article</a> and code that make it easy to launch Spark and Shark on Elastic MapReduce. The article includes examples of how to run both interactive Scala commands and SQL queries from Shark on data in S3. Head over to the <a href="http://aws.amazon.com/articles/Elastic-MapReduce/4926593393724923">Amazon article</a> for details. We&#8217;re very excited because, to our knowledge, this makes Spark the first non-Hadoop engine that you can launch with EMR.</p>
+ <div class="entry-content"><p>This weekend, Amazon posted an <a href="http://aws.amazon.com/articles/Elastic-MapReduce/4926593393724923">article</a> and code that make it easy to launch Spark and Shark on Elastic MapReduce. The article includes examples of how to run both interactive Scala commands and SQL queries from Shark on data in S3. Head over to the <a href="http://aws.amazon.com/articles/Elastic-MapReduce/4926593393724923">Amazon article</a> for details. We’re very excited because, to our knowledge, this makes Spark the first non-Hadoop engine that you can launch with EMR.</p>
</div>
</article>
@@ -442,7 +442,7 @@ Over 450 Spark developers and enthusiasts from 13 countries and more than 180 co
<h3 class="entry-title"><a href="/news/video-from-first-spark-development-meetup.html">Video up from first Spark development meetup</a></h3>
<div class="entry-date">December 21, 2012</div>
</header>
- <div class="entry-content"><p>On December 18th, we held the first of a series of Spark development meetups, for people interested in learning the Spark codebase and contributing to the project. There was quite a bit more demand than we anticipated, with over 80 people signing up and 64 attending. The first meetup was an <a href="http://www.meetup.com/spark-users/events/94101942/">introduction to Spark internals</a>. Thanks to one of the attendees, there&#8217;s now a <a href="http://www.youtube.com/watch?v=49Hr5xZyTEA">video of the meetup</a> on YouTube. We&#8217;ve also posted the <a href="http://files.meetup.com/3138542/dev-meetup-dec-2012.pptx">slides</a>. Look to see more development meetups on Spark and Shark in the future.</p>
+ <div class="entry-content"><p>On December 18th, we held the first of a series of Spark development meetups, for people interested in learning the Spark codebase and contributing to the project. There was quite a bit more demand than we anticipated, with over 80 people signing up and 64 attending. The first meetup was an <a href="http://www.meetup.com/spark-users/events/94101942/">introduction to Spark internals</a>. Thanks to one of the attendees, there’s now a <a href="http://www.youtube.com/watch?v=49Hr5xZyTEA">video of the meetup</a> on YouTube. We’ve also posted the <a href="http://files.meetup.com/3138542/dev-meetup-dec-2012.pptx">slides</a>. Look to see more development meetups on Spark and Shark in the future.</p>
</div>
</article>
@@ -452,7 +452,7 @@ Over 450 Spark developers and enthusiasts from 13 countries and more than 180 co
<h3 class="entry-title"><a href="/news/spark-and-shark-in-the-news.html">Spark and Shark in the news</a></h3>
<div class="entry-date">December 21, 2012</div>
</header>
- <div class="entry-content"><p>Recently, we&#8217;ve seen quite a bit of coverage of both Spark and <a href="http://shark.cs.berkeley.edu">Shark</a> in the news. I wanted to list some of the more recent articles, for readers interested in learning more.</p>
+ <div class="entry-content"><p>Recently, we’ve seen quite a bit of coverage of both Spark and <a href="http://shark.cs.berkeley.edu">Shark</a> in the news. I wanted to list some of the more recent articles, for readers interested in learning more.</p>
<ul>
<li>Curt Monash, editor of the popular DBMS2 blog, wrote a great <a href="http://www.dbms2.com/2012/12/13/introduction-to-spark-shark-bdas-and-amplab/">introduction to Spark and Shark</a>, as well as a more detailed <a href="http://www.dbms2.com/2012/12/13/spark-shark-and-rdds-technology-notes/">technical overview</a>.</li>
@@ -462,7 +462,7 @@ Over 450 Spark developers and enthusiasts from 13 countries and more than 180 co
<li><a href="http://data-informed.com/spark-an-open-source-engine-for-iterative-data-mining/">DataInformed</a> interviewed two Spark users and wrote about their applications in anomaly detection, predictive analytics and data mining.</li>
</ul>
-<p>In other news, there will be a full day of tutorials on Spark and Shark at the <a href="http://strataconf.com/strata2013">O&#8217;Reilly Strata conference</a> in February. They include a three-hour <a href="http://strataconf.com/strata2013/public/schedule/detail/27438">introduction to Spark, Shark and BDAS</a> Tuesday morning, and a three-hour <a href="http://strataconf.com/strata2013/public/schedule/detail/27440">hands-on exercise session</a>. </p>
+<p>In other news, there will be a full day of tutorials on Spark and Shark at the <a href="http://strataconf.com/strata2013">O’Reilly Strata conference</a> in February. They include a three-hour <a href="http://strataconf.com/strata2013/public/schedule/detail/27438">introduction to Spark, Shark and BDAS</a> Tuesday morning, and a three-hour <a href="http://strataconf.com/strata2013/public/schedule/detail/27440">hands-on exercise session</a>. </p>
</div>
</article>
@@ -472,7 +472,7 @@ Over 450 Spark developers and enthusiasts from 13 countries and more than 180 co
<h3 class="entry-title"><a href="/news/spark-0-6-1-and-0-5-2-released.html">Spark 0.6.1 and 0.5.2 out</a></h3>
<div class="entry-date">November 22, 2012</div>
</header>
- <div class="entry-content"><p>Today we&#8217;ve made available two maintenance releases for Spark: <a href="/releases/spark-release-0-6-1.html" title="Spark Release 0.6.1">0.6.1</a> and <a href="/releases/spark-release-0-5-2.html" title="Spark Release 0.5.2">0.5.2</a>. They both contain important bug fixes as well as some new features, such as the ability to build against Hadoop 2 distributions. We recommend that users update to the latest version for their branch; for new users, we recommend <a href="/releases/spark-release-0-6-1.html" title="Spark Release 0.6.1">0.6.1</a>.</p>
+ <div class="entry-content"><p>Today we’ve made available two maintenance releases for Spark: <a href="/releases/spark-release-0-6-1.html" title="Spark Release 0.6.1">0.6.1</a> and <a href="/releases/spark-release-0-5-2.html" title="Spark Release 0.5.2">0.5.2</a>. They both contain important bug fixes as well as some new features, such as the ability to build against Hadoop 2 distributions. We recommend that users update to the latest version for their branch; for new users, we recommend <a href="/releases/spark-release-0-6-1.html" title="Spark Release 0.6.1">0.6.1</a>.</p>
</div>
</article>
@@ -502,7 +502,7 @@ Over 450 Spark developers and enthusiasts from 13 countries and more than 180 co
<h3 class="entry-title"><a href="/news/spark-meetups.html">We've started hosting a Bay Area Spark User Meetup</a></h3>
<div class="entry-date">January 10, 2012</div>
</header>
- <div class="entry-content"><p>We&#8217;ve started hosting a regular <a href="http://www.meetup.com/spark-users/">Bay Area Spark User Meetup</a>. Sign up on the meetup.com page to be notified about events and meet other Spark developers and users.</p>
+ <div class="entry-content"><p>We’ve started hosting a regular <a href="http://www.meetup.com/spark-users/">Bay Area Spark User Meetup</a>. Sign up on the meetup.com page to be notified about events and meet other Spark developers and users.</p>
</div>
</article>
diff --git a/site/news/run-spark-and-shark-on-amazon-emr.html b/site/news/run-spark-and-shark-on-amazon-emr.html
index ec904cc23..568d995f7 100644
--- a/site/news/run-spark-and-shark-on-amazon-emr.html
+++ b/site/news/run-spark-and-shark-on-amazon-emr.html
@@ -160,7 +160,7 @@
<h2>Spark/Shark Tutorial for Amazon EMR</h2>
-<p>This weekend, Amazon posted an <a href="http://aws.amazon.com/articles/Elastic-MapReduce/4926593393724923">article</a> and code that make it easy to launch Spark and Shark on Elastic MapReduce. The article includes examples of how to run both interactive Scala commands and SQL queries from Shark on data in S3. Head over to the <a href="http://aws.amazon.com/articles/Elastic-MapReduce/4926593393724923">Amazon article</a> for details. We&#8217;re very excited because, to our knowledge, this makes Spark the first non-Hadoop engine that you can launch with EMR.</p>
+<p>This weekend, Amazon posted an <a href="http://aws.amazon.com/articles/Elastic-MapReduce/4926593393724923">article</a> and code that make it easy to launch Spark and Shark on Elastic MapReduce. The article includes examples of how to run both interactive Scala commands and SQL queries from Shark on data in S3. Head over to the <a href="http://aws.amazon.com/articles/Elastic-MapReduce/4926593393724923">Amazon article</a> for details. We’re very excited because, to our knowledge, this makes Spark the first non-Hadoop engine that you can launch with EMR.</p>
<p>
diff --git a/site/news/spark-0-6-1-and-0-5-2-released.html b/site/news/spark-0-6-1-and-0-5-2-released.html
index e46701043..4bdaf8543 100644
--- a/site/news/spark-0-6-1-and-0-5-2-released.html
+++ b/site/news/spark-0-6-1-and-0-5-2-released.html
@@ -160,7 +160,7 @@
<h2>Spark 0.6.1 and 0.5.2 out</h2>
-<p>Today we&#8217;ve made available two maintenance releases for Spark: <a href="/releases/spark-release-0-6-1.html" title="Spark Release 0.6.1">0.6.1</a> and <a href="/releases/spark-release-0-5-2.html" title="Spark Release 0.5.2">0.5.2</a>. They both contain important bug fixes as well as some new features, such as the ability to build against Hadoop 2 distributions. We recommend that users update to the latest version for their branch; for new users, we recommend <a href="/releases/spark-release-0-6-1.html" title="Spark Release 0.6.1">0.6.1</a>.</p>
+<p>Today we’ve made available two maintenance releases for Spark: <a href="/releases/spark-release-0-6-1.html" title="Spark Release 0.6.1">0.6.1</a> and <a href="/releases/spark-release-0-5-2.html" title="Spark Release 0.5.2">0.5.2</a>. They both contain important bug fixes as well as some new features, such as the ability to build against Hadoop 2 distributions. We recommend that users update to the latest version for their branch; for new users, we recommend <a href="/releases/spark-release-0-6-1.html" title="Spark Release 0.6.1">0.6.1</a>.</p>
<p>
diff --git a/site/news/spark-0-7-0-released.html b/site/news/spark-0-7-0-released.html
index 98164d5bf..2fa4e59ab 100644
--- a/site/news/spark-0-7-0-released.html
+++ b/site/news/spark-0-7-0-released.html
@@ -160,7 +160,7 @@
<h2>Spark 0.7.0 released</h2>
-<p>We&#8217;re proud to announce the release of <a href="/releases/spark-release-0-7-0.html" title="Spark Release 0.7.0">Spark 0.7.0</a>, a new major version of Spark that adds several key features, including a <a href="/docs/latest/python-programming-guide.html">Python API</a> for Spark and an <a href="/docs/latest/streaming-programming-guide.html">alpha of Spark Streaming</a>. This release is the result of the largest group of contributors yet behind a Spark release &#8211; 31 contributors from inside and outside Berkeley. Head over to the <a href="/releases/spark-release-0-7-0.html" title="Spark Release 0.7.0">release notes</a> to read more about the new features, or <a href="/downloads.html">download</a> the release today.</p>
+<p>We’re proud to announce the release of <a href="/releases/spark-release-0-7-0.html" title="Spark Release 0.7.0">Spark 0.7.0</a>, a new major version of Spark that adds several key features, including a <a href="/docs/latest/python-programming-guide.html">Python API</a> for Spark and an <a href="/docs/latest/streaming-programming-guide.html">alpha of Spark Streaming</a>. This release is the result of the largest group of contributors yet behind a Spark release – 31 contributors from inside and outside Berkeley. Head over to the <a href="/releases/spark-release-0-7-0.html" title="Spark Release 0.7.0">release notes</a> to read more about the new features, or <a href="/downloads.html">download</a> the release today.</p>
<p>
diff --git a/site/news/spark-0-7-2-released.html b/site/news/spark-0-7-2-released.html
index 2cd2fa119..7f2396b36 100644
--- a/site/news/spark-0-7-2-released.html
+++ b/site/news/spark-0-7-2-released.html
@@ -160,7 +160,7 @@
<h2>Spark 0.7.2 released</h2>
-<p>We&#8217;re happy to announce the release of <a href="/releases/spark-release-0-7-2.html" title="Spark Release 0.7.2">Spark 0.7.2</a>, a new maintenance release that includes several bug fixes and improvements, as well as new code examples and API features. We recommend that all users update to this release. Head over to the <a href="/releases/spark-release-0-7-2.html" title="Spark Release 0.7.2">release notes</a> to read about the new features, or <a href="/downloads.html">download</a> the release today.</p>
+<p>We’re happy to announce the release of <a href="/releases/spark-release-0-7-2.html" title="Spark Release 0.7.2">Spark 0.7.2</a>, a new maintenance release that includes several bug fixes and improvements, as well as new code examples and API features. We recommend that all users update to this release. Head over to the <a href="/releases/spark-release-0-7-2.html" title="Spark Release 0.7.2">release notes</a> to read about the new features, or <a href="/downloads.html">download</a> the release today.</p>
<p>
diff --git a/site/news/spark-0-7-3-released.html b/site/news/spark-0-7-3-released.html
index d620662d2..56b23f710 100644
--- a/site/news/spark-0-7-3-released.html
+++ b/site/news/spark-0-7-3-released.html
@@ -160,7 +160,7 @@
<h2>Spark 0.7.3 released</h2>
-<p>We&#8217;ve just posted <a href="/releases/spark-release-0-7-3.html" title="Spark Release 0.7.3">Spark Release 0.7.3</a>, a maintenance release that contains several fixes, including streaming API updates and new functionality for adding JARs to a <code>spark-shell</code> session. We recommend that all users update to this release. Visit the <a href="/releases/spark-release-0-7-3.html" title="Spark Release 0.7.3">release notes</a> to read about the new features, or <a href="/downloads.html">download</a> the release today.</p>
+<p>We’ve just posted <a href="/releases/spark-release-0-7-3.html" title="Spark Release 0.7.3">Spark Release 0.7.3</a>, a maintenance release that contains several fixes, including streaming API updates and new functionality for adding JARs to a <code>spark-shell</code> session. We recommend that all users update to this release. Visit the <a href="/releases/spark-release-0-7-3.html" title="Spark Release 0.7.3">release notes</a> to read about the new features, or <a href="/downloads.html">download</a> the release today.</p>
<p>
diff --git a/site/news/spark-0-8-0-released.html b/site/news/spark-0-8-0-released.html
index a13d7e794..f5991161b 100644
--- a/site/news/spark-0-8-0-released.html
+++ b/site/news/spark-0-8-0-released.html
@@ -160,7 +160,7 @@
<h2>Spark 0.8.0 released</h2>
-<p>We&#8217;re proud to announce the release of <a href="/releases/spark-release-0-8-0.html" title="Spark Release 0.8.0">Apache Spark 0.8.0</a>. Spark 0.8.0 is a major release that includes many new capabilities and usability improvements. It’s also our first release under the Apache incubator. It is the largest Spark release yet, with contributions from 67 developers and 24 companies. Major new features include an expanded monitoring framework and UI, a machine learning library, and support for running Spark inside of YARN.</p>
+<p>We’re proud to announce the release of <a href="/releases/spark-release-0-8-0.html" title="Spark Release 0.8.0">Apache Spark 0.8.0</a>. Spark 0.8.0 is a major release that includes many new capabilities and usability improvements. It’s also our first release under the Apache incubator. It is the largest Spark release yet, with contributions from 67 developers and 24 companies. Major new features include an expanded monitoring framework and UI, a machine learning library, and support for running Spark inside of YARN.</p>
<p>
diff --git a/site/news/spark-0-8-1-released.html b/site/news/spark-0-8-1-released.html
index 5a2871c0d..07270670c 100644
--- a/site/news/spark-0-8-1-released.html
+++ b/site/news/spark-0-8-1-released.html
@@ -160,7 +160,7 @@
<h2>Spark 0.8.1 released</h2>
-<p>We&#8217;ve just posted <a href="/releases/spark-release-0-8-1.html" title="Spark Release 0.8.1">Spark Release 0.8.1</a>, a maintenance and performance release for the Scala 2.9 version of Spark. 0.8.1 includes support for YARN 2.2, a high availability mode for the standalone scheduler, optimizations to the shuffle, and many other improvements. We recommend that all users update to this release. Visit the <a href="/releases/spark-release-0-8-1.html" title="Spark Release 0.8.1">release notes</a> to read about the new features, or <a href="/downloads.html">download</a> the release today.</p>
+<p>We’ve just posted <a href="/releases/spark-release-0-8-1.html" title="Spark Release 0.8.1">Spark Release 0.8.1</a>, a maintenance and performance release for the Scala 2.9 version of Spark. 0.8.1 includes support for YARN 2.2, a high availability mode for the standalone scheduler, optimizations to the shuffle, and many other improvements. We recommend that all users update to this release. Visit the <a href="/releases/spark-release-0-8-1.html" title="Spark Release 0.8.1">release notes</a> to read about the new features, or <a href="/downloads.html">download</a> the release today.</p>
<p>
diff --git a/site/news/spark-0-9-0-released.html b/site/news/spark-0-9-0-released.html
index 753b654f8..6e9849ad8 100644
--- a/site/news/spark-0-9-0-released.html
+++ b/site/news/spark-0-9-0-released.html
@@ -161,8 +161,8 @@
<p>We are happy to announce the availability of <a href="/releases/spark-release-0-9-0.html" title="Spark Release 0.9.0">
-Spark 0.9.0</a>! Spark 0.9.0 is a major release and Spark&#8217;s largest release ever, with contributions from 83 developers.
-This release expands Spark&#8217;s standard libraries, introducing a new graph computation package (GraphX) and adding several new features to the machine learning and stream-processing packages. It also makes major improvements to the core engine,
+Spark 0.9.0</a>! Spark 0.9.0 is a major release and Spark’s largest release ever, with contributions from 83 developers.
+This release expands Spark’s standard libraries, introducing a new graph computation package (GraphX) and adding several new features to the machine learning and stream-processing packages. It also makes major improvements to the core engine,
including external aggregations, a simplified H/A mode for long lived applications, and
hardened YARN support.</p>
diff --git a/site/news/spark-1-0-0-released.html b/site/news/spark-1-0-0-released.html
index 6c7075801..8fcb0a885 100644
--- a/site/news/spark-1-0-0-released.html
+++ b/site/news/spark-1-0-0-released.html
@@ -160,8 +160,8 @@
<h2>Spark 1.0.0 released</h2>
-<p>We are happy to announce the availability of <a href="/releases/spark-release-1-0-0.html" title="Spark Release 1.0.0">Spark 1.0.0</a>! Spark 1.0.0 is the first in the 1.0 line of releases, providing API stability for Spark&#8217;s core interfaces. It is Spark&#8217;s largest release ever, with contributions from 117 developers.
-This release expands Spark&#8217;s standard libraries, introducing a new SQL package (SparkSQL) which lets users integrate SQL queries into existing Spark workflows. MLlib, Spark&#8217;s machine learning library, is expanded with sparse vector support and several new algorithms. The GraphX and Streaming libraries also introduce new features and optimizations. Spark&#8217;s core engine adds support for secured YARN clusters, a unified tool for submitting Spark applications, and several performance and stability improvements.</p>
+<p>We are happy to announce the availability of <a href="/releases/spark-release-1-0-0.html" title="Spark Release 1.0.0">Spark 1.0.0</a>! Spark 1.0.0 is the first in the 1.X line of releases, providing API stability for Spark’s core interfaces. It is Spark’s largest release ever, with contributions from 117 developers.
+This release expands Spark’s standard libraries, introducing a new SQL package (Spark SQL) that lets users integrate SQL queries into existing Spark workflows. MLlib, Spark’s machine learning library, is expanded with sparse vector support and several new algorithms. The GraphX and Streaming libraries also introduce new features and optimizations. Spark’s core engine adds support for secured YARN clusters, a unified tool for submitting Spark applications, and several performance and stability improvements.</p>
<p>Visit the <a href="/releases/spark-release-1-0-0.html" title="Spark Release 1.0.0">release notes</a> to read about the new features, or <a href="/downloads.html">download</a> the release today.</p>
diff --git a/site/news/spark-and-shark-in-the-news.html b/site/news/spark-and-shark-in-the-news.html
index 986cde462..4f3b7d45d 100644
--- a/site/news/spark-and-shark-in-the-news.html
+++ b/site/news/spark-and-shark-in-the-news.html
@@ -160,7 +160,7 @@
<h2>Spark and Shark in the news</h2>
-<p>Recently, we&#8217;ve seen quite a bit of coverage of both Spark and <a href="http://shark.cs.berkeley.edu">Shark</a> in the news. I wanted to list some of the more recent articles, for readers interested in learning more.</p>
+<p>Recently, we’ve seen quite a bit of coverage of both Spark and <a href="http://shark.cs.berkeley.edu">Shark</a> in the news. I wanted to list some of the more recent articles, for readers interested in learning more.</p>
<ul>
<li>Curt Monash, editor of the popular DBMS2 blog, wrote a great <a href="http://www.dbms2.com/2012/12/13/introduction-to-spark-shark-bdas-and-amplab/">introduction to Spark and Shark</a>, as well as a more detailed <a href="http://www.dbms2.com/2012/12/13/spark-shark-and-rdds-technology-notes/">technical overview</a>.</li>
@@ -170,7 +170,7 @@
<li><a href="http://data-informed.com/spark-an-open-source-engine-for-iterative-data-mining/">DataInformed</a> interviewed two Spark users and wrote about their applications in anomaly detection, predictive analytics and data mining.</li>
</ul>
-<p>In other news, there will be a full day of tutorials on Spark and Shark at the <a href="http://strataconf.com/strata2013">O&#8217;Reilly Strata conference</a> in February. They include a three-hour <a href="http://strataconf.com/strata2013/public/schedule/detail/27438">introduction to Spark, Shark and BDAS</a> Tuesday morning, and a three-hour <a href="http://strataconf.com/strata2013/public/schedule/detail/27440">hands-on exercise session</a>. </p>
+<p>In other news, there will be a full day of tutorials on Spark and Shark at the <a href="http://strataconf.com/strata2013">O’Reilly Strata conference</a> in February. They include a three-hour <a href="http://strataconf.com/strata2013/public/schedule/detail/27438">introduction to Spark, Shark and BDAS</a> Tuesday morning, and a three-hour <a href="http://strataconf.com/strata2013/public/schedule/detail/27440">hands-on exercise session</a>. </p>
<p>
diff --git a/site/news/spark-becomes-tlp.html b/site/news/spark-becomes-tlp.html
index 498377eac..0b9319e31 100644
--- a/site/news/spark-becomes-tlp.html
+++ b/site/news/spark-becomes-tlp.html
@@ -160,9 +160,9 @@
<h2>Spark becomes top-level Apache project</h2>
-<p>The Apache Software Foundation <a href="https://blogs.apache.org/foundation/entry/the_apache_software_foundation_announces50">announced</a> today that Spark has graduated from the Apache Incubator to become a top-level Apache project, signifying that the project&#8217;s community and products have been well-governed under the ASF&#8217;s meritocratic process and principles. This is a major step for the community and we are very proud to share this news with users as we complete Spark&#8217;s move to Apache. Read more about Spark&#8217;s growth during the past year and from contributors and users in the ASF&#8217;s <a href="https://blogs.apache.org/foundation/entry/the_apache_software_foundation_announces50">press release</a>.</p>
+<p>The Apache Software Foundation <a href="https://blogs.apache.org/foundation/entry/the_apache_software_foundation_announces50">announced</a> today that Spark has graduated from the Apache Incubator to become a top-level Apache project, signifying that the project’s community and products have been well-governed under the ASF’s meritocratic process and principles. This is a major step for the community and we are very proud to share this news with users as we complete Spark’s move to Apache. Read more about Spark’s growth during the past year and from contributors and users in the ASF’s <a href="https://blogs.apache.org/foundation/entry/the_apache_software_foundation_announces50">press release</a>.</p>
-<p>As part of this change, note that Spark&#8217;s <a href="/community.html">mailing lists</a> have moved to <tt>@spark.apache.org</tt> addresses, although the old <tt>@spark.incubator.apache.org</tt> addresses also still work.</p>
+<p>As part of this change, note that Spark’s <a href="/community.html">mailing lists</a> have moved to <tt>@spark.apache.org</tt> addresses, although the old <tt>@spark.incubator.apache.org</tt> addresses also still work.</p>
<p>
diff --git a/site/news/spark-meetups.html b/site/news/spark-meetups.html
index 7a32542ba..0a3eda9a6 100644
--- a/site/news/spark-meetups.html
+++ b/site/news/spark-meetups.html
@@ -160,7 +160,7 @@
<h2>We've started hosting a Bay Area Spark User Meetup</h2>
-<p>We&#8217;ve started hosting a regular <a href="http://www.meetup.com/spark-users/">Bay Area Spark User Meetup</a>. Sign up on the meetup.com page to be notified about events and meet other Spark developers and users.</p>
+<p>We’ve started hosting a regular <a href="http://www.meetup.com/spark-users/">Bay Area Spark User Meetup</a>. Sign up on the meetup.com page to be notified about events and meet other Spark developers and users.</p>
<p>
diff --git a/site/news/spark-user-survey-and-powered-by-page.html b/site/news/spark-user-survey-and-powered-by-page.html
index c1889ce38..c7a77598f 100644
--- a/site/news/spark-user-survey-and-powered-by-page.html
+++ b/site/news/spark-user-survey-and-powered-by-page.html
@@ -160,9 +160,9 @@
<h2>Spark user survey and "Powered By" page</h2>
-<p>As we continue developing Spark, we would love to get feedback from users and hear what you&#8217;d like us to work on next. We&#8217;ve decided that a good way to do that is a survey &#8211; we hope to run this at regular intervals. If you have a few minutes to participate, <a href="https://docs.google.com/forms/d/1eMXp4GjcIXglxJe5vYYBzXKVm-6AiYt1KThJwhCjJiY/viewform">fill in the survey here</a>. Your time is greatly appreciated.</p>
+<p>As we continue developing Spark, we would love to get feedback from users and hear what you’d like us to work on next. We’ve decided that a good way to do that is a survey – we hope to run this at regular intervals. If you have a few minutes to participate, <a href="https://docs.google.com/forms/d/1eMXp4GjcIXglxJe5vYYBzXKVm-6AiYt1KThJwhCjJiY/viewform">fill in the survey here</a>. Your time is greatly appreciated.</p>
-<p>In parallel, we are starting a <a href="https://cwiki.apache.org/confluence/display/SPARK/Powered+By+Spark">&#8220;powered by&#8221; page</a> on the Apache Spark wiki for organizations that are using, or contributing to, Spark. Sign up if you&#8217;d like to support the project! This is a great way to let the world know you&#8217;re using Spark, and can also be helpful to generate leads for recruiting. You can also add yourself when you fill the survey.</p>
+<p>In parallel, we are starting a <a href="https://cwiki.apache.org/confluence/display/SPARK/Powered+By+Spark">“powered by” page</a> on the Apache Spark wiki for organizations that are using, or contributing to, Spark. Sign up if you’d like to support the project! This is a great way to let the world know you’re using Spark, and can also be helpful to generate leads for recruiting. You can also add yourself when you fill the survey.</p>
<p>Thanks for taking the time to give feedback.</p>
diff --git a/site/news/strata-exercises-now-available-online.html b/site/news/strata-exercises-now-available-online.html
index 402cdb0bf..41567683b 100644
--- a/site/news/strata-exercises-now-available-online.html
+++ b/site/news/strata-exercises-now-available-online.html
@@ -160,7 +160,7 @@
<h2>Strata exercises now available online</h2>
-<p>At this year&#8217;s <a href="http://strataconf.com/strata2013">Strata</a> conference, the AMP Lab hosted a full day of tutorials on Spark, Shark, and Spark Streaming, including online exercises on Amazon EC2. Those exercises are now <a href="http://ampcamp.berkeley.edu/big-data-mini-course/">available online</a>, letting you learn Spark and Shark at your own pace on an EC2 cluster with real data. They are a great resource for learning the systems. You can also find <a href="http://ampcamp.berkeley.edu/amp-camp-two-strata-2013/">slides</a> from the Strata tutorials online, as well as <a href="http://ampcamp.berkeley.edu/amp-camp-one-berkeley-2012/">videos</a> from the AMP Camp workshop we held at Berkeley in August.</p>
+<p>At this year’s <a href="http://strataconf.com/strata2013">Strata</a> conference, the AMP Lab hosted a full day of tutorials on Spark, Shark, and Spark Streaming, including online exercises on Amazon EC2. Those exercises are now <a href="http://ampcamp.berkeley.edu/big-data-mini-course/">available online</a>, letting you learn Spark and Shark at your own pace on an EC2 cluster with real data. They are a great resource for learning the systems. You can also find <a href="http://ampcamp.berkeley.edu/amp-camp-two-strata-2013/">slides</a> from the Strata tutorials online, as well as <a href="http://ampcamp.berkeley.edu/amp-camp-one-berkeley-2012/">videos</a> from the AMP Camp workshop we held at Berkeley in August.</p>
<p>
diff --git a/site/news/submit-talks-to-spark-summit-2014.html b/site/news/submit-talks-to-spark-summit-2014.html
index 90c8597a5..15ec7f3d6 100644
--- a/site/news/submit-talks-to-spark-summit-2014.html
+++ b/site/news/submit-talks-to-spark-summit-2014.html
@@ -160,12 +160,12 @@
<h2>Submissions and registration open for Spark Summit 2014</h2>
-<p>After last year&#8217;s successful <a href="http://spark-summit.org/2013">first Spark Summit</a>, registrations
+<p>After last year’s successful <a href="http://spark-summit.org/2013">first Spark Summit</a>, registrations
and talk submissions are now open for <a href="http://spark-summit.org/2014">Spark Summit 2014</a>.
This will be a 3-day event in San Francisco organized by multiple companies in the Spark community.
The event will run <strong>June 30th to July 2nd</strong> in San Francisco, CA.</p>
-<p>If you&#8217;d like to present at the Summit, <a href="http://spark-summit.org/submit">submit a talk</a>
+<p>If you’d like to present at the Summit, <a href="http://spark-summit.org/submit">submit a talk</a>
before April 11th, 2014. We welcome talks on use cases, open source development, and applications built
on Spark.</p>
diff --git a/site/news/video-from-first-spark-development-meetup.html b/site/news/video-from-first-spark-development-meetup.html
index 40fb9ef15..a7cf13206 100644
--- a/site/news/video-from-first-spark-development-meetup.html
+++ b/site/news/video-from-first-spark-development-meetup.html
@@ -160,7 +160,7 @@
<h2>Video up from first Spark development meetup</h2>
-<p>On December 18th, we held the first of a series of Spark development meetups, for people interested in learning the Spark codebase and contributing to the project. There was quite a bit more demand than we anticipated, with over 80 people signing up and 64 attending. The first meetup was an <a href="http://www.meetup.com/spark-users/events/94101942/">introduction to Spark internals</a>. Thanks to one of the attendees, there&#8217;s now a <a href="http://www.youtube.com/watch?v=49Hr5xZyTEA">video of the meetup</a> on YouTube. We&#8217;ve also posted the <a href="http://files.meetup.com/3138542/dev-meetup-dec-2012.pptx">slides</a>. Look to see more development meetups on Spark and Shark in the future.</p>
+<p>On December 18th, we held the first of a series of Spark development meetups, for people interested in learning the Spark codebase and contributing to the project. There was quite a bit more demand than we anticipated, with over 80 people signing up and 64 attending. The first meetup was an <a href="http://www.meetup.com/spark-users/events/94101942/">introduction to Spark internals</a>. Thanks to one of the attendees, there’s now a <a href="http://www.youtube.com/watch?v=49Hr5xZyTEA">video of the meetup</a> on YouTube. We’ve also posted the <a href="http://files.meetup.com/3138542/dev-meetup-dec-2012.pptx">slides</a>. Look to see more development meetups on Spark and Shark in the future.</p>
<p>
diff --git a/site/releases/spark-release-0-3.html b/site/releases/spark-release-0-3.html
index 391f70060..b69ea3857 100644
--- a/site/releases/spark-release-0-3.html
+++ b/site/releases/spark-release-0-3.html
@@ -176,7 +176,7 @@
<h3>Native Types for SequenceFiles</h3>
-<p>In working with SequenceFiles, which store objects that implement Hadoop&#8217;s Writable interface, Spark will now let you use native types for certain common Writable types, like IntWritable and Text. For example:</p>
+<p>In working with SequenceFiles, which store objects that implement Hadoop’s Writable interface, Spark will now let you use native types for certain common Writable types, like IntWritable and Text. For example:</p>
<div class="code">
<span class="comment">// Will read a SequenceFile of (IntWritable, Text)</span><br />
diff --git a/site/releases/spark-release-0-5-0.html b/site/releases/spark-release-0-5-0.html
index 80b168e87..5c92ea212 100644
--- a/site/releases/spark-release-0-5-0.html
+++ b/site/releases/spark-release-0-5-0.html
@@ -164,10 +164,10 @@
<h3>Mesos 0.9 Support</h3>
-<p>This release runs on <a href="http://www.mesosproject.org/">Apache Mesos 0.9</a>, the first Apache Incubator release of Mesos, which contains significant usability and stability improvements. Most notable are better memory accounting for applications with long-term memory use, easier access of old jobs&#8217; traces and logs (by keeping a history of executed tasks on the web UI), and simpler installation.</p>
+<p>This release runs on <a href="http://www.mesosproject.org/">Apache Mesos 0.9</a>, the first Apache Incubator release of Mesos, which contains significant usability and stability improvements. Most notable are better memory accounting for applications with long-term memory use, easier access of old jobs’ traces and logs (by keeping a history of executed tasks on the web UI), and simpler installation.</p>
<h3>Performance Improvements</h3>
-<p>Spark&#8217;s scheduling is more communication-efficient when sending out operations on RDDs with large lineage graphs. In addition, the cache replacement policy has been improved to more smartly replace data when an RDD does not fit in the cache, shuffles are more efficient, and the serializer used for shipping closures is now configurable, making it possible to use faster libraries than Java serialization there.</p>
+<p>Spark’s scheduling is more communication-efficient when sending out operations on RDDs with large lineage graphs. In addition, the cache replacement policy has been improved to more smartly replace data when an RDD does not fit in the cache, shuffles are more efficient, and the serializer used for shipping closures is now configurable, making it possible to use faster libraries than Java serialization there.</p>
<h3>Debug Improvements</h3>
@@ -179,11 +179,11 @@
<h3>EC2 Launch Script Improvements</h3>
-<p>Spark&#8217;s EC2 launch scripts are now included in the main package, and have the ability to discover and use the latest Spark AMI automatically instead of launching a hardcoded machine image ID.</p>
+<p>Spark’s EC2 launch scripts are now included in the main package, and have the ability to discover and use the latest Spark AMI automatically instead of launching a hardcoded machine image ID.</p>
<h3>New Hadoop API Support</h3>
-<p>You can now use Spark to read and write data to storage formats in the new <tt>org.apache.mapreduce</tt> packages (the &#8220;new Hadoop&#8221; API). In addition, this release fixes an issue caused by a HDFS initialization bug in some recent versions of HDFS.</p>
+<p>You can now use Spark to read and write data to storage formats in the new <tt>org.apache.mapreduce</tt> packages (the “new Hadoop” API). In addition, this release fixes an issue caused by a HDFS initialization bug in some recent versions of HDFS.</p>
<p>
diff --git a/site/releases/spark-release-0-5-1.html b/site/releases/spark-release-0-5-1.html
index 786dcc113..925d320e3 100644
--- a/site/releases/spark-release-0-5-1.html
+++ b/site/releases/spark-release-0-5-1.html
@@ -193,7 +193,7 @@
<h3>EC2 Improvements</h3>
-<p>Spark&#8217;s EC2 launch script now configures Spark&#8217;s memory limit automatically based on the machine&#8217;s available RAM.</p>
+<p>Spark’s EC2 launch script now configures Spark’s memory limit automatically based on the machine’s available RAM.</p>
<p>
diff --git a/site/releases/spark-release-0-6-0.html b/site/releases/spark-release-0-6-0.html
index 09ef0c179..016a19514 100644
--- a/site/releases/spark-release-0-6-0.html
+++ b/site/releases/spark-release-0-6-0.html
@@ -172,11 +172,11 @@
<h3>Java API</h3>
-<p>Java programmers can now use Spark through a new <a href="/docs/0.6.0/java-programming-guide.html">Java API layer</a>. This layer makes available all of Spark&#8217;s features, including parallel transformations, distributed datasets, broadcast variables, and accumulators, in a Java-friendly manner.</p>
+<p>Java programmers can now use Spark through a new <a href="/docs/0.6.0/java-programming-guide.html">Java API layer</a>. This layer makes available all of Spark’s features, including parallel transformations, distributed datasets, broadcast variables, and accumulators, in a Java-friendly manner.</p>
<h3>Expanded Documentation</h3>
-<p>Spark&#8217;s <a href="/docs/0.6.0/">documentation</a> has been expanded with a new <a href="/docs/0.6.0/quick-start.html">quick start guide</a>, additional deployment instructions, configuration guide, tuning guide, and improved <a href="/docs/0.6.0/api/core">Scaladoc</a> API documentation.</p>
+<p>Spark’s <a href="/docs/0.6.0/">documentation</a> has been expanded with a new <a href="/docs/0.6.0/quick-start.html">quick start guide</a>, additional deployment instructions, configuration guide, tuning guide, and improved <a href="/docs/0.6.0/api/core">Scaladoc</a> API documentation.</p>
<h3>Engine Changes</h3>
@@ -199,7 +199,7 @@
<h3>Enhanced Debugging</h3>
-<p>Spark&#8217;s log now prints which operation in your program each RDD and job described in your logs belongs to, making it easier to tie back to which parts of your code experience problems.</p>
+<p>Spark’s log now prints which operation in your program each RDD and job described in your logs belongs to, making it easier to tie back to which parts of your code experience problems.</p>
<h3>Maven Artifacts</h3>
diff --git a/site/releases/spark-release-0-7-0.html b/site/releases/spark-release-0-7-0.html
index 35fa8ae2f..d485f7858 100644
--- a/site/releases/spark-release-0-7-0.html
+++ b/site/releases/spark-release-0-7-0.html
@@ -186,7 +186,7 @@
<h3>New Operations</h3>
-<p>This release adds several RDD transformations, including <tt>keys</tt>, <tt>values</tt>, <tt>keyBy</tt>, <tt>subtract</tt>, <tt>coalesce</tt>, <tt>zip</tt>. It also adds <tt>SparkContext.hadoopConfiguration</tt> to allow programs to configure Hadoop input/output settings globally across operations. Finally, it adds the <tt>RDD.toDebugString()</tt> method, which can be used to print an RDD&#8217;s lineage graph for troubleshooting.</p>
+<p>This release adds several RDD transformations, including <tt>keys</tt>, <tt>values</tt>, <tt>keyBy</tt>, <tt>subtract</tt>, <tt>coalesce</tt>, <tt>zip</tt>. It also adds <tt>SparkContext.hadoopConfiguration</tt> to allow programs to configure Hadoop input/output settings globally across operations. Finally, it adds the <tt>RDD.toDebugString()</tt> method, which can be used to print an RDD’s lineage graph for troubleshooting.</p>
<h3>EC2 Improvements</h3>
@@ -223,7 +223,7 @@
<h3>Credits</h3>
-<p>Spark 0.7 was the work of many contributors from Berkeley and outside&#8212;in total, 31 different contributors, of which 20 were from outside Berkeley. Here are the people who contributed, along with areas they worked on:</p>
+<p>Spark 0.7 was the work of many contributors from Berkeley and outside—in total, 31 different contributors, of which 20 were from outside Berkeley. Here are the people who contributed, along with areas they worked on:</p>
<ul>
<li>Mikhail Bautin -- Maven build</li>
diff --git a/site/releases/spark-release-0-8-0.html b/site/releases/spark-release-0-8-0.html
index f89581470..a79417123 100644
--- a/site/releases/spark-release-0-8-0.html
+++ b/site/releases/spark-release-0-8-0.html
@@ -204,13 +204,13 @@
<li>The examples build has been isolated from the core build, substantially reducing the potential for dependency conflicts.</li>
<li>The Spark Streaming Twitter API has been updated to use OAuth authentication instead of the deprecated username/password authentication in Spark 0.7.0.</li>
<li>Several new example jobs have been added, including PageRank implementations in Java, Scala and Python, examples for accessing HBase and Cassandra, and MLlib examples.</li>
- <li>Support for running on Mesos has been improved &#8211; now you can deploy a Spark assembly JAR as part of the Mesos job, instead of having Spark pre-installed on each machine. The default Mesos version has also been updated to 0.13.</li>
+ <li>Support for running on Mesos has been improved – now you can deploy a Spark assembly JAR as part of the Mesos job, instead of having Spark pre-installed on each machine. The default Mesos version has also been updated to 0.13.</li>
<li>This release includes various optimizations to PySpark and to the job scheduler.</li>
</ul>
<h3 id="compatibility">Compatibility</h3>
<ul>
- <li><strong>This release changes Spark’s package name to &#8216;org.apache.spark&#8217;</strong>, so those upgrading from Spark 0.7 will need to adjust their imports accordingly. In addition, we’ve moved the <code>RDD</code> class to the org.apache.spark.rdd package (it was previously in the top-level package). The Spark artifacts published through Maven have also changed to the new package name.</li>
+ <li><strong>This release changes Spark’s package name to ‘org.apache.spark’</strong>, so those upgrading from Spark 0.7 will need to adjust their imports accordingly. In addition, we’ve moved the <code>RDD</code> class to the org.apache.spark.rdd package (it was previously in the top-level package). The Spark artifacts published through Maven have also changed to the new package name.</li>
<li>In the Java API, use of Scala’s <code>Option</code> class has been replaced with <code>Optional</code> from the Guava library.</li>
<li>Linking against Spark for arbitrary Hadoop versions is now possible by specifying a dependency on <code>hadoop-client</code>, instead of rebuilding <code>spark-core</code> against your version of Hadoop. See the documentation <a href="http://spark.incubator.apache.org/docs/0.8.0/scala-programming-guide.html#linking-with-spark">here</a> for details.</li>
<li>If you are building Spark, you’ll now need to run <code>sbt/sbt assembly</code> instead of <code>package</code>.</li>
@@ -220,73 +220,73 @@
<p>Spark 0.8.0 was the result of the largest team of contributors yet. The following developers contributed to this release:</p>
<ul>
- <li>Andrew Ash &#8211; documentation, code cleanup and logging improvements</li>
- <li>Mikhail Bautin &#8211; bug fix</li>
- <li>Konstantin Boudnik &#8211; Maven build, bug fixes, and documentation</li>
- <li>Ian Buss &#8211; sbt configuration improvement</li>
- <li>Evan Chan &#8211; API improvement, bug fix, and documentation</li>
- <li>Lian Cheng &#8211; bug fix</li>
- <li>Tathagata Das &#8211; performance improvement in streaming receiver and streaming bug fix</li>
- <li>Aaron Davidson &#8211; Python improvements, bug fix, and unit tests</li>
- <li>Giovanni Delussu &#8211; coalesced RDD feature</li>
- <li>Joseph E. Gonzalez &#8211; improvement to zipPartitions</li>
- <li>Karen Feng &#8211; several improvements to web UI</li>
- <li>Andy Feng &#8211; HDFS metrics</li>
- <li>Ali Ghodsi &#8211; configuration improvements and locality-aware coalesce</li>
- <li>Christoph Grothaus &#8211; bug fix</li>
- <li>Thomas Graves &#8211; support for secure YARN cluster and various YARN-related improvements</li>
- <li>Stephen Haberman &#8211; bug fix, documentation, and code cleanup</li>
- <li>Mark Hamstra &#8211; bug fixes and Maven build</li>
- <li>Benjamin Hindman &#8211; Mesos compatibility and documentation</li>
- <li>Liang-Chi Hsieh &#8211; bug fixes in build and in YARN mode</li>
- <li>Shane Huang &#8211; shuffle improvements, bug fix</li>
- <li>Ethan Jewett &#8211; Spark/HBase example</li>
- <li>Holden Karau &#8211; bug fix and EC2 improvement</li>
- <li>Kody Koeniger &#8211; JDBV RDD implementation</li>
- <li>Andy Konwinski &#8211; documentation</li>
- <li>Jey Kottalam &#8211; PySpark optimizations, Hadoop agnostic build (lead), and bug fixes</li>
- <li>Andrey Kouznetsov &#8211; Bug fix</li>
- <li>S. Kumar &#8211; Spark Streaming example</li>
- <li>Ryan LeCompte &#8211; topK method optimization and serialization improvements</li>
- <li>Gavin Li &#8211; compression codecs and pipe support</li>
- <li>Harold Lim &#8211; fair scheduler</li>
- <li>Dmitriy Lyubimov &#8211; bug fix</li>
- <li>Chris Mattmann &#8211; Apache mentor</li>
- <li>David McCauley &#8211; JSON API improvement</li>
- <li>Sean McNamara &#8211; added <code>takeOrdered</code> function, bug fixes, and a build fix</li>
- <li>Mridul Muralidharan &#8211; YARN integration (lead) and scheduler improvements</li>
- <li>Marc Mercer &#8211; improvements to UI json output</li>
- <li>Christopher Nguyen &#8211; bug fixes</li>
- <li>Erik van Oosten &#8211; example fix</li>
- <li>Kay Ousterhout &#8211; fix for scheduler regression and bug fixes</li>
- <li>Xinghao Pan &#8211; MLLib contributions</li>
- <li>Hiral Patel &#8211; bug fix</li>
- <li>James Phillpotts &#8211; updated Twitter API for Spark streaming</li>
- <li>Nick Pentreath &#8211; scala pageRank example, bagel improvement, and several Java examples</li>
- <li>Alexander Pivovarov &#8211; logging improvement and Maven build</li>
- <li>Mike Potts &#8211; configuration improvement</li>
- <li>Rohit Rai &#8211; Spark/Cassandra example</li>
- <li>Imran Rashid &#8211; bug fixes and UI improvement</li>
- <li>Charles Reiss &#8211; bug fixes, code cleanup, performance improvements</li>
- <li>Josh Rosen &#8211; Python API improvements, Java API improvements, EC2 scripts and bug fixes</li>
- <li>Henry Saputra &#8211; Apache mentor</li>
- <li>Jerry Shao &#8211; bug fixes, metrics system</li>
- <li>Prashant Sharma &#8211; documentation</li>
- <li>Mingfei Shi &#8211; joblogger and bug fix</li>
- <li>Andre Schumacher &#8211; several PySpark features</li>
- <li>Ginger Smith &#8211; MLLib contribution</li>
- <li>Evan Sparks &#8211; contributions to MLLib</li>
- <li>Ram Sriharsha &#8211; bug fix and RDD removal feature</li>
- <li>Ameet Talwalkar &#8211; MLlib contributions</li>
- <li>Roman Tkalenko &#8211; code refactoring and cleanup</li>
- <li>Chu Tong &#8211; Java PageRank algorithm and bug fix in bash scripts</li>
- <li>Shivaram Venkataraman &#8211; bug fixes, contributions to MLLib, netty shuffle fixes, and Java API additions</li>
- <li>Patrick Wendell &#8211; release manager, bug fixes, documentation, metrics system, and web UI</li>
- <li>Andrew Xia &#8211; fair scheduler (lead), metrics system, and ui improvements</li>
- <li>Reynold Xin &#8211; shuffle improvements, bug fixes, code refactoring, usability improvements, MLLib contributions</li>
- <li>Matei Zaharia &#8211; MLLib contributions, documentation, examples, UI improvements, PySpark improvements, and bug fixes</li>
- <li>Wu Zeming &#8211; bug fix in scheduler</li>
- <li>Bill Zhao &#8211; log message improvement</li>
+ <li>Andrew Ash – documentation, code cleanup and logging improvements</li>
+ <li>Mikhail Bautin – bug fix</li>
+ <li>Konstantin Boudnik – Maven build, bug fixes, and documentation</li>
+ <li>Ian Buss – sbt configuration improvement</li>
+ <li>Evan Chan – API improvement, bug fix, and documentation</li>
+ <li>Lian Cheng – bug fix</li>
+ <li>Tathagata Das – performance improvement in streaming receiver and streaming bug fix</li>
+ <li>Aaron Davidson – Python improvements, bug fix, and unit tests</li>
+ <li>Giovanni Delussu – coalesced RDD feature</li>
+ <li>Joseph E. Gonzalez – improvement to zipPartitions</li>
+ <li>Karen Feng – several improvements to web UI</li>
+ <li>Andy Feng – HDFS metrics</li>
+ <li>Ali Ghodsi – configuration improvements and locality-aware coalesce</li>
+ <li>Christoph Grothaus – bug fix</li>
+ <li>Thomas Graves – support for secure YARN cluster and various YARN-related improvements</li>
+ <li>Stephen Haberman – bug fix, documentation, and code cleanup</li>
+ <li>Mark Hamstra – bug fixes and Maven build</li>
+ <li>Benjamin Hindman – Mesos compatibility and documentation</li>
+ <li>Liang-Chi Hsieh – bug fixes in build and in YARN mode</li>
+ <li>Shane Huang – shuffle improvements, bug fix</li>
+ <li>Ethan Jewett – Spark/HBase example</li>
+ <li>Holden Karau – bug fix and EC2 improvement</li>
+ <li>Kody Koeniger – JDBV RDD implementation</li>
+ <li>Andy Konwinski – documentation</li>
+ <li>Jey Kottalam – PySpark optimizations, Hadoop agnostic build (lead), and bug fixes</li>
+ <li>Andrey Kouznetsov – Bug fix</li>
+ <li>S. Kumar – Spark Streaming example</li>
+ <li>Ryan LeCompte – topK method optimization and serialization improvements</li>
+ <li>Gavin Li – compression codecs and pipe support</li>
+ <li>Harold Lim – fair scheduler</li>
+ <li>Dmitriy Lyubimov – bug fix</li>
+ <li>Chris Mattmann – Apache mentor</li>
+ <li>David McCauley – JSON API improvement</li>
+ <li>Sean McNamara – added <code>takeOrdered</code> function, bug fixes, and a build fix</li>
+ <li>Mridul Muralidharan – YARN integration (lead) and scheduler improvements</li>
+ <li>Marc Mercer – improvements to UI json output</li>
+ <li>Christopher Nguyen – bug fixes</li>
+ <li>Erik van Oosten – example fix</li>
+ <li>Kay Ousterhout – fix for scheduler regression and bug fixes</li>
+ <li>Xinghao Pan – MLLib contributions</li>
+ <li>Hiral Patel – bug fix</li>
+ <li>James Phillpotts – updated Twitter API for Spark streaming</li>
+ <li>Nick Pentreath – scala pageRank example, bagel improvement, and several Java examples</li>
+ <li>Alexander Pivovarov – logging improvement and Maven build</li>
+ <li>Mike Potts – configuration improvement</li>
+ <li>Rohit Rai – Spark/Cassandra example</li>
+ <li>Imran Rashid – bug fixes and UI improvement</li>
+ <li>Charles Reiss – bug fixes, code cleanup, performance improvements</li>
+ <li>Josh Rosen – Python API improvements, Java API improvements, EC2 scripts and bug fixes</li>
+ <li>Henry Saputra – Apache mentor</li>
+ <li>Jerry Shao – bug fixes, metrics system</li>
+ <li>Prashant Sharma – documentation</li>
+ <li>Mingfei Shi – joblogger and bug fix</li>
+ <li>Andre Schumacher – several PySpark features</li>
+ <li>Ginger Smith – MLLib contribution</li>
+ <li>Evan Sparks – contributions to MLLib</li>
+ <li>Ram Sriharsha – bug fix and RDD removal feature</li>
+ <li>Ameet Talwalkar – MLlib contributions</li>
+ <li>Roman Tkalenko – code refactoring and cleanup</li>
+ <li>Chu Tong – Java PageRank algorithm and bug fix in bash scripts</li>
+ <li>Shivaram Venkataraman – bug fixes, contributions to MLLib, netty shuffle fixes, and Java API additions</li>
+ <li>Patrick Wendell – release manager, bug fixes, documentation, metrics system, and web UI</li>
+ <li>Andrew Xia – fair scheduler (lead), metrics system, and ui improvements</li>
+ <li>Reynold Xin – shuffle improvements, bug fixes, code refactoring, usability improvements, MLLib contributions</li>
+ <li>Matei Zaharia – MLLib contributions, documentation, examples, UI improvements, PySpark improvements, and bug fixes</li>
+ <li>Wu Zeming – bug fix in scheduler</li>
+ <li>Bill Zhao – log message improvement</li>
</ul>
<p>Thanks to everyone who contributed!
diff --git a/site/releases/spark-release-0-8-1.html b/site/releases/spark-release-0-8-1.html
index e3bcbbf3f..2017e7039 100644
--- a/site/releases/spark-release-0-8-1.html
+++ b/site/releases/spark-release-0-8-1.html
@@ -163,7 +163,7 @@
<p>Apache Spark 0.8.1 is a maintenance and performance release for the Scala 2.9 version of Spark. It also adds several new features, such as standalone mode high availability, that will appear in Spark 0.9 but developers wanted to have in Scala 2.9. Contributions to 0.8.1 came from 41 developers.</p>
<h3 id="yarn-22-support">YARN 2.2 Support</h3>
-<p>Support has been added for running Spark on YARN 2.2 and newer. Due to a change in the YARN API between previous versions and 2.2+, this was not supported in Spark 0.8.0. See the <a href="/docs/0.8.1/running-on-yarn.html">YARN documentation</a> for specific instructions on how to build Spark for YARN 2.2+. We&#8217;ve also included a pre-compiled binary for YARN 2.2.</p>
+<p>Support has been added for running Spark on YARN 2.2 and newer. Due to a change in the YARN API between previous versions and 2.2+, this was not supported in Spark 0.8.0. See the <a href="/docs/0.8.1/running-on-yarn.html">YARN documentation</a> for specific instructions on how to build Spark for YARN 2.2+. We’ve also included a pre-compiled binary for YARN 2.2.</p>
<h3 id="high-availability-mode-for-standalone-cluster-manager">High Availability Mode for Standalone Cluster Manager</h3>
<p>The standalone cluster manager now has a high availability (H/A) mode which can tolerate master failures. This is particularly useful for long-running applications such as streaming jobs and the shark server, where the scheduler master previously represented a single point of failure. Instructions for deploying H/A mode are included <a href="/docs/0.8.1/spark-standalone.html#high-availability">in the documentation</a>. The current implementation uses Zookeeper for coordination.</p>
@@ -174,7 +174,7 @@
<ul>
<li>Optimized hashtables for shuffle data - reduces memory and CPU consumption</li>
<li>Efficient encoding for JobConfs - improves latency for stages reading large numbers of blocks from HDFS, S3, and HBase</li>
- <li>Shuffle file consolidation (off by default) - reduces the number of files created in large shuffles for better filesystem performance. This change works best on filesystems newer than ext3 (we recommend ext4 or XFS), and it will be the default in Spark 0.9, but we’ve left it off by default for compatibility. We recommend users turn this on unless they are using ext3 by setting <code>spark.shuffle.consolidateFiles</code> to &#8220;true&#8221;.</li>
+ <li>Shuffle file consolidation (off by default) - reduces the number of files created in large shuffles for better filesystem performance. This change works best on filesystems newer than ext3 (we recommend ext4 or XFS), and it will be the default in Spark 0.9, but we’ve left it off by default for compatibility. We recommend users turn this on unless they are using ext3 by setting <code>spark.shuffle.consolidateFiles</code> to “true”.</li>
<li>Torrent broadcast (off by default) - a faster broadcast implementation for large objects.</li>
<li>Support for fetching large result sets - allows tasks to return large results without tuning Akka buffer sizes.</li>
</ul>
@@ -211,47 +211,47 @@
<h3 id="credits">Credits</h3>
<ul>
- <li>Michael Armbrust &#8211; build fix</li>
- <li>Pierre Borckmans &#8211; typo fix in documentation</li>
- <li>Evan Chan &#8211; <code>local://</code> scheme for dependency jars</li>
- <li>Ewen Cheslack-Postava &#8211; <code>add</code> method for python accumulators, support for setting config properties in python</li>
- <li>Mosharaf Chowdhury &#8211; optimized broadcast implementation</li>
- <li>Frank Dai &#8211; documentation fix</li>
- <li>Aaron Davidson &#8211; shuffle file consolidation, H/A mode for standalone scheduler, cleaned up representation of block IDs, several improvements and bug fixes</li>
- <li>Tathagata Das &#8211; new streaming operators, fix for kafka concurrency bug</li>
- <li>Ankur Dave &#8211; support for pausing spot clusters on EC2</li>
- <li>Harvey Feng &#8211; optimization to JobConf broadcasts, bug fixes, YARN 2.2 build</li>
- <li>Ali Ghodsi &#8211; YARN 2.2 build</li>
- <li>Thomas Graves &#8211; Spark YARN integration including secure HDFS access over YARN</li>
- <li>Li Guoqiang &#8211; fix for Maven build</li>
- <li>Stephen Haberman &#8211; bug fix</li>
- <li>Haidar Hadi &#8211; documentation fix</li>
- <li>Nathan Howell &#8211; bug fix relating to YARN</li>
- <li>Holden Karau &#8211; Java version of <code>mapPartitionsWithIndex</code></li>
- <li>Du Li &#8211; bug fix in make-distrubion.sh</li>
- <li>Raymond Liu &#8211; work on YARN 2.2 build</li>
- <li>Xi Liu &#8211; bug fix and code clean-up</li>
- <li>David McCauley &#8211; bug fix in standalone mode JSON output</li>
- <li>Michael (wannabeast) &#8211; bug fix in memory store</li>
- <li>Fabrizio Milo &#8211; typos in documentation, clean-up in DAGScheduler, typo in scaladoc</li>
- <li>Mridul Muralidharan &#8211; fixes to metadata cleaner and speculative execution</li>
- <li>Sundeep Narravula &#8211; build fix, bug fixes in scheduler and tests, code clean-up</li>
- <li>Kay Ousterhout &#8211; optimized result fetching, new information in UI, scheduler clean-up and bug fixes</li>
- <li>Nick Pentreath &#8211; implicit feedback variant of ALS algorithm</li>
- <li>Imran Rashid &#8211; improvement to executor launch</li>
- <li>Ahir Reddy &#8211; spark support for SIMR</li>
- <li>Josh Rosen &#8211; memory use optimization, clean up of BlockManager code, Java and Python clean-up/fixes</li>
- <li>Henry Saputra &#8211; build fix</li>
- <li>Jerry Shao &#8211; refactoring of fair scheduler, support for running Spark as a specific user, bug fix</li>
- <li>Mingfei Shi &#8211; documentation for JobLogger</li>
- <li>Andre Schumacher &#8211; sortByKey in PySpark and associated changes</li>
- <li>Karthik Tunga &#8211; bug fix in launch script</li>
- <li>Patrick Wendell &#8211; <code>repartition</code> operator, shuffle write metrics, various fixes and release management</li>
- <li>Neal Wiggins &#8211; import clean-up, documentation fixes</li>
- <li>Andrew Xia &#8211; bug fix in UI</li>
- <li>Reynold Xin &#8211; task killing, support for setting job properties in Spark shell, logging improvements, Kryo improvements, several bug fixes</li>
- <li>Matei Zaharia &#8211; optimized hashmap for shuffle data, PySpark documentation, optimizations to Kryo serializer</li>
- <li>Wu Zeming &#8211; bug fix in executors UI</li>
+ <li>Michael Armbrust – build fix</li>
+ <li>Pierre Borckmans – typo fix in documentation</li>
+ <li>Evan Chan – <code>local://</code> scheme for dependency jars</li>
+ <li>Ewen Cheslack-Postava – <code>add</code> method for python accumulators, support for setting config properties in python</li>
+ <li>Mosharaf Chowdhury – optimized broadcast implementation</li>
+ <li>Frank Dai – documentation fix</li>
+ <li>Aaron Davidson – shuffle file consolidation, H/A mode for standalone scheduler, cleaned up representation of block IDs, several improvements and bug fixes</li>
+ <li>Tathagata Das – new streaming operators, fix for kafka concurrency bug</li>
+ <li>Ankur Dave – support for pausing spot clusters on EC2</li>
+ <li>Harvey Feng – optimization to JobConf broadcasts, bug fixes, YARN 2.2 build</li>
+ <li>Ali Ghodsi – YARN 2.2 build</li>
+ <li>Thomas Graves – Spark YARN integration including secure HDFS access over YARN</li>
+ <li>Li Guoqiang – fix for Maven build</li>
+ <li>Stephen Haberman – bug fix</li>
+ <li>Haidar Hadi – documentation fix</li>
+ <li>Nathan Howell – bug fix relating to YARN</li>
+ <li>Holden Karau – Java version of <code>mapPartitionsWithIndex</code></li>
+ <li>Du Li – bug fix in make-distrubion.sh</li>
+ <li>Raymond Liu – work on YARN 2.2 build</li>
+ <li>Xi Liu – bug fix and code clean-up</li>
+ <li>David McCauley – bug fix in standalone mode JSON output</li>
+ <li>Michael (wannabeast) – bug fix in memory store</li>
+ <li>Fabrizio Milo – typos in documentation, clean-up in DAGScheduler, typo in scaladoc</li>
+ <li>Mridul Muralidharan – fixes to metadata cleaner and speculative execution</li>
+ <li>Sundeep Narravula – build fix, bug fixes in scheduler and tests, code clean-up</li>
+ <li>Kay Ousterhout – optimized result fetching, new information in UI, scheduler clean-up and bug fixes</li>
+ <li>Nick Pentreath – implicit feedback variant of ALS algorithm</li>
+ <li>Imran Rashid – improvement to executor launch</li>
+ <li>Ahir Reddy – spark support for SIMR</li>
+ <li>Josh Rosen – memory use optimization, clean up of BlockManager code, Java and Python clean-up/fixes</li>
+ <li>Henry Saputra – build fix</li>
+ <li>Jerry Shao – refactoring of fair scheduler, support for running Spark as a specific user, bug fix</li>
+ <li>Mingfei Shi – documentation for JobLogger</li>
+ <li>Andre Schumacher – sortByKey in PySpark and associated changes</li>
+ <li>Karthik Tunga – bug fix in launch script</li>
+ <li>Patrick Wendell – <code>repartition</code> operator, shuffle write metrics, various fixes and release management</li>
+ <li>Neal Wiggins – import clean-up, documentation fixes</li>
+ <li>Andrew Xia – bug fix in UI</li>
+ <li>Reynold Xin – task killing, support for setting job properties in Spark shell, logging improvements, Kryo improvements, several bug fixes</li>
+ <li>Matei Zaharia – optimized hashmap for shuffle data, PySpark documentation, optimizations to Kryo serializer</li>
+ <li>Wu Zeming – bug fix in executors UI</li>
</ul>
<p>Thanks to everyone who contributed!</p>
diff --git a/site/releases/spark-release-0-9-0.html b/site/releases/spark-release-0-9-0.html
index 5e3b6e0e8..12dc9594b 100644
--- a/site/releases/spark-release-0-9-0.html
+++ b/site/releases/spark-release-0-9-0.html
@@ -124,6 +124,9 @@
<h5>Latest News</h5>
<ul class="list-unstyled">
+ <li><a href="/news/spark-1-0-0-released.html">Spark 1.0.0 released</a>
+ <span class="small">(May 30, 2014)</span></li>
+
<li><a href="/news/spark-summit-agenda-posted.html">Spark Summit agenda posted</a>
<span class="small">(May 11, 2014)</span></li>
@@ -133,9 +136,6 @@
<li><a href="/news/submit-talks-to-spark-summit-2014.html">Submissions and registration open for Spark Summit 2014</a>
<span class="small">(Mar 20, 2014)</span></li>
- <li><a href="/news/spark-becomes-tlp.html">Spark becomes top-level Apache project</a>
- <span class="small">(Feb 27, 2014)</span></li>
-
</ul>
<p class="small" style="text-align: right;"><a href="/news/index.html">Archive</a></p>
</div>
@@ -265,87 +265,87 @@
<p>The following developers contributed to this release:</p>
<ul>
- <li>Andrew Ash &#8211; documentation improvements</li>
- <li>Pierre Borckmans &#8211; documentation fix</li>
- <li>Russell Cardullo &#8211; graphite sink for metrics</li>
- <li>Evan Chan &#8211; local:// URI feature</li>
- <li>Vadim Chekan &#8211; bug fix</li>
- <li>Lian Cheng &#8211; refactoring and code clean-up in several locations, bug fixes</li>
- <li>Ewen Cheslack-Postava &#8211; Spark EC2 and PySpark improvements</li>
- <li>Mosharaf Chowdhury &#8211; optimized broadcast</li>
- <li>Dan Crankshaw &#8211; GraphX contributions</li>
- <li>Haider Haidi &#8211; documentation fix</li>
- <li>Frank Dai &#8211; Naive Bayes classifier in MLlib, documentation improvements</li>
- <li>Tathagata Das &#8211; new operators, fixes, and improvements to Spark Streaming (lead)</li>
- <li>Ankur Dave &#8211; GraphX contributions</li>
- <li>Henry Davidge &#8211; warning for large tasks</li>
- <li>Aaron Davidson &#8211; shuffle file consolidation, H/A mode for standalone scheduler, various improvements and fixes</li>
- <li>Kyle Ellrott &#8211; GraphX contributions</li>
- <li>Hossein Falaki &#8211; new statistical operators, Scala and Python examples in MLlib</li>
- <li>Harvey Feng &#8211; hadoop file optimizations and YARN integration</li>
- <li>Ali Ghodsi &#8211; support for SIMR</li>
- <li>Joseph E. Gonzalez &#8211; GraphX contributions</li>
- <li>Thomas Graves &#8211; fixes and improvements for YARN support (lead)</li>
- <li>Rong Gu &#8211; documentation fix</li>
- <li>Stephen Haberman &#8211; bug fixes</li>
- <li>Walker Hamilton &#8211; bug fix</li>
- <li>Mark Hamstra &#8211; scheduler improvements and fixes, build fixes</li>
- <li>Damien Hardy &#8211; Debian build fix</li>
- <li>Nathan Howell &#8211; sbt upgrade</li>
- <li>Grace Huang &#8211; improvements to metrics code</li>
- <li>Shane Huang &#8211; separation of admin and user scripts:</li>
- <li>Prabeesh K &#8211; MQTT integration for Spark Streaming and code fix</li>
- <li>Holden Karau &#8211; sbt build improvements and Java API extensions</li>
- <li>KarthikTunga &#8211; bug fix</li>
- <li>Grega Kespret &#8211; bug fix</li>
- <li>Marek Kolodziej &#8211; optimized random number generator</li>
- <li>Jey Kottalam &#8211; EC2 script improvements</li>
- <li>Du Li &#8211; bug fixes</li>
- <li>Haoyuan Li &#8211; tachyon support in EC2</li>
- <li>LiGuoqiang &#8211; fixes to build and YARN integration</li>
- <li>Raymond Liu &#8211; build improvement and various fixes for YARN support</li>
- <li>George Loentiev &#8211; Maven build fixes</li>
- <li>Akihiro Matsukawa &#8211; GraphX contributions</li>
- <li>David McCauley &#8211; improvements to json endpoint</li>
- <li>Mike &#8211; bug fixes</li>
- <li>Fabrizio (Misto) Milo &#8211; bug fix</li>
- <li>Mridul Muralidharan &#8211; speculation improvements, several bug fixes</li>
- <li>Tor Myklebust &#8211; Python mllib bindings, instrumentation for task serailization</li>
- <li>Sundeep Narravula &#8211; bug fix</li>
- <li>Binh Nguyen &#8211; Java API improvements and version upgrades</li>
- <li>Adam Novak &#8211; bug fix</li>
- <li>Andrew Or &#8211; external sorting</li>
- <li>Kay Ousterhout &#8211; several bug fixes and improvements to Spark scheduler</li>
- <li>Sean Owen &#8211; style fixes</li>
- <li>Nick Pentreath &#8211; ALS implicit feedback algorithm</li>
- <li>Pillis &#8211; <code>Vector.random()</code> method</li>
- <li>Imran Rashid &#8211; bug fix</li>
- <li>Ahir Reddy &#8211; support for SIMR</li>
- <li>Luca Rosellini &#8211; script loading for Scala shell</li>
- <li>Josh Rosen &#8211; fixes, clean-up, and extensions to scala and Java API’s</li>
- <li>Henry Saputra &#8211; style improvements and clean-up</li>
- <li>Andre Schumacher &#8211; Python improvements and bug fixes</li>
- <li>Jerry Shao &#8211; multi-user support, various fixes and improvements</li>
- <li>Prashant Sharma &#8211; Scala 2.10 support, configuration system, several smaller fixes</li>
- <li>Shiyun &#8211; style fix</li>
- <li>Wangda Tan &#8211; UI improvement and bug fixes</li>
- <li>Matthew Taylor &#8211; bug fix</li>
- <li>Jyun-Fan Tsai &#8211; documentation fix</li>
- <li>Takuya Ueshin &#8211; bug fix</li>
- <li>Shivaram Venkataraman &#8211; sbt build optimization, EC2 improvements, Java and Python API</li>
- <li>Jianping J Wang &#8211; GraphX contributions</li>
- <li>Martin Weindel &#8211; build fix</li>
- <li>Patrick Wendell &#8211; standalone driver submission, various fixes, release manager</li>
- <li>Neal Wiggins &#8211; bug fix</li>
- <li>Andrew Xia &#8211; bug fixes and code cleanup</li>
- <li>Reynold Xin &#8211; GraphX contributions, task killing, various fixes, improvements and optimizations</li>
- <li>Dong Yan &#8211; bug fix</li>
- <li>Haitao Yao &#8211; bug fix</li>
- <li>Xusen Yin &#8211; bug fix</li>
- <li>Fengdong Yu &#8211; documentation fixes</li>
- <li>Matei Zaharia &#8211; new configuration system, Python MLlib bindings, scheduler improvements, various fixes and optimizations</li>
- <li>Wu Zeming &#8211; bug fix</li>
- <li>Nan Zhu &#8211; documentation improvements</li>
+ <li>Andrew Ash – documentation improvements</li>
+ <li>Pierre Borckmans – documentation fix</li>
+ <li>Russell Cardullo – graphite sink for metrics</li>
+ <li>Evan Chan – local:// URI feature</li>
+ <li>Vadim Chekan – bug fix</li>
+ <li>Lian Cheng – refactoring and code clean-up in several locations, bug fixes</li>
+ <li>Ewen Cheslack-Postava – Spark EC2 and PySpark improvements</li>
+ <li>Mosharaf Chowdhury – optimized broadcast</li>
+ <li>Dan Crankshaw – GraphX contributions</li>
+ <li>Haider Haidi – documentation fix</li>
+ <li>Frank Dai – Naive Bayes classifier in MLlib, documentation improvements</li>
+ <li>Tathagata Das – new operators, fixes, and improvements to Spark Streaming (lead)</li>
+ <li>Ankur Dave – GraphX contributions</li>
+ <li>Henry Davidge – warning for large tasks</li>
+ <li>Aaron Davidson – shuffle file consolidation, H/A mode for standalone scheduler, various improvements and fixes</li>
+ <li>Kyle Ellrott – GraphX contributions</li>
+ <li>Hossein Falaki – new statistical operators, Scala and Python examples in MLlib</li>
+ <li>Harvey Feng – hadoop file optimizations and YARN integration</li>
+ <li>Ali Ghodsi – support for SIMR</li>
+ <li>Joseph E. Gonzalez – GraphX contributions</li>
+ <li>Thomas Graves – fixes and improvements for YARN support (lead)</li>
+ <li>Rong Gu – documentation fix</li>
+ <li>Stephen Haberman – bug fixes</li>
+ <li>Walker Hamilton – bug fix</li>
+ <li>Mark Hamstra – scheduler improvements and fixes, build fixes</li>
+ <li>Damien Hardy – Debian build fix</li>
+ <li>Nathan Howell – sbt upgrade</li>
+ <li>Grace Huang – improvements to metrics code</li>
+ <li>Shane Huang – separation of admin and user scripts:</li>
+ <li>Prabeesh K – MQTT integration for Spark Streaming and code fix</li>
+ <li>Holden Karau – sbt build improvements and Java API extensions</li>
+ <li>KarthikTunga – bug fix</li>
+ <li>Grega Kespret – bug fix</li>
+ <li>Marek Kolodziej – optimized random number generator</li>
+ <li>Jey Kottalam – EC2 script improvements</li>
+ <li>Du Li – bug fixes</li>
+ <li>Haoyuan Li – tachyon support in EC2</li>
+ <li>LiGuoqiang – fixes to build and YARN integration</li>
+ <li>Raymond Liu – build improvement and various fixes for YARN support</li>
+ <li>George Loentiev – Maven build fixes</li>
+ <li>Akihiro Matsukawa – GraphX contributions</li>
+ <li>David McCauley – improvements to json endpoint</li>
+ <li>Mike – bug fixes</li>
+ <li>Fabrizio (Misto) Milo – bug fix</li>
+ <li>Mridul Muralidharan – speculation improvements, several bug fixes</li>
+ <li>Tor Myklebust – Python mllib bindings, instrumentation for task serailization</li>
+ <li>Sundeep Narravula – bug fix</li>
+ <li>Binh Nguyen – Java API improvements and version upgrades</li>
+ <li>Adam Novak – bug fix</li>
+ <li>Andrew Or – external sorting</li>
+ <li>Kay Ousterhout – several bug fixes and improvements to Spark scheduler</li>
+ <li>Sean Owen – style fixes</li>
+ <li>Nick Pentreath – ALS implicit feedback algorithm</li>
+ <li>Pillis – <code>Vector.random()</code> method</li>
+ <li>Imran Rashid – bug fix</li>
+ <li>Ahir Reddy – support for SIMR</li>
+ <li>Luca Rosellini – script loading for Scala shell</li>
+ <li>Josh Rosen – fixes, clean-up, and extensions to scala and Java API’s</li>
+ <li>Henry Saputra – style improvements and clean-up</li>
+ <li>Andre Schumacher – Python improvements and bug fixes</li>
+ <li>Jerry Shao – multi-user support, various fixes and improvements</li>
+ <li>Prashant Sharma – Scala 2.10 support, configuration system, several smaller fixes</li>
+ <li>Shiyun – style fix</li>
+ <li>Wangda Tan – UI improvement and bug fixes</li>
+ <li>Matthew Taylor – bug fix</li>
+ <li>Jyun-Fan Tsai – documentation fix</li>
+ <li>Takuya Ueshin – bug fix</li>
+ <li>Shivaram Venkataraman – sbt build optimization, EC2 improvements, Java and Python API</li>
+ <li>Jianping J Wang – GraphX contributions</li>
+ <li>Martin Weindel – build fix</li>
+ <li>Patrick Wendell – standalone driver submission, various fixes, release manager</li>
+ <li>Neal Wiggins – bug fix</li>
+ <li>Andrew Xia – bug fixes and code cleanup</li>
+ <li>Reynold Xin – GraphX contributions, task killing, various fixes, improvements and optimizations</li>
+ <li>Dong Yan – bug fix</li>
+ <li>Haitao Yao – bug fix</li>
+ <li>Xusen Yin – bug fix</li>
+ <li>Fengdong Yu – documentation fixes</li>
+ <li>Matei Zaharia – new configuration system, Python MLlib bindings, scheduler improvements, various fixes and optimizations</li>
+ <li>Wu Zeming – bug fix</li>
+ <li>Nan Zhu – documentation improvements</li>
</ul>
<p><em>Thanks to everyone who contributed!</em></p>
diff --git a/site/releases/spark-release-0-9-1.html b/site/releases/spark-release-0-9-1.html
index 586911339..3556e0668 100644
--- a/site/releases/spark-release-0-9-1.html
+++ b/site/releases/spark-release-0-9-1.html
@@ -235,7 +235,7 @@
<li>Andrew Tulloch - Minor updates to MLLib</li>
<li>Bijay Bisht - Fix for hadoop-client for Hadoop &lt; 1.0.1 and for bug in Spark on Mesos + CDH4.5.0</li>
<li>Bouke van der Bijl - Bug fix in Python depickling</li>
- <li>Bryn Keller - Support for HBase&#8217;s TableOutputFormat</li>
+ <li>Bryn Keller - Support for HBase’s TableOutputFormat</li>
<li>Chen Chao - Bug fix in spark-shell script, and improvements to streaming programming guide</li>
<li>Christian Lundgren - Support for C3 EC2 instance type</li>
<li>Diana Carroll - Improvements to PySpark programming guide</li>
@@ -245,7 +245,7 @@
<li>jianghan - Bug fixes in Java examples</li>
<li>Josh Rosen - Bug fix in PySpark string serialization and exception handling</li>
<li>Jyotiska NK - Improvements to PySpark doc and examples</li>
- <li>Kay Ousterhout - Multiple bug fixes in scheduler&#8217;s handling of task failures</li>
+ <li>Kay Ousterhout - Multiple bug fixes in scheduler’s handling of task failures</li>
<li>Kousuke Saruta - Use of https to access github</li>
<li>Mark Grover - Bug fix in distribution tar.gz</li>
<li>Matei Zaharia - Bug fixes in handling of task failures due to NPE, and cleaning up of scheduler data structures </li>
@@ -258,10 +258,10 @@
<li>Raymond Liu - Changed working directory in ZookeeperPersistenceEngine</li>
<li>Reynold Xin - Improvements to docs and test infrastructure</li>
<li>Sandy Ryza - Multiple important Yarn bug fixes and improvements</li>
- <li>Sean Owen - Bug fixes and improvements for MLLib&#8217;s ALS</li>
+ <li>Sean Owen - Bug fixes and improvements for MLLib’s ALS</li>
<li>Shixiong Zhu - Fixed thread-unsafe use of SimpleDateFormat</li>
<li>shiyun.wxm - UI bug fix</li>
- <li>Stevo Slavić - Bug fix in window&#8217;s run-example script</li>
+ <li>Stevo Slavić - Bug fix in window’s run-example script</li>
<li>Tathagata Das - Improvements to streaming docs</li>
<li>Tom Graves - Bug fixes in YARN deployment modes</li>
<li>Xiangrui Meng - Improvements to ALS and GLM, and MLLib programming guide</li>
diff --git a/site/releases/spark-release-1-0-0.html b/site/releases/spark-release-1-0-0.html
index 37630607b..51a4187b8 100644
--- a/site/releases/spark-release-1-0-0.html
+++ b/site/releases/spark-release-1-0-0.html
@@ -196,7 +196,7 @@
<h3 id="smaller-changes">Smaller Changes</h3>
<ul>
- <li>PySpark now works with more Python versions than before &#8211; Python 2.6+ instead of 2.7+, and NumPy 1.4+ instead of 1.7+.</li>
+ <li>PySpark now works with more Python versions than before – Python 2.6+ instead of 2.7+, and NumPy 1.4+ instead of 1.7+.</li>
<li>Spark has upgraded to Avro 1.7.6, adding support for Avro specific types.</li>
<li>Internal instrumentation has been added to allow applications to monitor and instrument Spark jobs.</li>
<li>Support for off-heap storage in Tachyon has been added via a special build target.</li>
@@ -213,123 +213,123 @@
<p>The following developers contributed to this release:</p>
<ul>
- <li>Aaron Davidson &#8211; packaging and deployment improvements, several bug fixes, local[*] mode</li>
- <li>Aaron Kimball &#8211; documentation improvements</li>
- <li>Abhishek Kumar &#8211; Python configuration fixes</li>
- <li>Ahir Reddy &#8211; PySpark build, fixes, and cancellation support</li>
- <li>Allan Douglas R. de Oliveira &#8211; Improvements to spark-ec2 scripts</li>
- <li>Andre Schumacher &#8211; Parquet support and optimizations</li>
- <li>Andrew Ash &#8211; Mesos documentation and other doc improvements, bug fixes</li>
- <li>Andrew Or &#8211; history server (lead), garbage collection (lead), spark-submit, PySpark and YARN improvements</li>
- <li>Andrew Tulloch &#8211; MLlib contributions and code clean-up</li>
- <li>Andy Konwinski &#8211; documentation fix</li>
- <li>Anita Tailor &#8211; Cassandra example</li>
- <li>Ankur Dave &#8211; GraphX (lead) optimizations, documentation, and usability</li>
- <li>Archer Shao &#8211; bug fixes</li>
- <li>Arun Ramakrishnan &#8211; improved random sampling</li>
- <li>Baishuo &#8211; test improvements</li>
- <li>Bernardo Gomez Palacio &#8211; spark-shell improvements and Mesos updates</li>
- <li>Bharath Bhushan &#8211; bug fix</li>
- <li>Bijay Bisht &#8211; bug fixes</li>
- <li>Binh Nguyen &#8211; dependency fix</li>
- <li>Bouke van der Bijl &#8211; fixes for PySpark on Mesos and other Mesos fixes</li>
- <li>Bryn Keller &#8211; improvement to HBase support and unit tests</li>
- <li>Chen Chao &#8211; documentation, bug fix, and code clean-up</li>
- <li>Cheng Hao &#8211; performance and feature improvements in Spark SQL</li>
- <li>Cheng Lian &#8211; column storage and other improvements in Spark SQL</li>
- <li>Christian Lundgren &#8211; improvement to spark-ec2 scripts</li>
- <li>DB Tsai &#8211; L-BGFS optimizer in MLlib, MLlib documentation and fixes</li>
- <li>Dan McClary &#8211; Improvement to stats counter</li>
- <li>Daniel Darabos &#8211; GraphX performance improvement</li>
- <li>Davis Shepherd &#8211; bug fix</li>
- <li>Diana Carroll &#8211; documentation and bug fix</li>
- <li>Egor Pakhomov &#8211; local iterator for RDD’s</li>
- <li>Emtiaz Ahmed &#8211; bug fix</li>
- <li>Erik Selin &#8211; bug fix</li>
- <li>Ethan Jewett &#8211; documentation improvement</li>
- <li>Evan Chan &#8211; automatic clean-up of application data</li>
- <li>Evan Sparks &#8211; MLlib optimizations and doc improvement</li>
- <li>Frank Dai &#8211; code clean-up in MLlib</li>
- <li>Guoquiang Li &#8211; build improvements and several bug fixes</li>
- <li>Ghidireac &#8211; bug fix</li>
- <li>Haoyuan Li &#8211; Tachyon storage level for RDD’s</li>
- <li>Harvey Feng &#8211; spark-ec2 update</li>
- <li>Henry Saputra &#8211; code clean-up</li>
- <li>Henry Cook &#8211; Spark SQL improvements</li>
- <li>Holden Karau &#8211; cross validation in MLlib, Python and core engine improvements</li>
- <li>Ivan Wick &#8211; Mesos bug fix</li>
- <li>Jey Kottalam &#8211; sbt build improvement</li>
- <li>Jerry Shao &#8211; Spark metrics and Spark SQL improvements</li>
- <li>Jiacheng Guo &#8211; bug fix</li>
- <li>Jianghan &#8211; bug fix</li>
- <li>Jianping J Wang &#8211; JBLAS support in MLlib</li>
- <li>Joseph E. Gonzalez &#8211; GraphX improvements, fixes, and documentation</li>
- <li>Josh Rosen &#8211; PySpark improvements and bug fixes</li>
- <li>Jyotiska NK &#8211; documentation, test improvements, and bug fix</li>
- <li>Kan Zhang &#8211; bug fixes in Spark core, SQL, and PySpark</li>
- <li>Kay Ousterhout &#8211; bug fixes and code refactoring in scheduler</li>
- <li>Kelvin Chu &#8211; automatic clean-up of application data</li>
- <li>Kevin Mader &#8211; example fix</li>
- <li>Koert Kuipers &#8211; code visibility fix</li>
- <li>Kousuke Saruta &#8211; documentation and build fixes</li>
- <li>Kyle Ellrott &#8211; improved memory usage for DISK_ONLY persistence</li>
- <li>Larva Boy &#8211; approximate counts in Spark SQL</li>
- <li>Madhu Siddalingaiah &#8211; ec2 fixes</li>
- <li>Manish Amde &#8211; decision trees in MLlib</li>
- <li>Marcelo Vanzin &#8211; improvements and fixes to YARN support, dependency clean-up</li>
- <li>Mark Grover &#8211; build fixes</li>
- <li>Mark Hamstra &#8211; build and dependency improvements, scheduler bug fixes</li>
- <li>Margin Jaggi &#8211; MLlib documentation improvements</li>
- <li>Matei Zaharia &#8211; Python versions of several MLlib algorithms, spark-submit improvements, bug fixes, and documentation improvements</li>
- <li>Michael Armbrust &#8211; Spark SQL (lead), including schema support for RDD’s, catalyst optimizer, and Hive support</li>
- <li>Mridul Muralidharan &#8211; code visibility changes and bug fixes</li>
- <li>Nan Zhu &#8211; bug and stability fixes, code clean-up, documentation, and new features</li>
- <li>Neville Li &#8211; bug fix</li>
- <li>Nick Lanham &#8211; Tachyon bundling in distribution script</li>
- <li>Nirmal Reddy &#8211; code clean-up</li>
- <li>OuYang Jin &#8211; local mode and json improvements</li>
- <li>Patrick Wendell &#8211; release manager, build improvements, bug fixes, and code clean-up</li>
- <li>Petko Nikolov &#8211; new utility functions</li>
- <li>Prabeesh K &#8211; typo fix</li>
- <li>Prabin Banka &#8211; new PySpark API’s</li>
- <li>Prashant Sharma &#8211; PySpark improvements, Java 8 lambda support, and build improvements</li>
- <li>Punya Biswal &#8211; Java API improvements</li>
- <li>Qiuzhuang Lian &#8211; bug fixes</li>
- <li>Rahul Singhal &#8211; build improvements, bug fixes</li>
- <li>Raymond Liu &#8211; YARN build fixes and UI improvements</li>
- <li>Reynold Xin &#8211; bug fixes, internal changes, Spark SQL improvements, build fixes, and style improvements</li>
- <li>Reza Zadeh &#8211; SVD implementation in MLlib and other MLlib contributions</li>
- <li>Roman Pastukhov &#8211; clean-up of broadcast files</li>
- <li>Rong Gu &#8211; Tachyon storage level for RDD’s</li>
- <li>Sandeep Sing &#8211; several bug fixes, MLLib improvements and fixes to Spark examples</li>
- <li>Sandy Ryza &#8211; spark-submit script and several YARN improvements</li>
- <li>Saurabh Rawat &#8211; Java API improvements</li>
- <li>Sean Owen &#8211; several build improvements, code clean-up, and MLlib fixes</li>
- <li>Semih Salihoglu &#8211; GraphX improvements</li>
- <li>Shaocun Tian &#8211; bug fix in MLlib</li>
- <li>Shivaram Venkataraman &#8211; bug fixes</li>
- <li>Shixiong Zhu &#8211; code style and correctness fixes</li>
- <li>Shiyun Wxm &#8211; typo fix</li>
- <li>Stevo Slavic &#8211; bug fix</li>
- <li>Sumedh Mungee &#8211; documentation fix</li>
- <li>Sundeep Narravula &#8211; “cancel” button in Spark UI</li>
- <li>Takayu Ueshin &#8211; bug fixes and improvements to Spark SQL</li>
- <li>Tathagata Das &#8211; web UI and other improvements to Spark Streaming (lead), bug fixes, state clean-up, and release manager</li>
- <li>Timothy Chen &#8211; Spark SQL improvements</li>
- <li>Ted Malaska &#8211; improved Flume support</li>
- <li>Tom Graves &#8211; Hadoop security integration (lead) and YARN support</li>
- <li>Tianshuo Deng &#8211; Bug fix</li>
- <li>Tor Myklebust &#8211; improvements to ALS</li>
- <li>Wangfei &#8211; Spark SQL docs</li>
- <li>Wang Tao &#8211; code clean-up</li>
- <li>William Bendon &#8211; JSON support changes and bug fixes</li>
- <li>Xiangrui Meng &#8211; several improvements to MLlib (lead)</li>
- <li>Xuan Nguyen &#8211; build fix</li>
- <li>Xusen Yin &#8211; MLlib contributions and bug fix</li>
- <li>Ye Xianjin &#8211; test fixes</li>
- <li>Yinan Li &#8211; addFile improvement</li>
- <li>Yin Hua &#8211; Spark SQL improvements</li>
- <li>Zheng Peng &#8211; bug fixes</li>
+ <li>Aaron Davidson – packaging and deployment improvements, several bug fixes, local[*] mode</li>
+ <li>Aaron Kimball – documentation improvements</li>
+ <li>Abhishek Kumar – Python configuration fixes</li>
+ <li>Ahir Reddy – PySpark build, fixes, and cancellation support</li>
+ <li>Allan Douglas R. de Oliveira – Improvements to spark-ec2 scripts</li>
+ <li>Andre Schumacher – Parquet support and optimizations</li>
+ <li>Andrew Ash – Mesos documentation and other doc improvements, bug fixes</li>
+ <li>Andrew Or – history server (lead), garbage collection (lead), spark-submit, PySpark and YARN improvements</li>
+ <li>Andrew Tulloch – MLlib contributions and code clean-up</li>
+ <li>Andy Konwinski – documentation fix</li>
+ <li>Anita Tailor – Cassandra example</li>
+ <li>Ankur Dave – GraphX (lead) optimizations, documentation, and usability</li>
+ <li>Archer Shao – bug fixes</li>
+ <li>Arun Ramakrishnan – improved random sampling</li>
+ <li>Baishuo – test improvements</li>
+ <li>Bernardo Gomez Palacio – spark-shell improvements and Mesos updates</li>
+ <li>Bharath Bhushan – bug fix</li>
+ <li>Bijay Bisht – bug fixes</li>
+ <li>Binh Nguyen – dependency fix</li>
+ <li>Bouke van der Bijl – fixes for PySpark on Mesos and other Mesos fixes</li>
+ <li>Bryn Keller – improvement to HBase support and unit tests</li>
+ <li>Chen Chao – documentation, bug fix, and code clean-up</li>
+ <li>Cheng Hao – performance and feature improvements in Spark SQL</li>
+ <li>Cheng Lian – column storage and other improvements in Spark SQL</li>
+ <li>Christian Lundgren – improvement to spark-ec2 scripts</li>
+ <li>DB Tsai – L-BGFS optimizer in MLlib, MLlib documentation and fixes</li>
+ <li>Dan McClary – Improvement to stats counter</li>
+ <li>Daniel Darabos – GraphX performance improvement</li>
+ <li>Davis Shepherd – bug fix</li>
+ <li>Diana Carroll – documentation and bug fix</li>
+ <li>Egor Pakhomov – local iterator for RDD’s</li>
+ <li>Emtiaz Ahmed – bug fix</li>
+ <li>Erik Selin – bug fix</li>
+ <li>Ethan Jewett – documentation improvement</li>
+ <li>Evan Chan – automatic clean-up of application data</li>
+ <li>Evan Sparks – MLlib optimizations and doc improvement</li>
+ <li>Frank Dai – code clean-up in MLlib</li>
+ <li>Guoquiang Li – build improvements and several bug fixes</li>
+ <li>Ghidireac – bug fix</li>
+ <li>Haoyuan Li – Tachyon storage level for RDD’s</li>
+ <li>Harvey Feng – spark-ec2 update</li>
+ <li>Henry Saputra – code clean-up</li>
+ <li>Henry Cook – Spark SQL improvements</li>
+ <li>Holden Karau – cross validation in MLlib, Python and core engine improvements</li>
+ <li>Ivan Wick – Mesos bug fix</li>
+ <li>Jey Kottalam – sbt build improvement</li>
+ <li>Jerry Shao – Spark metrics and Spark SQL improvements</li>
+ <li>Jiacheng Guo – bug fix</li>
+ <li>Jianghan – bug fix</li>
+ <li>Jianping J Wang – JBLAS support in MLlib</li>
+ <li>Joseph E. Gonzalez – GraphX improvements, fixes, and documentation</li>
+ <li>Josh Rosen – PySpark improvements and bug fixes</li>
+ <li>Jyotiska NK – documentation, test improvements, and bug fix</li>
+ <li>Kan Zhang – bug fixes in Spark core, SQL, and PySpark</li>
+ <li>Kay Ousterhout – bug fixes and code refactoring in scheduler</li>
+ <li>Kelvin Chu – automatic clean-up of application data</li>
+ <li>Kevin Mader – example fix</li>
+ <li>Koert Kuipers – code visibility fix</li>
+ <li>Kousuke Saruta – documentation and build fixes</li>
+ <li>Kyle Ellrott – improved memory usage for DISK_ONLY persistence</li>
+ <li>Larva Boy – approximate counts in Spark SQL</li>
+ <li>Madhu Siddalingaiah – ec2 fixes</li>
+ <li>Manish Amde – decision trees in MLlib</li>
+ <li>Marcelo Vanzin – improvements and fixes to YARN support, dependency clean-up</li>
+ <li>Mark Grover – build fixes</li>
+ <li>Mark Hamstra – build and dependency improvements, scheduler bug fixes</li>
+ <li>Margin Jaggi – MLlib documentation improvements</li>
+ <li>Matei Zaharia – Python versions of several MLlib algorithms, spark-submit improvements, bug fixes, and documentation improvements</li>
+ <li>Michael Armbrust – Spark SQL (lead), including schema support for RDD’s, catalyst optimizer, and Hive support</li>
+ <li>Mridul Muralidharan – code visibility changes and bug fixes</li>
+ <li>Nan Zhu – bug and stability fixes, code clean-up, documentation, and new features</li>
+ <li>Neville Li – bug fix</li>
+ <li>Nick Lanham – Tachyon bundling in distribution script</li>
+ <li>Nirmal Reddy – code clean-up</li>
+ <li>OuYang Jin – local mode and json improvements</li>
+ <li>Patrick Wendell – release manager, build improvements, bug fixes, and code clean-up</li>
+ <li>Petko Nikolov – new utility functions</li>
+ <li>Prabeesh K – typo fix</li>
+ <li>Prabin Banka – new PySpark API’s</li>
+ <li>Prashant Sharma – PySpark improvements, Java 8 lambda support, and build improvements</li>
+ <li>Punya Biswal – Java API improvements</li>
+ <li>Qiuzhuang Lian – bug fixes</li>
+ <li>Rahul Singhal – build improvements, bug fixes</li>
+ <li>Raymond Liu – YARN build fixes and UI improvements</li>
+ <li>Reynold Xin – bug fixes, internal changes, Spark SQL improvements, build fixes, and style improvements</li>
+ <li>Reza Zadeh – SVD implementation in MLlib and other MLlib contributions</li>
+ <li>Roman Pastukhov – clean-up of broadcast files</li>
+ <li>Rong Gu – Tachyon storage level for RDD’s</li>
+ <li>Sandeep Sing – several bug fixes, MLLib improvements and fixes to Spark examples</li>
+ <li>Sandy Ryza – spark-submit script and several YARN improvements</li>
+ <li>Saurabh Rawat – Java API improvements</li>
+ <li>Sean Owen – several build improvements, code clean-up, and MLlib fixes</li>
+ <li>Semih Salihoglu – GraphX improvements</li>
+ <li>Shaocun Tian – bug fix in MLlib</li>
+ <li>Shivaram Venkataraman – bug fixes</li>
+ <li>Shixiong Zhu – code style and correctness fixes</li>
+ <li>Shiyun Wxm – typo fix</li>
+ <li>Stevo Slavic – bug fix</li>
+ <li>Sumedh Mungee – documentation fix</li>
+ <li>Sundeep Narravula – “cancel” button in Spark UI</li>
+ <li>Takayu Ueshin – bug fixes and improvements to Spark SQL</li>
+ <li>Tathagata Das – web UI and other improvements to Spark Streaming (lead), bug fixes, state clean-up, and release manager</li>
+ <li>Timothy Chen – Spark SQL improvements</li>
+ <li>Ted Malaska – improved Flume support</li>
+ <li>Tom Graves – Hadoop security integration (lead) and YARN support</li>
+ <li>Tianshuo Deng – Bug fix</li>
+ <li>Tor Myklebust – improvements to ALS</li>
+ <li>Wangfei – Spark SQL docs</li>
+ <li>Wang Tao – code clean-up</li>
+ <li>William Bendon – JSON support changes and bug fixes</li>
+ <li>Xiangrui Meng – several improvements to MLlib (lead)</li>
+ <li>Xuan Nguyen – build fix</li>
+ <li>Xusen Yin – MLlib contributions and bug fix</li>
+ <li>Ye Xianjin – test fixes</li>
+ <li>Yinan Li – addFile improvement</li>
+ <li>Yin Hua – Spark SQL improvements</li>
+ <li>Zheng Peng – bug fixes</li>
</ul>
<p><em>Thanks to everyone who contributed!</em></p>