summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorSean R. Owen <srowen@apache.org>2015-05-08 13:59:51 +0000
committerSean R. Owen <srowen@apache.org>2015-05-08 13:59:51 +0000
commit090a6d74cea47ee2023e2f15aa207afe83194702 (patch)
treed9a77f60ec22facf4ea1740a0a50aee5804169aa
parent89f7bb0ead0b91237b156377f12b0216347d5b75 (diff)
downloadspark-website-090a6d74cea47ee2023e2f15aa207afe83194702.tar.gz
spark-website-090a6d74cea47ee2023e2f15aa207afe83194702.tar.bz2
spark-website-090a6d74cea47ee2023e2f15aa207afe83194702.zip
Reapply my past changes, which I had only applied to .html, to .md too, and add the changes from the regenerated .html too
-rw-r--r--community.md2
-rw-r--r--downloads.md6
-rw-r--r--examples.md24
-rw-r--r--faq.md4
-rw-r--r--index.md8
-rw-r--r--releases/_posts/2015-03-13-spark-release-1-3-0.md2
-rw-r--r--site/community.html2
-rw-r--r--site/downloads.html6
-rw-r--r--site/examples.html24
-rw-r--r--site/faq.html1
-rw-r--r--site/index.html8
-rw-r--r--site/releases/spark-release-1-3-0.html2
-rw-r--r--sql/index.md2
13 files changed, 48 insertions, 43 deletions
diff --git a/community.md b/community.md
index f32c249c4..23046902d 100644
--- a/community.md
+++ b/community.md
@@ -28,6 +28,8 @@ navigation:
</li>
</ul>
+<p>The StackOverflow tag <a href="http://stackoverflow.com/questions/tagged/apache-spark"><code>apache-spark</code></a> is an unofficial but active forum for Spark users' questions and answers.</p>
+
<a name="events"></a>
<h3>Events and Meetups</h3>
diff --git a/downloads.md b/downloads.md
index 5587d5021..c96ad69f4 100644
--- a/downloads.md
+++ b/downloads.md
@@ -20,13 +20,13 @@ The latest release of Spark is Spark 1.3.1, released on April 17, 2015
<a href="{{site.url}}releases/spark-release-1-3-1.html">(release notes)</a>
<a href="https://git-wip-us.apache.org/repos/asf?p=spark.git;a=commit;h=3e8391327ba586eaf54447043bd526d919043a44">(git tag)</a><br/>
-1. Chose a Spark release:
+1. Choose a Spark release:
<select id="sparkVersionSelect" onChange="javascript:onVersionSelect();"></select><br>
-2. Chose a package type:
+2. Choose a package type:
<select id="sparkPackageSelect" onChange="javascript:onPackageSelect();"></select><br>
-3. Chose a download type:
+3. Choose a download type:
<select id="sparkDownloadSelect" onChange="javascript:onDownloadSelect()"></select><br>
4. Download Spark: <span id="spanDownloadLink"></span>
diff --git a/examples.md b/examples.md
index 3658d1788..18d869353 100644
--- a/examples.md
+++ b/examples.md
@@ -26,8 +26,8 @@ In this example, we search through the error messages in a log file:
<div class="tab-content">
<div class="tab-pane tab-pane-python active">
<div class="code code-tab">
- file = spark.textFile(<span class="string">"hdfs://..."</span>)<br>
- errors = file.<span class="sparkop">filter</span>(<span class="closure">lambda line: "ERROR" in line</span>)<br>
+ text_file = spark.textFile(<span class="string">"hdfs://..."</span>)<br />
+ errors = text_file.<span class="sparkop">filter</span>(<span class="closure">lambda line: "ERROR" in line</span>)<br />
<span class="comment"># Count all the errors</span><br>
errors.<span class="sparkop">count</span>()<br>
<span class="comment"># Count errors mentioning MySQL</span><br>
@@ -38,8 +38,8 @@ In this example, we search through the error messages in a log file:
</div>
<div class="tab-pane tab-pane-scala">
<div class="code code-tab">
- <span class="keyword">val</span> file = spark.textFile(<span class="string">"hdfs://..."</span>)<br>
- <span class="keyword">val</span> errors = file.<span class="sparkop">filter</span>(<span class="closure">line =&gt; line.contains("ERROR")</span>)<br>
+ <span class="keyword">val</span> textFile = spark.textFile(<span class="string">"hdfs://..."</span>)<br>
+ <span class="keyword">val</span> errors = textFile.<span class="sparkop">filter</span>(<span class="closure">line =&gt; line.contains("ERROR")</span>)<br>
<span class="comment">// Count all the errors</span><br>
errors.<span class="sparkop">count</span>()<br>
<span class="comment">// Count errors mentioning MySQL</span><br>
@@ -50,8 +50,8 @@ In this example, we search through the error messages in a log file:
</div>
<div class="tab-pane tab-pane-java">
<div class="code code-tab">
- JavaRDD&lt;String&gt; file = spark.textFile(<span class="string">"hdfs://..."</span>);<br>
- JavaRDD&lt;String&gt; errors = file.<span class="sparkop">filter</span>(<span class="closure">new Function&lt;String, Boolean&gt;() {<br>
+ JavaRDD&lt;String&gt; textFile = spark.textFile(<span class="string">"hdfs://..."</span>);<br>
+ JavaRDD&lt;String&gt; errors = textFile.<span class="sparkop">filter</span>(<span class="closure">new Function&lt;String, Boolean&gt;() {<br>
&nbsp;&nbsp;public Boolean call(String s) { return s.contains("ERROR"); }<br>
}</span>);<br>
<span class="comment">// Count all the errors</span><br>
@@ -112,8 +112,8 @@ In this example, we search through the error messages in a log file:
<div class="tab-content">
<div class="tab-pane tab-pane-python active">
<div class="code code-tab">
- file = spark.textFile(<span class="string">"hdfs://..."</span>)<br>
- counts = file.<span class="sparkop">flatMap</span>(<span class="closure">lambda line: line.split(" ")</span>) \<br>
+ text_file = spark.textFile(<span class="string">"hdfs://..."</span>)<br>
+ counts = text_file.<span class="sparkop">flatMap</span>(<span class="closure">lambda line: line.split(" ")</span>) \<br>
&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;.<span class="sparkop">map</span>(<span class="closure">lambda word: (word, 1)</span>) \<br>
&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;.<span class="sparkop">reduceByKey</span>(<span class="closure">lambda a, b: a + b</span>)<br>
counts.<span class="sparkop">saveAsTextFile</span>(<span class="string">"hdfs://..."</span>)
@@ -121,8 +121,8 @@ In this example, we search through the error messages in a log file:
</div>
<div class="tab-pane tab-pane-scala">
<div class="code code-tab">
- <span class="keyword">val</span> file = spark.textFile(<span class="string">"hdfs://..."</span>)<br>
- <span class="keyword">val</span> counts = file.<span class="sparkop">flatMap</span>(<span class="closure">line =&gt; line.split(" ")</span>)<br>
+ <span class="keyword">val</span> textFile = spark.textFile(<span class="string">"hdfs://..."</span>)<br>
+ <span class="keyword">val</span> counts = textFile.<span class="sparkop">flatMap</span>(<span class="closure">line =&gt; line.split(" ")</span>)<br>
&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;.<span class="sparkop">map</span>(<span class="closure">word =&gt; (word, 1)</span>)<br>
&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;.<span class="sparkop">reduceByKey</span>(<span class="closure">_ + _</span>)<br>
counts.<span class="sparkop">saveAsTextFile</span>(<span class="string">"hdfs://..."</span>)
@@ -130,8 +130,8 @@ In this example, we search through the error messages in a log file:
</div>
<div class="tab-pane tab-pane-java">
<div class="code code-tab">
- JavaRDD&lt;String&gt; file = spark.textFile(<span class="string">"hdfs://..."</span>);<br>
- JavaRDD&lt;String&gt; words = file.<span class="sparkop">flatMap</span>(<span class="closure">new FlatMapFunction&lt;String, String&gt;() {<br>
+ JavaRDD&lt;String&gt; textFile = spark.textFile(<span class="string">"hdfs://..."</span>);<br>
+ JavaRDD&lt;String&gt; words = textFile.<span class="sparkop">flatMap</span>(<span class="closure">new FlatMapFunction&lt;String, String&gt;() {<br>
&nbsp;&nbsp;public Iterable&lt;String&gt; call(String s) { return Arrays.asList(s.split(" ")); }<br>
}</span>);<br>
JavaPairRDD&lt;String, Integer&gt; pairs = words.<span class="sparkop">mapToPair</span>(<span class="closure">new PairFunction&lt;String, String, Integer&gt;() {<br>
diff --git a/faq.md b/faq.md
index 4a6ecc66a..c3e6b2da0 100644
--- a/faq.md
+++ b/faq.md
@@ -53,8 +53,8 @@ Spark is a fast and general processing engine compatible with Hadoop data. It ca
<p class="answer">Starting in version 0.8, Spark is under the <a href="http://www.apache.org/licenses/LICENSE-2.0.html">Apache 2.0 license</a>. Previous versions used the <a href="https://github.com/mesos/spark/blob/branch-0.7/LICENSE">BSD license</a>.</p>
<p class="question">How can I contribute to Spark?</p>
-<p class="answer">Contact the <a href="{{site.url}}community.html">mailing list</a> or send us a pull request on <a href="https://github.com/apache/spark">GitHub</a> (instructions <a href="https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark">here</a>). We're glad to hear about your experience using Spark and to accept patches.</p>
-<p>If you would like to report an issue, post it to the <a href="https://issues.apache.org/jira/browse/SPARK">Spark issue tracker</a>.</p>
+
+<p class="answer">See the <a href="https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark">Contributing to Spark wiki</a> for more information.</p>
<p class="question">Where can I get more help?</p>
<p class="answer">Please post on the <a href="http://apache-spark-user-list.1001560.n3.nabble.com">Spark Users</a> mailing list. We'll be glad to help!</p>
diff --git a/index.md b/index.md
index 754bbf253..90304b64c 100644
--- a/index.md
+++ b/index.md
@@ -53,9 +53,9 @@ navigation:
<div class="col-md-5 col-sm-5 col-padded-top col-center">
<div style="text-align: left; display: inline-block;">
<div class="code">
- file = spark.textFile(<span class="string">"hdfs://..."</span>)<br/>
+ text_file = spark.textFile(<span class="string">"hdfs://..."</span>)<br/>
&nbsp;<br/>
- file.<span class="sparkop">flatMap</span>(<span class="closure">lambda line: line.split()</span>)<br/>
+ text_file.<span class="sparkop">flatMap</span>(<span class="closure">lambda&nbsp;line:&nbsp;line.split()</span>)<br/>
&nbsp;&nbsp;&nbsp;&nbsp;.<span class="sparkop">map</span>(<span class="closure">lambda word: (word, 1)</span>)<br/>
&nbsp;&nbsp;&nbsp;&nbsp;.<span class="sparkop">reduceByKey</span>(<span class="closure">lambda a, b: a+b</span>)
</div>
@@ -63,9 +63,9 @@ navigation:
</div>
<!--
<div class="code" style="margin-top: 20px; text-align: left; display: inline-block;">
- file = spark.textFile(<span class="string">"hdfs://..."</span>)<br/>
+ text_file = spark.textFile(<span class="string">"hdfs://..."</span>)<br/>
&nbsp;<br/>
- file.<span class="sparkop">filter</span>(<span class="closure">lambda line: "ERROR" in line</span>)<br/>
+ text_file.<span class="sparkop">filter</span>(<span class="closure">lambda line: "ERROR" in line</span>)<br/>
&nbsp;&nbsp;&nbsp;&nbsp;.<span class="sparkop">count</span>()
</div>
-->
diff --git a/releases/_posts/2015-03-13-spark-release-1-3-0.md b/releases/_posts/2015-03-13-spark-release-1-3-0.md
index a7c2a7924..bc9c4db84 100644
--- a/releases/_posts/2015-03-13-spark-release-1-3-0.md
+++ b/releases/_posts/2015-03-13-spark-release-1-3-0.md
@@ -36,7 +36,7 @@ GraphX adds a handful of utility functions in this release, including conversion
## Upgrading to Spark 1.3
Spark 1.3 is binary compatible with Spark 1.X releases, so no code changes are necessary. This excludes API’s marked explicitly as unstable.
-As part of stabilizing the Spark SQL API, the `SchemaRDD` class has been extended renamed to `DataFrame`. Spark SQL's [migration guide](http://spark.apache.org/docs/1.3.0/sql-programming-guide.html#migration-guide) describes the upgrade process in detail. Spark SQL also now requires that column identifiers which use reserved words (such as "string" or "table") be escaped using backticks.
+As part of stabilizing the Spark SQL API, the `SchemaRDD` class has been renamed to `DataFrame`. Spark SQL's [migration guide](http://spark.apache.org/docs/1.3.0/sql-programming-guide.html#migration-guide) describes the upgrade process in detail. Spark SQL also now requires that column identifiers which use reserved words (such as "string" or "table") be escaped using backticks.
### Known Issues
This release has few known issues which will be addressed in Spark 1.3.1:
diff --git a/site/community.html b/site/community.html
index 46fe44f2f..42701b422 100644
--- a/site/community.html
+++ b/site/community.html
@@ -188,6 +188,8 @@
</li>
</ul>
+<p>The StackOverflow tag <a href="http://stackoverflow.com/questions/tagged/apache-spark"><code>apache-spark</code></a> is an unofficial but active forum for Spark users' questions and answers.</p>
+
<p><a name="events"></a></p>
<h3>Events and Meetups</h3>
diff --git a/site/downloads.html b/site/downloads.html
index d9d4fa9f5..e8aec835f 100644
--- a/site/downloads.html
+++ b/site/downloads.html
@@ -182,15 +182,15 @@ $(document).ready(function() {
<ol>
<li>
- <p>Chose a Spark release:
+ <p>Choose a Spark release:
<select id="sparkVersionSelect" onchange="javascript:onVersionSelect();"></select><br /></p>
</li>
<li>
- <p>Chose a package type:
+ <p>Choose a package type:
<select id="sparkPackageSelect" onchange="javascript:onPackageSelect();"></select><br /></p>
</li>
<li>
- <p>Chose a download type:
+ <p>Choose a download type:
<select id="sparkDownloadSelect" onchange="javascript:onDownloadSelect()"></select><br /></p>
</li>
<li>
diff --git a/site/examples.html b/site/examples.html
index fca0b08aa..2176ba99a 100644
--- a/site/examples.html
+++ b/site/examples.html
@@ -187,8 +187,8 @@ previous ones, and <em>actions</em>, which kick off a job to execute on a cluste
<div class="tab-content">
<div class="tab-pane tab-pane-python active">
<div class="code code-tab">
- file = spark.textFile(<span class="string">"hdfs://..."</span>)<br />
- errors = file.<span class="sparkop">filter</span>(<span class="closure">lambda line: "ERROR" in line</span>)<br />
+ text_file = spark.textFile(<span class="string">"hdfs://..."</span>)<br />
+ errors = text_file.<span class="sparkop">filter</span>(<span class="closure">lambda line: "ERROR" in line</span>)<br />
<span class="comment"># Count all the errors</span><br />
errors.<span class="sparkop">count</span>()<br />
<span class="comment"># Count errors mentioning MySQL</span><br />
@@ -199,8 +199,8 @@ previous ones, and <em>actions</em>, which kick off a job to execute on a cluste
</div>
<div class="tab-pane tab-pane-scala">
<div class="code code-tab">
- <span class="keyword">val</span> file = spark.textFile(<span class="string">"hdfs://..."</span>)<br />
- <span class="keyword">val</span> errors = file.<span class="sparkop">filter</span>(<span class="closure">line =&gt; line.contains("ERROR")</span>)<br />
+ <span class="keyword">val</span> textFile = spark.textFile(<span class="string">"hdfs://..."</span>)<br />
+ <span class="keyword">val</span> errors = textFile.<span class="sparkop">filter</span>(<span class="closure">line =&gt; line.contains("ERROR")</span>)<br />
<span class="comment">// Count all the errors</span><br />
errors.<span class="sparkop">count</span>()<br />
<span class="comment">// Count errors mentioning MySQL</span><br />
@@ -211,8 +211,8 @@ previous ones, and <em>actions</em>, which kick off a job to execute on a cluste
</div>
<div class="tab-pane tab-pane-java">
<div class="code code-tab">
- JavaRDD&lt;String&gt; file = spark.textFile(<span class="string">"hdfs://..."</span>);<br />
- JavaRDD&lt;String&gt; errors = file.<span class="sparkop">filter</span>(<span class="closure">new Function&lt;String, Boolean&gt;() {<br />
+ JavaRDD&lt;String&gt; textFile = spark.textFile(<span class="string">"hdfs://..."</span>);<br />
+ JavaRDD&lt;String&gt; errors = textFile.<span class="sparkop">filter</span>(<span class="closure">new Function&lt;String, Boolean&gt;() {<br />
&nbsp;&nbsp;public Boolean call(String s) { return s.contains("ERROR"); }<br />
}</span>);<br />
<span class="comment">// Count all the errors</span><br />
@@ -272,8 +272,8 @@ previous ones, and <em>actions</em>, which kick off a job to execute on a cluste
<div class="tab-content">
<div class="tab-pane tab-pane-python active">
<div class="code code-tab">
- file = spark.textFile(<span class="string">"hdfs://..."</span>)<br />
- counts = file.<span class="sparkop">flatMap</span>(<span class="closure">lambda line: line.split(" ")</span>) \<br />
+ text_file = spark.textFile(<span class="string">"hdfs://..."</span>)<br />
+ counts = text_file.<span class="sparkop">flatMap</span>(<span class="closure">lambda line: line.split(" ")</span>) \<br />
&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;.<span class="sparkop">map</span>(<span class="closure">lambda word: (word, 1)</span>) \<br />
&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;.<span class="sparkop">reduceByKey</span>(<span class="closure">lambda a, b: a + b</span>)<br />
counts.<span class="sparkop">saveAsTextFile</span>(<span class="string">"hdfs://..."</span>)
@@ -281,8 +281,8 @@ previous ones, and <em>actions</em>, which kick off a job to execute on a cluste
</div>
<div class="tab-pane tab-pane-scala">
<div class="code code-tab">
- <span class="keyword">val</span> file = spark.textFile(<span class="string">"hdfs://..."</span>)<br />
- <span class="keyword">val</span> counts = file.<span class="sparkop">flatMap</span>(<span class="closure">line =&gt; line.split(" ")</span>)<br />
+ <span class="keyword">val</span> textFile = spark.textFile(<span class="string">"hdfs://..."</span>)<br />
+ <span class="keyword">val</span> counts = textFile.<span class="sparkop">flatMap</span>(<span class="closure">line =&gt; line.split(" ")</span>)<br />
&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;.<span class="sparkop">map</span>(<span class="closure">word =&gt; (word, 1)</span>)<br />
&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;&nbsp;.<span class="sparkop">reduceByKey</span>(<span class="closure">_ + _</span>)<br />
counts.<span class="sparkop">saveAsTextFile</span>(<span class="string">"hdfs://..."</span>)
@@ -290,8 +290,8 @@ previous ones, and <em>actions</em>, which kick off a job to execute on a cluste
</div>
<div class="tab-pane tab-pane-java">
<div class="code code-tab">
- JavaRDD&lt;String&gt; file = spark.textFile(<span class="string">"hdfs://..."</span>);<br />
- JavaRDD&lt;String&gt; words = file.<span class="sparkop">flatMap</span>(<span class="closure">new FlatMapFunction&lt;String, String&gt;() {<br />
+ JavaRDD&lt;String&gt; textFile = spark.textFile(<span class="string">"hdfs://..."</span>);<br />
+ JavaRDD&lt;String&gt; words = textFile.<span class="sparkop">flatMap</span>(<span class="closure">new FlatMapFunction&lt;String, String&gt;() {<br />
&nbsp;&nbsp;public Iterable&lt;String&gt; call(String s) { return Arrays.asList(s.split(" ")); }<br />
}</span>);<br />
JavaPairRDD&lt;String, Integer&gt; pairs = words.<span class="sparkop">mapToPair</span>(<span class="closure">new PairFunction&lt;String, String, Integer&gt;() {<br />
diff --git a/site/faq.html b/site/faq.html
index 61e88decd..eac53a5f7 100644
--- a/site/faq.html
+++ b/site/faq.html
@@ -213,6 +213,7 @@ Spark is a fast and general processing engine compatible with Hadoop data. It ca
<p class="answer">Starting in version 0.8, Spark is under the <a href="http://www.apache.org/licenses/LICENSE-2.0.html">Apache 2.0 license</a>. Previous versions used the <a href="https://github.com/mesos/spark/blob/branch-0.7/LICENSE">BSD license</a>.</p>
<p class="question">How can I contribute to Spark?</p>
+
<p class="answer">See the <a href="https://cwiki.apache.org/confluence/display/SPARK/Contributing+to+Spark">Contributing to Spark wiki</a> for more information.</p>
<p class="question">Where can I get more help?</p>
diff --git a/site/index.html b/site/index.html
index c8be9b5cd..762a2c0e0 100644
--- a/site/index.html
+++ b/site/index.html
@@ -212,9 +212,9 @@
<div class="col-md-5 col-sm-5 col-padded-top col-center">
<div style="text-align: left; display: inline-block;">
<div class="code">
- file = spark.textFile(<span class="string">"hdfs://..."</span>)<br />
+ text_file = spark.textFile(<span class="string">"hdfs://..."</span>)<br />
&nbsp;<br />
- file.<span class="sparkop">flatMap</span>(<span class="closure">lambda line: line.split()</span>)<br />
+ text_file.<span class="sparkop">flatMap</span>(<span class="closure">lambda&nbsp;line:&nbsp;line.split()</span>)<br />
&nbsp;&nbsp;&nbsp;&nbsp;.<span class="sparkop">map</span>(<span class="closure">lambda word: (word, 1)</span>)<br />
&nbsp;&nbsp;&nbsp;&nbsp;.<span class="sparkop">reduceByKey</span>(<span class="closure">lambda a, b: a+b</span>)
</div>
@@ -222,9 +222,9 @@
</div>
<!--
<div class="code" style="margin-top: 20px; text-align: left; display: inline-block;">
- file = spark.textFile(<span class="string">"hdfs://..."</span>)<br/>
+ text_file = spark.textFile(<span class="string">"hdfs://..."</span>)<br/>
&nbsp;<br/>
- file.<span class="sparkop">filter</span>(<span class="closure">lambda line: "ERROR" in line</span>)<br/>
+ text_file.<span class="sparkop">filter</span>(<span class="closure">lambda line: "ERROR" in line</span>)<br/>
&nbsp;&nbsp;&nbsp;&nbsp;.<span class="sparkop">count</span>()
</div>
-->
diff --git a/site/releases/spark-release-1-3-0.html b/site/releases/spark-release-1-3-0.html
index adda1fb3c..1a00a8e3a 100644
--- a/site/releases/spark-release-1-3-0.html
+++ b/site/releases/spark-release-1-3-0.html
@@ -195,7 +195,7 @@
<h2 id="upgrading-to-spark-13">Upgrading to Spark 1.3</h2>
<p>Spark 1.3 is binary compatible with Spark 1.X releases, so no code changes are necessary. This excludes API’s marked explicitly as unstable.</p>
-<p>As part of stabilizing the Spark SQL API, the <code>SchemaRDD</code> class has been extended renamed to <code>DataFrame</code>. Spark SQL&#8217;s <a href="http://spark.apache.org/docs/1.3.0/sql-programming-guide.html#migration-guide">migration guide</a> describes the upgrade process in detail. Spark SQL also now requires that column identifiers which use reserved words (such as &#8220;string&#8221; or &#8220;table&#8221;) be escaped using backticks.</p>
+<p>As part of stabilizing the Spark SQL API, the <code>SchemaRDD</code> class has been renamed to <code>DataFrame</code>. Spark SQL&#8217;s <a href="http://spark.apache.org/docs/1.3.0/sql-programming-guide.html#migration-guide">migration guide</a> describes the upgrade process in detail. Spark SQL also now requires that column identifiers which use reserved words (such as &#8220;string&#8221; or &#8220;table&#8221;) be escaped using backticks.</p>
<h3 id="known-issues">Known Issues</h3>
<p>This release has few known issues which will be addressed in Spark 1.3.1:</p>
diff --git a/sql/index.md b/sql/index.md
index 4f042235d..09ce9deaa 100644
--- a/sql/index.md
+++ b/sql/index.md
@@ -16,7 +16,7 @@ subproject: SQL
<div class="col-md-7 col-sm-7">
<h2>Integrated</h2>
<p class="lead">
- Seemlessly mix SQL queries with Spark programs.
+ Seamlessly mix SQL queries with Spark programs.
</p>
<p>
Spark SQL lets you query structured data as a distributed dataset (RDD) in Spark, with integrated APIs in Python, Scala and Java.