diff options
author | Patrick Wendell <pwendell@apache.org> | 2015-03-13 15:48:06 +0000 |
---|---|---|
committer | Patrick Wendell <pwendell@apache.org> | 2015-03-13 15:48:06 +0000 |
commit | c6275f3c1cc5e4e506f7e2f6c4356d0cf661d120 (patch) | |
tree | 4b961e9284b08829616bb52cc00c703de6931c00 /site/examples.html | |
parent | aaf670598f15d47784d8b91b34dd614d0f8162af (diff) | |
download | spark-website-c6275f3c1cc5e4e506f7e2f6c4356d0cf661d120.tar.gz spark-website-c6275f3c1cc5e4e506f7e2f6c4356d0cf661d120.tar.bz2 spark-website-c6275f3c1cc5e4e506f7e2f6c4356d0cf661d120.zip |
Initial 1.3.0 code
Diffstat (limited to 'site/examples.html')
-rw-r--r-- | site/examples.html | 24 |
1 files changed, 12 insertions, 12 deletions
diff --git a/site/examples.html b/site/examples.html index 01ba0ec87..5b7eab2e2 100644 --- a/site/examples.html +++ b/site/examples.html @@ -187,8 +187,8 @@ previous ones, and <em>actions</em>, which kick off a job to execute on a cluste <div class="tab-content"> <div class="tab-pane tab-pane-python active"> <div class="code code-tab"> - text_file = spark.textFile(<span class="string">"hdfs://..."</span>)<br /> - errors = text_file.<span class="sparkop">filter</span>(<span class="closure">lambda line: "ERROR" in line</span>)<br /> + file = spark.textFile(<span class="string">"hdfs://..."</span>)<br /> + errors = file.<span class="sparkop">filter</span>(<span class="closure">lambda line: "ERROR" in line</span>)<br /> <span class="comment"># Count all the errors</span><br /> errors.<span class="sparkop">count</span>()<br /> <span class="comment"># Count errors mentioning MySQL</span><br /> @@ -199,8 +199,8 @@ previous ones, and <em>actions</em>, which kick off a job to execute on a cluste </div> <div class="tab-pane tab-pane-scala"> <div class="code code-tab"> - <span class="keyword">val</span> textFile = spark.textFile(<span class="string">"hdfs://..."</span>)<br /> - <span class="keyword">val</span> errors = textFile.<span class="sparkop">filter</span>(<span class="closure">line => line.contains("ERROR")</span>)<br /> + <span class="keyword">val</span> file = spark.textFile(<span class="string">"hdfs://..."</span>)<br /> + <span class="keyword">val</span> errors = file.<span class="sparkop">filter</span>(<span class="closure">line => line.contains("ERROR")</span>)<br /> <span class="comment">// Count all the errors</span><br /> errors.<span class="sparkop">count</span>()<br /> <span class="comment">// Count errors mentioning MySQL</span><br /> @@ -211,8 +211,8 @@ previous ones, and <em>actions</em>, which kick off a job to execute on a cluste </div> <div class="tab-pane tab-pane-java"> <div class="code code-tab"> - JavaRDD<String> textFile = spark.textFile(<span class="string">"hdfs://..."</span>);<br /> - JavaRDD<String> errors = textFile.<span class="sparkop">filter</span>(<span class="closure">new Function<String, Boolean>() {<br /> + JavaRDD<String> file = spark.textFile(<span class="string">"hdfs://..."</span>);<br /> + JavaRDD<String> errors = file.<span class="sparkop">filter</span>(<span class="closure">new Function<String, Boolean>() {<br /> public Boolean call(String s) { return s.contains("ERROR"); }<br /> }</span>);<br /> <span class="comment">// Count all the errors</span><br /> @@ -272,8 +272,8 @@ previous ones, and <em>actions</em>, which kick off a job to execute on a cluste <div class="tab-content"> <div class="tab-pane tab-pane-python active"> <div class="code code-tab"> - text_file = spark.textFile(<span class="string">"hdfs://..."</span>)<br /> - counts = text_file.<span class="sparkop">flatMap</span>(<span class="closure">lambda line: line.split(" ")</span>) \<br /> + file = spark.textFile(<span class="string">"hdfs://..."</span>)<br /> + counts = file.<span class="sparkop">flatMap</span>(<span class="closure">lambda line: line.split(" ")</span>) \<br /> .<span class="sparkop">map</span>(<span class="closure">lambda word: (word, 1)</span>) \<br /> .<span class="sparkop">reduceByKey</span>(<span class="closure">lambda a, b: a + b</span>)<br /> counts.<span class="sparkop">saveAsTextFile</span>(<span class="string">"hdfs://..."</span>) @@ -281,8 +281,8 @@ previous ones, and <em>actions</em>, which kick off a job to execute on a cluste </div> <div class="tab-pane tab-pane-scala"> <div class="code code-tab"> - <span class="keyword">val</span> textFile = spark.textFile(<span class="string">"hdfs://..."</span>)<br /> - <span class="keyword">val</span> counts = textFile.<span class="sparkop">flatMap</span>(<span class="closure">line => line.split(" ")</span>)<br /> + <span class="keyword">val</span> file = spark.textFile(<span class="string">"hdfs://..."</span>)<br /> + <span class="keyword">val</span> counts = file.<span class="sparkop">flatMap</span>(<span class="closure">line => line.split(" ")</span>)<br /> .<span class="sparkop">map</span>(<span class="closure">word => (word, 1)</span>)<br /> .<span class="sparkop">reduceByKey</span>(<span class="closure">_ + _</span>)<br /> counts.<span class="sparkop">saveAsTextFile</span>(<span class="string">"hdfs://..."</span>) @@ -290,8 +290,8 @@ previous ones, and <em>actions</em>, which kick off a job to execute on a cluste </div> <div class="tab-pane tab-pane-java"> <div class="code code-tab"> - JavaRDD<String> textFile = spark.textFile(<span class="string">"hdfs://..."</span>);<br /> - JavaRDD<String> words = textFile.<span class="sparkop">flatMap</span>(<span class="closure">new FlatMapFunction<String, String>() {<br /> + JavaRDD<String> file = spark.textFile(<span class="string">"hdfs://..."</span>);<br /> + JavaRDD<String> words = file.<span class="sparkop">flatMap</span>(<span class="closure">new FlatMapFunction<String, String>() {<br /> public Iterable<String> call(String s) { return Arrays.asList(s.split(" ")); }<br /> }</span>);<br /> JavaPairRDD<String, Integer> pairs = words.<span class="sparkop">mapToPair</span>(<span class="closure">new PairFunction<String, String, Integer>() {<br /> |