summaryrefslogtreecommitdiff
path: root/site/docs/1.5.0/api/java/org/apache/spark/SparkContext.html
diff options
context:
space:
mode:
Diffstat (limited to 'site/docs/1.5.0/api/java/org/apache/spark/SparkContext.html')
-rw-r--r--site/docs/1.5.0/api/java/org/apache/spark/SparkContext.html137
1 files changed, 75 insertions, 62 deletions
diff --git a/site/docs/1.5.0/api/java/org/apache/spark/SparkContext.html b/site/docs/1.5.0/api/java/org/apache/spark/SparkContext.html
index 8096be542..ca6320f4f 100644
--- a/site/docs/1.5.0/api/java/org/apache/spark/SparkContext.html
+++ b/site/docs/1.5.0/api/java/org/apache/spark/SparkContext.html
@@ -2,9 +2,9 @@
<!-- NewPage -->
<html lang="en">
<head>
-<!-- Generated by javadoc (version 1.7.0_51) on Tue Sep 08 16:00:15 PDT 2015 -->
+<!-- Generated by javadoc (version 1.7.0_51) on Wed Sep 16 10:26:34 PDT 2015 -->
<title>SparkContext</title>
-<meta name="date" content="2015-09-08">
+<meta name="date" content="2015-09-16">
<link rel="stylesheet" type="text/css" href="../../../stylesheet.css" title="Style">
</head>
<body>
@@ -35,7 +35,7 @@
</div>
<div class="subNav">
<ul class="navList">
-<li><a href="../../../org/apache/spark/SparkConfSuite.html" title="class in org.apache.spark"><span class="strong">Prev Class</span></a></li>
+<li><a href="../../../org/apache/spark/SparkConf.html" title="class in org.apache.spark"><span class="strong">Prev Class</span></a></li>
<li><a href="../../../org/apache/spark/SparkContext.DoubleAccumulatorParam$.html" title="class in org.apache.spark"><span class="strong">Next Class</span></a></li>
</ul>
<ul class="navList">
@@ -631,14 +631,18 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#listenerBus()">listenerBus</a></strong>()</code>&nbsp;</td>
</tr>
<tr class="altColor">
+<td class="colFirst"><code>protected java.lang.InheritableThreadLocal&lt;java.util.Properties&gt;</code></td>
+<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#localProperties()">localProperties</a></strong>()</code>&nbsp;</td>
+</tr>
+<tr class="rowColor">
<td class="colFirst"><code>static org.apache.hadoop.io.LongWritable</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#longToLongWritable(long)">longToLongWritable</a></strong>(long&nbsp;l)</code>&nbsp;</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>static org.apache.spark.WritableConverter&lt;java.lang.Object&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#longWritableConverter()">longWritableConverter</a></strong>()</code>&nbsp;</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>&lt;T&gt;&nbsp;<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#makeRDD(scala.collection.Seq, int, scala.reflect.ClassTag)">makeRDD</a></strong>(scala.collection.Seq&lt;T&gt;&nbsp;seq,
int&nbsp;numSlices,
@@ -646,7 +650,7 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
<div class="block">Distribute a local Scala collection to form an RDD.</div>
</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>&lt;T&gt;&nbsp;<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#makeRDD(scala.collection.Seq, scala.reflect.ClassTag)">makeRDD</a></strong>(scala.collection.Seq&lt;scala.Tuple2&lt;T,scala.collection.Seq&lt;java.lang.String&gt;&gt;&gt;&nbsp;seq,
scala.reflect.ClassTag&lt;T&gt;&nbsp;evidence$3)</code>
@@ -654,15 +658,15 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
location preferences (hostnames of Spark nodes) for each object.</div>
</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>java.lang.String</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#master()">master</a></strong>()</code>&nbsp;</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>org.apache.spark.metrics.MetricsSystem</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#metricsSystem()">metricsSystem</a></strong>()</code>&nbsp;</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>&lt;K,V,F extends org.apache.hadoop.mapreduce.InputFormat&lt;K,V&gt;&gt;&nbsp;<br><a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;scala.Tuple2&lt;K,V&gt;&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#newAPIHadoopFile(java.lang.String, java.lang.Class, java.lang.Class, java.lang.Class, org.apache.hadoop.conf.Configuration)">newAPIHadoopFile</a></strong>(java.lang.String&nbsp;path,
java.lang.Class&lt;F&gt;&nbsp;fClass,
@@ -673,7 +677,7 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
and extra configuration options to pass to the input format.</div>
</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>&lt;K,V,F extends org.apache.hadoop.mapreduce.InputFormat&lt;K,V&gt;&gt;&nbsp;<br><a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;scala.Tuple2&lt;K,V&gt;&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#newAPIHadoopFile(java.lang.String, scala.reflect.ClassTag, scala.reflect.ClassTag, scala.reflect.ClassTag)">newAPIHadoopFile</a></strong>(java.lang.String&nbsp;path,
scala.reflect.ClassTag&lt;K&gt;&nbsp;km,
@@ -682,7 +686,7 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
<div class="block">Get an RDD for a Hadoop file with an arbitrary new API InputFormat.</div>
</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>&lt;K,V,F extends org.apache.hadoop.mapreduce.InputFormat&lt;K,V&gt;&gt;&nbsp;<br><a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;scala.Tuple2&lt;K,V&gt;&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#newAPIHadoopRDD(org.apache.hadoop.conf.Configuration, java.lang.Class, java.lang.Class, java.lang.Class)">newAPIHadoopRDD</a></strong>(org.apache.hadoop.conf.Configuration&nbsp;conf,
java.lang.Class&lt;F&gt;&nbsp;fClass,
@@ -692,12 +696,12 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
and extra configuration options to pass to the input format.</div>
</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>static &lt;T&gt;&nbsp;<a href="../../../org/apache/spark/rdd/DoubleRDDFunctions.html" title="class in org.apache.spark.rdd">DoubleRDDFunctions</a></code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#numericRDDToDoubleRDDFunctions(org.apache.spark.rdd.RDD, scala.math.Numeric)">numericRDDToDoubleRDDFunctions</a></strong>(<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;rdd,
scala.math.Numeric&lt;T&gt;&nbsp;num)</code>&nbsp;</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>&lt;T&gt;&nbsp;<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#objectFile(java.lang.String, int, scala.reflect.ClassTag)">objectFile</a></strong>(java.lang.String&nbsp;path,
int&nbsp;minPartitions,
@@ -706,7 +710,7 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
BytesWritable values that contain a serialized partition.</div>
</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>&lt;T&gt;&nbsp;<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#parallelize(scala.collection.Seq, int, scala.reflect.ClassTag)">parallelize</a></strong>(scala.collection.Seq&lt;T&gt;&nbsp;seq,
int&nbsp;numSlices,
@@ -714,15 +718,15 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
<div class="block">Distribute a local Scala collection to form an RDD.</div>
</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code><any></code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#persistentRdds()">persistentRdds</a></strong>()</code>&nbsp;</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>scala.collection.Map&lt;java.lang.String,scala.collection.Set&lt;<a href="../../../org/apache/spark/scheduler/SplitInfo.html" title="class in org.apache.spark.scheduler">SplitInfo</a>&gt;&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#preferredNodeLocationData()">preferredNodeLocationData</a></strong>()</code>&nbsp;</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code><a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;java.lang.Object&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#range(long, long, long, int)">range</a></strong>(long&nbsp;start,
long&nbsp;end,
@@ -732,34 +736,34 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
<code>step</code> every element.</div>
</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>static java.lang.String</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#RDD_SCOPE_KEY()">RDD_SCOPE_KEY</a></strong>()</code>&nbsp;</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>static java.lang.String</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#RDD_SCOPE_NO_OVERRIDE_KEY()">RDD_SCOPE_NO_OVERRIDE_KEY</a></strong>()</code>&nbsp;</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>static &lt;T&gt;&nbsp;<a href="../../../org/apache/spark/rdd/AsyncRDDActions.html" title="class in org.apache.spark.rdd">AsyncRDDActions</a>&lt;T&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#rddToAsyncRDDActions(org.apache.spark.rdd.RDD, scala.reflect.ClassTag)">rddToAsyncRDDActions</a></strong>(<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;rdd,
scala.reflect.ClassTag&lt;T&gt;&nbsp;evidence$22)</code>&nbsp;</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>static &lt;K,V&gt;&nbsp;<a href="../../../org/apache/spark/rdd/OrderedRDDFunctions.html" title="class in org.apache.spark.rdd">OrderedRDDFunctions</a>&lt;K,V,scala.Tuple2&lt;K,V&gt;&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#rddToOrderedRDDFunctions(org.apache.spark.rdd.RDD, scala.math.Ordering, scala.reflect.ClassTag, scala.reflect.ClassTag)">rddToOrderedRDDFunctions</a></strong>(<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;scala.Tuple2&lt;K,V&gt;&gt;&nbsp;rdd,
scala.math.Ordering&lt;K&gt;&nbsp;evidence$27,
scala.reflect.ClassTag&lt;K&gt;&nbsp;evidence$28,
scala.reflect.ClassTag&lt;V&gt;&nbsp;evidence$29)</code>&nbsp;</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>static &lt;K,V&gt;&nbsp;<a href="../../../org/apache/spark/rdd/PairRDDFunctions.html" title="class in org.apache.spark.rdd">PairRDDFunctions</a>&lt;K,V&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#rddToPairRDDFunctions(org.apache.spark.rdd.RDD, scala.reflect.ClassTag, scala.reflect.ClassTag, scala.math.Ordering)">rddToPairRDDFunctions</a></strong>(<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;scala.Tuple2&lt;K,V&gt;&gt;&nbsp;rdd,
scala.reflect.ClassTag&lt;K&gt;&nbsp;kt,
scala.reflect.ClassTag&lt;V&gt;&nbsp;vt,
scala.math.Ordering&lt;K&gt;&nbsp;ord)</code>&nbsp;</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>static &lt;K,V&gt;&nbsp;<a href="../../../org/apache/spark/rdd/SequenceFileRDDFunctions.html" title="class in org.apache.spark.rdd">SequenceFileRDDFunctions</a>&lt;K,V&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#rddToSequenceFileRDDFunctions(org.apache.spark.rdd.RDD, scala.Function1, scala.reflect.ClassTag, scala.Function1, scala.reflect.ClassTag)">rddToSequenceFileRDDFunctions</a></strong>(<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;scala.Tuple2&lt;K,V&gt;&gt;&nbsp;rdd,
scala.Function1&lt;K,org.apache.hadoop.io.Writable&gt;&nbsp;evidence$23,
@@ -767,14 +771,14 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
scala.Function1&lt;V,org.apache.hadoop.io.Writable&gt;&nbsp;evidence$25,
scala.reflect.ClassTag&lt;V&gt;&nbsp;evidence$26)</code>&nbsp;</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>boolean</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#requestExecutors(int)">requestExecutors</a></strong>(int&nbsp;numAdditionalExecutors)</code>
<div class="block">:: DeveloperApi ::
Request an additional number of executors from the cluster manager.</div>
</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>&lt;T,U,R&gt;&nbsp;<a href="../../../org/apache/spark/partial/PartialResult.html" title="class in org.apache.spark.partial">PartialResult</a>&lt;R&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#runApproximateJob(org.apache.spark.rdd.RDD, scala.Function2, , long)">runApproximateJob</a></strong>(<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;rdd,
scala.Function2&lt;<a href="../../../org/apache/spark/TaskContext.html" title="class in org.apache.spark">TaskContext</a>,scala.collection.Iterator&lt;T&gt;,U&gt;&nbsp;func,
@@ -784,7 +788,7 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
Run a job that can return approximate results.</div>
</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>&lt;T,U&gt;&nbsp;java.lang.Object</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#runJob(org.apache.spark.rdd.RDD, scala.Function1, scala.reflect.ClassTag)">runJob</a></strong>(<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;rdd,
scala.Function1&lt;scala.collection.Iterator&lt;T&gt;,U&gt;&nbsp;func,
@@ -792,7 +796,7 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
<div class="block">Run a job on all partitions in an RDD and return the results in an array.</div>
</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>&lt;T,U&gt;&nbsp;void</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#runJob(org.apache.spark.rdd.RDD, scala.Function1, scala.Function2, scala.reflect.ClassTag)">runJob</a></strong>(<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;rdd,
scala.Function1&lt;scala.collection.Iterator&lt;T&gt;,U&gt;&nbsp;processPartition,
@@ -801,7 +805,7 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
<div class="block">Run a job on all partitions in an RDD and pass the results to a handler function.</div>
</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>&lt;T,U&gt;&nbsp;java.lang.Object</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#runJob(org.apache.spark.rdd.RDD, scala.Function1, scala.collection.Seq, boolean, scala.reflect.ClassTag)">runJob</a></strong>(<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;rdd,
scala.Function1&lt;scala.collection.Iterator&lt;T&gt;,U&gt;&nbsp;func,
@@ -812,7 +816,7 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
<code>Iterator[T] =&gt; U</code> instead of <code>(TaskContext, Iterator[T]) =&gt; U</code>.</div>
</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>&lt;T,U&gt;&nbsp;java.lang.Object</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#runJob(org.apache.spark.rdd.RDD, scala.Function1, scala.collection.Seq, scala.reflect.ClassTag)">runJob</a></strong>(<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;rdd,
scala.Function1&lt;scala.collection.Iterator&lt;T&gt;,U&gt;&nbsp;func,
@@ -822,7 +826,7 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
<code>Iterator[T] =&gt; U</code> instead of <code>(TaskContext, Iterator[T]) =&gt; U</code>.</div>
</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>&lt;T,U&gt;&nbsp;java.lang.Object</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#runJob(org.apache.spark.rdd.RDD, scala.Function2, scala.reflect.ClassTag)">runJob</a></strong>(<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;rdd,
scala.Function2&lt;<a href="../../../org/apache/spark/TaskContext.html" title="class in org.apache.spark">TaskContext</a>,scala.collection.Iterator&lt;T&gt;,U&gt;&nbsp;func,
@@ -830,7 +834,7 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
<div class="block">Run a job on all partitions in an RDD and return the results in an array.</div>
</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>&lt;T,U&gt;&nbsp;void</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#runJob(org.apache.spark.rdd.RDD, scala.Function2, scala.Function2, scala.reflect.ClassTag)">runJob</a></strong>(<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;rdd,
scala.Function2&lt;<a href="../../../org/apache/spark/TaskContext.html" title="class in org.apache.spark">TaskContext</a>,scala.collection.Iterator&lt;T&gt;,U&gt;&nbsp;processPartition,
@@ -839,7 +843,7 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
<div class="block">Run a job on all partitions in an RDD and pass the results to a handler function.</div>
</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>&lt;T,U&gt;&nbsp;java.lang.Object</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#runJob(org.apache.spark.rdd.RDD, scala.Function2, scala.collection.Seq, boolean, scala.reflect.ClassTag)">runJob</a></strong>(<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;rdd,
scala.Function2&lt;<a href="../../../org/apache/spark/TaskContext.html" title="class in org.apache.spark">TaskContext</a>,scala.collection.Iterator&lt;T&gt;,U&gt;&nbsp;func,
@@ -849,7 +853,7 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
<div class="block">Run a function on a given set of partitions in an RDD and return the results as an array.</div>
</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>&lt;T,U&gt;&nbsp;void</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#runJob(org.apache.spark.rdd.RDD, scala.Function2, scala.collection.Seq, boolean, scala.Function2, scala.reflect.ClassTag)">runJob</a></strong>(<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;rdd,
scala.Function2&lt;<a href="../../../org/apache/spark/TaskContext.html" title="class in org.apache.spark">TaskContext</a>,scala.collection.Iterator&lt;T&gt;,U&gt;&nbsp;func,
@@ -861,7 +865,7 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
handler function.</div>
</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>&lt;T,U&gt;&nbsp;java.lang.Object</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#runJob(org.apache.spark.rdd.RDD, scala.Function2, scala.collection.Seq, scala.reflect.ClassTag)">runJob</a></strong>(<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;rdd,
scala.Function2&lt;<a href="../../../org/apache/spark/TaskContext.html" title="class in org.apache.spark">TaskContext</a>,scala.collection.Iterator&lt;T&gt;,U&gt;&nbsp;func,
@@ -870,7 +874,7 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
<div class="block">Run a function on a given set of partitions in an RDD and return the results as an array.</div>
</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>&lt;T,U&gt;&nbsp;void</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#runJob(org.apache.spark.rdd.RDD, scala.Function2, scala.collection.Seq, scala.Function2, scala.reflect.ClassTag)">runJob</a></strong>(<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;rdd,
scala.Function2&lt;<a href="../../../org/apache/spark/TaskContext.html" title="class in org.apache.spark">TaskContext</a>,scala.collection.Iterator&lt;T&gt;,U&gt;&nbsp;func,
@@ -881,7 +885,7 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
handler function.</div>
</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>&lt;K,V&gt;&nbsp;<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;scala.Tuple2&lt;K,V&gt;&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#sequenceFile(java.lang.String, java.lang.Class, java.lang.Class)">sequenceFile</a></strong>(java.lang.String&nbsp;path,
java.lang.Class&lt;K&gt;&nbsp;keyClass,
@@ -889,7 +893,7 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
<div class="block">Get an RDD for a Hadoop SequenceFile with given key and value types.</div>
</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>&lt;K,V&gt;&nbsp;<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;scala.Tuple2&lt;K,V&gt;&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#sequenceFile(java.lang.String, java.lang.Class, java.lang.Class, int)">sequenceFile</a></strong>(java.lang.String&nbsp;path,
java.lang.Class&lt;K&gt;&nbsp;keyClass,
@@ -898,7 +902,7 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
<div class="block">Get an RDD for a Hadoop SequenceFile with given key and value types.</div>
</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>&lt;K,V&gt;&nbsp;<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;scala.Tuple2&lt;K,V&gt;&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#sequenceFile(java.lang.String, int, scala.reflect.ClassTag, scala.reflect.ClassTag, scala.Function0, scala.Function0)">sequenceFile</a></strong>(java.lang.String&nbsp;path,
int&nbsp;minPartitions,
@@ -910,26 +914,26 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
WritableConverter.</div>
</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>void</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#setCallSite(java.lang.String)">setCallSite</a></strong>(java.lang.String&nbsp;shortCallSite)</code>
<div class="block">Set the thread-local property for overriding the call sites
of actions and RDDs.</div>
</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>void</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#setCheckpointDir(java.lang.String)">setCheckpointDir</a></strong>(java.lang.String&nbsp;directory)</code>
<div class="block">Set the directory under which RDDs are going to be checkpointed.</div>
</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>void</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#setJobDescription(java.lang.String)">setJobDescription</a></strong>(java.lang.String&nbsp;value)</code>
<div class="block">Set a human readable description of the current job.</div>
</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>void</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#setJobGroup(java.lang.String, java.lang.String, boolean)">setJobGroup</a></strong>(java.lang.String&nbsp;groupId,
java.lang.String&nbsp;description,
@@ -938,7 +942,7 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
different value or cleared.</div>
</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>void</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#setLocalProperty(java.lang.String, java.lang.String)">setLocalProperty</a></strong>(java.lang.String&nbsp;key,
java.lang.String&nbsp;value)</code>
@@ -946,49 +950,49 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
Spark fair scheduler pool.</div>
</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>void</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#setLogLevel(java.lang.String)">setLogLevel</a></strong>(java.lang.String&nbsp;logLevel)</code>
<div class="block">Control our logLevel.</div>
</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>static java.lang.String</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#SPARK_JOB_DESCRIPTION()">SPARK_JOB_DESCRIPTION</a></strong>()</code>&nbsp;</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>static java.lang.String</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#SPARK_JOB_GROUP_ID()">SPARK_JOB_GROUP_ID</a></strong>()</code>&nbsp;</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>static java.lang.String</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#SPARK_JOB_INTERRUPT_ON_CANCEL()">SPARK_JOB_INTERRUPT_ON_CANCEL</a></strong>()</code>&nbsp;</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>java.lang.String</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#sparkUser()">sparkUser</a></strong>()</code>&nbsp;</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>long</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#startTime()">startTime</a></strong>()</code>&nbsp;</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code><a href="../../../org/apache/spark/SparkStatusTracker.html" title="class in org.apache.spark">SparkStatusTracker</a></code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#statusTracker()">statusTracker</a></strong>()</code>&nbsp;</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>void</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#stop()">stop</a></strong>()</code>&nbsp;</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>static org.apache.hadoop.io.Text</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#stringToText(java.lang.String)">stringToText</a></strong>(java.lang.String&nbsp;s)</code>&nbsp;</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>static org.apache.spark.WritableConverter&lt;java.lang.String&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#stringWritableConverter()">stringWritableConverter</a></strong>()</code>&nbsp;</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>&lt;T,U,R&gt;&nbsp;<a href="../../../org/apache/spark/SimpleFutureAction.html" title="class in org.apache.spark">SimpleFutureAction</a>&lt;R&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#submitJob(org.apache.spark.rdd.RDD, scala.Function1, scala.collection.Seq, scala.Function2, scala.Function0)">submitJob</a></strong>(<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;rdd,
scala.Function1&lt;scala.collection.Iterator&lt;T&gt;,U&gt;&nbsp;processPartition,
@@ -999,11 +1003,11 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
Submit a job for execution and return a FutureJob holding the result.</div>
</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>java.lang.String</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#tachyonFolderName()">tachyonFolderName</a></strong>()</code>&nbsp;</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code><a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;java.lang.String&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#textFile(java.lang.String, int)">textFile</a></strong>(java.lang.String&nbsp;path,
int&nbsp;minPartitions)</code>
@@ -1011,7 +1015,7 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
Hadoop-supported file system URI, and return it as an RDD of Strings.</div>
</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>&lt;T&gt;&nbsp;<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#union(org.apache.spark.rdd.RDD, scala.collection.Seq, scala.reflect.ClassTag)">union</a></strong>(<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&nbsp;first,
scala.collection.Seq&lt;<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&gt;&nbsp;rest,
@@ -1019,20 +1023,20 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
<div class="block">Build the union of a list of RDDs passed as variable-length arguments.</div>
</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code>&lt;T&gt;&nbsp;<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#union(scala.collection.Seq, scala.reflect.ClassTag)">union</a></strong>(scala.collection.Seq&lt;<a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;T&gt;&gt;&nbsp;rdds,
scala.reflect.ClassTag&lt;T&gt;&nbsp;evidence$6)</code>
<div class="block">Build the union of a list of RDDs.</div>
</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>java.lang.String</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#version()">version</a></strong>()</code>
<div class="block">The version of Spark on which this application is running.</div>
</td>
</tr>
-<tr class="rowColor">
+<tr class="altColor">
<td class="colFirst"><code><a href="../../../org/apache/spark/rdd/RDD.html" title="class in org.apache.spark.rdd">RDD</a>&lt;scala.Tuple2&lt;java.lang.String,java.lang.String&gt;&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#wholeTextFiles(java.lang.String, int)">wholeTextFiles</a></strong>(java.lang.String&nbsp;path,
int&nbsp;minPartitions)</code>
@@ -1040,7 +1044,7 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
Hadoop-supported file system URI.</div>
</td>
</tr>
-<tr class="altColor">
+<tr class="rowColor">
<td class="colFirst"><code>static &lt;T extends org.apache.hadoop.io.Writable&gt;&nbsp;<br>org.apache.spark.WritableConverter&lt;T&gt;</code></td>
<td class="colLast"><code><strong><a href="../../../org/apache/spark/SparkContext.html#writableWritableConverter()">writableWritableConverter</a></strong>()</code>&nbsp;</td>
</tr>
@@ -1696,6 +1700,15 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
<pre>public&nbsp;scala.Option&lt;java.lang.String&gt;&nbsp;checkpointDir()</pre>
</li>
</ul>
+<a name="localProperties()">
+<!-- -->
+</a>
+<ul class="blockList">
+<li class="blockList">
+<h4>localProperties</h4>
+<pre>protected&nbsp;java.lang.InheritableThreadLocal&lt;java.util.Properties&gt;&nbsp;localProperties()</pre>
+</li>
+</ul>
<a name="setLogLevel(java.lang.String)">
<!-- -->
</a>
@@ -2926,7 +2939,7 @@ implements <a href="../../../org/apache/spark/Logging.html" title="interface in
</div>
<div class="subNav">
<ul class="navList">
-<li><a href="../../../org/apache/spark/SparkConfSuite.html" title="class in org.apache.spark"><span class="strong">Prev Class</span></a></li>
+<li><a href="../../../org/apache/spark/SparkConf.html" title="class in org.apache.spark"><span class="strong">Prev Class</span></a></li>
<li><a href="../../../org/apache/spark/SparkContext.DoubleAccumulatorParam$.html" title="class in org.apache.spark"><span class="strong">Next Class</span></a></li>
</ul>
<ul class="navList">