aboutsummaryrefslogtreecommitdiff
path: root/python/pyspark
diff options
context:
space:
mode:
Diffstat (limited to 'python/pyspark')
-rw-r--r--python/pyspark/context.py3
-rw-r--r--python/pyspark/rdd.py4
2 files changed, 3 insertions, 4 deletions
diff --git a/python/pyspark/context.py b/python/pyspark/context.py
index 43bde5ae41..eb5b0bbbda 100644
--- a/python/pyspark/context.py
+++ b/python/pyspark/context.py
@@ -913,8 +913,7 @@ class SparkContext(object):
# by runJob() in order to avoid having to pass a Python lambda into
# SparkContext#runJob.
mappedRDD = rdd.mapPartitions(partitionFunc)
- port = self._jvm.PythonRDD.runJob(self._jsc.sc(), mappedRDD._jrdd, partitions,
- allowLocal)
+ port = self._jvm.PythonRDD.runJob(self._jsc.sc(), mappedRDD._jrdd, partitions)
return list(_load_from_socket(port, mappedRDD._jrdd_deserializer))
def show_profiles(self):
diff --git a/python/pyspark/rdd.py b/python/pyspark/rdd.py
index 7e788148d9..fa8e0a0574 100644
--- a/python/pyspark/rdd.py
+++ b/python/pyspark/rdd.py
@@ -1293,7 +1293,7 @@ class RDD(object):
taken += 1
p = range(partsScanned, min(partsScanned + numPartsToTry, totalParts))
- res = self.context.runJob(self, takeUpToNumLeft, p, True)
+ res = self.context.runJob(self, takeUpToNumLeft, p)
items += res
partsScanned += numPartsToTry
@@ -2193,7 +2193,7 @@ class RDD(object):
values = self.filter(lambda kv: kv[0] == key).values()
if self.partitioner is not None:
- return self.ctx.runJob(values, lambda x: x, [self.partitioner(key)], False)
+ return self.ctx.runJob(values, lambda x: x, [self.partitioner(key)])
return values.collect()