aboutsummaryrefslogtreecommitdiff
path: root/python/pyspark/context.py
diff options
context:
space:
mode:
authorzsxwing <zsxwing@gmail.com>2015-06-17 13:59:39 -0700
committerAndrew Or <andrew@databricks.com>2015-06-17 13:59:39 -0700
commit0fc4b96f3e3bf81724ac133a6acc97c1b77271b4 (patch)
tree7d5e783967ac602ce979fe0f15a3d3613f44a4f8 /python/pyspark/context.py
parent2837e067099921dd4ab6639ac5f6e89f789d4ff4 (diff)
downloadspark-0fc4b96f3e3bf81724ac133a6acc97c1b77271b4.tar.gz
spark-0fc4b96f3e3bf81724ac133a6acc97c1b77271b4.tar.bz2
spark-0fc4b96f3e3bf81724ac133a6acc97c1b77271b4.zip
[SPARK-8373] [PYSPARK] Add emptyRDD to pyspark and fix the issue when calling sum on an empty RDD
This PR fixes the sum issue and also adds `emptyRDD` so that it's easy to create a test case. Author: zsxwing <zsxwing@gmail.com> Closes #6826 from zsxwing/python-emptyRDD and squashes the following commits: b36993f [zsxwing] Update the return type to JavaRDD[T] 71df047 [zsxwing] Add emptyRDD to pyspark and fix the issue when calling sum on an empty RDD
Diffstat (limited to 'python/pyspark/context.py')
-rw-r--r--python/pyspark/context.py6
1 files changed, 6 insertions, 0 deletions
diff --git a/python/pyspark/context.py b/python/pyspark/context.py
index 44d90f1437..90b2fffbb9 100644
--- a/python/pyspark/context.py
+++ b/python/pyspark/context.py
@@ -324,6 +324,12 @@ class SparkContext(object):
with SparkContext._lock:
SparkContext._active_spark_context = None
+ def emptyRDD(self):
+ """
+ Create an RDD that has no partitions or elements.
+ """
+ return RDD(self._jsc.emptyRDD(), self, NoOpSerializer())
+
def range(self, start, end=None, step=1, numSlices=None):
"""
Create a new RDD of int containing elements from `start` to `end`