aboutsummaryrefslogtreecommitdiff
path: root/python/pyspark/rdd.py
Commit message (Expand)AuthorAgeFilesLines
* [SPARK-8738] [SQL] [PYSPARK] capture SQL AnalysisException in Python APIDavies Liu2015-06-301-1/+2
* [SPARK-7810] [PYSPARK] solve python rdd socket connection problemAi He2015-06-291-3/+15
* [SPARK-8541] [PYSPARK] test the absolute error in approx doctestsScott Taylor2015-06-221-2/+2
* [SPARK-8373] [PYSPARK] Add emptyRDD to pyspark and fix the issue when calling...zsxwing2015-06-171-1/+1
* [SPARK-6416] [DOCS] RDD.fold() requires the operator to be commutativeSean Owen2015-05-211-2/+10
* [SPARK-6216] [PYSPARK] check python version of worker with driverDavies Liu2015-05-181-2/+2
* [SPARK-7438] [SPARK CORE] Fixed validation of relativeSD in countApproxDistinctVinod K C2015-05-091-2/+0
* [SPARK-6949] [SQL] [PySpark] Support Date/Timestamp in Column expressionDavies Liu2015-04-211-0/+3
* [SPARK-4897] [PySpark] Python 3 supportDavies Liu2015-04-161-80/+109
* [SPARK-6886] [PySpark] fix big closure with shuffleDavies Liu2015-04-151-10/+5
* [SPARK-6216] [PySpark] check the python version in workerDavies Liu2015-04-101-1/+1
* [SPARK-5969][PySpark] Fix descending pyspark.rdd.sortByKey.Milan Straka2015-04-101-1/+1
* [SPARK-3074] [PySpark] support groupByKey() with single huge keyDavies Liu2015-04-091-12/+36
* [SPARK-6667] [PySpark] remove setReuseAddressDavies Liu2015-04-021-0/+1
* [SPARK-6370][core] Documentation: Improve all 3 docs for RDD.samplembonaci2015-03-201-0/+6
* [SPARK-6194] [SPARK-677] [PySpark] fix memory leak in collect()Davies Liu2015-03-091-16/+14
* [SPARK-5944] [PySpark] fix version in Python API docsDavies Liu2015-02-251-0/+4
* [SPARK-5973] [PySpark] fix zip with two RDDs with AutoBatchedSerializerDavies Liu2015-02-241-1/+1
* [SPARK-5785] [PySpark] narrow dependency for cogroup/join in PySparkDavies Liu2015-02-171-16/+33
* SPARK-5633 pyspark saveAsTextFile support for compression codecVladimir Vladimirov2015-02-061-2/+20
* [SPARK-5577] Python udf for DataFrameDavies Liu2015-02-041-16/+22
* [SPARK-5430] move treeReduce and treeAggregate from mllib to coreXiangrui Meng2015-01-281-1/+90
* [SPARK-4387][PySpark] Refactoring python profiling code to make it extensibleYandu Oppacher2015-01-281-6/+9
* [SPARK-5440][pyspark] Add toLocalIterator to pyspark rddMichael Nazario2015-01-281-0/+14
* SPARK-5458. Refer to aggregateByKey instead of combineByKey in docsSandy Ryza2015-01-281-2/+2
* [SPARK-5063] More helpful error messages for several invalid operationsJosh Rosen2015-01-231-0/+11
* SPARK-5270 [CORE] Provide isEmpty() function in RDD APISean Owen2015-01-191-0/+12
* [SPARK-4822] Use sphinx tags for Python doc annotationslewuathe2014-12-171-4/+4
* [SPARK-4841] fix zip with textFile()Davies Liu2014-12-151-14/+11
* [SPARK-4477] [PySpark] remove numpy from RDDSamplerDavies Liu2014-11-201-4/+6
* [SPARK-4327] [PySpark] Python API for RDD.randomSplit()Davies Liu2014-11-181-3/+27
* [SPARK-4304] [PySpark] Fix sort on empty RDDDavies Liu2014-11-071-0/+2
* [SPARK-3886] [PySpark] simplify serializer, use AutoBatchedSerializer by defa...Davies Liu2014-11-031-54/+37
* [SPARK-4148][PySpark] fix seed distribution and add some tests for rdd.sampleXiangrui Meng2014-11-031-3/+0
* [SPARK-4150][PySpark] return self in rdd.setNameXiangrui Meng2014-10-311-2/+2
* [Spark] RDD take() method: overestimate too muchyingjieMiao2014-10-131-1/+4
* [SPARK-3909][PySpark][Doc] A corrupted format in Sphinx documents and buildin...cocoatomo2014-10-111-1/+1
* [SPARK-3412] [PySpark] Replace Epydoc with Sphinx to generate Python API docsDavies Liu2014-10-071-26/+26
* [SPARK-3773][PySpark][Doc] Sphinx build warningcocoatomo2014-10-061-0/+1
* [SPARK-3749] [PySpark] fix bugs in broadcast large closure of RDDDavies Liu2014-10-011-3/+9
* [SPARK-3478] [PySpark] Profile the Python tasksDavies Liu2014-09-301-2/+8
* Revert "[SPARK-3478] [PySpark] Profile the Python tasks"Josh Rosen2014-09-261-8/+2
* [SPARK-3478] [PySpark] Profile the Python tasksDavies Liu2014-09-261-2/+8
* [SPARK-546] Add full outer join to RDD and DStream.Aaron Staple2014-09-241-2/+23
* [SPARK-3491] [MLlib] [PySpark] use pickle to serialize data in MLlibDavies Liu2014-09-191-5/+5
* [SPARK-3554] [PySpark] use broadcast automatically for large closureDavies Liu2014-09-181-0/+4
* [SPARK-3519] add distinct(n) to PySparkMatthew Farrellee2014-09-161-2/+2
* [SPARK-1087] Move python traceback utilities into new traceback_utils.py file.Aaron Staple2014-09-151-55/+3
* [PySpark] Add blank line so that Python RDD.top() docstring renders correctlyRJ Nowling2014-09-121-0/+1
* SPARK-2978. Transformation with MR shuffle semanticsSandy Ryza2014-09-081-0/+24