aboutsummaryrefslogtreecommitdiff
path: root/python/pyspark/rdd.py
Commit message (Expand)AuthorAgeFilesLines
* Make Python function/line appear in the UI.Tor Myklebust2013-12-281-11/+55
* Merge pull request #276 from shivaram/collectPartitionReynold Xin2013-12-191-1/+6
|\
| * Make collectPartitions take an array of partitionsShivaram Venkataraman2013-12-191-1/+6
| * Add collectPartition to JavaRDD interface.Shivaram Venkataraman2013-12-181-1/+1
* | Add toString to Java RDD, and __repr__ to Python RDDNick Pentreath2013-12-191-0/+3
|/
* Merge branch 'master' into akka-bug-fixPrashant Sharma2013-12-111-1/+4
|\
| * Fix UnicodeEncodeError in PySpark saveAsTextFile().Josh Rosen2013-11-281-1/+4
* | Merge branch 'master' into wip-scala-2.10Prashant Sharma2013-11-271-43/+54
|\|
| * FramedSerializer: _dumps => dumps, _loads => loads.Josh Rosen2013-11-101-2/+2
| * Send PySpark commands as bytes insetad of strings.Josh Rosen2013-11-101-6/+6
| * Add custom serializer support to PySpark.Josh Rosen2013-11-101-39/+47
| * Remove Pickle-wrapping of Java objects in PySpark.Josh Rosen2013-11-031-4/+7
* | Merge branch 'master' of github.com:apache/incubator-spark into scala-2.10Prashant Sharma2013-10-101-7/+53
|\|
| * Fix PySpark docs and an overly long line of code after fdbae41eMatei Zaharia2013-10-091-8/+8
| * SPARK-705: implement sortByKey() in PySparkAndre Schumacher2013-10-071-1/+47
* | Merge branch 'master' into wip-merge-masterPrashant Sharma2013-10-081-4/+6
|\|
| * Fixing SPARK-602: PythonPartitionerAndre Schumacher2013-10-041-4/+6
* | Merge branch 'master' of git://github.com/mesos/spark into scala-2.10Prashant Sharma2013-09-151-0/+19
|\|
| * Export StorageLevel and refactorAaron Davidson2013-09-071-1/+2
| * SPARK-660: Add StorageLevel support in PythonAaron Davidson2013-09-051-0/+18
* | Merged with masterPrashant Sharma2013-09-061-20/+188
|\|
| * Merge pull request #861 from AndreSchumacher/pyspark_sampling_functionMatei Zaharia2013-08-311-7/+55
| |\
| | * RDD sample() and takeSample() prototypes for PySparkAndre Schumacher2013-08-281-7/+55
| * | PySpark: implementing subtractByKey(), subtract() and keyBy()Andre Schumacher2013-08-281-0/+37
| |/
| * Implementing SPARK-838: Add DoubleRDDFunctions methods to PySparkAndre Schumacher2013-08-211-1/+59
| * Implementing SPARK-878 for PySpark: adding zip and egg files to context and p...Andre Schumacher2013-08-161-1/+3
| * Do not inherit master's PYTHONPATH on workers.Josh Rosen2013-07-291-3/+2
| * Use None instead of empty string as it's slightly smaller/fasterMatei Zaharia2013-07-291-1/+1
| * Optimize Python foreach() to not return as many objectsMatei Zaharia2013-07-291-1/+5
| * Optimize Python take() to not compute entire first partitionMatei Zaharia2013-07-291-6/+9
| * Add Apache license headers and LICENSE and NOTICE filesMatei Zaharia2013-07-161-0/+17
* | PySpark: replacing class manifest by class tag for Scala 2.10.2 inside rdd.pyAndre Schumacher2013-08-301-2/+2
|/
* Fix Python saveAsTextFile doctest to not expect order to be preservedJey Kottalam2013-04-021-1/+1
* Change numSplits to numPartitions in PySpark.Josh Rosen2013-02-241-28/+28
* Add commutative requirement for 'reduce' to Python docstring.Mark Hamstra2013-02-091-2/+2
* Fetch fewer objects in PySpark's take() method.Josh Rosen2013-02-031-0/+4
* Fix reporting of PySpark doctest failures.Josh Rosen2013-02-031-1/+3
* Use spark.local.dir for PySpark temp files (SPARK-580).Josh Rosen2013-02-011-6/+1
* Do not launch JavaGateways on workers (SPARK-674).Josh Rosen2013-02-011-6/+6
* Merge pull request #389 from JoshRosen/python_rdd_checkpointingMatei Zaharia2013-01-201-1/+34
|\
| * Clean up setup code in PySpark checkpointing testsJosh Rosen2013-01-201-2/+1
| * Update checkpointing API docs in Python/Java.Josh Rosen2013-01-201-12/+5
| * Add checkpointFile() and more tests to PySpark.Josh Rosen2013-01-201-1/+8
| * Add RDD checkpointing to Python API.Josh Rosen2013-01-201-0/+34
* | Fix PythonPartitioner equality; see SPARK-654.Josh Rosen2013-01-201-6/+11
|/
* Added accumulators to PySparkMatei Zaharia2013-01-201-1/+1
* Add mapPartitionsWithSplit() to PySpark.Josh Rosen2013-01-081-11/+22
* Change PySpark RDD.take() to not call iterator().Josh Rosen2013-01-031-6/+5
* Rename top-level 'pyspark' directory to 'python'Josh Rosen2013-01-011-0/+713