aboutsummaryrefslogtreecommitdiff
path: root/python
Commit message (Expand)AuthorAgeFilesLines
* [SPARK-3681] [SQL] [PySpark] fix serialization of List and Map in SchemaRDDDavies Liu2014-09-272-27/+34
* Revert "[SPARK-3478] [PySpark] Profile the Python tasks"Josh Rosen2014-09-266-108/+7
* [SPARK-3478] [PySpark] Profile the Python tasksDavies Liu2014-09-266-7/+108
* [SPARK-546] Add full outer join to RDD and DStream.Aaron Staple2014-09-242-2/+39
* [SPARK-3679] [PySpark] pickle the exact globals of functionsDavies Liu2014-09-242-6/+54
* [SPARK-3634] [PySpark] User's module should take precedence over system modulesDavies Liu2014-09-244-8/+48
* [PySpark] remove unnecessary use of numSlices from pyspark testsMatthew Farrellee2014-09-201-2/+2
* [SPARK-3592] [SQL] [PySpark] support applySchema to RDD of RowDavies Liu2014-09-192-4/+20
* [SPARK-3491] [MLlib] [PySpark] use pickle to serialize data in MLlibDavies Liu2014-09-1916-917/+650
* [SPARK-3554] [PySpark] use broadcast automatically for large closureDavies Liu2014-09-184-3/+19
* [SPARK-3430] [PySpark] [Doc] generate PySpark API docs using SphinxDavies Liu2014-09-1613-5/+944
* [SPARK-2314][SQL] Override collect and take in python library, and count in j...Aaron Staple2014-09-161-5/+42
* [SPARK-3519] add distinct(n) to PySparkMatthew Farrellee2014-09-163-4/+24
* [SPARK-1087] Move python traceback utilities into new traceback_utils.py file.Aaron Staple2014-09-153-61/+83
* [SPARK-2951] [PySpark] support unpickle array.array for Python 2.6Davies Liu2014-09-152-2/+1
* [SPARK-3516] [mllib] DecisionTree: Add minInstancesPerNode, minInfoGain param...qiping.lqp2014-09-151-4/+12
* [SPARK-3463] [PySpark] aggregate and show spilled bytes in PythonDavies Liu2014-09-133-14/+34
* [SPARK-3030] [PySpark] Reuse Python workerDavies Liu2014-09-135-28/+70
* [SPARK-3500] [SQL] use JavaSchemaRDD as SchemaRDD._jschema_rddDavies Liu2014-09-122-20/+46
* [SPARK-3094] [PySpark] compatitable with PyPyDavies Liu2014-09-125-118/+172
* [PySpark] Add blank line so that Python RDD.top() docstring renders correctlyRJ Nowling2014-09-121-0/+1
* [SPARK-3047] [PySpark] add an option to use str in textFileRDDDavies Liu2014-09-112-11/+23
* [SPARK-3458] enable python "with" statements for SparkContextMatthew Farrellee2014-09-092-0/+43
* [SPARK-3443][MLLIB] update default values of tree:Xiangrui Meng2014-09-081-2/+2
* [SPARK-3417] Use new-style classes in PySparkMatthew Rocklin2014-09-084-4/+4
* Provide a default PYSPARK_PYTHON for python/run_testsMatthew Farrellee2014-09-081-0/+2
* SPARK-2978. Transformation with MR shuffle semanticsSandy Ryza2014-09-082-0/+32
* SPARK-3337 Paranoid quoting in shell to allow install dirs with spaces within.Prashant Sharma2014-09-081-2/+4
* [SPARK-3415] [PySpark] removes SerializingAdapter codeWard Viaene2014-09-072-5/+12
* [SPARK-2334] fix AttributeError when call PipelineRDD.id()Davies Liu2014-09-063-4/+20
* [SPARK-3273][SPARK-3301]We should read the version information from the same ...GuoQiang Li2014-09-061-2/+2
* Spark-3406 add a default storage level to python RDD persist APIHolden Karau2014-09-062-2/+8
* SPARK-3211 .take() is OOM-prone with empty partitionsAndrew Ash2014-09-051-4/+4
* [SPARK-3378] [DOCS] Replace the word "SparkSQL" with right word "Spark SQL"Kousuke Saruta2014-09-042-4/+4
* [SPARK-3401][PySpark] Wrong usage of tee command in python/run-testsKousuke Saruta2014-09-041-1/+1
* [SPARK-2435] Add shutdown hook to pysparkMatthew Farrellee2014-09-031-0/+2
* [SPARK-3335] [SQL] [PySpark] support broadcast in Python UDFDavies Liu2014-09-032-8/+31
* [SPARK-3309] [PySpark] Put all public API in __all__Davies Liu2014-09-0317-26/+81
* [SPARK-2871] [PySpark] add countApproxDistinct() APIDavies Liu2014-09-022-5/+50
* SPARK-3318: Documentation update in addFile on how to use SparkFiles.getHolden Karau2014-08-301-2/+2
* [SPARK-3307] [PySpark] Fix doc string of SparkContext.broadcast()Davies Liu2014-08-291-2/+0
* [SPARK-2871] [PySpark] add RDD.lookup(key)Davies Liu2014-08-271-132/+79
* [SPARK-3167] Handle special driver configs in WindowsAndrew Or2014-08-261-0/+17
* [SPARK-3073] [PySpark] use external sort in sortBy() and sortByKey()Davies Liu2014-08-264-11/+1021
* [SPARK-2969][SQL] Make ScalaReflection be able to handle ArrayType.containsNu...Takuya UESHIN2014-08-261-3/+3
* [SPARK-2871] [PySpark] add histgram() APIDavies Liu2014-08-262-1/+232
* [SPARK-2871] [PySpark] add zipWithIndex() and zipWithUniqueId()Davies Liu2014-08-241-0/+47
* [SPARK-2871] [PySpark] add approx API for RDDDavies Liu2014-08-231-0/+81
* [SPARK-2871] [PySpark] add `key` argument for max(), min() and top(n)Davies Liu2014-08-231-17/+27
* [SPARK-3140] Clarify confusing PySpark exception messageAndrew Or2014-08-201-3/+10