diff options
author | Matei Zaharia <matei@databricks.com> | 2013-12-29 20:15:07 -0500 |
---|---|---|
committer | Matei Zaharia <matei@databricks.com> | 2013-12-29 20:15:07 -0500 |
commit | eaa8a68ff08304f713f4f75d39c61c020e0e691d (patch) | |
tree | 5543260c25af21555673154c0305a07e46f4ff6c /python/pyspark/__init__.py | |
parent | 11540b798d622f3883cb40b20cc30ea7d894790a (diff) | |
download | spark-eaa8a68ff08304f713f4f75d39c61c020e0e691d.tar.gz spark-eaa8a68ff08304f713f4f75d39c61c020e0e691d.tar.bz2 spark-eaa8a68ff08304f713f4f75d39c61c020e0e691d.zip |
Fix some Python docs and make sure to unset SPARK_TESTING in Python
tests so we don't get the test spark.conf on the classpath.
Diffstat (limited to 'python/pyspark/__init__.py')
-rw-r--r-- | python/pyspark/__init__.py | 31 |
1 files changed, 17 insertions, 14 deletions
diff --git a/python/pyspark/__init__.py b/python/pyspark/__init__.py index f1b95acf09..2b2c3a061a 100644 --- a/python/pyspark/__init__.py +++ b/python/pyspark/__init__.py @@ -20,21 +20,24 @@ PySpark is the Python API for Spark. Public classes: - - L{SparkContext<pyspark.context.SparkContext>} - Main entry point for Spark functionality. - - L{RDD<pyspark.rdd.RDD>} - A Resilient Distributed Dataset (RDD), the basic abstraction in Spark. - - L{Broadcast<pyspark.broadcast.Broadcast>} - A broadcast variable that gets reused across tasks. - - L{Accumulator<pyspark.accumulators.Accumulator>} - An "add-only" shared variable that tasks can only add values to. - - L{SparkConf<pyspark.conf.SparkConf} - Configuration for a Spark application. - - L{SparkFiles<pyspark.files.SparkFiles>} - Access files shipped with jobs. - - L{StorageLevel<pyspark.storagelevel.StorageLevel>} - Finer-grained cache persistence levels. + - L{SparkContext<pyspark.context.SparkContext>} + Main entry point for Spark functionality. + - L{RDD<pyspark.rdd.RDD>} + A Resilient Distributed Dataset (RDD), the basic abstraction in Spark. + - L{Broadcast<pyspark.broadcast.Broadcast>} + A broadcast variable that gets reused across tasks. + - L{Accumulator<pyspark.accumulators.Accumulator>} + An "add-only" shared variable that tasks can only add values to. + - L{SparkConf<pyspark.conf.SparkConf>} + For configuring Spark. + - L{SparkFiles<pyspark.files.SparkFiles>} + Access files shipped with jobs. + - L{StorageLevel<pyspark.storagelevel.StorageLevel>} + Finer-grained cache persistence levels. """ + + + import sys import os sys.path.insert(0, os.path.join(os.environ["SPARK_HOME"], "python/lib/py4j0.7.egg")) |