aboutsummaryrefslogtreecommitdiff
path: root/python/pyspark/__init__.py
diff options
context:
space:
mode:
Diffstat (limited to 'python/pyspark/__init__.py')
-rw-r--r--python/pyspark/__init__.py18
1 files changed, 17 insertions, 1 deletions
diff --git a/python/pyspark/__init__.py b/python/pyspark/__init__.py
index a51d5af79b..73fe7378ff 100644
--- a/python/pyspark/__init__.py
+++ b/python/pyspark/__init__.py
@@ -34,6 +34,19 @@ Public classes:
Access files shipped with jobs.
- L{StorageLevel<pyspark.storagelevel.StorageLevel>}
Finer-grained cache persistence levels.
+
+Spark SQL:
+ - L{SQLContext<pyspark.sql.SQLContext>}
+ Main entry point for SQL functionality.
+ - L{SchemaRDD<pyspark.sql.SchemaRDD>}
+ A Resilient Distributed Dataset (RDD) with Schema information for the data contained. In
+ addition to normal RDD operations, SchemaRDDs also support SQL.
+ - L{Row<pyspark.sql.Row>}
+ A Row of data returned by a Spark SQL query.
+
+Hive:
+ - L{HiveContext<pyspark.context.HiveContext>}
+ Main entry point for accessing data stored in Apache Hive..
"""
@@ -45,9 +58,12 @@ sys.path.insert(0, os.path.join(os.environ["SPARK_HOME"], "python/lib/py4j-0.8.1
from pyspark.conf import SparkConf
from pyspark.context import SparkContext
+from pyspark.sql import SQLContext
from pyspark.rdd import RDD
+from pyspark.sql import SchemaRDD
+from pyspark.sql import Row
from pyspark.files import SparkFiles
from pyspark.storagelevel import StorageLevel
-__all__ = ["SparkConf", "SparkContext", "RDD", "SparkFiles", "StorageLevel"]
+__all__ = ["SparkConf", "SparkContext", "SQLContext", "RDD", "SchemaRDD", "SparkFiles", "StorageLevel", "Row"]