aboutsummaryrefslogtreecommitdiff
path: root/python/pyspark/sql/__init__.py
diff options
context:
space:
mode:
authorAndrew Or <andrew@databricks.com>2016-04-28 10:55:48 -0700
committerReynold Xin <rxin@databricks.com>2016-04-28 10:55:48 -0700
commit89addd40abdacd65cc03ac8aa5f9cf3dd4a4c19b (patch)
tree5ecd3d9a736333c7951de6159eefef86129e3744 /python/pyspark/sql/__init__.py
parent5743352a28fffbfbaca2201208ce7a1d7893f813 (diff)
downloadspark-89addd40abdacd65cc03ac8aa5f9cf3dd4a4c19b.tar.gz
spark-89addd40abdacd65cc03ac8aa5f9cf3dd4a4c19b.tar.bz2
spark-89addd40abdacd65cc03ac8aa5f9cf3dd4a4c19b.zip
[SPARK-14945][PYTHON] SparkSession Python API
## What changes were proposed in this pull request? ``` Welcome to ____ __ / __/__ ___ _____/ /__ _\ \/ _ \/ _ `/ __/ '_/ /__ / .__/\_,_/_/ /_/\_\ version 2.0.0-SNAPSHOT /_/ Using Python version 2.7.5 (default, Mar 9 2014 22:15:05) SparkSession available as 'spark'. >>> spark <pyspark.sql.session.SparkSession object at 0x101f3bfd0> >>> spark.sql("SHOW TABLES").show() ... +---------+-----------+ |tableName|isTemporary| +---------+-----------+ | src| false| +---------+-----------+ >>> spark.range(1, 10, 2).show() +---+ | id| +---+ | 1| | 3| | 5| | 7| | 9| +---+ ``` **Note**: This API is NOT complete in its current state. In particular, for now I left out the `conf` and `catalog` APIs, which were added later in Scala. These will be added later before 2.0. ## How was this patch tested? Python tests. Author: Andrew Or <andrew@databricks.com> Closes #12746 from andrewor14/python-spark-session.
Diffstat (limited to 'python/pyspark/sql/__init__.py')
-rw-r--r--python/pyspark/sql/__init__.py5
1 files changed, 3 insertions, 2 deletions
diff --git a/python/pyspark/sql/__init__.py b/python/pyspark/sql/__init__.py
index 0b06c8339f..cff73ff192 100644
--- a/python/pyspark/sql/__init__.py
+++ b/python/pyspark/sql/__init__.py
@@ -46,6 +46,7 @@ from __future__ import absolute_import
from pyspark.sql.types import Row
from pyspark.sql.context import SQLContext, HiveContext
+from pyspark.sql.session import SparkSession
from pyspark.sql.column import Column
from pyspark.sql.dataframe import DataFrame, DataFrameNaFunctions, DataFrameStatFunctions
from pyspark.sql.group import GroupedData
@@ -54,7 +55,7 @@ from pyspark.sql.window import Window, WindowSpec
__all__ = [
- 'SQLContext', 'HiveContext', 'DataFrame', 'GroupedData', 'Column', 'Row',
- 'DataFrameNaFunctions', 'DataFrameStatFunctions', 'Window', 'WindowSpec',
+ 'SparkSession', 'SQLContext', 'HiveContext', 'DataFrame', 'GroupedData', 'Column',
+ 'Row', 'DataFrameNaFunctions', 'DataFrameStatFunctions', 'Window', 'WindowSpec',
'DataFrameReader', 'DataFrameWriter'
]