aboutsummaryrefslogtreecommitdiff
path: root/python/pyspark
diff options
context:
space:
mode:
authorHolden Karau <holden@us.ibm.com>2016-04-14 09:42:15 +0100
committerSean Owen <sowen@cloudera.com>2016-04-14 09:42:15 +0100
commit478af2f45595913c9b8f560d13e8d88447486f99 (patch)
tree0c078a5b722da4df50339dae6eacd18e3d2bd2cb /python/pyspark
parentb4819404a65f9b97c1f8deb1fcb8419969831574 (diff)
downloadspark-478af2f45595913c9b8f560d13e8d88447486f99.tar.gz
spark-478af2f45595913c9b8f560d13e8d88447486f99.tar.bz2
spark-478af2f45595913c9b8f560d13e8d88447486f99.zip
[SPARK-14573][PYSPARK][BUILD] Fix PyDoc Makefile & highlighting issues
## What changes were proposed in this pull request? The PyDoc Makefile used "=" rather than "?=" for setting env variables so it overwrote the user values. This ignored the environment variables we set for linting allowing warnings through. This PR also fixes the warnings that had been introduced. ## How was this patch tested? manual local export & make Author: Holden Karau <holden@us.ibm.com> Closes #12336 from holdenk/SPARK-14573-fix-pydoc-makefile.
Diffstat (limited to 'python/pyspark')
-rw-r--r--python/pyspark/ml/regression.py2
-rw-r--r--python/pyspark/sql/context.py2
-rw-r--r--python/pyspark/sql/dataframe.py2
3 files changed, 3 insertions, 3 deletions
diff --git a/python/pyspark/ml/regression.py b/python/pyspark/ml/regression.py
index 316d7e30bc..c064fe500c 100644
--- a/python/pyspark/ml/regression.py
+++ b/python/pyspark/ml/regression.py
@@ -28,7 +28,7 @@ from pyspark.sql import DataFrame
__all__ = ['AFTSurvivalRegression', 'AFTSurvivalRegressionModel',
'DecisionTreeRegressor', 'DecisionTreeRegressionModel',
'GBTRegressor', 'GBTRegressionModel',
- 'GeneralizedLinearRegression', 'GeneralizedLinearRegressionModel'
+ 'GeneralizedLinearRegression', 'GeneralizedLinearRegressionModel',
'IsotonicRegression', 'IsotonicRegressionModel',
'LinearRegression', 'LinearRegressionModel',
'LinearRegressionSummary', 'LinearRegressionTrainingSummary',
diff --git a/python/pyspark/sql/context.py b/python/pyspark/sql/context.py
index 4008332c84..11dfcfe13e 100644
--- a/python/pyspark/sql/context.py
+++ b/python/pyspark/sql/context.py
@@ -405,7 +405,7 @@ class SQLContext(object):
>>> sqlContext.createDataFrame(rdd, "boolean").collect() # doctest: +IGNORE_EXCEPTION_DETAIL
Traceback (most recent call last):
...
- Py4JJavaError:...
+ Py4JJavaError: ...
"""
if isinstance(data, DataFrame):
raise TypeError("data is already a DataFrame")
diff --git a/python/pyspark/sql/dataframe.py b/python/pyspark/sql/dataframe.py
index d473d6b534..b4fa836893 100644
--- a/python/pyspark/sql/dataframe.py
+++ b/python/pyspark/sql/dataframe.py
@@ -60,7 +60,7 @@ class DataFrame(object):
people = sqlContext.read.parquet("...")
department = sqlContext.read.parquet("...")
- people.filter(people.age > 30).join(department, people.deptId == department.id)) \
+ people.filter(people.age > 30).join(department, people.deptId == department.id)\
.groupBy(department.name, "gender").agg({"salary": "avg", "age": "max"})
.. note:: Experimental