aboutsummaryrefslogtreecommitdiff
path: root/core
diff options
context:
space:
mode:
authorSean Owen <sowen@cloudera.com>2016-08-24 20:04:09 +0100
committerSean Owen <sowen@cloudera.com>2016-08-24 20:04:09 +0100
commit0b3a4be92ca6b38eef32ea5ca240d9f91f68aa65 (patch)
treefceae4d22e14f83548bd505535cd67e71b030b3d /core
parent2fbdb606392631b1dff88ec86f388cc2559c28f5 (diff)
downloadspark-0b3a4be92ca6b38eef32ea5ca240d9f91f68aa65.tar.gz
spark-0b3a4be92ca6b38eef32ea5ca240d9f91f68aa65.tar.bz2
spark-0b3a4be92ca6b38eef32ea5ca240d9f91f68aa65.zip
[SPARK-16781][PYSPARK] java launched by PySpark as gateway may not be the same java used in the spark environment
## What changes were proposed in this pull request? Update to py4j 0.10.3 to enable JAVA_HOME support ## How was this patch tested? Pyspark tests Author: Sean Owen <sowen@cloudera.com> Closes #14748 from srowen/SPARK-16781.
Diffstat (limited to 'core')
-rw-r--r--core/pom.xml2
-rw-r--r--core/src/main/scala/org/apache/spark/api/python/PythonUtils.scala2
2 files changed, 2 insertions, 2 deletions
diff --git a/core/pom.xml b/core/pom.xml
index 04b94a258c..ab6c3ce805 100644
--- a/core/pom.xml
+++ b/core/pom.xml
@@ -326,7 +326,7 @@
<dependency>
<groupId>net.sf.py4j</groupId>
<artifactId>py4j</artifactId>
- <version>0.10.1</version>
+ <version>0.10.3</version>
</dependency>
<dependency>
<groupId>org.apache.spark</groupId>
diff --git a/core/src/main/scala/org/apache/spark/api/python/PythonUtils.scala b/core/src/main/scala/org/apache/spark/api/python/PythonUtils.scala
index 64cf498171..701097ace8 100644
--- a/core/src/main/scala/org/apache/spark/api/python/PythonUtils.scala
+++ b/core/src/main/scala/org/apache/spark/api/python/PythonUtils.scala
@@ -32,7 +32,7 @@ private[spark] object PythonUtils {
val pythonPath = new ArrayBuffer[String]
for (sparkHome <- sys.env.get("SPARK_HOME")) {
pythonPath += Seq(sparkHome, "python", "lib", "pyspark.zip").mkString(File.separator)
- pythonPath += Seq(sparkHome, "python", "lib", "py4j-0.10.1-src.zip").mkString(File.separator)
+ pythonPath += Seq(sparkHome, "python", "lib", "py4j-0.10.3-src.zip").mkString(File.separator)
}
pythonPath ++= SparkContext.jarOfObject(this)
pythonPath.mkString(File.pathSeparator)