aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
-rw-r--r--core/src/main/scala/org/apache/spark/api/python/SerDeUtil.scala2
-rw-r--r--python/pyspark/tests.py4
2 files changed, 3 insertions, 3 deletions
diff --git a/core/src/main/scala/org/apache/spark/api/python/SerDeUtil.scala b/core/src/main/scala/org/apache/spark/api/python/SerDeUtil.scala
index 1c632ebdf9..6e4eab4b80 100644
--- a/core/src/main/scala/org/apache/spark/api/python/SerDeUtil.scala
+++ b/core/src/main/scala/org/apache/spark/api/python/SerDeUtil.scala
@@ -137,7 +137,7 @@ private[spark] object SerDeUtil extends Logging {
* Convert an RDD of Java objects to an RDD of serialized Python objects, that is usable by
* PySpark.
*/
- private[spark] def javaToPython(jRDD: JavaRDD[_]): JavaRDD[Array[Byte]] = {
+ def javaToPython(jRDD: JavaRDD[_]): JavaRDD[Array[Byte]] = {
jRDD.rdd.mapPartitions { iter => new AutoBatchedPickler(iter) }
}
diff --git a/python/pyspark/tests.py b/python/pyspark/tests.py
index 97ea39dde0..222c5ca5f4 100644
--- a/python/pyspark/tests.py
+++ b/python/pyspark/tests.py
@@ -960,13 +960,13 @@ class RDDTests(ReusedPySparkTestCase):
]
data_rdd = self.sc.parallelize(data)
data_java_rdd = data_rdd._to_java_object_rdd()
- data_python_rdd = self.sc._jvm.SerDe.javaToPython(data_java_rdd)
+ data_python_rdd = self.sc._jvm.SerDeUtil.javaToPython(data_java_rdd)
converted_rdd = RDD(data_python_rdd, self.sc)
self.assertEqual(2, converted_rdd.count())
# conversion between python and java RDD threw exceptions
data_java_rdd = converted_rdd._to_java_object_rdd()
- data_python_rdd = self.sc._jvm.SerDe.javaToPython(data_java_rdd)
+ data_python_rdd = self.sc._jvm.SerDeUtil.javaToPython(data_java_rdd)
converted_rdd = RDD(data_python_rdd, self.sc)
self.assertEqual(2, converted_rdd.count())