aboutsummaryrefslogtreecommitdiff
path: root/sql/catalyst
diff options
context:
space:
mode:
authorJeff Zhang <zjffdu@apache.org>2016-10-14 15:50:35 -0700
committerMichael Armbrust <michael@databricks.com>2016-10-14 15:50:35 -0700
commitf00df40cfefef0f3fc73f16ada1006e4dcfa5a39 (patch)
tree17197231fa9c82e3158b039d950c021383e94885 /sql/catalyst
parent5aeb7384c7aa5f487f031f9ae07d3f1653399d14 (diff)
downloadspark-f00df40cfefef0f3fc73f16ada1006e4dcfa5a39.tar.gz
spark-f00df40cfefef0f3fc73f16ada1006e4dcfa5a39.tar.bz2
spark-f00df40cfefef0f3fc73f16ada1006e4dcfa5a39.zip
[SPARK-11775][PYSPARK][SQL] Allow PySpark to register Java UDF
Currently pyspark can only call the builtin java UDF, but can not call custom java UDF. It would be better to allow that. 2 benefits: * Leverage the power of rich third party java library * Improve the performance. Because if we use python UDF, python daemons will be started on worker which will affect the performance. Author: Jeff Zhang <zjffdu@apache.org> Closes #9766 from zjffdu/SPARK-11775.
Diffstat (limited to 'sql/catalyst')
-rw-r--r--sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/JavaTypeInference.scala2
1 files changed, 1 insertions, 1 deletions
diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/JavaTypeInference.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/JavaTypeInference.scala
index e6f61b00eb..04f0cfce88 100644
--- a/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/JavaTypeInference.scala
+++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/catalyst/JavaTypeInference.scala
@@ -59,7 +59,7 @@ object JavaTypeInference {
* @param typeToken Java type
* @return (SQL data type, nullable)
*/
- private def inferDataType(typeToken: TypeToken[_]): (DataType, Boolean) = {
+ private[sql] def inferDataType(typeToken: TypeToken[_]): (DataType, Boolean) = {
typeToken.getRawType match {
case c: Class[_] if c.isAnnotationPresent(classOf[SQLUserDefinedType]) =>
(c.getAnnotation(classOf[SQLUserDefinedType]).udt().newInstance(), true)