aboutsummaryrefslogtreecommitdiff
path: root/sql/hive/src/test/scala/org/apache
diff options
context:
space:
mode:
authorXiao Li <gatorsmile@gmail.com>2017-04-12 09:01:26 -0700
committerXiao Li <gatorsmile@gmail.com>2017-04-12 09:01:26 -0700
commit504e62e2f4b7df7e002ea014a855cebe1ff95193 (patch)
treee24fb9939cf38f1be5590882175a7d33961b81b8 /sql/hive/src/test/scala/org/apache
parentceaf77ae43a14e993ac6d1ff34b50256eacd6abb (diff)
downloadspark-504e62e2f4b7df7e002ea014a855cebe1ff95193.tar.gz
spark-504e62e2f4b7df7e002ea014a855cebe1ff95193.tar.bz2
spark-504e62e2f4b7df7e002ea014a855cebe1ff95193.zip
[SPARK-20303][SQL] Rename createTempFunction to registerFunction
### What changes were proposed in this pull request? Session catalog API `createTempFunction` is being used by Hive build-in functions, persistent functions, and temporary functions. Thus, the name is confusing. This PR is to rename it by `registerFunction`. Also we can move construction of `FunctionBuilder` and `ExpressionInfo` into the new `registerFunction`, instead of duplicating the logics everywhere. In the next PRs, the remaining Function-related APIs also need cleanups. ### How was this patch tested? Existing test cases. Author: Xiao Li <gatorsmile@gmail.com> Closes #17615 from gatorsmile/cleanupCreateTempFunction.
Diffstat (limited to 'sql/hive/src/test/scala/org/apache')
-rw-r--r--sql/hive/src/test/scala/org/apache/spark/sql/execution/benchmark/ObjectHashAggregateExecBenchmark.scala10
1 files changed, 6 insertions, 4 deletions
diff --git a/sql/hive/src/test/scala/org/apache/spark/sql/execution/benchmark/ObjectHashAggregateExecBenchmark.scala b/sql/hive/src/test/scala/org/apache/spark/sql/execution/benchmark/ObjectHashAggregateExecBenchmark.scala
index 197110f491..73383ae4d4 100644
--- a/sql/hive/src/test/scala/org/apache/spark/sql/execution/benchmark/ObjectHashAggregateExecBenchmark.scala
+++ b/sql/hive/src/test/scala/org/apache/spark/sql/execution/benchmark/ObjectHashAggregateExecBenchmark.scala
@@ -22,7 +22,9 @@ import scala.concurrent.duration._
import org.apache.hadoop.hive.ql.udf.generic.GenericUDAFPercentileApprox
import org.apache.spark.sql.Column
-import org.apache.spark.sql.catalyst.expressions.{ExpressionInfo, Literal}
+import org.apache.spark.sql.catalyst.FunctionIdentifier
+import org.apache.spark.sql.catalyst.catalog.CatalogFunction
+import org.apache.spark.sql.catalyst.expressions.Literal
import org.apache.spark.sql.catalyst.expressions.aggregate.ApproximatePercentile
import org.apache.spark.sql.hive.HiveSessionCatalog
import org.apache.spark.sql.hive.execution.TestingTypedCount
@@ -217,9 +219,9 @@ class ObjectHashAggregateExecBenchmark extends BenchmarkBase with TestHiveSingle
private def registerHiveFunction(functionName: String, clazz: Class[_]): Unit = {
val sessionCatalog = sparkSession.sessionState.catalog.asInstanceOf[HiveSessionCatalog]
- val builder = sessionCatalog.makeFunctionBuilder(functionName, clazz.getName)
- val info = new ExpressionInfo(clazz.getName, functionName)
- sessionCatalog.createTempFunction(functionName, info, builder, ignoreIfExists = false)
+ val functionIdentifier = FunctionIdentifier(functionName, database = None)
+ val func = CatalogFunction(functionIdentifier, clazz.getName, resources = Nil)
+ sessionCatalog.registerFunction(func, ignoreIfExists = false)
}
private def percentile_approx(