diff options
author | Reynold Xin <rxin@databricks.com> | 2015-01-16 21:09:06 -0800 |
---|---|---|
committer | Reynold Xin <rxin@databricks.com> | 2015-01-16 21:09:06 -0800 |
commit | 61b427d4b1c4934bd70ed4da844b64f0e9a377aa (patch) | |
tree | 5068b31119fa7e2256422d4fdf18703ae64d7ab2 /sql/hive/src/test | |
parent | ee1c1f3a04dfe80843432e349f01178e47f02443 (diff) | |
download | spark-61b427d4b1c4934bd70ed4da844b64f0e9a377aa.tar.gz spark-61b427d4b1c4934bd70ed4da844b64f0e9a377aa.tar.bz2 spark-61b427d4b1c4934bd70ed4da844b64f0e9a377aa.zip |
[SPARK-5193][SQL] Remove Spark SQL Java-specific API.
After the following patches, the main (Scala) API is now usable for Java users directly.
https://github.com/apache/spark/pull/4056
https://github.com/apache/spark/pull/4054
https://github.com/apache/spark/pull/4049
https://github.com/apache/spark/pull/4030
https://github.com/apache/spark/pull/3965
https://github.com/apache/spark/pull/3958
Author: Reynold Xin <rxin@databricks.com>
Closes #4065 from rxin/sql-java-api and squashes the following commits:
b1fd860 [Reynold Xin] Fix Mima
6d86578 [Reynold Xin] Ok one more attempt in fixing Python...
e8f1455 [Reynold Xin] Fix Python again...
3e53f91 [Reynold Xin] Fixed Python.
83735da [Reynold Xin] Fix BigDecimal test.
e9f1de3 [Reynold Xin] Use scala BigDecimal.
500d2c4 [Reynold Xin] Fix Decimal.
ba3bfa2 [Reynold Xin] Updated javadoc for RowFactory.
c4ae1c5 [Reynold Xin] [SPARK-5193][SQL] Remove Spark SQL Java-specific API.
Diffstat (limited to 'sql/hive/src/test')
-rw-r--r-- | sql/hive/src/test/scala/org/apache/spark/sql/hive/api/java/JavaHiveQLSuite.scala | 91 |
1 files changed, 0 insertions, 91 deletions
diff --git a/sql/hive/src/test/scala/org/apache/spark/sql/hive/api/java/JavaHiveQLSuite.scala b/sql/hive/src/test/scala/org/apache/spark/sql/hive/api/java/JavaHiveQLSuite.scala deleted file mode 100644 index ca78dfba4f..0000000000 --- a/sql/hive/src/test/scala/org/apache/spark/sql/hive/api/java/JavaHiveQLSuite.scala +++ /dev/null @@ -1,91 +0,0 @@ -/* - * Licensed to the Apache Software Foundation (ASF) under one or more - * contributor license agreements. See the NOTICE file distributed with - * this work for additional information regarding copyright ownership. - * The ASF licenses this file to You under the Apache License, Version 2.0 - * (the "License"); you may not use this file except in compliance with - * the License. You may obtain a copy of the License at - * - * http://www.apache.org/licenses/LICENSE-2.0 - * - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -package org.apache.spark.sql.hive.api.java - -import scala.util.Try - -import org.scalatest.FunSuite - -import org.apache.spark.api.java.JavaSparkContext -import org.apache.spark.sql.api.java.{JavaSQLContext, JavaSchemaRDD} -import org.apache.spark.sql.execution.ExplainCommand -import org.apache.spark.sql.hive.test.TestHive - -// Implicits -import scala.collection.JavaConversions._ - -class JavaHiveQLSuite extends FunSuite { - lazy val javaCtx = new JavaSparkContext(TestHive.sparkContext) - - // There is a little trickery here to avoid instantiating two HiveContexts in the same JVM - lazy val javaHiveCtx = new JavaHiveContext(TestHive) - - test("SELECT * FROM src") { - assert( - javaHiveCtx.sql("SELECT * FROM src").collect().map(_.getInt(0)) === - TestHive.sql("SELECT * FROM src").collect().map(_.getInt(0)).toSeq) - } - - def isExplanation(result: JavaSchemaRDD) = { - val explanation = result.collect().map(_.getString(0)) - explanation.size > 1 && explanation.head.startsWith("== Physical Plan ==") - } - - test("Query Hive native command execution result") { - val tableName = "test_native_commands" - - assertResult(0) { - javaHiveCtx.sql(s"DROP TABLE IF EXISTS $tableName").count() - } - - assertResult(0) { - javaHiveCtx.sql(s"CREATE TABLE $tableName(key INT, value STRING)").count() - } - - assert( - javaHiveCtx - .sql("SHOW TABLES") - .collect() - .map(_.getString(0)) - .contains(tableName)) - - assertResult(Array(Array("key", "int"), Array("value", "string"))) { - javaHiveCtx - .sql(s"describe $tableName") - .collect() - .map(row => Array(row.get(0).asInstanceOf[String], row.get(1).asInstanceOf[String])) - .toArray - } - - assert(isExplanation(javaHiveCtx.sql( - s"EXPLAIN SELECT key, COUNT(*) FROM $tableName GROUP BY key"))) - - TestHive.reset() - } - - test("Exactly once semantics for DDL and command statements") { - val tableName = "test_exactly_once" - val q0 = javaHiveCtx.sql(s"CREATE TABLE $tableName(key INT, value STRING)") - - // If the table was not created, the following assertion would fail - assert(Try(TestHive.table(tableName)).isSuccess) - - // If the CREATE TABLE command got executed again, the following assertion would fail - assert(Try(q0.count()).isSuccess) - } -} |