aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorMichael Armbrust <michael@databricks.com>2014-08-02 18:27:04 -0700
committerMichael Armbrust <michael@databricks.com>2014-08-02 18:27:04 -0700
commit1a8043739dc1d9435def6ea3c6341498ba52b708 (patch)
treefe499822fa58fc9416e8664b76839b4e198679a2
parentd210022e96804e59e42ab902e53637e50884a9ab (diff)
downloadspark-1a8043739dc1d9435def6ea3c6341498ba52b708.tar.gz
spark-1a8043739dc1d9435def6ea3c6341498ba52b708.tar.bz2
spark-1a8043739dc1d9435def6ea3c6341498ba52b708.zip
[SPARK-2739][SQL] Rename registerAsTable to registerTempTable
There have been user complaints that the difference between `registerAsTable` and `saveAsTable` is too subtle. This PR addresses this by renaming `registerAsTable` to `registerTempTable`, which more clearly reflects what is happening. `registerAsTable` remains, but will cause a deprecation warning. Author: Michael Armbrust <michael@databricks.com> Closes #1743 from marmbrus/registerTempTable and squashes the following commits: d031348 [Michael Armbrust] Merge remote-tracking branch 'apache/master' into registerTempTable 4dff086 [Michael Armbrust] Fix .java files too 89a2f12 [Michael Armbrust] Merge remote-tracking branch 'apache/master' into registerTempTable 0b7b71e [Michael Armbrust] Rename registerAsTable to registerTempTable
-rw-r--r--dev/audit-release/sbt_app_sql/src/main/scala/SqlApp.scala2
-rw-r--r--docs/sql-programming-guide.md18
-rw-r--r--examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java8
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/sql/RDDRelation.scala4
-rw-r--r--examples/src/main/scala/org/apache/spark/examples/sql/hive/HiveFromSpark.scala2
-rw-r--r--python/pyspark/sql.py12
-rw-r--r--sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala4
-rw-r--r--sql/core/src/main/scala/org/apache/spark/sql/SchemaRDD.scala2
-rw-r--r--sql/core/src/main/scala/org/apache/spark/sql/SchemaRDDLike.scala5
-rw-r--r--sql/core/src/main/scala/org/apache/spark/sql/api/java/JavaSQLContext.scala2
-rw-r--r--sql/core/src/test/java/org/apache/spark/sql/api/java/JavaApplySchemaSuite.java6
-rw-r--r--sql/core/src/test/scala/org/apache/spark/sql/CachedTableSuite.scala2
-rw-r--r--sql/core/src/test/scala/org/apache/spark/sql/InsertIntoSuite.scala4
-rw-r--r--sql/core/src/test/scala/org/apache/spark/sql/JoinSuite.scala4
-rw-r--r--sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala6
-rw-r--r--sql/core/src/test/scala/org/apache/spark/sql/ScalaReflectionRelationSuite.scala8
-rw-r--r--sql/core/src/test/scala/org/apache/spark/sql/TestData.scala28
-rw-r--r--sql/core/src/test/scala/org/apache/spark/sql/api/java/JavaSQLSuite.scala10
-rw-r--r--sql/core/src/test/scala/org/apache/spark/sql/json/JsonSuite.scala22
-rw-r--r--sql/core/src/test/scala/org/apache/spark/sql/parquet/ParquetQuerySuite.scala26
-rw-r--r--sql/hive/src/test/scala/org/apache/spark/sql/hive/InsertIntoHiveTableSuite.scala2
-rw-r--r--sql/hive/src/test/scala/org/apache/spark/sql/hive/api/java/JavaHiveQLSuite.scala4
-rw-r--r--sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveQuerySuite.scala6
-rw-r--r--sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveResolutionSuite.scala4
-rw-r--r--sql/hive/src/test/scala/org/apache/spark/sql/parquet/HiveParquetSuite.scala8
25 files changed, 103 insertions, 96 deletions
diff --git a/dev/audit-release/sbt_app_sql/src/main/scala/SqlApp.scala b/dev/audit-release/sbt_app_sql/src/main/scala/SqlApp.scala
index 50af90c213..d888de929f 100644
--- a/dev/audit-release/sbt_app_sql/src/main/scala/SqlApp.scala
+++ b/dev/audit-release/sbt_app_sql/src/main/scala/SqlApp.scala
@@ -38,7 +38,7 @@ object SparkSqlExample {
import sqlContext._
val people = sc.makeRDD(1 to 100, 10).map(x => Person(s"Name$x", x))
- people.registerAsTable("people")
+ people.registerTempTable("people")
val teenagers = sql("SELECT name FROM people WHERE age >= 13 AND age <= 19")
val teenagerNames = teenagers.map(t => "Name: " + t(0)).collect()
teenagerNames.foreach(println)
diff --git a/docs/sql-programming-guide.md b/docs/sql-programming-guide.md
index 7261badd41..0465468084 100644
--- a/docs/sql-programming-guide.md
+++ b/docs/sql-programming-guide.md
@@ -142,7 +142,7 @@ case class Person(name: String, age: Int)
// Create an RDD of Person objects and register it as a table.
val people = sc.textFile("examples/src/main/resources/people.txt").map(_.split(",")).map(p => Person(p(0), p(1).trim.toInt))
-people.registerAsTable("people")
+people.registerTempTable("people")
// SQL statements can be run by using the sql methods provided by sqlContext.
val teenagers = sqlContext.sql("SELECT name FROM people WHERE age >= 13 AND age <= 19")
@@ -210,7 +210,7 @@ JavaRDD<Person> people = sc.textFile("examples/src/main/resources/people.txt").m
// Apply a schema to an RDD of JavaBeans and register it as a table.
JavaSchemaRDD schemaPeople = sqlContext.applySchema(people, Person.class);
-schemaPeople.registerAsTable("people");
+schemaPeople.registerTempTable("people");
// SQL can be run over RDDs that have been registered as tables.
JavaSchemaRDD teenagers = sqlContext.sql("SELECT name FROM people WHERE age >= 13 AND age <= 19")
@@ -248,7 +248,7 @@ people = parts.map(lambda p: {"name": p[0], "age": int(p[1])})
# In future versions of PySpark we would like to add support for registering RDDs with other
# datatypes as tables
schemaPeople = sqlContext.inferSchema(people)
-schemaPeople.registerAsTable("people")
+schemaPeople.registerTempTable("people")
# SQL can be run over SchemaRDDs that have been registered as a table.
teenagers = sqlContext.sql("SELECT name FROM people WHERE age >= 13 AND age <= 19")
@@ -292,7 +292,7 @@ people.saveAsParquetFile("people.parquet")
val parquetFile = sqlContext.parquetFile("people.parquet")
//Parquet files can also be registered as tables and then used in SQL statements.
-parquetFile.registerAsTable("parquetFile")
+parquetFile.registerTempTable("parquetFile")
val teenagers = sqlContext.sql("SELECT name FROM parquetFile WHERE age >= 13 AND age <= 19")
teenagers.map(t => "Name: " + t(0)).collect().foreach(println)
{% endhighlight %}
@@ -314,7 +314,7 @@ schemaPeople.saveAsParquetFile("people.parquet");
JavaSchemaRDD parquetFile = sqlContext.parquetFile("people.parquet");
//Parquet files can also be registered as tables and then used in SQL statements.
-parquetFile.registerAsTable("parquetFile");
+parquetFile.registerTempTable("parquetFile");
JavaSchemaRDD teenagers = sqlContext.sql("SELECT name FROM parquetFile WHERE age >= 13 AND age <= 19");
List<String> teenagerNames = teenagers.map(new Function<Row, String>() {
public String call(Row row) {
@@ -340,7 +340,7 @@ schemaPeople.saveAsParquetFile("people.parquet")
parquetFile = sqlContext.parquetFile("people.parquet")
# Parquet files can also be registered as tables and then used in SQL statements.
-parquetFile.registerAsTable("parquetFile");
+parquetFile.registerTempTable("parquetFile");
teenagers = sqlContext.sql("SELECT name FROM parquetFile WHERE age >= 13 AND age <= 19")
teenNames = teenagers.map(lambda p: "Name: " + p.name)
for teenName in teenNames.collect():
@@ -378,7 +378,7 @@ people.printSchema()
// |-- name: StringType
// Register this SchemaRDD as a table.
-people.registerAsTable("people")
+people.registerTempTable("people")
// SQL statements can be run by using the sql methods provided by sqlContext.
val teenagers = sqlContext.sql("SELECT name FROM people WHERE age >= 13 AND age <= 19")
@@ -416,7 +416,7 @@ people.printSchema();
// |-- name: StringType
// Register this JavaSchemaRDD as a table.
-people.registerAsTable("people");
+people.registerTempTable("people");
// SQL statements can be run by using the sql methods provided by sqlContext.
JavaSchemaRDD teenagers = sqlContext.sql("SELECT name FROM people WHERE age >= 13 AND age <= 19");
@@ -455,7 +455,7 @@ people.printSchema()
# |-- name: StringType
# Register this SchemaRDD as a table.
-people.registerAsTable("people")
+people.registerTempTable("people")
# SQL statements can be run by using the sql methods provided by sqlContext.
teenagers = sqlContext.sql("SELECT name FROM people WHERE age >= 13 AND age <= 19")
diff --git a/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java b/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java
index 607df3eddd..898297dc65 100644
--- a/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java
+++ b/examples/src/main/java/org/apache/spark/examples/sql/JavaSparkSQL.java
@@ -74,7 +74,7 @@ public class JavaSparkSQL {
// Apply a schema to an RDD of Java Beans and register it as a table.
JavaSchemaRDD schemaPeople = sqlCtx.applySchema(people, Person.class);
- schemaPeople.registerAsTable("people");
+ schemaPeople.registerTempTable("people");
// SQL can be run over RDDs that have been registered as tables.
JavaSchemaRDD teenagers = sqlCtx.sql("SELECT name FROM people WHERE age >= 13 AND age <= 19");
@@ -100,7 +100,7 @@ public class JavaSparkSQL {
JavaSchemaRDD parquetFile = sqlCtx.parquetFile("people.parquet");
//Parquet files can also be registered as tables and then used in SQL statements.
- parquetFile.registerAsTable("parquetFile");
+ parquetFile.registerTempTable("parquetFile");
JavaSchemaRDD teenagers2 =
sqlCtx.sql("SELECT name FROM parquetFile WHERE age >= 13 AND age <= 19");
teenagerNames = teenagers2.map(new Function<Row, String>() {
@@ -128,7 +128,7 @@ public class JavaSparkSQL {
// |-- name: StringType
// Register this JavaSchemaRDD as a table.
- peopleFromJsonFile.registerAsTable("people");
+ peopleFromJsonFile.registerTempTable("people");
// SQL statements can be run by using the sql methods provided by sqlCtx.
JavaSchemaRDD teenagers3 = sqlCtx.sql("SELECT name FROM people WHERE age >= 13 AND age <= 19");
@@ -158,7 +158,7 @@ public class JavaSparkSQL {
// | |-- state: StringType
// |-- name: StringType
- peopleFromJsonRDD.registerAsTable("people2");
+ peopleFromJsonRDD.registerTempTable("people2");
JavaSchemaRDD peopleWithCity = sqlCtx.sql("SELECT name, address.city FROM people2");
List<String> nameAndCity = peopleWithCity.map(new Function<Row, String>() {
diff --git a/examples/src/main/scala/org/apache/spark/examples/sql/RDDRelation.scala b/examples/src/main/scala/org/apache/spark/examples/sql/RDDRelation.scala
index 63db688bfb..d56d64c564 100644
--- a/examples/src/main/scala/org/apache/spark/examples/sql/RDDRelation.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/sql/RDDRelation.scala
@@ -36,7 +36,7 @@ object RDDRelation {
val rdd = sc.parallelize((1 to 100).map(i => Record(i, s"val_$i")))
// Any RDD containing case classes can be registered as a table. The schema of the table is
// automatically inferred using scala reflection.
- rdd.registerAsTable("records")
+ rdd.registerTempTable("records")
// Once tables have been registered, you can run SQL queries over them.
println("Result of SELECT *:")
@@ -66,7 +66,7 @@ object RDDRelation {
parquetFile.where('key === 1).select('value as 'a).collect().foreach(println)
// These files can also be registered as tables.
- parquetFile.registerAsTable("parquetFile")
+ parquetFile.registerTempTable("parquetFile")
sql("SELECT * FROM parquetFile").collect().foreach(println)
}
}
diff --git a/examples/src/main/scala/org/apache/spark/examples/sql/hive/HiveFromSpark.scala b/examples/src/main/scala/org/apache/spark/examples/sql/hive/HiveFromSpark.scala
index dc5290fb4f..12530c8490 100644
--- a/examples/src/main/scala/org/apache/spark/examples/sql/hive/HiveFromSpark.scala
+++ b/examples/src/main/scala/org/apache/spark/examples/sql/hive/HiveFromSpark.scala
@@ -56,7 +56,7 @@ object HiveFromSpark {
// You can also register RDDs as temporary tables within a HiveContext.
val rdd = sc.parallelize((1 to 100).map(i => Record(i, s"val_$i")))
- rdd.registerAsTable("records")
+ rdd.registerTempTable("records")
// Queries can then join RDD data with data stored in Hive.
println("Result of SELECT *:")
diff --git a/python/pyspark/sql.py b/python/pyspark/sql.py
index 36e50e49c9..42b738e112 100644
--- a/python/pyspark/sql.py
+++ b/python/pyspark/sql.py
@@ -909,7 +909,7 @@ class SQLContext:
... b=True, list=[1, 2, 3], dict={"s": 0}, row=Row(a=1),
... time=datetime(2014, 8, 1, 14, 1, 5))])
>>> srdd = sqlCtx.inferSchema(allTypes)
- >>> srdd.registerAsTable("allTypes")
+ >>> srdd.registerTempTable("allTypes")
>>> sqlCtx.sql('select i+1, d+1, not b, list[1], dict["s"], time, row.a '
... 'from allTypes where b and i > 0').collect()
[Row(c0=2, c1=2.0, c2=False, c3=2, c4=0...8, 1, 14, 1, 5), a=1)]
@@ -1486,19 +1486,23 @@ class SchemaRDD(RDD):
"""
self._jschema_rdd.saveAsParquetFile(path)
- def registerAsTable(self, name):
+ def registerTempTable(self, name):
"""Registers this RDD as a temporary table using the given name.
The lifetime of this temporary table is tied to the L{SQLContext}
that was used to create this SchemaRDD.
>>> srdd = sqlCtx.inferSchema(rdd)
- >>> srdd.registerAsTable("test")
+ >>> srdd.registerTempTable("test")
>>> srdd2 = sqlCtx.sql("select * from test")
>>> sorted(srdd.collect()) == sorted(srdd2.collect())
True
"""
- self._jschema_rdd.registerAsTable(name)
+ self._jschema_rdd.registerTempTable(name)
+
+ def registerAsTable(self, name):
+ warnings.warn("Use registerTempTable instead of registerAsTable.", DeprecationWarning)
+ self.registerTempTable(name)
def insertInto(self, tableName, overwrite=False):
"""Inserts the contents of this SchemaRDD into the specified table.
diff --git a/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala b/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala
index 33931e5d99..567f4dca99 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/SQLContext.scala
@@ -116,7 +116,7 @@ class SQLContext(@transient val sparkContext: SparkContext)
* // |-- name: string (nullable = false)
* // |-- age: integer (nullable = true)
*
- * peopleSchemaRDD.registerAsTable("people")
+ * peopleSchemaRDD.registerTempTable("people")
* sqlContext.sql("select name from people").collect.foreach(println)
* }}}
*
@@ -212,7 +212,7 @@ class SQLContext(@transient val sparkContext: SparkContext)
* import sqlContext._
*
* case class Person(name: String, age: Int)
- * createParquetFile[Person]("path/to/file.parquet").registerAsTable("people")
+ * createParquetFile[Person]("path/to/file.parquet").registerTempTable("people")
* sql("INSERT INTO people SELECT 'michael', 29")
* }}}
*
diff --git a/sql/core/src/main/scala/org/apache/spark/sql/SchemaRDD.scala b/sql/core/src/main/scala/org/apache/spark/sql/SchemaRDD.scala
index d34f62dc88..57df79321b 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/SchemaRDD.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/SchemaRDD.scala
@@ -67,7 +67,7 @@ import org.apache.spark.api.java.JavaRDD
* val rdd = sc.parallelize((1 to 100).map(i => Record(i, s"val_$i")))
* // Any RDD containing case classes can be registered as a table. The schema of the table is
* // automatically inferred using scala reflection.
- * rdd.registerAsTable("records")
+ * rdd.registerTempTable("records")
*
* val results: SchemaRDD = sql("SELECT * FROM records")
* }}}
diff --git a/sql/core/src/main/scala/org/apache/spark/sql/SchemaRDDLike.scala b/sql/core/src/main/scala/org/apache/spark/sql/SchemaRDDLike.scala
index 6a20def475..2f3033a5f9 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/SchemaRDDLike.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/SchemaRDDLike.scala
@@ -83,10 +83,13 @@ private[sql] trait SchemaRDDLike {
*
* @group schema
*/
- def registerAsTable(tableName: String): Unit = {
+ def registerTempTable(tableName: String): Unit = {
sqlContext.registerRDDAsTable(baseSchemaRDD, tableName)
}
+ @deprecated("Use registerTempTable instead of registerAsTable.", "1.1")
+ def registerAsTable(tableName: String): Unit = registerTempTable(tableName)
+
/**
* :: Experimental ::
* Adds the rows from this RDD to the specified table, optionally overwriting the existing data.
diff --git a/sql/core/src/main/scala/org/apache/spark/sql/api/java/JavaSQLContext.scala b/sql/core/src/main/scala/org/apache/spark/sql/api/java/JavaSQLContext.scala
index ae45193ed1..dbaa16e8b0 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/api/java/JavaSQLContext.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/api/java/JavaSQLContext.scala
@@ -52,7 +52,7 @@ class JavaSQLContext(val sqlContext: SQLContext) extends UDFRegistration {
* {{{
* JavaSQLContext sqlCtx = new JavaSQLContext(...)
*
- * sqlCtx.createParquetFile(Person.class, "path/to/file.parquet").registerAsTable("people")
+ * sqlCtx.createParquetFile(Person.class, "path/to/file.parquet").registerTempTable("people")
* sqlCtx.sql("INSERT INTO people SELECT 'michael', 29")
* }}}
*
diff --git a/sql/core/src/test/java/org/apache/spark/sql/api/java/JavaApplySchemaSuite.java b/sql/core/src/test/java/org/apache/spark/sql/api/java/JavaApplySchemaSuite.java
index 3c92906d82..33e5020bc6 100644
--- a/sql/core/src/test/java/org/apache/spark/sql/api/java/JavaApplySchemaSuite.java
+++ b/sql/core/src/test/java/org/apache/spark/sql/api/java/JavaApplySchemaSuite.java
@@ -98,7 +98,7 @@ public class JavaApplySchemaSuite implements Serializable {
StructType schema = DataType.createStructType(fields);
JavaSchemaRDD schemaRDD = javaSqlCtx.applySchema(rowRDD, schema);
- schemaRDD.registerAsTable("people");
+ schemaRDD.registerTempTable("people");
List<Row> actual = javaSqlCtx.sql("SELECT * FROM people").collect();
List<Row> expected = new ArrayList<Row>(2);
@@ -149,14 +149,14 @@ public class JavaApplySchemaSuite implements Serializable {
JavaSchemaRDD schemaRDD1 = javaSqlCtx.jsonRDD(jsonRDD);
StructType actualSchema1 = schemaRDD1.schema();
Assert.assertEquals(expectedSchema, actualSchema1);
- schemaRDD1.registerAsTable("jsonTable1");
+ schemaRDD1.registerTempTable("jsonTable1");
List<Row> actual1 = javaSqlCtx.sql("select * from jsonTable1").collect();
Assert.assertEquals(expectedResult, actual1);
JavaSchemaRDD schemaRDD2 = javaSqlCtx.jsonRDD(jsonRDD, expectedSchema);
StructType actualSchema2 = schemaRDD2.schema();
Assert.assertEquals(expectedSchema, actualSchema2);
- schemaRDD1.registerAsTable("jsonTable2");
+ schemaRDD1.registerTempTable("jsonTable2");
List<Row> actual2 = javaSqlCtx.sql("select * from jsonTable2").collect();
Assert.assertEquals(expectedResult, actual2);
}
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/CachedTableSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/CachedTableSuite.scala
index c3c0dcb1aa..fbf9bd9dbc 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/CachedTableSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/CachedTableSuite.scala
@@ -78,7 +78,7 @@ class CachedTableSuite extends QueryTest {
}
test("SELECT Star Cached Table") {
- TestSQLContext.sql("SELECT * FROM testData").registerAsTable("selectStar")
+ TestSQLContext.sql("SELECT * FROM testData").registerTempTable("selectStar")
TestSQLContext.cacheTable("selectStar")
TestSQLContext.sql("SELECT * FROM selectStar WHERE key = 1").collect()
TestSQLContext.uncacheTable("selectStar")
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/InsertIntoSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/InsertIntoSuite.scala
index 23a711d08c..c87d762751 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/InsertIntoSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/InsertIntoSuite.scala
@@ -31,7 +31,7 @@ class InsertIntoSuite extends QueryTest {
testFilePath.delete()
testFilePath.deleteOnExit()
val testFile = createParquetFile[TestData](testFilePath.getCanonicalPath)
- testFile.registerAsTable("createAndInsertTest")
+ testFile.registerTempTable("createAndInsertTest")
// Add some data.
testData.insertInto("createAndInsertTest")
@@ -86,7 +86,7 @@ class InsertIntoSuite extends QueryTest {
testFilePath.delete()
testFilePath.deleteOnExit()
val testFile = createParquetFile[TestData](testFilePath.getCanonicalPath)
- testFile.registerAsTable("createAndInsertSQLTest")
+ testFile.registerTempTable("createAndInsertSQLTest")
sql("INSERT INTO createAndInsertSQLTest SELECT * FROM testData")
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/JoinSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/JoinSuite.scala
index 2fc8058818..6c7697ece8 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/JoinSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/JoinSuite.scala
@@ -285,8 +285,8 @@ class JoinSuite extends QueryTest with BeforeAndAfterEach {
}
test("full outer join") {
- upperCaseData.where('N <= 4).registerAsTable("left")
- upperCaseData.where('N >= 3).registerAsTable("right")
+ upperCaseData.where('N <= 4).registerTempTable("left")
+ upperCaseData.where('N >= 3).registerTempTable("right")
val left = UnresolvedRelation(None, "left", None)
val right = UnresolvedRelation(None, "right", None)
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala
index 5c571d35d1..9b2a36d33f 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/SQLQuerySuite.scala
@@ -461,7 +461,7 @@ class SQLQuerySuite extends QueryTest {
}
val schemaRDD1 = applySchema(rowRDD1, schema1)
- schemaRDD1.registerAsTable("applySchema1")
+ schemaRDD1.registerTempTable("applySchema1")
checkAnswer(
sql("SELECT * FROM applySchema1"),
(1, "A1", true, null) ::
@@ -491,7 +491,7 @@ class SQLQuerySuite extends QueryTest {
}
val schemaRDD2 = applySchema(rowRDD2, schema2)
- schemaRDD2.registerAsTable("applySchema2")
+ schemaRDD2.registerTempTable("applySchema2")
checkAnswer(
sql("SELECT * FROM applySchema2"),
(Seq(1, true), Map("A1" -> null)) ::
@@ -516,7 +516,7 @@ class SQLQuerySuite extends QueryTest {
}
val schemaRDD3 = applySchema(rowRDD3, schema2)
- schemaRDD3.registerAsTable("applySchema3")
+ schemaRDD3.registerTempTable("applySchema3")
checkAnswer(
sql("SELECT f1.f11, f2['D4'] FROM applySchema3"),
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/ScalaReflectionRelationSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/ScalaReflectionRelationSuite.scala
index f2934da9a0..5b84c658db 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/ScalaReflectionRelationSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/ScalaReflectionRelationSuite.scala
@@ -61,7 +61,7 @@ class ScalaReflectionRelationSuite extends FunSuite {
val data = ReflectData("a", 1, 1L, 1.toFloat, 1.toDouble, 1.toShort, 1.toByte, true,
BigDecimal(1), new Timestamp(12345), Seq(1,2,3))
val rdd = sparkContext.parallelize(data :: Nil)
- rdd.registerAsTable("reflectData")
+ rdd.registerTempTable("reflectData")
assert(sql("SELECT * FROM reflectData").collect().head === data.productIterator.toSeq)
}
@@ -69,7 +69,7 @@ class ScalaReflectionRelationSuite extends FunSuite {
test("query case class RDD with nulls") {
val data = NullReflectData(null, null, null, null, null, null, null)
val rdd = sparkContext.parallelize(data :: Nil)
- rdd.registerAsTable("reflectNullData")
+ rdd.registerTempTable("reflectNullData")
assert(sql("SELECT * FROM reflectNullData").collect().head === Seq.fill(7)(null))
}
@@ -77,7 +77,7 @@ class ScalaReflectionRelationSuite extends FunSuite {
test("query case class RDD with Nones") {
val data = OptionalReflectData(None, None, None, None, None, None, None)
val rdd = sparkContext.parallelize(data :: Nil)
- rdd.registerAsTable("reflectOptionalData")
+ rdd.registerTempTable("reflectOptionalData")
assert(sql("SELECT * FROM reflectOptionalData").collect().head === Seq.fill(7)(null))
}
@@ -85,7 +85,7 @@ class ScalaReflectionRelationSuite extends FunSuite {
// Equality is broken for Arrays, so we test that separately.
test("query binary data") {
val rdd = sparkContext.parallelize(ReflectBinary(Array[Byte](1)) :: Nil)
- rdd.registerAsTable("reflectBinary")
+ rdd.registerTempTable("reflectBinary")
val result = sql("SELECT data FROM reflectBinary").collect().head(0).asInstanceOf[Array[Byte]]
assert(result.toSeq === Seq[Byte](1))
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/TestData.scala b/sql/core/src/test/scala/org/apache/spark/sql/TestData.scala
index 088e6e3c84..c3ec82fb69 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/TestData.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/TestData.scala
@@ -30,7 +30,7 @@ case class TestData(key: Int, value: String)
object TestData {
val testData: SchemaRDD = TestSQLContext.sparkContext.parallelize(
(1 to 100).map(i => TestData(i, i.toString)))
- testData.registerAsTable("testData")
+ testData.registerTempTable("testData")
case class LargeAndSmallInts(a: Int, b: Int)
val largeAndSmallInts: SchemaRDD =
@@ -41,7 +41,7 @@ object TestData {
LargeAndSmallInts(2, 2) ::
LargeAndSmallInts(2147483646, 1) ::
LargeAndSmallInts(3, 2) :: Nil)
- largeAndSmallInts.registerAsTable("largeAndSmallInts")
+ largeAndSmallInts.registerTempTable("largeAndSmallInts")
case class TestData2(a: Int, b: Int)
val testData2: SchemaRDD =
@@ -52,7 +52,7 @@ object TestData {
TestData2(2, 2) ::
TestData2(3, 1) ::
TestData2(3, 2) :: Nil)
- testData2.registerAsTable("testData2")
+ testData2.registerTempTable("testData2")
// TODO: There is no way to express null primitives as case classes currently...
val testData3 =
@@ -71,7 +71,7 @@ object TestData {
UpperCaseData(4, "D") ::
UpperCaseData(5, "E") ::
UpperCaseData(6, "F") :: Nil)
- upperCaseData.registerAsTable("upperCaseData")
+ upperCaseData.registerTempTable("upperCaseData")
case class LowerCaseData(n: Int, l: String)
val lowerCaseData =
@@ -80,14 +80,14 @@ object TestData {
LowerCaseData(2, "b") ::
LowerCaseData(3, "c") ::
LowerCaseData(4, "d") :: Nil)
- lowerCaseData.registerAsTable("lowerCaseData")
+ lowerCaseData.registerTempTable("lowerCaseData")
case class ArrayData(data: Seq[Int], nestedData: Seq[Seq[Int]])
val arrayData =
TestSQLContext.sparkContext.parallelize(
ArrayData(Seq(1,2,3), Seq(Seq(1,2,3))) ::
ArrayData(Seq(2,3,4), Seq(Seq(2,3,4))) :: Nil)
- arrayData.registerAsTable("arrayData")
+ arrayData.registerTempTable("arrayData")
case class MapData(data: Map[Int, String])
val mapData =
@@ -97,18 +97,18 @@ object TestData {
MapData(Map(1 -> "a3", 2 -> "b3", 3 -> "c3")) ::
MapData(Map(1 -> "a4", 2 -> "b4")) ::
MapData(Map(1 -> "a5")) :: Nil)
- mapData.registerAsTable("mapData")
+ mapData.registerTempTable("mapData")
case class StringData(s: String)
val repeatedData =
TestSQLContext.sparkContext.parallelize(List.fill(2)(StringData("test")))
- repeatedData.registerAsTable("repeatedData")
+ repeatedData.registerTempTable("repeatedData")
val nullableRepeatedData =
TestSQLContext.sparkContext.parallelize(
List.fill(2)(StringData(null)) ++
List.fill(2)(StringData("test")))
- nullableRepeatedData.registerAsTable("nullableRepeatedData")
+ nullableRepeatedData.registerTempTable("nullableRepeatedData")
case class NullInts(a: Integer)
val nullInts =
@@ -118,7 +118,7 @@ object TestData {
NullInts(3) ::
NullInts(null) :: Nil
)
- nullInts.registerAsTable("nullInts")
+ nullInts.registerTempTable("nullInts")
val allNulls =
TestSQLContext.sparkContext.parallelize(
@@ -126,7 +126,7 @@ object TestData {
NullInts(null) ::
NullInts(null) ::
NullInts(null) :: Nil)
- allNulls.registerAsTable("allNulls")
+ allNulls.registerTempTable("allNulls")
case class NullStrings(n: Int, s: String)
val nullStrings =
@@ -134,10 +134,10 @@ object TestData {
NullStrings(1, "abc") ::
NullStrings(2, "ABC") ::
NullStrings(3, null) :: Nil)
- nullStrings.registerAsTable("nullStrings")
+ nullStrings.registerTempTable("nullStrings")
case class TableName(tableName: String)
- TestSQLContext.sparkContext.parallelize(TableName("test") :: Nil).registerAsTable("tableName")
+ TestSQLContext.sparkContext.parallelize(TableName("test") :: Nil).registerTempTable("tableName")
val unparsedStrings =
TestSQLContext.sparkContext.parallelize(
@@ -150,5 +150,5 @@ object TestData {
val timestamps = TestSQLContext.sparkContext.parallelize((1 to 3).map { i =>
TimestampField(new Timestamp(i))
})
- timestamps.registerAsTable("timestamps")
+ timestamps.registerTempTable("timestamps")
}
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/api/java/JavaSQLSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/api/java/JavaSQLSuite.scala
index 020baf0c7e..203ff847e9 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/api/java/JavaSQLSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/api/java/JavaSQLSuite.scala
@@ -59,7 +59,7 @@ class JavaSQLSuite extends FunSuite {
val rdd = javaCtx.parallelize(person :: Nil)
val schemaRDD = javaSqlCtx.applySchema(rdd, classOf[PersonBean])
- schemaRDD.registerAsTable("people")
+ schemaRDD.registerTempTable("people")
javaSqlCtx.sql("SELECT * FROM people").collect()
}
@@ -76,7 +76,7 @@ class JavaSQLSuite extends FunSuite {
val rdd = javaCtx.parallelize(bean :: Nil)
val schemaRDD = javaSqlCtx.applySchema(rdd, classOf[AllTypesBean])
- schemaRDD.registerAsTable("allTypes")
+ schemaRDD.registerTempTable("allTypes")
assert(
javaSqlCtx.sql(
@@ -101,7 +101,7 @@ class JavaSQLSuite extends FunSuite {
val rdd = javaCtx.parallelize(bean :: Nil)
val schemaRDD = javaSqlCtx.applySchema(rdd, classOf[AllTypesBean])
- schemaRDD.registerAsTable("allTypes")
+ schemaRDD.registerTempTable("allTypes")
assert(
javaSqlCtx.sql(
@@ -127,7 +127,7 @@ class JavaSQLSuite extends FunSuite {
var schemaRDD = javaSqlCtx.jsonRDD(rdd)
- schemaRDD.registerAsTable("jsonTable1")
+ schemaRDD.registerTempTable("jsonTable1")
assert(
javaSqlCtx.sql("select * from jsonTable1").collect.head.row ===
@@ -144,7 +144,7 @@ class JavaSQLSuite extends FunSuite {
rdd.saveAsTextFile(path)
schemaRDD = javaSqlCtx.jsonFile(path)
- schemaRDD.registerAsTable("jsonTable2")
+ schemaRDD.registerTempTable("jsonTable2")
assert(
javaSqlCtx.sql("select * from jsonTable2").collect.head.row ===
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/json/JsonSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/json/JsonSuite.scala
index 9d9cfdd7c9..75c0589eb2 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/json/JsonSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/json/JsonSuite.scala
@@ -183,7 +183,7 @@ class JsonSuite extends QueryTest {
assert(expectedSchema === jsonSchemaRDD.schema)
- jsonSchemaRDD.registerAsTable("jsonTable")
+ jsonSchemaRDD.registerTempTable("jsonTable")
checkAnswer(
sql("select * from jsonTable"),
@@ -223,7 +223,7 @@ class JsonSuite extends QueryTest {
assert(expectedSchema === jsonSchemaRDD.schema)
- jsonSchemaRDD.registerAsTable("jsonTable")
+ jsonSchemaRDD.registerTempTable("jsonTable")
// Access elements of a primitive array.
checkAnswer(
@@ -291,7 +291,7 @@ class JsonSuite extends QueryTest {
ignore("Complex field and type inferring (Ignored)") {
val jsonSchemaRDD = jsonRDD(complexFieldAndType)
- jsonSchemaRDD.registerAsTable("jsonTable")
+ jsonSchemaRDD.registerTempTable("jsonTable")
// Right now, "field1" and "field2" are treated as aliases. We should fix it.
checkAnswer(
@@ -320,7 +320,7 @@ class JsonSuite extends QueryTest {
assert(expectedSchema === jsonSchemaRDD.schema)
- jsonSchemaRDD.registerAsTable("jsonTable")
+ jsonSchemaRDD.registerTempTable("jsonTable")
checkAnswer(
sql("select * from jsonTable"),
@@ -374,7 +374,7 @@ class JsonSuite extends QueryTest {
ignore("Type conflict in primitive field values (Ignored)") {
val jsonSchemaRDD = jsonRDD(primitiveFieldValueTypeConflict)
- jsonSchemaRDD.registerAsTable("jsonTable")
+ jsonSchemaRDD.registerTempTable("jsonTable")
// Right now, the analyzer does not promote strings in a boolean expreesion.
// Number and Boolean conflict: resolve the type as boolean in this query.
@@ -445,7 +445,7 @@ class JsonSuite extends QueryTest {
assert(expectedSchema === jsonSchemaRDD.schema)
- jsonSchemaRDD.registerAsTable("jsonTable")
+ jsonSchemaRDD.registerTempTable("jsonTable")
checkAnswer(
sql("select * from jsonTable"),
@@ -466,7 +466,7 @@ class JsonSuite extends QueryTest {
assert(expectedSchema === jsonSchemaRDD.schema)
- jsonSchemaRDD.registerAsTable("jsonTable")
+ jsonSchemaRDD.registerTempTable("jsonTable")
checkAnswer(
sql("select * from jsonTable"),
@@ -494,7 +494,7 @@ class JsonSuite extends QueryTest {
assert(expectedSchema === jsonSchemaRDD.schema)
- jsonSchemaRDD.registerAsTable("jsonTable")
+ jsonSchemaRDD.registerTempTable("jsonTable")
}
test("Loading a JSON dataset from a text file") {
@@ -514,7 +514,7 @@ class JsonSuite extends QueryTest {
assert(expectedSchema === jsonSchemaRDD.schema)
- jsonSchemaRDD.registerAsTable("jsonTable")
+ jsonSchemaRDD.registerTempTable("jsonTable")
checkAnswer(
sql("select * from jsonTable"),
@@ -546,7 +546,7 @@ class JsonSuite extends QueryTest {
assert(schema === jsonSchemaRDD1.schema)
- jsonSchemaRDD1.registerAsTable("jsonTable1")
+ jsonSchemaRDD1.registerTempTable("jsonTable1")
checkAnswer(
sql("select * from jsonTable1"),
@@ -563,7 +563,7 @@ class JsonSuite extends QueryTest {
assert(schema === jsonSchemaRDD2.schema)
- jsonSchemaRDD2.registerAsTable("jsonTable2")
+ jsonSchemaRDD2.registerTempTable("jsonTable2")
checkAnswer(
sql("select * from jsonTable2"),
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/parquet/ParquetQuerySuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/parquet/ParquetQuerySuite.scala
index 8955455ec9..9933575038 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/parquet/ParquetQuerySuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/parquet/ParquetQuerySuite.scala
@@ -101,9 +101,9 @@ class ParquetQuerySuite extends QueryTest with FunSuiteLike with BeforeAndAfterA
ParquetTestData.writeNestedFile3()
ParquetTestData.writeNestedFile4()
testRDD = parquetFile(ParquetTestData.testDir.toString)
- testRDD.registerAsTable("testsource")
+ testRDD.registerTempTable("testsource")
parquetFile(ParquetTestData.testFilterDir.toString)
- .registerAsTable("testfiltersource")
+ .registerTempTable("testfiltersource")
}
override def afterAll() {
@@ -247,7 +247,7 @@ class ParquetQuerySuite extends QueryTest with FunSuiteLike with BeforeAndAfterA
test("Creating case class RDD table") {
TestSQLContext.sparkContext.parallelize((1 to 100))
.map(i => TestRDDEntry(i, s"val_$i"))
- .registerAsTable("tmp")
+ .registerTempTable("tmp")
val rdd = sql("SELECT * FROM tmp").collect().sortBy(_.getInt(0))
var counter = 1
rdd.foreach {
@@ -266,7 +266,7 @@ class ParquetQuerySuite extends QueryTest with FunSuiteLike with BeforeAndAfterA
.map(i => TestRDDEntry(i, s"val_$i"))
rdd.saveAsParquetFile(path)
val readFile = parquetFile(path)
- readFile.registerAsTable("tmpx")
+ readFile.registerTempTable("tmpx")
val rdd_copy = sql("SELECT * FROM tmpx").collect()
val rdd_orig = rdd.collect()
for(i <- 0 to 99) {
@@ -280,9 +280,9 @@ class ParquetQuerySuite extends QueryTest with FunSuiteLike with BeforeAndAfterA
val dirname = Utils.createTempDir()
val source_rdd = TestSQLContext.sparkContext.parallelize((1 to 100))
.map(i => TestRDDEntry(i, s"val_$i"))
- source_rdd.registerAsTable("source")
+ source_rdd.registerTempTable("source")
val dest_rdd = createParquetFile[TestRDDEntry](dirname.toString)
- dest_rdd.registerAsTable("dest")
+ dest_rdd.registerTempTable("dest")
sql("INSERT OVERWRITE INTO dest SELECT * FROM source").collect()
val rdd_copy1 = sql("SELECT * FROM dest").collect()
assert(rdd_copy1.size === 100)
@@ -547,7 +547,7 @@ class ParquetQuerySuite extends QueryTest with FunSuiteLike with BeforeAndAfterA
val data = nestedParserSqlContext
.parquetFile(ParquetTestData.testNestedDir1.toString)
.toSchemaRDD
- data.registerAsTable("data")
+ data.registerTempTable("data")
val query = nestedParserSqlContext.sql("SELECT owner, contacts[1].name FROM data")
val tmp = query.collect()
assert(tmp.size === 2)
@@ -562,7 +562,7 @@ class ParquetQuerySuite extends QueryTest with FunSuiteLike with BeforeAndAfterA
val data = nestedParserSqlContext
.parquetFile(ParquetTestData.testNestedDir2.toString)
.toSchemaRDD
- data.registerAsTable("data")
+ data.registerTempTable("data")
val result1 = nestedParserSqlContext.sql("SELECT entries[0].value FROM data").collect()
assert(result1.size === 1)
assert(result1(0).size === 1)
@@ -589,7 +589,7 @@ class ParquetQuerySuite extends QueryTest with FunSuiteLike with BeforeAndAfterA
val data = nestedParserSqlContext
.parquetFile(ParquetTestData.testNestedDir3.toString)
.toSchemaRDD
- data.registerAsTable("data")
+ data.registerTempTable("data")
val result1 = nestedParserSqlContext.sql("SELECT booleanNumberPairs[0].value[0].truth FROM data").collect()
assert(result1.size === 1)
assert(result1(0).size === 1)
@@ -608,7 +608,7 @@ class ParquetQuerySuite extends QueryTest with FunSuiteLike with BeforeAndAfterA
val data = TestSQLContext
.parquetFile(ParquetTestData.testNestedDir4.toString)
.toSchemaRDD
- data.registerAsTable("mapTable")
+ data.registerTempTable("mapTable")
val result1 = sql("SELECT data1 FROM mapTable").collect()
assert(result1.size === 1)
assert(result1(0)(0)
@@ -625,7 +625,7 @@ class ParquetQuerySuite extends QueryTest with FunSuiteLike with BeforeAndAfterA
val data = nestedParserSqlContext
.parquetFile(ParquetTestData.testNestedDir4.toString)
.toSchemaRDD
- data.registerAsTable("mapTable")
+ data.registerTempTable("mapTable")
val result1 = nestedParserSqlContext.sql("SELECT data2 FROM mapTable").collect()
assert(result1.size === 1)
val entry1 = result1(0)(0)
@@ -658,7 +658,7 @@ class ParquetQuerySuite extends QueryTest with FunSuiteLike with BeforeAndAfterA
nestedParserSqlContext
.parquetFile(tmpdir.toString)
.toSchemaRDD
- .registerAsTable("tmpcopy")
+ .registerTempTable("tmpcopy")
val tmpdata = nestedParserSqlContext.sql("SELECT owner, contacts[1].name FROM tmpcopy").collect()
assert(tmpdata.size === 2)
assert(tmpdata(0).size === 2)
@@ -679,7 +679,7 @@ class ParquetQuerySuite extends QueryTest with FunSuiteLike with BeforeAndAfterA
nestedParserSqlContext
.parquetFile(tmpdir.toString)
.toSchemaRDD
- .registerAsTable("tmpmapcopy")
+ .registerTempTable("tmpmapcopy")
val result1 = nestedParserSqlContext.sql("""SELECT data1["key2"] FROM tmpmapcopy""").collect()
assert(result1.size === 1)
assert(result1(0)(0) === 2)
diff --git a/sql/hive/src/test/scala/org/apache/spark/sql/hive/InsertIntoHiveTableSuite.scala b/sql/hive/src/test/scala/org/apache/spark/sql/hive/InsertIntoHiveTableSuite.scala
index 833f350215..7e323146f9 100644
--- a/sql/hive/src/test/scala/org/apache/spark/sql/hive/InsertIntoHiveTableSuite.scala
+++ b/sql/hive/src/test/scala/org/apache/spark/sql/hive/InsertIntoHiveTableSuite.scala
@@ -28,7 +28,7 @@ case class TestData(key: Int, value: String)
class InsertIntoHiveTableSuite extends QueryTest {
val testData = TestHive.sparkContext.parallelize(
(1 to 100).map(i => TestData(i, i.toString)))
- testData.registerAsTable("testData")
+ testData.registerTempTable("testData")
test("insertInto() HiveTable") {
createTable[TestData]("createAndInsertTest")
diff --git a/sql/hive/src/test/scala/org/apache/spark/sql/hive/api/java/JavaHiveQLSuite.scala b/sql/hive/src/test/scala/org/apache/spark/sql/hive/api/java/JavaHiveQLSuite.scala
index 10c8069a62..578f27574a 100644
--- a/sql/hive/src/test/scala/org/apache/spark/sql/hive/api/java/JavaHiveQLSuite.scala
+++ b/sql/hive/src/test/scala/org/apache/spark/sql/hive/api/java/JavaHiveQLSuite.scala
@@ -63,7 +63,7 @@ class JavaHiveQLSuite extends FunSuite {
javaHiveCtx.hql(s"CREATE TABLE $tableName(key INT, value STRING)").count()
}
- javaHiveCtx.hql("SHOW TABLES").registerAsTable("show_tables")
+ javaHiveCtx.hql("SHOW TABLES").registerTempTable("show_tables")
assert(
javaHiveCtx
@@ -73,7 +73,7 @@ class JavaHiveQLSuite extends FunSuite {
.contains(tableName))
assertResult(Array(Array("key", "int", "None"), Array("value", "string", "None"))) {
- javaHiveCtx.hql(s"DESCRIBE $tableName").registerAsTable("describe_table")
+ javaHiveCtx.hql(s"DESCRIBE $tableName").registerTempTable("describe_table")
javaHiveCtx
.hql("SELECT result FROM describe_table")
diff --git a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveQuerySuite.scala b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveQuerySuite.scala
index 89cc589fb8..4ed41550cf 100644
--- a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveQuerySuite.scala
+++ b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveQuerySuite.scala
@@ -247,7 +247,7 @@ class HiveQuerySuite extends HiveComparisonTest {
TestHive.sparkContext.parallelize(
TestData(1, "str1") ::
TestData(2, "str2") :: Nil)
- testData.registerAsTable("REGisteredTABle")
+ testData.registerTempTable("REGisteredTABle")
assertResult(Array(Array(2, "str2"))) {
hql("SELECT tablealias.A, TABLEALIAS.b FROM reGisteredTABle TableAlias " +
@@ -272,7 +272,7 @@ class HiveQuerySuite extends HiveComparisonTest {
test("SPARK-2180: HAVING support in GROUP BY clauses (positive)") {
val fixture = List(("foo", 2), ("bar", 1), ("foo", 4), ("bar", 3))
.zipWithIndex.map {case Pair(Pair(value, attr), key) => HavingRow(key, value, attr)}
- TestHive.sparkContext.parallelize(fixture).registerAsTable("having_test")
+ TestHive.sparkContext.parallelize(fixture).registerTempTable("having_test")
val results =
hql("SELECT value, max(attr) AS attr FROM having_test GROUP BY value HAVING attr > 3")
.collect()
@@ -401,7 +401,7 @@ class HiveQuerySuite extends HiveComparisonTest {
TestHive.sparkContext.parallelize(
TestData(1, "str1") ::
TestData(1, "str2") :: Nil)
- testData.registerAsTable("test_describe_commands2")
+ testData.registerTempTable("test_describe_commands2")
assertResult(
Array(
diff --git a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveResolutionSuite.scala b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveResolutionSuite.scala
index fb03db12a0..2455c18925 100644
--- a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveResolutionSuite.scala
+++ b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveResolutionSuite.scala
@@ -54,14 +54,14 @@ class HiveResolutionSuite extends HiveComparisonTest {
test("case insensitivity with scala reflection") {
// Test resolution with Scala Reflection
TestHive.sparkContext.parallelize(Data(1, 2, Nested(1,2), Seq(Nested(1,2))) :: Nil)
- .registerAsTable("caseSensitivityTest")
+ .registerTempTable("caseSensitivityTest")
hql("SELECT a, b, A, B, n.a, n.b, n.A, n.B FROM caseSensitivityTest")
}
test("nested repeated resolution") {
TestHive.sparkContext.parallelize(Data(1, 2, Nested(1,2), Seq(Nested(1,2))) :: Nil)
- .registerAsTable("nestedRepeatedTest")
+ .registerTempTable("nestedRepeatedTest")
assert(hql("SELECT nestedArray[0].a FROM nestedRepeatedTest").collect().head(0) === 1)
}
diff --git a/sql/hive/src/test/scala/org/apache/spark/sql/parquet/HiveParquetSuite.scala b/sql/hive/src/test/scala/org/apache/spark/sql/parquet/HiveParquetSuite.scala
index 47526e3596..6545e8d7dc 100644
--- a/sql/hive/src/test/scala/org/apache/spark/sql/parquet/HiveParquetSuite.scala
+++ b/sql/hive/src/test/scala/org/apache/spark/sql/parquet/HiveParquetSuite.scala
@@ -41,7 +41,7 @@ class HiveParquetSuite extends FunSuite with BeforeAndAfterAll with BeforeAndAft
// write test data
ParquetTestData.writeFile()
testRDD = parquetFile(ParquetTestData.testDir.toString)
- testRDD.registerAsTable("testsource")
+ testRDD.registerTempTable("testsource")
}
override def afterAll() {
@@ -67,7 +67,7 @@ class HiveParquetSuite extends FunSuite with BeforeAndAfterAll with BeforeAndAft
.map(i => Cases(i, i))
.saveAsParquetFile(tempFile.getCanonicalPath)
- parquetFile(tempFile.getCanonicalPath).registerAsTable("cases")
+ parquetFile(tempFile.getCanonicalPath).registerTempTable("cases")
hql("SELECT upper FROM cases").collect().map(_.getString(0)) === (1 to 10).map(_.toString)
hql("SELECT LOWER FROM cases").collect().map(_.getString(0)) === (1 to 10).map(_.toString)
}
@@ -86,7 +86,7 @@ class HiveParquetSuite extends FunSuite with BeforeAndAfterAll with BeforeAndAft
test("Converting Hive to Parquet Table via saveAsParquetFile") {
hql("SELECT * FROM src").saveAsParquetFile(dirname.getAbsolutePath)
- parquetFile(dirname.getAbsolutePath).registerAsTable("ptable")
+ parquetFile(dirname.getAbsolutePath).registerTempTable("ptable")
val rddOne = hql("SELECT * FROM src").collect().sortBy(_.getInt(0))
val rddTwo = hql("SELECT * from ptable").collect().sortBy(_.getInt(0))
compareRDDs(rddOne, rddTwo, "src (Hive)", Seq("key:Int", "value:String"))
@@ -94,7 +94,7 @@ class HiveParquetSuite extends FunSuite with BeforeAndAfterAll with BeforeAndAft
test("INSERT OVERWRITE TABLE Parquet table") {
hql("SELECT * FROM testsource").saveAsParquetFile(dirname.getAbsolutePath)
- parquetFile(dirname.getAbsolutePath).registerAsTable("ptable")
+ parquetFile(dirname.getAbsolutePath).registerTempTable("ptable")
// let's do three overwrites for good measure
hql("INSERT OVERWRITE TABLE ptable SELECT * FROM testsource").collect()
hql("INSERT OVERWRITE TABLE ptable SELECT * FROM testsource").collect()