aboutsummaryrefslogtreecommitdiff
path: root/sql/core/src/test
diff options
context:
space:
mode:
authorwindpiger <songjun@outlook.com>2017-02-28 11:59:18 -0800
committerWenchen Fan <wenchen@databricks.com>2017-02-28 11:59:18 -0800
commitce233f18e381fa1ea00be74ca26e97d35baa6c9c (patch)
tree3ea6727866bbd54ff34e57c93e91d7b6481f2ce9 /sql/core/src/test
parent9734a928a75d29ea202e9f309f92ca4637d35671 (diff)
downloadspark-ce233f18e381fa1ea00be74ca26e97d35baa6c9c.tar.gz
spark-ce233f18e381fa1ea00be74ca26e97d35baa6c9c.tar.bz2
spark-ce233f18e381fa1ea00be74ca26e97d35baa6c9c.zip
[SPARK-19463][SQL] refresh cache after the InsertIntoHadoopFsRelationCommand
## What changes were proposed in this pull request? If we first cache a DataSource table, then we insert some data into the table, we should refresh the data in the cache after the insert command. ## How was this patch tested? unit test added Author: windpiger <songjun@outlook.com> Closes #16809 from windpiger/refreshCacheAfterInsert.
Diffstat (limited to 'sql/core/src/test')
-rw-r--r--sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetQuerySuite.scala4
-rw-r--r--sql/core/src/test/scala/org/apache/spark/sql/sources/InsertSuite.scala18
2 files changed, 9 insertions, 13 deletions
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetQuerySuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetQuerySuite.scala
index d7d7176c48..200e356c72 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetQuerySuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetQuerySuite.scala
@@ -77,8 +77,6 @@ class ParquetQuerySuite extends QueryTest with ParquetTest with SharedSQLContext
val df = spark.read.parquet(path).cache()
assert(df.count() == 1000)
spark.range(10).write.mode("overwrite").parquet(path)
- assert(df.count() == 1000)
- spark.catalog.refreshByPath(path)
assert(df.count() == 10)
assert(spark.read.parquet(path).count() == 10)
}
@@ -91,8 +89,6 @@ class ParquetQuerySuite extends QueryTest with ParquetTest with SharedSQLContext
val df = spark.read.parquet(path).cache()
assert(df.count() == 1000)
spark.range(10).write.mode("append").parquet(path)
- assert(df.count() == 1000)
- spark.catalog.refreshByPath(path)
assert(df.count() == 1010)
assert(spark.read.parquet(path).count() == 1010)
}
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/sources/InsertSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/sources/InsertSuite.scala
index 19835cd184..2eae66dda8 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/sources/InsertSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/sources/InsertSuite.scala
@@ -281,15 +281,15 @@ class InsertSuite extends DataSourceTest with SharedSQLContext {
""".stripMargin)
// jsonTable should be recached.
assertCached(sql("SELECT * FROM jsonTable"))
- // TODO we need to invalidate the cached data in InsertIntoHadoopFsRelation
-// // The cached data is the new data.
-// checkAnswer(
-// sql("SELECT a, b FROM jsonTable"),
-// sql("SELECT a * 2, b FROM jt").collect())
-//
-// // Verify uncaching
-// spark.catalog.uncacheTable("jsonTable")
-// assertCached(sql("SELECT * FROM jsonTable"), 0)
+
+ // The cached data is the new data.
+ checkAnswer(
+ sql("SELECT a, b FROM jsonTable"),
+ sql("SELECT a * 2, b FROM jt").collect())
+
+ // Verify uncaching
+ spark.catalog.uncacheTable("jsonTable")
+ assertCached(sql("SELECT * FROM jsonTable"), 0)
}
test("it's not allowed to insert into a relation that is not an InsertableRelation") {