aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorMichael Armbrust <michael@databricks.com>2015-08-25 10:22:54 -0700
committerMichael Armbrust <michael@databricks.com>2015-08-25 10:23:08 -0700
commit0402f1297c697bfbe8b5c7bfc170fcdc6b2c9de5 (patch)
tree0bac3dc1381f0341e719a6ea3566f76df2706012
parentbdcc8e608d9a1160db988faa76808149c28a3b50 (diff)
downloadspark-0402f1297c697bfbe8b5c7bfc170fcdc6b2c9de5.tar.gz
spark-0402f1297c697bfbe8b5c7bfc170fcdc6b2c9de5.tar.bz2
spark-0402f1297c697bfbe8b5c7bfc170fcdc6b2c9de5.zip
[SPARK-10198] [SQL] Turn off partition verification by default
Author: Michael Armbrust <michael@databricks.com> Closes #8404 from marmbrus/turnOffPartitionVerification. (cherry picked from commit 5c08c86bfa43462fb2ca5f7c5980ddfb44dd57f8) Signed-off-by: Michael Armbrust <michael@databricks.com>
-rw-r--r--sql/core/src/main/scala/org/apache/spark/sql/SQLConf.scala2
-rw-r--r--sql/hive/src/test/scala/org/apache/spark/sql/hive/QueryPartitionSuite.scala64
2 files changed, 35 insertions, 31 deletions
diff --git a/sql/core/src/main/scala/org/apache/spark/sql/SQLConf.scala b/sql/core/src/main/scala/org/apache/spark/sql/SQLConf.scala
index e9de14f025..2974055a14 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/SQLConf.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/SQLConf.scala
@@ -312,7 +312,7 @@ private[spark] object SQLConf {
doc = "When true, enable filter pushdown for ORC files.")
val HIVE_VERIFY_PARTITION_PATH = booleanConf("spark.sql.hive.verifyPartitionPath",
- defaultValue = Some(true),
+ defaultValue = Some(false),
doc = "<TODO>")
val HIVE_METASTORE_PARTITION_PRUNING = booleanConf("spark.sql.hive.metastorePartitionPruning",
diff --git a/sql/hive/src/test/scala/org/apache/spark/sql/hive/QueryPartitionSuite.scala b/sql/hive/src/test/scala/org/apache/spark/sql/hive/QueryPartitionSuite.scala
index 017bc2adc1..1cc8a93e83 100644
--- a/sql/hive/src/test/scala/org/apache/spark/sql/hive/QueryPartitionSuite.scala
+++ b/sql/hive/src/test/scala/org/apache/spark/sql/hive/QueryPartitionSuite.scala
@@ -18,50 +18,54 @@
package org.apache.spark.sql.hive
import com.google.common.io.Files
+import org.apache.spark.sql.test.SQLTestUtils
import org.apache.spark.sql.{QueryTest, _}
import org.apache.spark.util.Utils
-class QueryPartitionSuite extends QueryTest {
+class QueryPartitionSuite extends QueryTest with SQLTestUtils {
private lazy val ctx = org.apache.spark.sql.hive.test.TestHive
import ctx.implicits._
- import ctx.sql
+
+ protected def _sqlContext = ctx
test("SPARK-5068: query data when path doesn't exist"){
- val testData = ctx.sparkContext.parallelize(
- (1 to 10).map(i => TestData(i, i.toString))).toDF()
- testData.registerTempTable("testData")
+ withSQLConf((SQLConf.HIVE_VERIFY_PARTITION_PATH.key, "true")) {
+ val testData = ctx.sparkContext.parallelize(
+ (1 to 10).map(i => TestData(i, i.toString))).toDF()
+ testData.registerTempTable("testData")
- val tmpDir = Files.createTempDir()
- // create the table for test
- sql(s"CREATE TABLE table_with_partition(key int,value string) " +
- s"PARTITIONED by (ds string) location '${tmpDir.toURI.toString}' ")
- sql("INSERT OVERWRITE TABLE table_with_partition partition (ds='1') " +
- "SELECT key,value FROM testData")
- sql("INSERT OVERWRITE TABLE table_with_partition partition (ds='2') " +
- "SELECT key,value FROM testData")
- sql("INSERT OVERWRITE TABLE table_with_partition partition (ds='3') " +
- "SELECT key,value FROM testData")
- sql("INSERT OVERWRITE TABLE table_with_partition partition (ds='4') " +
- "SELECT key,value FROM testData")
+ val tmpDir = Files.createTempDir()
+ // create the table for test
+ sql(s"CREATE TABLE table_with_partition(key int,value string) " +
+ s"PARTITIONED by (ds string) location '${tmpDir.toURI.toString}' ")
+ sql("INSERT OVERWRITE TABLE table_with_partition partition (ds='1') " +
+ "SELECT key,value FROM testData")
+ sql("INSERT OVERWRITE TABLE table_with_partition partition (ds='2') " +
+ "SELECT key,value FROM testData")
+ sql("INSERT OVERWRITE TABLE table_with_partition partition (ds='3') " +
+ "SELECT key,value FROM testData")
+ sql("INSERT OVERWRITE TABLE table_with_partition partition (ds='4') " +
+ "SELECT key,value FROM testData")
- // test for the exist path
- checkAnswer(sql("select key,value from table_with_partition"),
- testData.toDF.collect ++ testData.toDF.collect
- ++ testData.toDF.collect ++ testData.toDF.collect)
+ // test for the exist path
+ checkAnswer(sql("select key,value from table_with_partition"),
+ testData.toDF.collect ++ testData.toDF.collect
+ ++ testData.toDF.collect ++ testData.toDF.collect)
- // delete the path of one partition
- tmpDir.listFiles
- .find { f => f.isDirectory && f.getName().startsWith("ds=") }
- .foreach { f => Utils.deleteRecursively(f) }
+ // delete the path of one partition
+ tmpDir.listFiles
+ .find { f => f.isDirectory && f.getName().startsWith("ds=") }
+ .foreach { f => Utils.deleteRecursively(f) }
- // test for after delete the path
- checkAnswer(sql("select key,value from table_with_partition"),
- testData.toDF.collect ++ testData.toDF.collect ++ testData.toDF.collect)
+ // test for after delete the path
+ checkAnswer(sql("select key,value from table_with_partition"),
+ testData.toDF.collect ++ testData.toDF.collect ++ testData.toDF.collect)
- sql("DROP TABLE table_with_partition")
- sql("DROP TABLE createAndInsertTest")
+ sql("DROP TABLE table_with_partition")
+ sql("DROP TABLE createAndInsertTest")
+ }
}
}