aboutsummaryrefslogtreecommitdiff
path: root/sql
diff options
context:
space:
mode:
Diffstat (limited to 'sql')
-rw-r--r--sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/rules.scala7
-rw-r--r--sql/core/src/test/scala/org/apache/spark/sql/execution/command/DDLSuite.scala7
2 files changed, 13 insertions, 1 deletions
diff --git a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/rules.scala b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/rules.scala
index 5b96206ba8..fbf4063ff6 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/rules.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/execution/datasources/rules.scala
@@ -97,7 +97,12 @@ case class PreprocessDDL(conf: SQLConf) extends Rule[LogicalPlan] {
// * sort columns' type must be orderable.
case c @ CreateTable(tableDesc, mode, query) if c.childrenResolved =>
val schema = if (query.isDefined) query.get.schema else tableDesc.schema
- checkDuplication(schema.map(_.name), "table definition of " + tableDesc.identifier)
+ val columnNames = if (conf.caseSensitiveAnalysis) {
+ schema.map(_.name)
+ } else {
+ schema.map(_.name.toLowerCase)
+ }
+ checkDuplication(columnNames, "table definition of " + tableDesc.identifier)
val partitionColsChecked = checkPartitionColumns(schema, tableDesc)
val bucketColsChecked = checkBucketColumns(schema, partitionColsChecked)
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/command/DDLSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/command/DDLSuite.scala
index fd35c987ca..05f826a11b 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/command/DDLSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/command/DDLSuite.scala
@@ -371,6 +371,13 @@ class DDLSuite extends QueryTest with SharedSQLContext with BeforeAndAfterEach {
sql("CREATE TABLE tbl(a int, a string) USING json")
}
assert(e.message == "Found duplicate column(s) in table definition of `tbl`: a")
+
+ withSQLConf(SQLConf.CASE_SENSITIVE.key -> "false") {
+ val e2 = intercept[AnalysisException] {
+ sql("CREATE TABLE tbl(a int, A string) USING json")
+ }
+ assert(e2.message == "Found duplicate column(s) in table definition of `tbl`: a")
+ }
}
test("create table - partition column names not in table definition") {