aboutsummaryrefslogtreecommitdiff
path: root/docs
diff options
context:
space:
mode:
authorDongjoon Hyun <dongjoon@apache.org>2017-01-07 18:55:01 -0800
committergatorsmile <gatorsmile@gmail.com>2017-01-07 18:55:01 -0800
commit923e594844a7ad406195b91877f0fb374d5a454b (patch)
tree160631e62a498791bff9817ebe95764dcb8c397d /docs
parent54138f6e89abfc17101b4f2812715784a2b98331 (diff)
downloadspark-923e594844a7ad406195b91877f0fb374d5a454b.tar.gz
spark-923e594844a7ad406195b91877f0fb374d5a454b.tar.bz2
spark-923e594844a7ad406195b91877f0fb374d5a454b.zip
[SPARK-18941][SQL][DOC] Add a new behavior document on `CREATE/DROP TABLE` with `LOCATION`
## What changes were proposed in this pull request? This PR adds a new behavior change description on `CREATE TABLE ... LOCATION` at `sql-programming-guide.md` clearly under `Upgrading From Spark SQL 1.6 to 2.0`. This change is introduced at Apache Spark 2.0.0 as [SPARK-15276](https://issues.apache.org/jira/browse/SPARK-15276). ## How was this patch tested? ``` SKIP_API=1 jekyll build ``` **Newly Added Description** <img width="913" alt="new" src="https://cloud.githubusercontent.com/assets/9700541/21743606/7efe2b12-d4ba-11e6-8a0d-551222718ea2.png"> Author: Dongjoon Hyun <dongjoon@apache.org> Closes #16400 from dongjoon-hyun/SPARK-18941.
Diffstat (limited to 'docs')
-rw-r--r--docs/sql-programming-guide.md8
1 files changed, 8 insertions, 0 deletions
diff --git a/docs/sql-programming-guide.md b/docs/sql-programming-guide.md
index 0f6e344655..f4c89e58fa 100644
--- a/docs/sql-programming-guide.md
+++ b/docs/sql-programming-guide.md
@@ -1413,6 +1413,14 @@ options.
- Dataset and DataFrame API `explode` has been deprecated, alternatively, use `functions.explode()` with `select` or `flatMap`
- Dataset and DataFrame API `registerTempTable` has been deprecated and replaced by `createOrReplaceTempView`
+ - Changes to `CREATE TABLE ... LOCATION` behavior for Hive tables.
+ - From Spark 2.0, `CREATE TABLE ... LOCATION` is equivalent to `CREATE EXTERNAL TABLE ... LOCATION`
+ in order to prevent accidental dropping the existing data in the user-provided locations.
+ That means, a Hive table created in Spark SQL with the user-specified location is always a Hive external table.
+ Dropping external tables will not remove the data. Users are not allowed to specify the location for Hive managed tables.
+ Note that this is different from the Hive behavior.
+ - As a result, `DROP TABLE` statements on those tables will not remove the data.
+
## Upgrading From Spark SQL 1.5 to 1.6
- From Spark 1.6, by default the Thrift server runs in multi-session mode. Which means each JDBC/ODBC