aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorJihong MA <linlin200605@gmail.com>2015-06-19 14:05:11 +0200
committerSean Owen <sowen@cloudera.com>2015-06-19 14:06:49 +0200
commitebd363aecde977511469d47fb1ea7cb5df3c3541 (patch)
tree9f11887b87623f627febfcd8eb7abcf6bd507885
parent93360dc3cd6186e9d33c762d153a829a5882b72b (diff)
downloadspark-ebd363aecde977511469d47fb1ea7cb5df3c3541.tar.gz
spark-ebd363aecde977511469d47fb1ea7cb5df3c3541.tar.bz2
spark-ebd363aecde977511469d47fb1ea7cb5df3c3541.zip
[SPARK-7265] Improving documentation for Spark SQL Hive support
Please review this pull request. Author: Jihong MA <linlin200605@gmail.com> Closes #5933 from JihongMA/SPARK-7265 and squashes the following commits: dfaa971 [Jihong MA] SPARK-7265 minor fix of the content ace454d [Jihong MA] SPARK-7265 take out PySpark on YARN limitation 9ea0832 [Jihong MA] Merge remote-tracking branch 'upstream/master' d5bf3f5 [Jihong MA] Merge remote-tracking branch 'upstream/master' 7b842e6 [Jihong MA] Merge remote-tracking branch 'upstream/master' 9c84695 [Jihong MA] SPARK-7265 address review comment a399aa6 [Jihong MA] SPARK-7265 Improving documentation for Spark SQL Hive support
-rw-r--r--docs/sql-programming-guide.md7
1 files changed, 6 insertions, 1 deletions
diff --git a/docs/sql-programming-guide.md b/docs/sql-programming-guide.md
index 9b5ea394a6..26c036f664 100644
--- a/docs/sql-programming-guide.md
+++ b/docs/sql-programming-guide.md
@@ -1445,7 +1445,12 @@ This command builds a new assembly jar that includes Hive. Note that this Hive a
on all of the worker nodes, as they will need access to the Hive serialization and deserialization libraries
(SerDes) in order to access data stored in Hive.
-Configuration of Hive is done by placing your `hive-site.xml` file in `conf/`.
+Configuration of Hive is done by placing your `hive-site.xml` file in `conf/`. Please note when running
+the query on a YARN cluster (`yarn-cluster` mode), the `datanucleus` jars under the `lib_managed/jars` directory
+and `hive-site.xml` under `conf/` directory need to be available on the driver and all executors launched by the
+YARN cluster. The convenient way to do this is adding them through the `--jars` option and `--file` option of the
+`spark-submit` command.
+
<div class="codetabs">