aboutsummaryrefslogtreecommitdiff
path: root/sql
diff options
context:
space:
mode:
authorMichael Armbrust <michael@databricks.com>2014-07-31 11:26:43 -0700
committerMichael Armbrust <michael@databricks.com>2014-07-31 11:26:43 -0700
commit72cfb13987bab07461266905930f84619b3a0068 (patch)
treef54cb89724c43628ba5b0fc8306c9bd53f74826e /sql
parent3072b96026fa3e63e8eef780f2b04dd81f11ea27 (diff)
downloadspark-72cfb13987bab07461266905930f84619b3a0068.tar.gz
spark-72cfb13987bab07461266905930f84619b3a0068.tar.bz2
spark-72cfb13987bab07461266905930f84619b3a0068.zip
[SPARK-2397][SQL] Deprecate LocalHiveContext
LocalHiveContext is redundant with HiveContext. The only difference is it creates `./metastore` instead of `./metastore_db`. Author: Michael Armbrust <michael@databricks.com> Closes #1641 from marmbrus/localHiveContext and squashes the following commits: e5ec497 [Michael Armbrust] Add deprecation version 626e056 [Michael Armbrust] Don't remove from imports yet 905cc5f [Michael Armbrust] Merge remote-tracking branch 'apache/master' into localHiveContext 1c2727e [Michael Armbrust] Deprecate LocalHiveContext
Diffstat (limited to 'sql')
-rw-r--r--sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveContext.scala7
-rw-r--r--sql/hive/src/main/scala/org/apache/spark/sql/hive/TestHive.scala15
2 files changed, 17 insertions, 5 deletions
diff --git a/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveContext.scala b/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveContext.scala
index b413373345..27b444daba 100644
--- a/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveContext.scala
+++ b/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveContext.scala
@@ -42,9 +42,12 @@ import org.apache.spark.sql.execution.{Command => PhysicalCommand}
import org.apache.spark.sql.hive.execution.DescribeHiveTableCommand
/**
- * Starts up an instance of hive where metadata is stored locally. An in-process metadata data is
- * created with data stored in ./metadata. Warehouse data is stored in in ./warehouse.
+ * DEPRECATED: Use HiveContext instead.
*/
+@deprecated("""
+ Use HiveContext instead. It will still create a local metastore if one is not specified.
+ However, note that the default directory is ./metastore_db, not ./metastore
+ """, "1.1")
class LocalHiveContext(sc: SparkContext) extends HiveContext(sc) {
lazy val metastorePath = new File("metastore").getCanonicalPath
diff --git a/sql/hive/src/main/scala/org/apache/spark/sql/hive/TestHive.scala b/sql/hive/src/main/scala/org/apache/spark/sql/hive/TestHive.scala
index 9386008d02..c50e8c4b5c 100644
--- a/sql/hive/src/main/scala/org/apache/spark/sql/hive/TestHive.scala
+++ b/sql/hive/src/main/scala/org/apache/spark/sql/hive/TestHive.scala
@@ -53,15 +53,24 @@ object TestHive
* hive metastore seems to lead to weird non-deterministic failures. Therefore, the execution of
* test cases that rely on TestHive must be serialized.
*/
-class TestHiveContext(sc: SparkContext) extends LocalHiveContext(sc) {
+class TestHiveContext(sc: SparkContext) extends HiveContext(sc) {
self =>
// By clearing the port we force Spark to pick a new one. This allows us to rerun tests
// without restarting the JVM.
System.clearProperty("spark.hostPort")
- override lazy val warehousePath = getTempFilePath("sparkHiveWarehouse").getCanonicalPath
- override lazy val metastorePath = getTempFilePath("sparkHiveMetastore").getCanonicalPath
+ lazy val warehousePath = getTempFilePath("sparkHiveWarehouse").getCanonicalPath
+ lazy val metastorePath = getTempFilePath("sparkHiveMetastore").getCanonicalPath
+
+ /** Sets up the system initially or after a RESET command */
+ protected def configure() {
+ set("javax.jdo.option.ConnectionURL",
+ s"jdbc:derby:;databaseName=$metastorePath;create=true")
+ set("hive.metastore.warehouse.dir", warehousePath)
+ }
+
+ configure() // Must be called before initializing the catalog below.
/** The location of the compiled hive distribution */
lazy val hiveHome = envVarToFile("HIVE_HOME")