aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorCheng Lian <lian@databricks.com>2015-04-02 12:59:38 +0800
committerCheng Lian <lian@databricks.com>2015-04-02 12:59:38 +0800
commit1160cc9e1883fb4704c46e7983799671313f9f39 (patch)
tree43ac5e2bb592bec1e98cbdc0806a7f86f09bcc84
parent223dd3f5df204297a02154389c14921524387477 (diff)
downloadspark-1160cc9e1883fb4704c46e7983799671313f9f39.tar.gz
spark-1160cc9e1883fb4704c46e7983799671313f9f39.tar.bz2
spark-1160cc9e1883fb4704c46e7983799671313f9f39.zip
Revert "[SPARK-6618][SQL] HiveMetastoreCatalog.lookupRelation should use fine-grained lock"
This reverts commit fd600cec0c8cf9e14c3d5d5f63b1de94413ffba8.
-rw-r--r--sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveMetastoreCatalog.scala12
-rw-r--r--sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/SQLQuerySuite.scala11
2 files changed, 3 insertions, 20 deletions
diff --git a/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveMetastoreCatalog.scala b/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveMetastoreCatalog.scala
index 2b5d031741..f0076cef13 100644
--- a/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveMetastoreCatalog.scala
+++ b/sql/hive/src/main/scala/org/apache/spark/sql/hive/HiveMetastoreCatalog.scala
@@ -173,16 +173,12 @@ private[hive] class HiveMetastoreCatalog(hive: HiveContext) extends Catalog with
def lookupRelation(
tableIdentifier: Seq[String],
- alias: Option[String]): LogicalPlan = {
+ alias: Option[String]): LogicalPlan = synchronized {
val tableIdent = processTableIdentifier(tableIdentifier)
val databaseName = tableIdent.lift(tableIdent.size - 2).getOrElse(
hive.sessionState.getCurrentDatabase)
val tblName = tableIdent.last
- val table = try {
- synchronized {
- client.getTable(databaseName, tblName)
- }
- } catch {
+ val table = try client.getTable(databaseName, tblName) catch {
case te: org.apache.hadoop.hive.ql.metadata.InvalidTableException =>
throw new NoSuchTableException
}
@@ -204,9 +200,7 @@ private[hive] class HiveMetastoreCatalog(hive: HiveContext) extends Catalog with
} else {
val partitions: Seq[Partition] =
if (table.isPartitioned) {
- synchronized {
- HiveShim.getAllPartitionsOf(client, table).toSeq
- }
+ HiveShim.getAllPartitionsOf(client, table).toSeq
} else {
Nil
}
diff --git a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/SQLQuerySuite.scala b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/SQLQuerySuite.scala
index efb57f7dcb..ec791447c7 100644
--- a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/SQLQuerySuite.scala
+++ b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/SQLQuerySuite.scala
@@ -450,15 +450,4 @@ class SQLQuerySuite extends QueryTest {
dropTempTable("data")
setConf("spark.sql.hive.convertCTAS", originalConf)
}
-
- test("sanity test for SPARK-6618") {
- (1 to 100).par.map { i =>
- val tableName = s"SPARK_6618_table_$i"
- sql(s"CREATE TABLE $tableName (col1 string)")
- catalog.lookupRelation(Seq(tableName))
- table(tableName)
- tables()
- sql(s"DROP TABLE $tableName")
- }
- }
}