diff options
author | DoingDone9 <799203320@qq.com> | 2015-03-26 17:04:19 +0800 |
---|---|---|
committer | Cheng Lian <lian@databricks.com> | 2015-03-26 17:04:19 +0800 |
commit | 855cba8fe59ffe17b51ed00fbbb5d3d7cf17ade9 (patch) | |
tree | 912194634520cb9df98dd870251458ec917a739d /sql/hive/src | |
parent | 5bbcd1304cfebba31ec6857a80d3825a40d02e83 (diff) | |
download | spark-855cba8fe59ffe17b51ed00fbbb5d3d7cf17ade9.tar.gz spark-855cba8fe59ffe17b51ed00fbbb5d3d7cf17ade9.tar.bz2 spark-855cba8fe59ffe17b51ed00fbbb5d3d7cf17ade9.zip |
[SPARK-6546][Build] Using the wrong code that will make spark compile failed!!
wrong code : val tmpDir = Files.createTempDir()
not Files should Utils
Author: DoingDone9 <799203320@qq.com>
Closes #5198 from DoingDone9/FilesBug and squashes the following commits:
6e0140d [DoingDone9] Update InsertIntoHiveTableSuite.scala
e57d23f [DoingDone9] Update InsertIntoHiveTableSuite.scala
802261c [DoingDone9] Merge pull request #7 from apache/master
d00303b [DoingDone9] Merge pull request #6 from apache/master
98b134f [DoingDone9] Merge pull request #5 from apache/master
161cae3 [DoingDone9] Merge pull request #4 from apache/master
c87e8b6 [DoingDone9] Merge pull request #3 from apache/master
cb1852d [DoingDone9] Merge pull request #2 from apache/master
c3f046f [DoingDone9] Merge pull request #1 from apache/master
Diffstat (limited to 'sql/hive/src')
-rw-r--r-- | sql/hive/src/test/scala/org/apache/spark/sql/hive/InsertIntoHiveTableSuite.scala | 2 |
1 files changed, 1 insertions, 1 deletions
diff --git a/sql/hive/src/test/scala/org/apache/spark/sql/hive/InsertIntoHiveTableSuite.scala b/sql/hive/src/test/scala/org/apache/spark/sql/hive/InsertIntoHiveTableSuite.scala index aa6fb42de7..8011952e0d 100644 --- a/sql/hive/src/test/scala/org/apache/spark/sql/hive/InsertIntoHiveTableSuite.scala +++ b/sql/hive/src/test/scala/org/apache/spark/sql/hive/InsertIntoHiveTableSuite.scala @@ -198,7 +198,7 @@ class InsertIntoHiveTableSuite extends QueryTest with BeforeAndAfter { val testDatawithNull = TestHive.sparkContext.parallelize( (1 to 10).map(i => ThreeCloumntable(i, i.toString,null))).toDF() - val tmpDir = Files.createTempDir() + val tmpDir = Utils.createTempDir() sql(s"CREATE TABLE table_with_partition(key int,value string) PARTITIONED by (ds string) location '${tmpDir.toURI.toString}' ") sql("INSERT OVERWRITE TABLE table_with_partition partition (ds='1') SELECT key,value FROM testData") |