diff options
Diffstat (limited to 'sql/core')
4 files changed, 19 insertions, 18 deletions
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/ColumnExpressionSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/ColumnExpressionSuite.scala index b0339a88fb..ee280a313c 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/ColumnExpressionSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/ColumnExpressionSuite.scala @@ -542,7 +542,7 @@ class ColumnExpressionSuite extends QueryTest with SharedSQLContext { val q = spark.read.parquet(dir.getCanonicalPath).select( input_file_name(), expr("input_file_block_start()"), expr("input_file_block_length()")) val firstRow = q.head() - assert(firstRow.getString(0).contains(dir.getCanonicalPath)) + assert(firstRow.getString(0).contains(dir.toURI.getPath)) assert(firstRow.getLong(1) == 0) assert(firstRow.getLong(2) > 0) @@ -566,7 +566,7 @@ class ColumnExpressionSuite extends QueryTest with SharedSQLContext { val q = df.select( input_file_name(), expr("input_file_block_start()"), expr("input_file_block_length()")) val firstRow = q.head() - assert(firstRow.getString(0).contains(dir.getCanonicalPath)) + assert(firstRow.getString(0).contains(dir.toURI.getPath)) assert(firstRow.getLong(1) == 0) assert(firstRow.getLong(2) > 0) @@ -595,7 +595,7 @@ class ColumnExpressionSuite extends QueryTest with SharedSQLContext { val q = df.select( input_file_name(), expr("input_file_block_start()"), expr("input_file_block_length()")) val firstRow = q.head() - assert(firstRow.getString(0).contains(dir.getCanonicalPath)) + assert(firstRow.getString(0).contains(dir.toURI.getPath)) assert(firstRow.getLong(1) == 0) assert(firstRow.getLong(2) > 0) diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/GlobalTempViewSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/GlobalTempViewSuite.scala index 391bcb8b35..5c63c6a414 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/execution/GlobalTempViewSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/GlobalTempViewSuite.scala @@ -93,7 +93,7 @@ class GlobalTempViewSuite extends QueryTest with SharedSQLContext { withTempPath { path => try { Seq(1 -> "a").toDF("i", "j").write.parquet(path.getAbsolutePath) - sql(s"CREATE GLOBAL TEMP VIEW src USING parquet OPTIONS (PATH '${path.getAbsolutePath}')") + sql(s"CREATE GLOBAL TEMP VIEW src USING parquet OPTIONS (PATH '${path.toURI}')") checkAnswer(spark.table(s"$globalTempDB.src"), Row(1, "a")) sql(s"INSERT INTO $globalTempDB.src SELECT 2, 'b'") checkAnswer(spark.table(s"$globalTempDB.src"), Row(1, "a") :: Row(2, "b") :: Nil) diff --git a/sql/core/src/test/scala/org/apache/spark/sql/sources/CreateTableAsSelectSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/sources/CreateTableAsSelectSuite.scala index 61939fe5ef..99da1969fc 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/sources/CreateTableAsSelectSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/sources/CreateTableAsSelectSuite.scala @@ -70,7 +70,7 @@ class CreateTableAsSelectSuite |CREATE TABLE jsonTable |USING json |OPTIONS ( - | path '${path.toString}' + | path '${path.toURI}' |) AS |SELECT a, b FROM jt """.stripMargin) @@ -94,7 +94,7 @@ class CreateTableAsSelectSuite |CREATE TABLE jsonTable |USING json |OPTIONS ( - | path '${childPath.toString}' + | path '${childPath.toURI}' |) AS |SELECT a, b FROM jt """.stripMargin) @@ -112,7 +112,7 @@ class CreateTableAsSelectSuite |CREATE TABLE jsonTable |USING json |OPTIONS ( - | path '${path.toString}' + | path '${path.toURI}' |) AS |SELECT a, b FROM jt """.stripMargin) @@ -127,7 +127,7 @@ class CreateTableAsSelectSuite |CREATE TABLE IF NOT EXISTS jsonTable |USING json |OPTIONS ( - | path '${path.toString}' + | path '${path.toURI}' |) AS |SELECT a * 4 FROM jt """.stripMargin) @@ -145,7 +145,7 @@ class CreateTableAsSelectSuite |CREATE TABLE jsonTable |USING json |OPTIONS ( - | path '${path.toString}' + | path '${path.toURI}' |) AS |SELECT b FROM jt """.stripMargin) @@ -162,7 +162,7 @@ class CreateTableAsSelectSuite sql( s""" |CREATE TEMPORARY TABLE t USING PARQUET - |OPTIONS (PATH '${path.toString}') + |OPTIONS (PATH '${path.toURI}') |PARTITIONED BY (a) |AS SELECT 1 AS a, 2 AS b """.stripMargin @@ -179,7 +179,7 @@ class CreateTableAsSelectSuite sql( s""" |CREATE EXTERNAL TABLE t USING PARQUET - |OPTIONS (PATH '${path.toString}') + |OPTIONS (PATH '${path.toURI}') |AS SELECT 1 AS a, 2 AS b """.stripMargin ) @@ -196,7 +196,7 @@ class CreateTableAsSelectSuite sql( s""" |CREATE TABLE t USING PARQUET - |OPTIONS (PATH '${path.toString}') + |OPTIONS (PATH '${path.toURI}') |PARTITIONED BY (a) |AS SELECT 1 AS a, 2 AS b """.stripMargin @@ -212,7 +212,7 @@ class CreateTableAsSelectSuite sql( s""" |CREATE TABLE t USING PARQUET - |OPTIONS (PATH '${path.toString}') + |OPTIONS (PATH '${path.toURI}') |CLUSTERED BY (a) SORTED BY (b) INTO 5 BUCKETS |AS SELECT 1 AS a, 2 AS b """.stripMargin @@ -228,7 +228,7 @@ class CreateTableAsSelectSuite sql( s""" |CREATE TABLE t USING PARQUET - |OPTIONS (PATH '${path.toString}') + |OPTIONS (PATH '${path.toURI}') |CLUSTERED BY (a) SORTED BY (b) INTO 0 BUCKETS |AS SELECT 1 AS a, 2 AS b """.stripMargin diff --git a/sql/core/src/test/scala/org/apache/spark/sql/streaming/test/DataStreamReaderWriterSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/streaming/test/DataStreamReaderWriterSuite.scala index 097dd6e367..30a957ef81 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/streaming/test/DataStreamReaderWriterSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/streaming/test/DataStreamReaderWriterSuite.scala @@ -22,6 +22,7 @@ import java.util.concurrent.TimeUnit import scala.concurrent.duration._ +import org.apache.hadoop.fs.Path import org.mockito.Mockito._ import org.scalatest.{BeforeAndAfter, PrivateMethodTester} import org.scalatest.PrivateMethodTester.PrivateMethod @@ -355,7 +356,7 @@ class DataStreamReaderWriterSuite extends StreamTest with BeforeAndAfter with Pr test("source metadataPath") { LastOptions.clear() - val checkpointLocation = newMetadataDir + val checkpointLocationURI = new Path(newMetadataDir).toUri val df1 = spark.readStream .format("org.apache.spark.sql.streaming.test") @@ -367,21 +368,21 @@ class DataStreamReaderWriterSuite extends StreamTest with BeforeAndAfter with Pr val q = df1.union(df2).writeStream .format("org.apache.spark.sql.streaming.test") - .option("checkpointLocation", checkpointLocation) + .option("checkpointLocation", checkpointLocationURI.toString) .trigger(ProcessingTime(10.seconds)) .start() q.stop() verify(LastOptions.mockStreamSourceProvider).createSource( spark.sqlContext, - checkpointLocation + "/sources/0", + s"$checkpointLocationURI/sources/0", None, "org.apache.spark.sql.streaming.test", Map.empty) verify(LastOptions.mockStreamSourceProvider).createSource( spark.sqlContext, - checkpointLocation + "/sources/1", + s"$checkpointLocationURI/sources/1", None, "org.apache.spark.sql.streaming.test", Map.empty) |