From d17e5f2f123eecd5a7a1d87f5ce75a0fc44552b4 Mon Sep 17 00:00:00 2001 From: Dongjoon Hyun Date: Fri, 1 Jul 2016 15:35:19 -0700 Subject: [SPARK-16233][R][TEST] ORC test should be enabled only when HiveContext is available. ## What changes were proposed in this pull request? ORC test should be enabled only when HiveContext is available. ## How was this patch tested? Manual. ``` $ R/run-tests.sh ... 1. create DataFrame from RDD (test_sparkSQL.R#200) - Hive is not build with SparkSQL, skipped 2. test HiveContext (test_sparkSQL.R#1021) - Hive is not build with SparkSQL, skipped 3. read/write ORC files (test_sparkSQL.R#1728) - Hive is not build with SparkSQL, skipped 4. enableHiveSupport on SparkSession (test_sparkSQL.R#2448) - Hive is not build with SparkSQL, skipped 5. sparkJars tag in SparkContext (test_Windows.R#21) - This test is only for Windows, skipped DONE =========================================================================== Tests passed. ``` Author: Dongjoon Hyun Closes #14019 from dongjoon-hyun/SPARK-16233. --- R/pkg/inst/tests/testthat/test_sparkSQL.R | 2 ++ 1 file changed, 2 insertions(+) (limited to 'R') diff --git a/R/pkg/inst/tests/testthat/test_sparkSQL.R b/R/pkg/inst/tests/testthat/test_sparkSQL.R index 3b8d5707f4..a3aa26d9e7 100644 --- a/R/pkg/inst/tests/testthat/test_sparkSQL.R +++ b/R/pkg/inst/tests/testthat/test_sparkSQL.R @@ -1725,6 +1725,7 @@ test_that("mutate(), transform(), rename() and names()", { }) test_that("read/write ORC files", { + setHiveContext(sc) df <- read.df(jsonPath, "json") # Test write.df and read.df @@ -1741,6 +1742,7 @@ test_that("read/write ORC files", { expect_equal(count(orcDF), count(df)) unlink(orcPath2) + unsetHiveContext() }) test_that("read/write Parquet files", { -- cgit v1.2.3