diff options
author | Felix Cheung <felixcheung_m@hotmail.com> | 2016-06-26 13:10:43 -0700 |
---|---|---|
committer | Shivaram Venkataraman <shivaram@cs.berkeley.edu> | 2016-06-26 13:10:43 -0700 |
commit | 30b182bcc088aef161585211c517f473b9ee6632 (patch) | |
tree | 6681d46cb578fb214902a46109c0124565cac281 /R/pkg/inst/tests/testthat/test_sparkSQL.R | |
parent | e87741589a24821b5fe73e5d9ee2164247998580 (diff) | |
download | spark-30b182bcc088aef161585211c517f473b9ee6632.tar.gz spark-30b182bcc088aef161585211c517f473b9ee6632.tar.bz2 spark-30b182bcc088aef161585211c517f473b9ee6632.zip |
[SPARK-16184][SPARKR] conf API for SparkSession
## What changes were proposed in this pull request?
Add `conf` method to get Runtime Config from SparkSession
## How was this patch tested?
unit tests, manual tests
This is how it works in sparkR shell:
```
SparkSession available as 'spark'.
> conf()
$hive.metastore.warehouse.dir
[1] "file:/opt/spark-2.0.0-bin-hadoop2.6/R/spark-warehouse"
$spark.app.id
[1] "local-1466749575523"
$spark.app.name
[1] "SparkR"
$spark.driver.host
[1] "10.0.2.1"
$spark.driver.port
[1] "45629"
$spark.executorEnv.LD_LIBRARY_PATH
[1] "$LD_LIBRARY_PATH:/usr/lib/R/lib:/usr/lib/x86_64-linux-gnu:/usr/lib/jvm/default-java/jre/lib/amd64/server"
$spark.executor.id
[1] "driver"
$spark.home
[1] "/opt/spark-2.0.0-bin-hadoop2.6"
$spark.master
[1] "local[*]"
$spark.sql.catalogImplementation
[1] "hive"
$spark.submit.deployMode
[1] "client"
> conf("spark.master")
$spark.master
[1] "local[*]"
```
Author: Felix Cheung <felixcheung_m@hotmail.com>
Closes #13885 from felixcheung/rconf.
Diffstat (limited to 'R/pkg/inst/tests/testthat/test_sparkSQL.R')
-rw-r--r-- | R/pkg/inst/tests/testthat/test_sparkSQL.R | 12 |
1 files changed, 6 insertions, 6 deletions
diff --git a/R/pkg/inst/tests/testthat/test_sparkSQL.R b/R/pkg/inst/tests/testthat/test_sparkSQL.R index 9378c7afac..74def5ce42 100644 --- a/R/pkg/inst/tests/testthat/test_sparkSQL.R +++ b/R/pkg/inst/tests/testthat/test_sparkSQL.R @@ -2365,7 +2365,7 @@ test_that("randomSplit", { expect_true(all(sapply(abs(counts / num - weights / sum(weights)), function(e) { e < 0.05 }))) }) -test_that("Change config on SparkSession", { +test_that("Setting and getting config on SparkSession", { # first, set it to a random but known value conf <- callJMethod(sparkSession, "conf") property <- paste0("spark.testing.", as.character(runif(1))) @@ -2378,17 +2378,17 @@ test_that("Change config on SparkSession", { names(l) <- property sparkR.session(sparkConfig = l) - conf <- callJMethod(sparkSession, "conf") - newValue <- callJMethod(conf, "get", property, "") + newValue <- unlist(sparkR.conf(property, ""), use.names = FALSE) expect_equal(value2, newValue) value <- as.character(runif(1)) sparkR.session(spark.app.name = "sparkSession test", spark.testing.r.session.r = value) - conf <- callJMethod(sparkSession, "conf") - appNameValue <- callJMethod(conf, "get", "spark.app.name", "") - testValue <- callJMethod(conf, "get", "spark.testing.r.session.r", "") + allconf <- sparkR.conf() + appNameValue <- allconf[["spark.app.name"]] + testValue <- allconf[["spark.testing.r.session.r"]] expect_equal(appNameValue, "sparkSession test") expect_equal(testValue, value) + expect_error(sparkR.conf("completely.dummy"), "Config 'completely.dummy' is not set") }) test_that("enableHiveSupport on SparkSession", { |