aboutsummaryrefslogtreecommitdiff
path: root/R/pkg
diff options
context:
space:
mode:
authorhyukjinkwon <gurwls223@gmail.com>2016-11-01 22:14:53 -0700
committerFelix Cheung <felixcheung@apache.org>2016-11-01 22:14:53 -0700
commit1ecfafa0869cb3a3e367bda8be252a69874dc4de (patch)
treef24f5ae60f65310bde6978590726acfdc16dad1d /R/pkg
parentad4832a9faf2c0c869bbcad9d71afe1cecbd3ec8 (diff)
downloadspark-1ecfafa0869cb3a3e367bda8be252a69874dc4de.tar.gz
spark-1ecfafa0869cb3a3e367bda8be252a69874dc4de.tar.bz2
spark-1ecfafa0869cb3a3e367bda8be252a69874dc4de.zip
[SPARK-17838][SPARKR] Check named arguments for options and use formatted R friendly message from JVM exception message
## What changes were proposed in this pull request? This PR proposes to - improve the R-friendly error messages rather than raw JVM exception one. As `read.json`, `read.text`, `read.orc`, `read.parquet` and `read.jdbc` are executed in the same path with `read.df`, and `write.json`, `write.text`, `write.orc`, `write.parquet` and `write.jdbc` shares the same path with `write.df`, it seems it is safe to call `handledCallJMethod` to handle JVM messages. - prevent `zero-length variable name` and prints the ignored options as an warning message. **Before** ``` r > read.json("path", a = 1, 2, 3, "a") Error in env[[name]] <- value : zero-length variable name ``` ``` r > read.json("arbitrary_path") Error in invokeJava(isStatic = FALSE, objId$id, methodName, ...) : org.apache.spark.sql.AnalysisException: Path does not exist: file:/...; at org.apache.spark.sql.execution.datasources.DataSource$$anonfun$12.apply(DataSource.scala:398) ... > read.orc("arbitrary_path") Error in invokeJava(isStatic = FALSE, objId$id, methodName, ...) : org.apache.spark.sql.AnalysisException: Path does not exist: file:/...; at org.apache.spark.sql.execution.datasources.DataSource$$anonfun$12.apply(DataSource.scala:398) ... > read.text("arbitrary_path") Error in invokeJava(isStatic = FALSE, objId$id, methodName, ...) : org.apache.spark.sql.AnalysisException: Path does not exist: file:/...; at org.apache.spark.sql.execution.datasources.DataSource$$anonfun$12.apply(DataSource.scala:398) ... > read.parquet("arbitrary_path") Error in invokeJava(isStatic = FALSE, objId$id, methodName, ...) : org.apache.spark.sql.AnalysisException: Path does not exist: file:/...; at org.apache.spark.sql.execution.datasources.DataSource$$anonfun$12.apply(DataSource.scala:398) ... ``` ``` r > write.json(df, "existing_path") Error in invokeJava(isStatic = FALSE, objId$id, methodName, ...) : org.apache.spark.sql.AnalysisException: path file:/... already exists.; at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand.run(InsertIntoHadoopFsRelationCommand.scala:68) > write.orc(df, "existing_path") Error in invokeJava(isStatic = FALSE, objId$id, methodName, ...) : org.apache.spark.sql.AnalysisException: path file:/... already exists.; at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand.run(InsertIntoHadoopFsRelationCommand.scala:68) > write.text(df, "existing_path") Error in invokeJava(isStatic = FALSE, objId$id, methodName, ...) : org.apache.spark.sql.AnalysisException: path file:/... already exists.; at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand.run(InsertIntoHadoopFsRelationCommand.scala:68) > write.parquet(df, "existing_path") Error in invokeJava(isStatic = FALSE, objId$id, methodName, ...) : org.apache.spark.sql.AnalysisException: path file:/... already exists.; at org.apache.spark.sql.execution.datasources.InsertIntoHadoopFsRelationCommand.run(InsertIntoHadoopFsRelationCommand.scala:68) ``` **After** ``` r read.json("arbitrary_path", a = 1, 2, 3, "a") Unnamed arguments ignored: 2, 3, a. ``` ``` r > read.json("arbitrary_path") Error in json : analysis error - Path does not exist: file:/... > read.orc("arbitrary_path") Error in orc : analysis error - Path does not exist: file:/... > read.text("arbitrary_path") Error in text : analysis error - Path does not exist: file:/... > read.parquet("arbitrary_path") Error in parquet : analysis error - Path does not exist: file:/... ``` ``` r > write.json(df, "existing_path") Error in json : analysis error - path file:/... already exists.; > write.orc(df, "existing_path") Error in orc : analysis error - path file:/... already exists.; > write.text(df, "existing_path") Error in text : analysis error - path file:/... already exists.; > write.parquet(df, "existing_path") Error in parquet : analysis error - path file:/... already exists.; ``` ## How was this patch tested? Unit tests in `test_utils.R` and `test_sparkSQL.R`. Author: hyukjinkwon <gurwls223@gmail.com> Closes #15608 from HyukjinKwon/SPARK-17838.
Diffstat (limited to 'R/pkg')
-rw-r--r--R/pkg/R/DataFrame.R10
-rw-r--r--R/pkg/R/SQLContext.R17
-rw-r--r--R/pkg/R/utils.R44
-rw-r--r--R/pkg/inst/tests/testthat/test_sparkSQL.R16
-rw-r--r--R/pkg/inst/tests/testthat/test_utils.R2
5 files changed, 64 insertions, 25 deletions
diff --git a/R/pkg/R/DataFrame.R b/R/pkg/R/DataFrame.R
index 1df8bbf9fe..1cf9b38ea6 100644
--- a/R/pkg/R/DataFrame.R
+++ b/R/pkg/R/DataFrame.R
@@ -788,7 +788,7 @@ setMethod("write.json",
function(x, path, mode = "error", ...) {
write <- callJMethod(x@sdf, "write")
write <- setWriteOptions(write, mode = mode, ...)
- invisible(callJMethod(write, "json", path))
+ invisible(handledCallJMethod(write, "json", path))
})
#' Save the contents of SparkDataFrame as an ORC file, preserving the schema.
@@ -819,7 +819,7 @@ setMethod("write.orc",
function(x, path, mode = "error", ...) {
write <- callJMethod(x@sdf, "write")
write <- setWriteOptions(write, mode = mode, ...)
- invisible(callJMethod(write, "orc", path))
+ invisible(handledCallJMethod(write, "orc", path))
})
#' Save the contents of SparkDataFrame as a Parquet file, preserving the schema.
@@ -851,7 +851,7 @@ setMethod("write.parquet",
function(x, path, mode = "error", ...) {
write <- callJMethod(x@sdf, "write")
write <- setWriteOptions(write, mode = mode, ...)
- invisible(callJMethod(write, "parquet", path))
+ invisible(handledCallJMethod(write, "parquet", path))
})
#' @rdname write.parquet
@@ -895,7 +895,7 @@ setMethod("write.text",
function(x, path, mode = "error", ...) {
write <- callJMethod(x@sdf, "write")
write <- setWriteOptions(write, mode = mode, ...)
- invisible(callJMethod(write, "text", path))
+ invisible(handledCallJMethod(write, "text", path))
})
#' Distinct
@@ -3342,7 +3342,7 @@ setMethod("write.jdbc",
jprops <- varargsToJProperties(...)
write <- callJMethod(x@sdf, "write")
write <- callJMethod(write, "mode", jmode)
- invisible(callJMethod(write, "jdbc", url, tableName, jprops))
+ invisible(handledCallJMethod(write, "jdbc", url, tableName, jprops))
})
#' randomSplit
diff --git a/R/pkg/R/SQLContext.R b/R/pkg/R/SQLContext.R
index 216ca51666..38d83c6e5c 100644
--- a/R/pkg/R/SQLContext.R
+++ b/R/pkg/R/SQLContext.R
@@ -350,7 +350,7 @@ read.json.default <- function(path, ...) {
paths <- as.list(suppressWarnings(normalizePath(path)))
read <- callJMethod(sparkSession, "read")
read <- callJMethod(read, "options", options)
- sdf <- callJMethod(read, "json", paths)
+ sdf <- handledCallJMethod(read, "json", paths)
dataFrame(sdf)
}
@@ -422,7 +422,7 @@ read.orc <- function(path, ...) {
path <- suppressWarnings(normalizePath(path))
read <- callJMethod(sparkSession, "read")
read <- callJMethod(read, "options", options)
- sdf <- callJMethod(read, "orc", path)
+ sdf <- handledCallJMethod(read, "orc", path)
dataFrame(sdf)
}
@@ -444,7 +444,7 @@ read.parquet.default <- function(path, ...) {
paths <- as.list(suppressWarnings(normalizePath(path)))
read <- callJMethod(sparkSession, "read")
read <- callJMethod(read, "options", options)
- sdf <- callJMethod(read, "parquet", paths)
+ sdf <- handledCallJMethod(read, "parquet", paths)
dataFrame(sdf)
}
@@ -496,7 +496,7 @@ read.text.default <- function(path, ...) {
paths <- as.list(suppressWarnings(normalizePath(path)))
read <- callJMethod(sparkSession, "read")
read <- callJMethod(read, "options", options)
- sdf <- callJMethod(read, "text", paths)
+ sdf <- handledCallJMethod(read, "text", paths)
dataFrame(sdf)
}
@@ -914,12 +914,13 @@ read.jdbc <- function(url, tableName,
} else {
numPartitions <- numToInt(numPartitions)
}
- sdf <- callJMethod(read, "jdbc", url, tableName, as.character(partitionColumn),
- numToInt(lowerBound), numToInt(upperBound), numPartitions, jprops)
+ sdf <- handledCallJMethod(read, "jdbc", url, tableName, as.character(partitionColumn),
+ numToInt(lowerBound), numToInt(upperBound), numPartitions, jprops)
} else if (length(predicates) > 0) {
- sdf <- callJMethod(read, "jdbc", url, tableName, as.list(as.character(predicates)), jprops)
+ sdf <- handledCallJMethod(read, "jdbc", url, tableName, as.list(as.character(predicates)),
+ jprops)
} else {
- sdf <- callJMethod(read, "jdbc", url, tableName, jprops)
+ sdf <- handledCallJMethod(read, "jdbc", url, tableName, jprops)
}
dataFrame(sdf)
}
diff --git a/R/pkg/R/utils.R b/R/pkg/R/utils.R
index c4e78cbb80..20004549cc 100644
--- a/R/pkg/R/utils.R
+++ b/R/pkg/R/utils.R
@@ -338,21 +338,41 @@ varargsToEnv <- function(...) {
# into string.
varargsToStrEnv <- function(...) {
pairs <- list(...)
+ nameList <- names(pairs)
env <- new.env()
- for (name in names(pairs)) {
- value <- pairs[[name]]
- if (!(is.logical(value) || is.numeric(value) || is.character(value) || is.null(value))) {
- stop(paste0("Unsupported type for ", name, " : ", class(value),
- ". Supported types are logical, numeric, character and NULL."))
- }
- if (is.logical(value)) {
- env[[name]] <- tolower(as.character(value))
- } else if (is.null(value)) {
- env[[name]] <- value
- } else {
- env[[name]] <- as.character(value)
+ ignoredNames <- list()
+
+ if (is.null(nameList)) {
+ # When all arguments are not named, names(..) returns NULL.
+ ignoredNames <- pairs
+ } else {
+ for (i in seq_along(pairs)) {
+ name <- nameList[i]
+ value <- pairs[i]
+ if (identical(name, "")) {
+ # When some of arguments are not named, name is "".
+ ignoredNames <- append(ignoredNames, value)
+ } else {
+ value <- pairs[[name]]
+ if (!(is.logical(value) || is.numeric(value) || is.character(value) || is.null(value))) {
+ stop(paste0("Unsupported type for ", name, " : ", class(value),
+ ". Supported types are logical, numeric, character and NULL."), call. = FALSE)
+ }
+ if (is.logical(value)) {
+ env[[name]] <- tolower(as.character(value))
+ } else if (is.null(value)) {
+ env[[name]] <- value
+ } else {
+ env[[name]] <- as.character(value)
+ }
+ }
}
}
+
+ if (length(ignoredNames) != 0) {
+ warning(paste0("Unnamed arguments ignored: ", paste(ignoredNames, collapse = ", "), "."),
+ call. = FALSE)
+ }
env
}
diff --git a/R/pkg/inst/tests/testthat/test_sparkSQL.R b/R/pkg/inst/tests/testthat/test_sparkSQL.R
index 9289db57b6..806019d752 100644
--- a/R/pkg/inst/tests/testthat/test_sparkSQL.R
+++ b/R/pkg/inst/tests/testthat/test_sparkSQL.R
@@ -2660,6 +2660,14 @@ test_that("Call DataFrameWriter.save() API in Java without path and check argume
# DataFrameWriter.save() without path.
expect_error(write.df(df, source = "csv"),
"Error in save : illegal argument - 'path' is not specified")
+ expect_error(write.json(df, jsonPath),
+ "Error in json : analysis error - path file:.*already exists")
+ expect_error(write.text(df, jsonPath),
+ "Error in text : analysis error - path file:.*already exists")
+ expect_error(write.orc(df, jsonPath),
+ "Error in orc : analysis error - path file:.*already exists")
+ expect_error(write.parquet(df, jsonPath),
+ "Error in parquet : analysis error - path file:.*already exists")
# Arguments checking in R side.
expect_error(write.df(df, "data.tmp", source = c(1, 2)),
@@ -2679,6 +2687,11 @@ test_that("Call DataFrameWriter.load() API in Java without path and check argume
paste("Error in loadDF : analysis error - Unable to infer schema for JSON at .",
"It must be specified manually"))
expect_error(read.df("arbitrary_path"), "Error in loadDF : analysis error - Path does not exist")
+ expect_error(read.json("arbitrary_path"), "Error in json : analysis error - Path does not exist")
+ expect_error(read.text("arbitrary_path"), "Error in text : analysis error - Path does not exist")
+ expect_error(read.orc("arbitrary_path"), "Error in orc : analysis error - Path does not exist")
+ expect_error(read.parquet("arbitrary_path"),
+ "Error in parquet : analysis error - Path does not exist")
# Arguments checking in R side.
expect_error(read.df(path = c(3)),
@@ -2686,6 +2699,9 @@ test_that("Call DataFrameWriter.load() API in Java without path and check argume
expect_error(read.df(jsonPath, source = c(1, 2)),
paste("source should be character, NULL or omitted. It is the datasource specified",
"in 'spark.sql.sources.default' configuration by default."))
+
+ expect_warning(read.json(jsonPath, a = 1, 2, 3, "a"),
+ "Unnamed arguments ignored: 2, 3, a.")
})
unlink(parquetPath)
diff --git a/R/pkg/inst/tests/testthat/test_utils.R b/R/pkg/inst/tests/testthat/test_utils.R
index a20254e9b3..607c407f04 100644
--- a/R/pkg/inst/tests/testthat/test_utils.R
+++ b/R/pkg/inst/tests/testthat/test_utils.R
@@ -224,6 +224,8 @@ test_that("varargsToStrEnv", {
expect_error(varargsToStrEnv(a = list(1, "a")),
paste0("Unsupported type for a : list. Supported types are logical, ",
"numeric, character and NULL."))
+ expect_warning(varargsToStrEnv(a = 1, 2, 3, 4), "Unnamed arguments ignored: 2, 3, 4.")
+ expect_warning(varargsToStrEnv(1, 2, 3, 4), "Unnamed arguments ignored: 1, 2, 3, 4.")
})
sparkR.session.stop()