aboutsummaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorDongjoon Hyun <dongjoon@apache.org>2016-06-20 14:24:41 -0700
committerShivaram Venkataraman <shivaram@cs.berkeley.edu>2016-06-20 14:24:41 -0700
commitd0eddb80eca04e4f5f8af3b5143096cf67200277 (patch)
tree9b6a8eadd22124981344ba2cd934bcd252a0d4e1
parent92514232e52af0f5f0413ed97b9571b1b9daaa90 (diff)
downloadspark-d0eddb80eca04e4f5f8af3b5143096cf67200277.tar.gz
spark-d0eddb80eca04e4f5f8af3b5143096cf67200277.tar.bz2
spark-d0eddb80eca04e4f5f8af3b5143096cf67200277.zip
[SPARK-14995][R] Add `since` tag in Roxygen documentation for SparkR API methods
## What changes were proposed in this pull request? This PR adds `since` tags to Roxygen documentation according to the previous documentation archive. https://home.apache.org/~dongjoon/spark-2.0.0-docs/api/R/ ## How was this patch tested? Manual. Author: Dongjoon Hyun <dongjoon@apache.org> Closes #13734 from dongjoon-hyun/SPARK-14995.
-rw-r--r--R/pkg/R/DataFrame.R93
-rw-r--r--R/pkg/R/SQLContext.R42
-rw-r--r--R/pkg/R/WindowSpec.R8
-rw-r--r--R/pkg/R/column.R10
-rw-r--r--R/pkg/R/context.R3
-rw-r--r--R/pkg/R/functions.R153
-rw-r--r--R/pkg/R/group.R6
-rw-r--r--R/pkg/R/jobj.R1
-rw-r--r--R/pkg/R/mllib.R24
-rw-r--r--R/pkg/R/schema.R5
-rw-r--r--R/pkg/R/sparkR.R18
-rw-r--r--R/pkg/R/stats.R6
-rw-r--r--R/pkg/R/utils.R1
-rw-r--r--R/pkg/R/window.R4
14 files changed, 340 insertions, 34 deletions
diff --git a/R/pkg/R/DataFrame.R b/R/pkg/R/DataFrame.R
index 583d3aefd4..ecdcd6e29f 100644
--- a/R/pkg/R/DataFrame.R
+++ b/R/pkg/R/DataFrame.R
@@ -25,7 +25,7 @@ setOldClass("structType")
#' S4 class that represents a SparkDataFrame
#'
-#' DataFrames can be created using functions like \link{createDataFrame},
+#' SparkDataFrames can be created using functions like \link{createDataFrame},
#' \link{read.json}, \link{table} etc.
#'
#' @family SparkDataFrame functions
@@ -42,6 +42,7 @@ setOldClass("structType")
#' sparkR.session()
#' df <- createDataFrame(faithful)
#'}
+#' @note SparkDataFrame since 2.0.0
setClass("SparkDataFrame",
slots = list(env = "environment",
sdf = "jobj"))
@@ -81,6 +82,7 @@ dataFrame <- function(sdf, isCached = FALSE) {
#' df <- read.json(path)
#' printSchema(df)
#'}
+#' @note printSchema since 1.4.0
setMethod("printSchema",
signature(x = "SparkDataFrame"),
function(x) {
@@ -105,6 +107,7 @@ setMethod("printSchema",
#' df <- read.json(path)
#' dfSchema <- schema(df)
#'}
+#' @note schema since 1.4.0
setMethod("schema",
signature(x = "SparkDataFrame"),
function(x) {
@@ -128,6 +131,7 @@ setMethod("schema",
#' df <- read.json(path)
#' explain(df, TRUE)
#'}
+#' @note explain since 1.4.0
setMethod("explain",
signature(x = "SparkDataFrame"),
function(x, extended = FALSE) {
@@ -158,6 +162,7 @@ setMethod("explain",
#' df <- read.json(path)
#' isLocal(df)
#'}
+#' @note isLocal since 1.4.0
setMethod("isLocal",
signature(x = "SparkDataFrame"),
function(x) {
@@ -182,6 +187,7 @@ setMethod("isLocal",
#' df <- read.json(path)
#' showDF(df)
#'}
+#' @note showDF since 1.4.0
setMethod("showDF",
signature(x = "SparkDataFrame"),
function(x, numRows = 20, truncate = TRUE) {
@@ -206,6 +212,7 @@ setMethod("showDF",
#' df <- read.json(path)
#' df
#'}
+#' @note show(SparkDataFrame) since 1.4.0
setMethod("show", "SparkDataFrame",
function(object) {
cols <- lapply(dtypes(object), function(l) {
@@ -232,6 +239,7 @@ setMethod("show", "SparkDataFrame",
#' df <- read.json(path)
#' dtypes(df)
#'}
+#' @note dtypes since 1.4.0
setMethod("dtypes",
signature(x = "SparkDataFrame"),
function(x) {
@@ -259,6 +267,7 @@ setMethod("dtypes",
#' columns(df)
#' colnames(df)
#'}
+#' @note columns since 1.4.0
setMethod("columns",
signature(x = "SparkDataFrame"),
function(x) {
@@ -269,6 +278,7 @@ setMethod("columns",
#' @rdname columns
#' @name names
+#' @note names since 1.5.0
setMethod("names",
signature(x = "SparkDataFrame"),
function(x) {
@@ -277,6 +287,7 @@ setMethod("names",
#' @rdname columns
#' @name names<-
+#' @note names<- since 1.5.0
setMethod("names<-",
signature(x = "SparkDataFrame"),
function(x, value) {
@@ -288,6 +299,7 @@ setMethod("names<-",
#' @rdname columns
#' @name colnames
+#' @note colnames since 1.6.0
setMethod("colnames",
signature(x = "SparkDataFrame"),
function(x) {
@@ -296,6 +308,7 @@ setMethod("colnames",
#' @rdname columns
#' @name colnames<-
+#' @note colnames<- since 1.6.0
setMethod("colnames<-",
signature(x = "SparkDataFrame"),
function(x, value) {
@@ -338,6 +351,7 @@ setMethod("colnames<-",
#' irisDF <- createDataFrame(iris)
#' coltypes(irisDF)
#'}
+#' @note coltypes since 1.6.0
setMethod("coltypes",
signature(x = "SparkDataFrame"),
function(x) {
@@ -393,6 +407,7 @@ setMethod("coltypes",
#' coltypes(df) <- c("character", "integer")
#' coltypes(df) <- c(NA, "numeric")
#'}
+#' @note coltypes<- since 1.6.0
setMethod("coltypes<-",
signature(x = "SparkDataFrame", value = "character"),
function(x, value) {
@@ -440,6 +455,7 @@ setMethod("coltypes<-",
#' createOrReplaceTempView(df, "json_df")
#' new_df <- sql("SELECT * FROM json_df")
#'}
+#' @note createOrReplaceTempView since 2.0.0
setMethod("createOrReplaceTempView",
signature(x = "SparkDataFrame", viewName = "character"),
function(x, viewName) {
@@ -464,6 +480,7 @@ setMethod("createOrReplaceTempView",
#' registerTempTable(df, "json_df")
#' new_df <- sql("SELECT * FROM json_df")
#'}
+#' @note registerTempTable since 1.4.0
setMethod("registerTempTable",
signature(x = "SparkDataFrame", tableName = "character"),
function(x, tableName) {
@@ -492,6 +509,7 @@ setMethod("registerTempTable",
#' createOrReplaceTempView(df, "table1")
#' insertInto(df2, "table1", overwrite = TRUE)
#'}
+#' @note insertInto since 1.4.0
setMethod("insertInto",
signature(x = "SparkDataFrame", tableName = "character"),
function(x, tableName, overwrite = FALSE) {
@@ -518,6 +536,7 @@ setMethod("insertInto",
#' df <- read.json(path)
#' cache(df)
#'}
+#' @note cache since 1.4.0
setMethod("cache",
signature(x = "SparkDataFrame"),
function(x) {
@@ -545,6 +564,7 @@ setMethod("cache",
#' df <- read.json(path)
#' persist(df, "MEMORY_AND_DISK")
#'}
+#' @note persist since 1.4.0
setMethod("persist",
signature(x = "SparkDataFrame", newLevel = "character"),
function(x, newLevel) {
@@ -573,6 +593,7 @@ setMethod("persist",
#' persist(df, "MEMORY_AND_DISK")
#' unpersist(df)
#'}
+#' @note unpersist since 1.4.0
setMethod("unpersist",
signature(x = "SparkDataFrame"),
function(x, blocking = TRUE) {
@@ -609,6 +630,7 @@ setMethod("unpersist",
#' newDF <- repartition(df, col = df$"col1", df$"col2")
#' newDF <- repartition(df, 3L, col = df$"col1", df$"col2")
#'}
+#' @note repartition since 1.4.0
setMethod("repartition",
signature(x = "SparkDataFrame"),
function(x, numPartitions = NULL, col = NULL, ...) {
@@ -675,6 +697,7 @@ setMethod("toJSON",
#' df <- read.json(path)
#' write.json(df, "/tmp/sparkr-tmp/")
#'}
+#' @note write.json since 1.6.0
setMethod("write.json",
signature(x = "SparkDataFrame", path = "character"),
function(x, path) {
@@ -729,6 +752,7 @@ setMethod("write.orc",
#' write.parquet(df, "/tmp/sparkr-tmp1/")
#' saveAsParquetFile(df, "/tmp/sparkr-tmp2/")
#'}
+#' @note write.parquet since 1.6.0
setMethod("write.parquet",
signature(x = "SparkDataFrame", path = "character"),
function(x, path) {
@@ -739,6 +763,7 @@ setMethod("write.parquet",
#' @rdname write.parquet
#' @name saveAsParquetFile
#' @export
+#' @note saveAsParquetFile since 1.4.0
setMethod("saveAsParquetFile",
signature(x = "SparkDataFrame", path = "character"),
function(x, path) {
@@ -766,6 +791,7 @@ setMethod("saveAsParquetFile",
#' df <- read.text(path)
#' write.text(df, "/tmp/sparkr-tmp/")
#'}
+#' @note write.text since 2.0.0
setMethod("write.text",
signature(x = "SparkDataFrame", path = "character"),
function(x, path) {
@@ -790,6 +816,7 @@ setMethod("write.text",
#' df <- read.json(path)
#' distinctDF <- distinct(df)
#'}
+#' @note distinct since 1.4.0
setMethod("distinct",
signature(x = "SparkDataFrame"),
function(x) {
@@ -799,6 +826,7 @@ setMethod("distinct",
#' @rdname distinct
#' @name unique
+#' @note unique since 1.5.0
setMethod("unique",
signature(x = "SparkDataFrame"),
function(x) {
@@ -826,6 +854,7 @@ setMethod("unique",
#' collect(sample(df, FALSE, 0.5))
#' collect(sample(df, TRUE, 0.5))
#'}
+#' @note sample since 1.4.0
setMethod("sample",
signature(x = "SparkDataFrame", withReplacement = "logical",
fraction = "numeric"),
@@ -843,6 +872,7 @@ setMethod("sample",
#' @rdname sample
#' @name sample_frac
+#' @note sample_frac since 1.4.0
setMethod("sample_frac",
signature(x = "SparkDataFrame", withReplacement = "logical",
fraction = "numeric"),
@@ -865,6 +895,7 @@ setMethod("sample_frac",
#' df <- read.json(path)
#' count(df)
#' }
+#' @note count since 1.4.0
setMethod("count",
signature(x = "SparkDataFrame"),
function(x) {
@@ -873,6 +904,7 @@ setMethod("count",
#' @name nrow
#' @rdname nrow
+#' @note nrow since 1.5.0
setMethod("nrow",
signature(x = "SparkDataFrame"),
function(x) {
@@ -894,6 +926,7 @@ setMethod("nrow",
#' df <- read.json(path)
#' ncol(df)
#' }
+#' @note ncol since 1.5.0
setMethod("ncol",
signature(x = "SparkDataFrame"),
function(x) {
@@ -916,6 +949,7 @@ setMethod("ncol",
#' df <- read.json(path)
#' dim(df)
#' }
+#' @note dim since 1.5.0
setMethod("dim",
signature(x = "SparkDataFrame"),
function(x) {
@@ -940,6 +974,7 @@ setMethod("dim",
#' collected <- collect(df)
#' firstName <- collected[[1]]$name
#' }
+#' @note collect since 1.4.0
setMethod("collect",
signature(x = "SparkDataFrame"),
function(x, stringsAsFactors = FALSE) {
@@ -1010,6 +1045,7 @@ setMethod("collect",
#' df <- read.json(path)
#' limitedDF <- limit(df, 10)
#' }
+#' @note limit since 1.4.0
setMethod("limit",
signature(x = "SparkDataFrame", num = "numeric"),
function(x, num) {
@@ -1030,6 +1066,7 @@ setMethod("limit",
#' df <- read.json(path)
#' take(df, 2)
#' }
+#' @note take since 1.4.0
setMethod("take",
signature(x = "SparkDataFrame", num = "numeric"),
function(x, num) {
@@ -1058,6 +1095,7 @@ setMethod("take",
#' df <- read.json(path)
#' head(df)
#' }
+#' @note head since 1.4.0
setMethod("head",
signature(x = "SparkDataFrame"),
function(x, num = 6L) {
@@ -1080,6 +1118,7 @@ setMethod("head",
#' df <- read.json(path)
#' first(df)
#' }
+#' @note first(SparkDataFrame) since 1.4.0
setMethod("first",
signature(x = "SparkDataFrame"),
function(x) {
@@ -1131,6 +1170,7 @@ setMethod("toRDD",
#' # Compute the max age and average salary, grouped by department and gender.
#' agg(groupBy(df, "department", "gender"), salary="avg", "age" -> "max")
#' }
+#' @note groupBy since 1.4.0
setMethod("groupBy",
signature(x = "SparkDataFrame"),
function(x, ...) {
@@ -1146,6 +1186,7 @@ setMethod("groupBy",
#' @rdname groupBy
#' @name group_by
+#' @note group_by since 1.4.0
setMethod("group_by",
signature(x = "SparkDataFrame"),
function(x, ...) {
@@ -1161,6 +1202,7 @@ setMethod("group_by",
#' @rdname agg
#' @name agg
#' @export
+#' @note agg since 1.4.0
setMethod("agg",
signature(x = "SparkDataFrame"),
function(x, ...) {
@@ -1169,6 +1211,7 @@ setMethod("agg",
#' @rdname agg
#' @name summarize
+#' @note summarize since 1.4.0
setMethod("summarize",
signature(x = "SparkDataFrame"),
function(x, ...) {
@@ -1233,6 +1276,7 @@ dapplyInternal <- function(x, func, schema) {
#' # 1 2 2 2 3
#' # 2 3 3 3 4
#' }
+#' @note dapply since 2.0.0
setMethod("dapply",
signature(x = "SparkDataFrame", func = "function", schema = "structType"),
function(x, func, schema) {
@@ -1273,6 +1317,7 @@ setMethod("dapply",
#' # 2 2 2 3
#' # 3 3 3 4
#' }
+#' @note dapplyCollect since 2.0.0
setMethod("dapplyCollect",
signature(x = "SparkDataFrame", func = "function"),
function(x, func) {
@@ -1361,6 +1406,7 @@ setMethod("dapplyCollect",
#' 3 2.351890 0.6548350 0.2375602 0.2521257
#'
#'}
+#' @note gapply(SparkDataFrame) since 2.0.0
setMethod("gapply",
signature(x = "SparkDataFrame"),
function(x, cols, func, schema) {
@@ -1444,6 +1490,7 @@ getColumn <- function(x, c) {
#' @rdname select
#' @name $
+#' @note $ since 1.4.0
setMethod("$", signature(x = "SparkDataFrame"),
function(x, name) {
getColumn(x, name)
@@ -1451,6 +1498,7 @@ setMethod("$", signature(x = "SparkDataFrame"),
#' @rdname select
#' @name $<-
+#' @note $<- since 1.4.0
setMethod("$<-", signature(x = "SparkDataFrame"),
function(x, name, value) {
stopifnot(class(value) == "Column" || is.null(value))
@@ -1468,6 +1516,7 @@ setClassUnion("numericOrcharacter", c("numeric", "character"))
#' @rdname subset
#' @name [[
+#' @note [[ since 1.4.0
setMethod("[[", signature(x = "SparkDataFrame", i = "numericOrcharacter"),
function(x, i) {
if (is.numeric(i)) {
@@ -1479,6 +1528,7 @@ setMethod("[[", signature(x = "SparkDataFrame", i = "numericOrcharacter"),
#' @rdname subset
#' @name [
+#' @note [ since 1.4.0
setMethod("[", signature(x = "SparkDataFrame"),
function(x, i, j, ..., drop = F) {
# Perform filtering first if needed
@@ -1543,6 +1593,7 @@ setMethod("[", signature(x = "SparkDataFrame"),
#' subset(df, df$age %in% c(19), select = c(1,2))
#' subset(df, select = c(1,2))
#' }
+#' @note subset since 1.5.0
setMethod("subset", signature(x = "SparkDataFrame"),
function(x, subset, select, drop = F, ...) {
if (missing(subset)) {
@@ -1573,6 +1624,7 @@ setMethod("subset", signature(x = "SparkDataFrame"),
#' # Similar to R data frames columns can also be selected using `$`
#' df[,df$age]
#' }
+#' @note select(SparkDataFrame, character) since 1.4.0
setMethod("select", signature(x = "SparkDataFrame", col = "character"),
function(x, col, ...) {
if (length(col) > 1) {
@@ -1590,6 +1642,7 @@ setMethod("select", signature(x = "SparkDataFrame", col = "character"),
#' @family SparkDataFrame functions
#' @rdname select
#' @export
+#' @note select(SparkDataFrame, Column) since 1.4.0
setMethod("select", signature(x = "SparkDataFrame", col = "Column"),
function(x, col, ...) {
jcols <- lapply(list(col, ...), function(c) {
@@ -1602,6 +1655,7 @@ setMethod("select", signature(x = "SparkDataFrame", col = "Column"),
#' @family SparkDataFrame functions
#' @rdname select
#' @export
+#' @note select(SparkDataFrame, list) since 1.4.0
setMethod("select",
signature(x = "SparkDataFrame", col = "list"),
function(x, col) {
@@ -1635,6 +1689,7 @@ setMethod("select",
#' df <- read.json(path)
#' selectExpr(df, "col1", "(col2 * 5) as newCol")
#' }
+#' @note selectExpr since 1.4.0
setMethod("selectExpr",
signature(x = "SparkDataFrame", expr = "character"),
function(x, expr, ...) {
@@ -1666,6 +1721,7 @@ setMethod("selectExpr",
#' # Replace an existing column
#' newDF2 <- withColumn(newDF, "newCol", newDF$col1)
#' }
+#' @note withColumn since 1.4.0
setMethod("withColumn",
signature(x = "SparkDataFrame", colName = "character", col = "Column"),
function(x, colName, col) {
@@ -1698,6 +1754,7 @@ setMethod("withColumn",
#' # Replace the "age" column
#' df1 <- mutate(df, age = df$age + 1L)
#' }
+#' @note mutate since 1.4.0
setMethod("mutate",
signature(.data = "SparkDataFrame"),
function(.data, ...) {
@@ -1756,6 +1813,7 @@ setMethod("mutate",
#' @export
#' @rdname mutate
#' @name transform
+#' @note transform since 1.5.0
setMethod("transform",
signature(`_data` = "SparkDataFrame"),
function(`_data`, ...) {
@@ -1782,6 +1840,7 @@ setMethod("transform",
#' df <- read.json(path)
#' newDF <- withColumnRenamed(df, "col1", "newCol1")
#' }
+#' @note withColumnRenamed since 1.4.0
setMethod("withColumnRenamed",
signature(x = "SparkDataFrame", existingCol = "character", newCol = "character"),
function(x, existingCol, newCol) {
@@ -1806,6 +1865,7 @@ setMethod("withColumnRenamed",
#' df <- read.json(path)
#' newDF <- rename(df, col1 = df$newCol1)
#' }
+#' @note rename since 1.4.0
setMethod("rename",
signature(x = "SparkDataFrame"),
function(x, ...) {
@@ -1852,6 +1912,7 @@ setClassUnion("characterOrColumn", c("character", "Column"))
#' arrange(df, "col1", decreasing = TRUE)
#' arrange(df, "col1", "col2", decreasing = c(TRUE, FALSE))
#' }
+#' @note arrange(SparkDataFrame, Column) since 1.4.0
setMethod("arrange",
signature(x = "SparkDataFrame", col = "Column"),
function(x, col, ...) {
@@ -1866,6 +1927,7 @@ setMethod("arrange",
#' @rdname arrange
#' @name arrange
#' @export
+#' @note arrange(SparkDataFrame, character) since 1.4.0
setMethod("arrange",
signature(x = "SparkDataFrame", col = "character"),
function(x, col, ..., decreasing = FALSE) {
@@ -1898,6 +1960,7 @@ setMethod("arrange",
#' @rdname arrange
#' @name orderBy
#' @export
+#' @note orderBy(SparkDataFrame, characterOrColumn) since 1.4.0
setMethod("orderBy",
signature(x = "SparkDataFrame", col = "characterOrColumn"),
function(x, col, ...) {
@@ -1925,6 +1988,7 @@ setMethod("orderBy",
#' filter(df, "col1 > 0")
#' filter(df, df$col2 != "abcdefg")
#' }
+#' @note filter since 1.4.0
setMethod("filter",
signature(x = "SparkDataFrame", condition = "characterOrColumn"),
function(x, condition) {
@@ -1938,6 +2002,7 @@ setMethod("filter",
#' @family SparkDataFrame functions
#' @rdname filter
#' @name where
+#' @note where since 1.4.0
setMethod("where",
signature(x = "SparkDataFrame", condition = "characterOrColumn"),
function(x, condition) {
@@ -1966,6 +2031,7 @@ setMethod("where",
#' dropDuplicates(df, "col1", "col2")
#' dropDuplicates(df, c("col1", "col2"))
#' }
+#' @note dropDuplicates since 2.0.0
setMethod("dropDuplicates",
signature(x = "SparkDataFrame"),
function(x, ...) {
@@ -2012,6 +2078,7 @@ setMethod("dropDuplicates",
#' join(df1, df2, df1$col1 == df2$col2) # Performs an inner join based on expression
#' join(df1, df2, df1$col1 == df2$col2, "right_outer")
#' }
+#' @note join since 1.4.0
setMethod("join",
signature(x = "SparkDataFrame", y = "SparkDataFrame"),
function(x, y, joinExpr = NULL, joinType = NULL) {
@@ -2073,6 +2140,7 @@ setMethod("join",
#' merge(df1, df2, by.x = "col1", by.y = "col2", all = TRUE, sort = FALSE)
#' merge(df1, df2, by = "col1", all = TRUE, suffixes = c("-X", "-Y"))
#' }
+#' @note merge since 1.5.0
setMethod("merge",
signature(x = "SparkDataFrame", y = "SparkDataFrame"),
function(x, y, by = intersect(names(x), names(y)), by.x = by, by.y = by,
@@ -2160,6 +2228,7 @@ setMethod("merge",
#' @param suffix a suffix for the column name
#' @return list of columns
#'
+#' @note generateAliasesForIntersectedCols since 1.6.0
generateAliasesForIntersectedCols <- function (x, intersectedColNames, suffix) {
allColNames <- names(x)
# sets alias for making colnames unique in dataframe 'x'
@@ -2198,6 +2267,7 @@ generateAliasesForIntersectedCols <- function (x, intersectedColNames, suffix) {
#' df2 <- read.json(path2)
#' unioned <- unionAll(df, df2)
#' }
+#' @note unionAll since 1.4.0
setMethod("unionAll",
signature(x = "SparkDataFrame", y = "SparkDataFrame"),
function(x, y) {
@@ -2212,6 +2282,7 @@ setMethod("unionAll",
#' @rdname rbind
#' @name rbind
#' @export
+#' @note rbind since 1.5.0
setMethod("rbind",
signature(... = "SparkDataFrame"),
function(x, ..., deparse.level = 1) {
@@ -2241,6 +2312,7 @@ setMethod("rbind",
#' df2 <- read.json(path2)
#' intersectDF <- intersect(df, df2)
#' }
+#' @note intersect since 1.4.0
setMethod("intersect",
signature(x = "SparkDataFrame", y = "SparkDataFrame"),
function(x, y) {
@@ -2269,6 +2341,7 @@ setMethod("intersect",
#' }
#' @rdname except
#' @export
+#' @note except since 1.4.0
setMethod("except",
signature(x = "SparkDataFrame", y = "SparkDataFrame"),
function(x, y) {
@@ -2310,6 +2383,7 @@ setMethod("except",
#' write.df(df, "myfile", "parquet", "overwrite")
#' saveDF(df, parquetPath2, "parquet", mode = saveMode, mergeSchema = mergeSchema)
#' }
+#' @note write.df since 1.4.0
setMethod("write.df",
signature(df = "SparkDataFrame", path = "character"),
function(df, path, source = NULL, mode = "error", ...){
@@ -2330,6 +2404,7 @@ setMethod("write.df",
#' @rdname write.df
#' @name saveDF
#' @export
+#' @note saveDF since 1.4.0
setMethod("saveDF",
signature(df = "SparkDataFrame", path = "character"),
function(df, path, source = NULL, mode = "error", ...){
@@ -2367,6 +2442,7 @@ setMethod("saveDF",
#' df <- read.json(path)
#' saveAsTable(df, "myfile")
#' }
+#' @note saveAsTable since 1.4.0
setMethod("saveAsTable",
signature(df = "SparkDataFrame", tableName = "character"),
function(df, tableName, source = NULL, mode="error", ...){
@@ -2405,6 +2481,7 @@ setMethod("saveAsTable",
#' describe(df, "col1")
#' describe(df, "col1", "col2")
#' }
+#' @note describe(SparkDataFrame, character) since 1.4.0
setMethod("describe",
signature(x = "SparkDataFrame", col = "character"),
function(x, col, ...) {
@@ -2415,6 +2492,7 @@ setMethod("describe",
#' @rdname summary
#' @name describe
+#' @note describe(SparkDataFrame) since 1.4.0
setMethod("describe",
signature(x = "SparkDataFrame"),
function(x) {
@@ -2425,6 +2503,7 @@ setMethod("describe",
#' @rdname summary
#' @name summary
+#' @note summary(SparkDataFrame) since 1.5.0
setMethod("summary",
signature(object = "SparkDataFrame"),
function(object, ...) {
@@ -2458,6 +2537,7 @@ setMethod("summary",
#' df <- read.json(path)
#' dropna(df)
#' }
+#' @note dropna since 1.4.0
setMethod("dropna",
signature(x = "SparkDataFrame"),
function(x, how = c("any", "all"), minNonNulls = NULL, cols = NULL) {
@@ -2478,6 +2558,7 @@ setMethod("dropna",
#' @rdname nafunctions
#' @name na.omit
#' @export
+#' @note na.omit since 1.5.0
setMethod("na.omit",
signature(object = "SparkDataFrame"),
function(object, how = c("any", "all"), minNonNulls = NULL, cols = NULL) {
@@ -2512,6 +2593,7 @@ setMethod("na.omit",
#' fillna(df, 1)
#' fillna(df, list("age" = 20, "name" = "unknown"))
#' }
+#' @note fillna since 1.4.0
setMethod("fillna",
signature(x = "SparkDataFrame"),
function(x, value, cols = NULL) {
@@ -2571,6 +2653,7 @@ setMethod("fillna",
#' irisDF <- createDataFrame(iris)
#' df <- as.data.frame(irisDF[irisDF$Species == "setosa", ])
#' }
+#' @note as.data.frame since 1.6.0
setMethod("as.data.frame",
signature(x = "SparkDataFrame"),
function(x, row.names = NULL, optional = FALSE, ...) {
@@ -2597,6 +2680,7 @@ setMethod("as.data.frame",
#' summary(Sepal_Width)
#' }
#' @seealso \link{detach}
+#' @note attach since 1.6.0
setMethod("attach",
signature(what = "SparkDataFrame"),
function(what, pos = 2, name = deparse(substitute(what)), warn.conflicts = TRUE) {
@@ -2622,6 +2706,7 @@ setMethod("attach",
#' with(irisDf, nrow(Sepal_Width))
#' }
#' @seealso \link{attach}
+#' @note with since 1.6.0
setMethod("with",
signature(data = "SparkDataFrame"),
function(data, expr, ...) {
@@ -2645,6 +2730,7 @@ setMethod("with",
#' # Show the structure of the SparkDataFrame
#' str(irisDF)
#' }
+#' @note str since 1.6.1
setMethod("str",
signature(object = "SparkDataFrame"),
function(object) {
@@ -2726,6 +2812,7 @@ setMethod("str",
#' drop(df, c("col1", "col2"))
#' drop(df, df$col1)
#' }
+#' @note drop since 2.0.0
setMethod("drop",
signature(x = "SparkDataFrame"),
function(x, col) {
@@ -2775,6 +2862,7 @@ setMethod("drop",
#' geom_bar(stat = "identity") +
#' xlab("Sepal_Length") + ylab("Frequency")
#' }
+#' @note histogram since 2.0.0
setMethod("histogram",
signature(df = "SparkDataFrame", col = "characterOrColumn"),
function(df, col, nbins = 10) {
@@ -2896,6 +2984,7 @@ setMethod("histogram",
#' jdbcUrl <- "jdbc:mysql://localhost:3306/databasename"
#' write.jdbc(df, jdbcUrl, "table", user = "username", password = "password")
#' }
+#' @note write.jdbc since 2.0.0
setMethod("write.jdbc",
signature(x = "SparkDataFrame", url = "character", tableName = "character"),
function(x, url, tableName, mode = "error", ...){
@@ -2926,7 +3015,7 @@ setMethod("write.jdbc",
#' # df_list contains 3 SparkDataFrames with each having about 200, 300 and 500 rows respectively
#' sapply(df_list, count)
#' }
-#' @note since 2.0.0
+#' @note randomSplit since 2.0.0
setMethod("randomSplit",
signature(x = "SparkDataFrame", weights = "numeric"),
function(x, weights, seed) {
diff --git a/R/pkg/R/SQLContext.R b/R/pkg/R/SQLContext.R
index b846b3dec2..8d2c4ac7ce 100644
--- a/R/pkg/R/SQLContext.R
+++ b/R/pkg/R/SQLContext.R
@@ -67,6 +67,7 @@ dispatchFunc <- function(newFuncSig, x, ...) {
}
#' return the SparkSession
+#' @note getSparkSession since 2.0.0
getSparkSession <- function() {
if (exists(".sparkRsession", envir = .sparkREnv)) {
get(".sparkRsession", envir = .sparkREnv)
@@ -76,6 +77,7 @@ getSparkSession <- function() {
}
#' infer the SQL type
+#' @note infer_type since 1.4.0
infer_type <- function(x) {
if (is.null(x)) {
stop("can not infer type from NULL")
@@ -133,7 +135,7 @@ getDefaultSqlSource <- function() {
#' }
#' @name createDataFrame
#' @method createDataFrame default
-
+#' @note createDataFrame since 1.4.0
# TODO(davies): support sampling and infer type from NA
createDataFrame.default <- function(data, schema = NULL, samplingRatio = 1.0) {
sparkSession <- getSparkSession()
@@ -218,7 +220,7 @@ createDataFrame <- function(x, ...) {
#' @aliases createDataFrame
#' @export
#' @method as.DataFrame default
-
+#' @note as.DataFrame since 1.6.0
as.DataFrame.default <- function(data, schema = NULL, samplingRatio = 1.0) {
createDataFrame(data, schema, samplingRatio)
}
@@ -266,7 +268,7 @@ setMethod("toDF", signature(x = "RDD"),
#' }
#' @name read.json
#' @method read.json default
-
+#' @note read.json since 1.6.0
read.json.default <- function(path) {
sparkSession <- getSparkSession()
# Allow the user to have a more flexible definiton of the text file path
@@ -284,7 +286,7 @@ read.json <- function(x, ...) {
#' @name jsonFile
#' @export
#' @method jsonFile default
-
+#' @note jsonFile since 1.4.0
jsonFile.default <- function(path) {
.Deprecated("read.json")
read.json(path)
@@ -355,7 +357,7 @@ read.orc <- function(path) {
#' @export
#' @name read.parquet
#' @method read.parquet default
-
+#' @note read.parquet since 1.6.0
read.parquet.default <- function(path) {
sparkSession <- getSparkSession()
# Allow the user to have a more flexible definiton of the Parquet file path
@@ -373,7 +375,7 @@ read.parquet <- function(x, ...) {
#' @name parquetFile
#' @export
#' @method parquetFile default
-
+#' @note parquetFile since 1.4.0
parquetFile.default <- function(...) {
.Deprecated("read.parquet")
read.parquet(unlist(list(...)))
@@ -403,7 +405,7 @@ parquetFile <- function(x, ...) {
#' }
#' @name read.text
#' @method read.text default
-
+#' @note read.text since 1.6.1
read.text.default <- function(path) {
sparkSession <- getSparkSession()
# Allow the user to have a more flexible definiton of the text file path
@@ -435,7 +437,7 @@ read.text <- function(x, ...) {
#' }
#' @name sql
#' @method sql default
-
+#' @note sql since 1.4.0
sql.default <- function(sqlQuery) {
sparkSession <- getSparkSession()
sdf <- callJMethod(sparkSession, "sql", sqlQuery)
@@ -464,8 +466,7 @@ sql <- function(x, ...) {
#' createOrReplaceTempView(df, "table")
#' new_df <- tableToDF("table")
#' }
-#' @note since 2.0.0
-
+#' @note tableToDF since 2.0.0
tableToDF <- function(tableName) {
sparkSession <- getSparkSession()
sdf <- callJMethod(sparkSession, "table", tableName)
@@ -487,7 +488,7 @@ tableToDF <- function(tableName) {
#' }
#' @name tables
#' @method tables default
-
+#' @note tables since 1.4.0
tables.default <- function(databaseName = NULL) {
sparkSession <- getSparkSession()
jdf <- callJStatic("org.apache.spark.sql.api.r.SQLUtils", "getTables", sparkSession, databaseName)
@@ -513,7 +514,7 @@ tables <- function(x, ...) {
#' }
#' @name tableNames
#' @method tableNames default
-
+#' @note tableNames since 1.4.0
tableNames.default <- function(databaseName = NULL) {
sparkSession <- getSparkSession()
callJStatic("org.apache.spark.sql.api.r.SQLUtils",
@@ -544,7 +545,7 @@ tableNames <- function(x, ...) {
#' }
#' @name cacheTable
#' @method cacheTable default
-
+#' @note cacheTable since 1.4.0
cacheTable.default <- function(tableName) {
sparkSession <- getSparkSession()
catalog <- callJMethod(sparkSession, "catalog")
@@ -573,7 +574,7 @@ cacheTable <- function(x, ...) {
#' }
#' @name uncacheTable
#' @method uncacheTable default
-
+#' @note uncacheTable since 1.4.0
uncacheTable.default <- function(tableName) {
sparkSession <- getSparkSession()
catalog <- callJMethod(sparkSession, "catalog")
@@ -596,7 +597,7 @@ uncacheTable <- function(x, ...) {
#' }
#' @name clearCache
#' @method clearCache default
-
+#' @note clearCache since 1.4.0
clearCache.default <- function() {
sparkSession <- getSparkSession()
catalog <- callJMethod(sparkSession, "catalog")
@@ -625,7 +626,7 @@ clearCache <- function() {
#' }
#' @name dropTempTable
#' @method dropTempTable default
-
+#' @note dropTempTable since 1.4.0
dropTempTable.default <- function(tableName) {
if (class(tableName) != "character") {
stop("tableName must be a string.")
@@ -691,7 +692,7 @@ dropTempView <- function(viewName) {
#' }
#' @name read.df
#' @method read.df default
-
+#' @note read.df since 1.4.0
read.df.default <- function(path = NULL, source = NULL, schema = NULL, ...) {
sparkSession <- getSparkSession()
options <- varargsToEnv(...)
@@ -719,7 +720,7 @@ read.df <- function(x, ...) {
#' @rdname read.df
#' @name loadDF
#' @method loadDF default
-
+#' @note loadDF since 1.6.0
loadDF.default <- function(path = NULL, source = NULL, schema = NULL, ...) {
read.df(path, source, schema, ...)
}
@@ -750,7 +751,7 @@ loadDF <- function(x, ...) {
#' }
#' @name createExternalTable
#' @method createExternalTable default
-
+#' @note createExternalTable since 1.4.0
createExternalTable.default <- function(tableName, path = NULL, source = NULL, ...) {
sparkSession <- getSparkSession()
options <- varargsToEnv(...)
@@ -798,8 +799,7 @@ createExternalTable <- function(x, ...) {
#' df2 <- read.jdbc(jdbcUrl, "table2", partitionColumn = "index", lowerBound = 0,
#' upperBound = 10000, user = "username", password = "password")
#' }
-#' @note since 2.0.0
-
+#' @note read.jdbc since 2.0.0
read.jdbc <- function(url, tableName,
partitionColumn = NULL, lowerBound = NULL, upperBound = NULL,
numPartitions = 0L, predicates = list(), ...) {
diff --git a/R/pkg/R/WindowSpec.R b/R/pkg/R/WindowSpec.R
index d8405420d0..9f3b1e4be5 100644
--- a/R/pkg/R/WindowSpec.R
+++ b/R/pkg/R/WindowSpec.R
@@ -29,6 +29,7 @@ NULL
#'
#' @param sws A Java object reference to the backing Scala WindowSpec
#' @export
+#' @note WindowSpec since 2.0.0
setClass("WindowSpec",
slots = list(sws = "jobj"))
@@ -43,6 +44,7 @@ windowSpec <- function(sws) {
}
#' @rdname show
+#' @note show(WindowSpec) since 2.0.0
setMethod("show", "WindowSpec",
function(object) {
cat("WindowSpec", callJMethod(object@sws, "toString"), "\n")
@@ -63,6 +65,7 @@ setMethod("show", "WindowSpec",
#' partitionBy(ws, "col1", "col2")
#' partitionBy(ws, df$col1, df$col2)
#' }
+#' @note partitionBy(WindowSpec) since 2.0.0
setMethod("partitionBy",
signature(x = "WindowSpec"),
function(x, col, ...) {
@@ -93,6 +96,7 @@ setMethod("partitionBy",
#' orderBy(ws, "col1", "col2")
#' orderBy(ws, df$col1, df$col2)
#' }
+#' @note orderBy(WindowSpec, character) since 2.0.0
setMethod("orderBy",
signature(x = "WindowSpec", col = "character"),
function(x, col, ...) {
@@ -102,6 +106,7 @@ setMethod("orderBy",
#' @rdname arrange
#' @name orderBy
#' @export
+#' @note orderBy(WindowSpec, Column) since 2.0.0
setMethod("orderBy",
signature(x = "WindowSpec", col = "Column"),
function(x, col, ...) {
@@ -133,6 +138,7 @@ setMethod("orderBy",
#' \dontrun{
#' rowsBetween(ws, 0, 3)
#' }
+#' @note rowsBetween since 2.0.0
setMethod("rowsBetween",
signature(x = "WindowSpec", start = "numeric", end = "numeric"),
function(x, start, end) {
@@ -163,6 +169,7 @@ setMethod("rowsBetween",
#' \dontrun{
#' rangeBetween(ws, 0, 3)
#' }
+#' @note rangeBetween since 2.0.0
setMethod("rangeBetween",
signature(x = "WindowSpec", start = "numeric", end = "numeric"),
function(x, start, end) {
@@ -182,6 +189,7 @@ setMethod("rangeBetween",
#' @name over
#' @family colum_func
#' @export
+#' @note over since 2.0.0
setMethod("over",
signature(x = "Column", window = "WindowSpec"),
function(x, window) {
diff --git a/R/pkg/R/column.R b/R/pkg/R/column.R
index cc2876ed94..1af65d5d6e 100644
--- a/R/pkg/R/column.R
+++ b/R/pkg/R/column.R
@@ -30,6 +30,7 @@ setOldClass("jobj")
#'
#' @slot jc reference to JVM SparkDataFrame column
#' @export
+#' @note Column since 1.4.0
setClass("Column",
slots = list(jc = "jobj"))
@@ -46,6 +47,7 @@ setMethod("column",
#' @rdname show
#' @name show
+#' @note show(Column) since 1.4.0
setMethod("show", "Column",
function(object) {
cat("Column", callJMethod(object@jc, "toString"), "\n")
@@ -127,6 +129,7 @@ createMethods()
#' @name alias
#' @family colum_func
#' @export
+#' @note alias since 1.4.0
setMethod("alias",
signature(object = "Column"),
function(object, data) {
@@ -147,6 +150,7 @@ setMethod("alias",
#'
#' @param start starting position
#' @param stop ending position
+#' @note substr since 1.4.0
setMethod("substr", signature(x = "Column"),
function(x, start, stop) {
jc <- callJMethod(x@jc, "substr", as.integer(start - 1), as.integer(stop - start + 1))
@@ -164,6 +168,7 @@ setMethod("substr", signature(x = "Column"),
#'
#' @param x vector of character string whose “starts” are considered
#' @param prefix character vector (often of length one)
+#' @note startsWith since 1.4.0
setMethod("startsWith", signature(x = "Column"),
function(x, prefix) {
jc <- callJMethod(x@jc, "startsWith", as.vector(prefix))
@@ -181,6 +186,7 @@ setMethod("startsWith", signature(x = "Column"),
#'
#' @param x vector of character string whose “ends” are considered
#' @param suffix character vector (often of length one)
+#' @note endsWith since 1.4.0
setMethod("endsWith", signature(x = "Column"),
function(x, suffix) {
jc <- callJMethod(x@jc, "endsWith", as.vector(suffix))
@@ -196,6 +202,7 @@ setMethod("endsWith", signature(x = "Column"),
#' @family colum_func
#'
#' @param bounds lower and upper bounds
+#' @note between since 1.5.0
setMethod("between", signature(x = "Column"),
function(x, bounds) {
if (is.vector(bounds) && length(bounds) == 2) {
@@ -216,6 +223,7 @@ setMethod("between", signature(x = "Column"),
#' cast(df$age, "string")
#' cast(df$name, list(type="array", elementType="byte", containsNull = TRUE))
#' }
+#' @note cast since 1.4.0
setMethod("cast",
signature(x = "Column"),
function(x, dataType) {
@@ -242,6 +250,7 @@ setMethod("cast",
#' filter(df, "age in (10, 30)")
#' where(df, df$age %in% c(10, 30))
#' }
+#' @note \%in\% since 1.5.0
setMethod("%in%",
signature(x = "Column"),
function(x, table) {
@@ -258,6 +267,7 @@ setMethod("%in%",
#' @name otherwise
#' @family colum_func
#' @export
+#' @note otherwise since 1.5.0
setMethod("otherwise",
signature(x = "Column", value = "ANY"),
function(x, value) {
diff --git a/R/pkg/R/context.R b/R/pkg/R/context.R
index 968a9d2251..42f89c806b 100644
--- a/R/pkg/R/context.R
+++ b/R/pkg/R/context.R
@@ -261,6 +261,7 @@ setCheckpointDir <- function(sc, dirName) {
#' sparkR.session()
#' doubled <- spark.lapply(1:10, function(x){2 * x})
#'}
+#' @note spark.lapply since 2.0.0
spark.lapply <- function(list, func) {
if (!exists(".sparkRjsc", envir = .sparkREnv)) {
stop("SparkR has not been initialized. Please call sparkR.session()")
@@ -283,7 +284,7 @@ spark.lapply <- function(list, func) {
#'\dontrun{
#' setLogLevel("ERROR")
#'}
-
+#' @note setLogLevel since 2.0.0
setLogLevel <- function(level) {
if (!exists(".sparkRjsc", envir = .sparkREnv)) {
stop("SparkR has not been initialized. Please call sparkR.session()")
diff --git a/R/pkg/R/functions.R b/R/pkg/R/functions.R
index c26f963258..ce2386998c 100644
--- a/R/pkg/R/functions.R
+++ b/R/pkg/R/functions.R
@@ -33,6 +33,7 @@ NULL
#' select(df, lit("x"))
#' select(df, lit("2015-01-01"))
#'}
+#' @note lit since 1.5.0
setMethod("lit", signature("ANY"),
function(x) {
jc <- callJStatic("org.apache.spark.sql.functions",
@@ -50,6 +51,7 @@ setMethod("lit", signature("ANY"),
#' @family normal_funcs
#' @export
#' @examples \dontrun{abs(df$c)}
+#' @note abs since 1.5.0
setMethod("abs",
signature(x = "Column"),
function(x) {
@@ -67,6 +69,7 @@ setMethod("abs",
#' @family math_funcs
#' @export
#' @examples \dontrun{acos(df$c)}
+#' @note acos since 1.5.0
setMethod("acos",
signature(x = "Column"),
function(x) {
@@ -83,6 +86,7 @@ setMethod("acos",
#' @family agg_funcs
#' @export
#' @examples \dontrun{approxCountDistinct(df$c)}
+#' @note approxCountDistinct(Column) since 1.4.0
setMethod("approxCountDistinct",
signature(x = "Column"),
function(x) {
@@ -100,6 +104,7 @@ setMethod("approxCountDistinct",
#' @family string_funcs
#' @export
#' @examples \dontrun{\dontrun{ascii(df$c)}}
+#' @note ascii since 1.5.0
setMethod("ascii",
signature(x = "Column"),
function(x) {
@@ -117,6 +122,7 @@ setMethod("ascii",
#' @family math_funcs
#' @export
#' @examples \dontrun{asin(df$c)}
+#' @note asin since 1.5.0
setMethod("asin",
signature(x = "Column"),
function(x) {
@@ -133,6 +139,7 @@ setMethod("asin",
#' @family math_funcs
#' @export
#' @examples \dontrun{atan(df$c)}
+#' @note atan since 1.5.0
setMethod("atan",
signature(x = "Column"),
function(x) {
@@ -149,6 +156,7 @@ setMethod("atan",
#' @family agg_funcs
#' @export
#' @examples \dontrun{avg(df$c)}
+#' @note avg since 1.4.0
setMethod("avg",
signature(x = "Column"),
function(x) {
@@ -166,6 +174,7 @@ setMethod("avg",
#' @family string_funcs
#' @export
#' @examples \dontrun{base64(df$c)}
+#' @note base64 since 1.5.0
setMethod("base64",
signature(x = "Column"),
function(x) {
@@ -183,6 +192,7 @@ setMethod("base64",
#' @family math_funcs
#' @export
#' @examples \dontrun{bin(df$c)}
+#' @note bin since 1.5.0
setMethod("bin",
signature(x = "Column"),
function(x) {
@@ -199,6 +209,7 @@ setMethod("bin",
#' @family normal_funcs
#' @export
#' @examples \dontrun{bitwiseNOT(df$c)}
+#' @note bitwiseNOT since 1.5.0
setMethod("bitwiseNOT",
signature(x = "Column"),
function(x) {
@@ -215,6 +226,7 @@ setMethod("bitwiseNOT",
#' @family math_funcs
#' @export
#' @examples \dontrun{cbrt(df$c)}
+#' @note cbrt since 1.4.0
setMethod("cbrt",
signature(x = "Column"),
function(x) {
@@ -231,6 +243,7 @@ setMethod("cbrt",
#' @family math_funcs
#' @export
#' @examples \dontrun{ceil(df$c)}
+#' @note ceil since 1.5.0
setMethod("ceil",
signature(x = "Column"),
function(x) {
@@ -254,6 +267,7 @@ col <- function(x) {
#' @family normal_funcs
#' @export
#' @examples \dontrun{column(df)}
+#' @note column since 1.6.0
setMethod("column",
signature(x = "character"),
function(x) {
@@ -268,6 +282,7 @@ setMethod("column",
#' @family math_funcs
#' @export
#' @examples \dontrun{corr(df$c, df$d)}
+#' @note corr since 1.6.0
setMethod("corr", signature(x = "Column"),
function(x, col2) {
stopifnot(class(col2) == "Column")
@@ -290,6 +305,7 @@ setMethod("corr", signature(x = "Column"),
#' covar_samp(df$c, df$d)
#' covar_samp("c", "d")
#' }
+#' @note cov since 1.6.0
setMethod("cov", signature(x = "characterOrColumn"),
function(x, col2) {
stopifnot(is(class(col2), "characterOrColumn"))
@@ -298,6 +314,7 @@ setMethod("cov", signature(x = "characterOrColumn"),
#' @rdname cov
#' @name covar_samp
+#' @note covar_samp since 2.0.0
setMethod("covar_samp", signature(col1 = "characterOrColumn", col2 = "characterOrColumn"),
function(col1, col2) {
stopifnot(class(col1) == class(col2))
@@ -322,6 +339,7 @@ setMethod("covar_samp", signature(col1 = "characterOrColumn", col2 = "characterO
#' covar_pop(df$c, df$d)
#' covar_pop("c", "d")
#' }
+#' @note covar_pop since 2.0.0
setMethod("covar_pop", signature(col1 = "characterOrColumn", col2 = "characterOrColumn"),
function(col1, col2) {
stopifnot(class(col1) == class(col2))
@@ -342,6 +360,7 @@ setMethod("covar_pop", signature(col1 = "characterOrColumn", col2 = "characterOr
#' @family math_funcs
#' @export
#' @examples \dontrun{cos(df$c)}
+#' @note cos since 1.5.0
setMethod("cos",
signature(x = "Column"),
function(x) {
@@ -358,6 +377,7 @@ setMethod("cos",
#' @family math_funcs
#' @export
#' @examples \dontrun{cosh(df$c)}
+#' @note cosh since 1.5.0
setMethod("cosh",
signature(x = "Column"),
function(x) {
@@ -374,6 +394,7 @@ setMethod("cosh",
#' @family agg_funcs
#' @export
#' @examples \dontrun{count(df$c)}
+#' @note count since 1.4.0
setMethod("count",
signature(x = "Column"),
function(x) {
@@ -391,6 +412,7 @@ setMethod("count",
#' @family misc_funcs
#' @export
#' @examples \dontrun{crc32(df$c)}
+#' @note crc32 since 1.5.0
setMethod("crc32",
signature(x = "Column"),
function(x) {
@@ -407,6 +429,7 @@ setMethod("crc32",
#' @family misc_funcs
#' @export
#' @examples \dontrun{hash(df$c)}
+#' @note hash since 2.0.0
setMethod("hash",
signature(x = "Column"),
function(x, ...) {
@@ -427,6 +450,7 @@ setMethod("hash",
#' @family datetime_funcs
#' @export
#' @examples \dontrun{dayofmonth(df$c)}
+#' @note dayofmonth since 1.5.0
setMethod("dayofmonth",
signature(x = "Column"),
function(x) {
@@ -443,6 +467,7 @@ setMethod("dayofmonth",
#' @family datetime_funcs
#' @export
#' @examples \dontrun{dayofyear(df$c)}
+#' @note dayofyear since 1.5.0
setMethod("dayofyear",
signature(x = "Column"),
function(x) {
@@ -460,6 +485,7 @@ setMethod("dayofyear",
#' @family string_funcs
#' @export
#' @examples \dontrun{decode(df$c, "UTF-8")}
+#' @note decode since 1.6.0
setMethod("decode",
signature(x = "Column", charset = "character"),
function(x, charset) {
@@ -477,6 +503,7 @@ setMethod("decode",
#' @family string_funcs
#' @export
#' @examples \dontrun{encode(df$c, "UTF-8")}
+#' @note encode since 1.6.0
setMethod("encode",
signature(x = "Column", charset = "character"),
function(x, charset) {
@@ -493,6 +520,7 @@ setMethod("encode",
#' @family math_funcs
#' @export
#' @examples \dontrun{exp(df$c)}
+#' @note exp since 1.5.0
setMethod("exp",
signature(x = "Column"),
function(x) {
@@ -509,6 +537,7 @@ setMethod("exp",
#' @family math_funcs
#' @export
#' @examples \dontrun{expm1(df$c)}
+#' @note expm1 since 1.5.0
setMethod("expm1",
signature(x = "Column"),
function(x) {
@@ -525,6 +554,7 @@ setMethod("expm1",
#' @family math_funcs
#' @export
#' @examples \dontrun{factorial(df$c)}
+#' @note factorial since 1.5.0
setMethod("factorial",
signature(x = "Column"),
function(x) {
@@ -548,6 +578,7 @@ setMethod("factorial",
#' first(df$c)
#' first(df$c, TRUE)
#' }
+#' @note first(characterOrColumn) since 1.4.0
setMethod("first",
signature(x = "characterOrColumn"),
function(x, na.rm = FALSE) {
@@ -569,6 +600,7 @@ setMethod("first",
#' @family math_funcs
#' @export
#' @examples \dontrun{floor(df$c)}
+#' @note floor since 1.5.0
setMethod("floor",
signature(x = "Column"),
function(x) {
@@ -585,6 +617,7 @@ setMethod("floor",
#' @family math_funcs
#' @export
#' @examples \dontrun{hex(df$c)}
+#' @note hex since 1.5.0
setMethod("hex",
signature(x = "Column"),
function(x) {
@@ -601,6 +634,7 @@ setMethod("hex",
#' @family datetime_funcs
#' @export
#' @examples \dontrun{hour(df$c)}
+#' @note hour since 1.5.0
setMethod("hour",
signature(x = "Column"),
function(x) {
@@ -620,6 +654,7 @@ setMethod("hour",
#' @family string_funcs
#' @export
#' @examples \dontrun{initcap(df$c)}
+#' @note initcap since 1.5.0
setMethod("initcap",
signature(x = "Column"),
function(x) {
@@ -640,6 +675,7 @@ setMethod("initcap",
#' is.nan(df$c)
#' isnan(df$c)
#' }
+#' @note is.nan since 2.0.0
setMethod("is.nan",
signature(x = "Column"),
function(x) {
@@ -648,6 +684,7 @@ setMethod("is.nan",
#' @rdname is.nan
#' @name isnan
+#' @note isnan since 2.0.0
setMethod("isnan",
signature(x = "Column"),
function(x) {
@@ -664,6 +701,7 @@ setMethod("isnan",
#' @family agg_funcs
#' @export
#' @examples \dontrun{kurtosis(df$c)}
+#' @note kurtosis since 1.6.0
setMethod("kurtosis",
signature(x = "Column"),
function(x) {
@@ -687,6 +725,7 @@ setMethod("kurtosis",
#' last(df$c)
#' last(df$c, TRUE)
#' }
+#' @note last since 1.4.0
setMethod("last",
signature(x = "characterOrColumn"),
function(x, na.rm = FALSE) {
@@ -710,6 +749,7 @@ setMethod("last",
#' @family datetime_funcs
#' @export
#' @examples \dontrun{last_day(df$c)}
+#' @note last_day since 1.5.0
setMethod("last_day",
signature(x = "Column"),
function(x) {
@@ -726,6 +766,7 @@ setMethod("last_day",
#' @family string_funcs
#' @export
#' @examples \dontrun{length(df$c)}
+#' @note length since 1.5.0
setMethod("length",
signature(x = "Column"),
function(x) {
@@ -742,6 +783,7 @@ setMethod("length",
#' @family math_funcs
#' @export
#' @examples \dontrun{log(df$c)}
+#' @note log since 1.5.0
setMethod("log",
signature(x = "Column"),
function(x) {
@@ -758,6 +800,7 @@ setMethod("log",
#' @family math_funcs
#' @export
#' @examples \dontrun{log10(df$c)}
+#' @note log10 since 1.5.0
setMethod("log10",
signature(x = "Column"),
function(x) {
@@ -774,6 +817,7 @@ setMethod("log10",
#' @family math_funcs
#' @export
#' @examples \dontrun{log1p(df$c)}
+#' @note log1p since 1.5.0
setMethod("log1p",
signature(x = "Column"),
function(x) {
@@ -790,6 +834,7 @@ setMethod("log1p",
#' @family math_funcs
#' @export
#' @examples \dontrun{log2(df$c)}
+#' @note log2 since 1.5.0
setMethod("log2",
signature(x = "Column"),
function(x) {
@@ -806,6 +851,7 @@ setMethod("log2",
#' @family string_funcs
#' @export
#' @examples \dontrun{lower(df$c)}
+#' @note lower since 1.4.0
setMethod("lower",
signature(x = "Column"),
function(x) {
@@ -822,6 +868,7 @@ setMethod("lower",
#' @family string_funcs
#' @export
#' @examples \dontrun{ltrim(df$c)}
+#' @note ltrim since 1.5.0
setMethod("ltrim",
signature(x = "Column"),
function(x) {
@@ -838,6 +885,7 @@ setMethod("ltrim",
#' @family agg_funcs
#' @export
#' @examples \dontrun{max(df$c)}
+#' @note max since 1.5.0
setMethod("max",
signature(x = "Column"),
function(x) {
@@ -855,6 +903,7 @@ setMethod("max",
#' @family misc_funcs
#' @export
#' @examples \dontrun{md5(df$c)}
+#' @note md5 since 1.5.0
setMethod("md5",
signature(x = "Column"),
function(x) {
@@ -872,6 +921,7 @@ setMethod("md5",
#' @family agg_funcs
#' @export
#' @examples \dontrun{mean(df$c)}
+#' @note mean since 1.5.0
setMethod("mean",
signature(x = "Column"),
function(x) {
@@ -888,6 +938,7 @@ setMethod("mean",
#' @family agg_funcs
#' @export
#' @examples \dontrun{min(df$c)}
+#' @note min since 1.5.0
setMethod("min",
signature(x = "Column"),
function(x) {
@@ -904,6 +955,7 @@ setMethod("min",
#' @family datetime_funcs
#' @export
#' @examples \dontrun{minute(df$c)}
+#' @note minute since 1.5.0
setMethod("minute",
signature(x = "Column"),
function(x) {
@@ -947,6 +999,7 @@ setMethod("monotonically_increasing_id",
#' @family datetime_funcs
#' @export
#' @examples \dontrun{month(df$c)}
+#' @note month since 1.5.0
setMethod("month",
signature(x = "Column"),
function(x) {
@@ -963,6 +1016,7 @@ setMethod("month",
#' @family normal_funcs
#' @export
#' @examples \dontrun{negate(df$c)}
+#' @note negate since 1.5.0
setMethod("negate",
signature(x = "Column"),
function(x) {
@@ -979,6 +1033,7 @@ setMethod("negate",
#' @family datetime_funcs
#' @export
#' @examples \dontrun{quarter(df$c)}
+#' @note quarter since 1.5.0
setMethod("quarter",
signature(x = "Column"),
function(x) {
@@ -995,6 +1050,7 @@ setMethod("quarter",
#' @family string_funcs
#' @export
#' @examples \dontrun{reverse(df$c)}
+#' @note reverse since 1.5.0
setMethod("reverse",
signature(x = "Column"),
function(x) {
@@ -1012,6 +1068,7 @@ setMethod("reverse",
#' @family math_funcs
#' @export
#' @examples \dontrun{rint(df$c)}
+#' @note rint since 1.5.0
setMethod("rint",
signature(x = "Column"),
function(x) {
@@ -1028,6 +1085,7 @@ setMethod("rint",
#' @family math_funcs
#' @export
#' @examples \dontrun{round(df$c)}
+#' @note round since 1.5.0
setMethod("round",
signature(x = "Column"),
function(x) {
@@ -1047,6 +1105,7 @@ setMethod("round",
#' @family math_funcs
#' @export
#' @examples \dontrun{bround(df$c, 0)}
+#' @note bround since 2.0.0
setMethod("bround",
signature(x = "Column"),
function(x, scale = 0) {
@@ -1064,6 +1123,7 @@ setMethod("bround",
#' @family string_funcs
#' @export
#' @examples \dontrun{rtrim(df$c)}
+#' @note rtrim since 1.5.0
setMethod("rtrim",
signature(x = "Column"),
function(x) {
@@ -1086,6 +1146,7 @@ setMethod("rtrim",
#'select(df, stddev(df$age))
#'agg(df, sd(df$age))
#'}
+#' @note sd since 1.6.0
setMethod("sd",
signature(x = "Column"),
function(x) {
@@ -1102,6 +1163,7 @@ setMethod("sd",
#' @family datetime_funcs
#' @export
#' @examples \dontrun{second(df$c)}
+#' @note second since 1.5.0
setMethod("second",
signature(x = "Column"),
function(x) {
@@ -1119,6 +1181,7 @@ setMethod("second",
#' @family misc_funcs
#' @export
#' @examples \dontrun{sha1(df$c)}
+#' @note sha1 since 1.5.0
setMethod("sha1",
signature(x = "Column"),
function(x) {
@@ -1135,6 +1198,7 @@ setMethod("sha1",
#' @family math_funcs
#' @export
#' @examples \dontrun{signum(df$c)}
+#' @note signum since 1.5.0
setMethod("signum",
signature(x = "Column"),
function(x) {
@@ -1151,6 +1215,7 @@ setMethod("signum",
#' @family math_funcs
#' @export
#' @examples \dontrun{sin(df$c)}
+#' @note sin since 1.5.0
setMethod("sin",
signature(x = "Column"),
function(x) {
@@ -1167,6 +1232,7 @@ setMethod("sin",
#' @family math_funcs
#' @export
#' @examples \dontrun{sinh(df$c)}
+#' @note sinh since 1.5.0
setMethod("sinh",
signature(x = "Column"),
function(x) {
@@ -1183,6 +1249,7 @@ setMethod("sinh",
#' @family agg_funcs
#' @export
#' @examples \dontrun{skewness(df$c)}
+#' @note skewness since 1.6.0
setMethod("skewness",
signature(x = "Column"),
function(x) {
@@ -1199,6 +1266,7 @@ setMethod("skewness",
#' @family string_funcs
#' @export
#' @examples \dontrun{soundex(df$c)}
+#' @note soundex since 1.5.0
setMethod("soundex",
signature(x = "Column"),
function(x) {
@@ -1229,6 +1297,7 @@ setMethod("spark_partition_id",
#' @rdname sd
#' @name stddev
+#' @note stddev since 1.6.0
setMethod("stddev",
signature(x = "Column"),
function(x) {
@@ -1246,6 +1315,7 @@ setMethod("stddev",
#' @seealso \link{sd}, \link{stddev_samp}
#' @export
#' @examples \dontrun{stddev_pop(df$c)}
+#' @note stddev_pop since 1.6.0
setMethod("stddev_pop",
signature(x = "Column"),
function(x) {
@@ -1263,6 +1333,7 @@ setMethod("stddev_pop",
#' @seealso \link{stddev_pop}, \link{sd}
#' @export
#' @examples \dontrun{stddev_samp(df$c)}
+#' @note stddev_samp since 1.6.0
setMethod("stddev_samp",
signature(x = "Column"),
function(x) {
@@ -1283,6 +1354,7 @@ setMethod("stddev_samp",
#' struct(df$c, df$d)
#' struct("col1", "col2")
#' }
+#' @note struct since 1.6.0
setMethod("struct",
signature(x = "characterOrColumn"),
function(x, ...) {
@@ -1304,6 +1376,7 @@ setMethod("struct",
#' @family math_funcs
#' @export
#' @examples \dontrun{sqrt(df$c)}
+#' @note sqrt since 1.5.0
setMethod("sqrt",
signature(x = "Column"),
function(x) {
@@ -1320,6 +1393,7 @@ setMethod("sqrt",
#' @family agg_funcs
#' @export
#' @examples \dontrun{sum(df$c)}
+#' @note sum since 1.5.0
setMethod("sum",
signature(x = "Column"),
function(x) {
@@ -1336,6 +1410,7 @@ setMethod("sum",
#' @family agg_funcs
#' @export
#' @examples \dontrun{sumDistinct(df$c)}
+#' @note sumDistinct since 1.4.0
setMethod("sumDistinct",
signature(x = "Column"),
function(x) {
@@ -1352,6 +1427,7 @@ setMethod("sumDistinct",
#' @family math_funcs
#' @export
#' @examples \dontrun{tan(df$c)}
+#' @note tan since 1.5.0
setMethod("tan",
signature(x = "Column"),
function(x) {
@@ -1368,6 +1444,7 @@ setMethod("tan",
#' @family math_funcs
#' @export
#' @examples \dontrun{tanh(df$c)}
+#' @note tanh since 1.5.0
setMethod("tanh",
signature(x = "Column"),
function(x) {
@@ -1384,6 +1461,7 @@ setMethod("tanh",
#' @family math_funcs
#' @export
#' @examples \dontrun{toDegrees(df$c)}
+#' @note toDegrees since 1.4.0
setMethod("toDegrees",
signature(x = "Column"),
function(x) {
@@ -1400,6 +1478,7 @@ setMethod("toDegrees",
#' @family math_funcs
#' @export
#' @examples \dontrun{toRadians(df$c)}
+#' @note toRadians since 1.4.0
setMethod("toRadians",
signature(x = "Column"),
function(x) {
@@ -1416,6 +1495,7 @@ setMethod("toRadians",
#' @family datetime_funcs
#' @export
#' @examples \dontrun{to_date(df$c)}
+#' @note to_date since 1.5.0
setMethod("to_date",
signature(x = "Column"),
function(x) {
@@ -1432,6 +1512,7 @@ setMethod("to_date",
#' @family string_funcs
#' @export
#' @examples \dontrun{trim(df$c)}
+#' @note trim since 1.5.0
setMethod("trim",
signature(x = "Column"),
function(x) {
@@ -1449,6 +1530,7 @@ setMethod("trim",
#' @family string_funcs
#' @export
#' @examples \dontrun{unbase64(df$c)}
+#' @note unbase64 since 1.5.0
setMethod("unbase64",
signature(x = "Column"),
function(x) {
@@ -1466,6 +1548,7 @@ setMethod("unbase64",
#' @family math_funcs
#' @export
#' @examples \dontrun{unhex(df$c)}
+#' @note unhex since 1.5.0
setMethod("unhex",
signature(x = "Column"),
function(x) {
@@ -1482,6 +1565,7 @@ setMethod("unhex",
#' @family string_funcs
#' @export
#' @examples \dontrun{upper(df$c)}
+#' @note upper since 1.4.0
setMethod("upper",
signature(x = "Column"),
function(x) {
@@ -1504,6 +1588,7 @@ setMethod("upper",
#'select(df, var_pop(df$age))
#'agg(df, var(df$age))
#'}
+#' @note var since 1.6.0
setMethod("var",
signature(x = "Column"),
function(x) {
@@ -1513,6 +1598,7 @@ setMethod("var",
#' @rdname var
#' @name variance
+#' @note variance since 1.6.0
setMethod("variance",
signature(x = "Column"),
function(x) {
@@ -1530,6 +1616,7 @@ setMethod("variance",
#' @seealso \link{var}, \link{var_samp}
#' @export
#' @examples \dontrun{var_pop(df$c)}
+#' @note var_pop since 1.5.0
setMethod("var_pop",
signature(x = "Column"),
function(x) {
@@ -1547,6 +1634,7 @@ setMethod("var_pop",
#' @seealso \link{var_pop}, \link{var}
#' @export
#' @examples \dontrun{var_samp(df$c)}
+#' @note var_samp since 1.6.0
setMethod("var_samp",
signature(x = "Column"),
function(x) {
@@ -1563,6 +1651,7 @@ setMethod("var_samp",
#' @family datetime_funcs
#' @export
#' @examples \dontrun{weekofyear(df$c)}
+#' @note weekofyear since 1.5.0
setMethod("weekofyear",
signature(x = "Column"),
function(x) {
@@ -1579,6 +1668,7 @@ setMethod("weekofyear",
#' @family datetime_funcs
#' @export
#' @examples \dontrun{year(df$c)}
+#' @note year since 1.5.0
setMethod("year",
signature(x = "Column"),
function(x) {
@@ -1596,6 +1686,7 @@ setMethod("year",
#' @family math_funcs
#' @export
#' @examples \dontrun{atan2(df$c, x)}
+#' @note atan2 since 1.5.0
setMethod("atan2", signature(y = "Column"),
function(y, x) {
if (class(x) == "Column") {
@@ -1614,6 +1705,7 @@ setMethod("atan2", signature(y = "Column"),
#' @family datetime_funcs
#' @export
#' @examples \dontrun{datediff(df$c, x)}
+#' @note datediff since 1.5.0
setMethod("datediff", signature(y = "Column"),
function(y, x) {
if (class(x) == "Column") {
@@ -1632,6 +1724,7 @@ setMethod("datediff", signature(y = "Column"),
#' @family math_funcs
#' @export
#' @examples \dontrun{hypot(df$c, x)}
+#' @note hypot since 1.4.0
setMethod("hypot", signature(y = "Column"),
function(y, x) {
if (class(x) == "Column") {
@@ -1650,6 +1743,7 @@ setMethod("hypot", signature(y = "Column"),
#' @family string_funcs
#' @export
#' @examples \dontrun{levenshtein(df$c, x)}
+#' @note levenshtein since 1.5.0
setMethod("levenshtein", signature(y = "Column"),
function(y, x) {
if (class(x) == "Column") {
@@ -1668,6 +1762,7 @@ setMethod("levenshtein", signature(y = "Column"),
#' @family datetime_funcs
#' @export
#' @examples \dontrun{months_between(df$c, x)}
+#' @note months_between since 1.5.0
setMethod("months_between", signature(y = "Column"),
function(y, x) {
if (class(x) == "Column") {
@@ -1687,6 +1782,7 @@ setMethod("months_between", signature(y = "Column"),
#' @family normal_funcs
#' @export
#' @examples \dontrun{nanvl(df$c, x)}
+#' @note nanvl since 1.5.0
setMethod("nanvl", signature(y = "Column"),
function(y, x) {
if (class(x) == "Column") {
@@ -1706,6 +1802,7 @@ setMethod("nanvl", signature(y = "Column"),
#' @family math_funcs
#' @export
#' @examples \dontrun{pmod(df$c, x)}
+#' @note pmod since 1.5.0
setMethod("pmod", signature(y = "Column"),
function(y, x) {
if (class(x) == "Column") {
@@ -1724,6 +1821,7 @@ setMethod("pmod", signature(y = "Column"),
#' @return the approximate number of distinct items in a group.
#' @export
#' @examples \dontrun{approxCountDistinct(df$c, 0.02)}
+#' @note approxCountDistinct(Column, numeric) since 1.4.0
setMethod("approxCountDistinct",
signature(x = "Column"),
function(x, rsd = 0.05) {
@@ -1739,6 +1837,7 @@ setMethod("approxCountDistinct",
#' @return the number of distinct items in a group.
#' @export
#' @examples \dontrun{countDistinct(df$c)}
+#' @note countDistinct since 1.4.0
setMethod("countDistinct",
signature(x = "Column"),
function(x, ...) {
@@ -1761,6 +1860,7 @@ setMethod("countDistinct",
#' @name concat
#' @export
#' @examples \dontrun{concat(df$strings, df$strings2)}
+#' @note concat since 1.5.0
setMethod("concat",
signature(x = "Column"),
function(x, ...) {
@@ -1782,6 +1882,7 @@ setMethod("concat",
#' @name greatest
#' @export
#' @examples \dontrun{greatest(df$c, df$d)}
+#' @note greatest since 1.5.0
setMethod("greatest",
signature(x = "Column"),
function(x, ...) {
@@ -1804,6 +1905,7 @@ setMethod("greatest",
#' @name least
#' @export
#' @examples \dontrun{least(df$c, df$d)}
+#' @note least since 1.5.0
setMethod("least",
signature(x = "Column"),
function(x, ...) {
@@ -1824,6 +1926,7 @@ setMethod("least",
#' @name ceiling
#' @export
#' @examples \dontrun{ceiling(df$c)}
+#' @note ceiling since 1.5.0
setMethod("ceiling",
signature(x = "Column"),
function(x) {
@@ -1838,6 +1941,7 @@ setMethod("ceiling",
#' @name sign
#' @export
#' @examples \dontrun{sign(df$c)}
+#' @note sign since 1.5.0
setMethod("sign", signature(x = "Column"),
function(x) {
signum(x)
@@ -1851,6 +1955,7 @@ setMethod("sign", signature(x = "Column"),
#' @name n_distinct
#' @export
#' @examples \dontrun{n_distinct(df$c)}
+#' @note n_distinct since 1.4.0
setMethod("n_distinct", signature(x = "Column"),
function(x, ...) {
countDistinct(x, ...)
@@ -1864,6 +1969,7 @@ setMethod("n_distinct", signature(x = "Column"),
#' @name n
#' @export
#' @examples \dontrun{n(df$c)}
+#' @note n since 1.4.0
setMethod("n", signature(x = "Column"),
function(x) {
count(x)
@@ -1885,6 +1991,7 @@ setMethod("n", signature(x = "Column"),
#' @name date_format
#' @export
#' @examples \dontrun{date_format(df$t, 'MM/dd/yyy')}
+#' @note date_format since 1.5.0
setMethod("date_format", signature(y = "Column", x = "character"),
function(y, x) {
jc <- callJStatic("org.apache.spark.sql.functions", "date_format", y@jc, x)
@@ -1900,6 +2007,7 @@ setMethod("date_format", signature(y = "Column", x = "character"),
#' @name from_utc_timestamp
#' @export
#' @examples \dontrun{from_utc_timestamp(df$t, 'PST')}
+#' @note from_utc_timestamp since 1.5.0
setMethod("from_utc_timestamp", signature(y = "Column", x = "character"),
function(y, x) {
jc <- callJStatic("org.apache.spark.sql.functions", "from_utc_timestamp", y@jc, x)
@@ -1919,6 +2027,7 @@ setMethod("from_utc_timestamp", signature(y = "Column", x = "character"),
#' @name instr
#' @export
#' @examples \dontrun{instr(df$c, 'b')}
+#' @note instr since 1.5.0
setMethod("instr", signature(y = "Column", x = "character"),
function(y, x) {
jc <- callJStatic("org.apache.spark.sql.functions", "instr", y@jc, x)
@@ -1945,6 +2054,7 @@ setMethod("instr", signature(y = "Column", x = "character"),
#'next_day(df$d, 'Sun')
#'next_day(df$d, 'Sunday')
#'}
+#' @note next_day since 1.5.0
setMethod("next_day", signature(y = "Column", x = "character"),
function(y, x) {
jc <- callJStatic("org.apache.spark.sql.functions", "next_day", y@jc, x)
@@ -1960,6 +2070,7 @@ setMethod("next_day", signature(y = "Column", x = "character"),
#' @name to_utc_timestamp
#' @export
#' @examples \dontrun{to_utc_timestamp(df$t, 'PST')}
+#' @note to_utc_timestamp since 1.5.0
setMethod("to_utc_timestamp", signature(y = "Column", x = "character"),
function(y, x) {
jc <- callJStatic("org.apache.spark.sql.functions", "to_utc_timestamp", y@jc, x)
@@ -1975,6 +2086,7 @@ setMethod("to_utc_timestamp", signature(y = "Column", x = "character"),
#' @rdname add_months
#' @export
#' @examples \dontrun{add_months(df$d, 1)}
+#' @note add_months since 1.5.0
setMethod("add_months", signature(y = "Column", x = "numeric"),
function(y, x) {
jc <- callJStatic("org.apache.spark.sql.functions", "add_months", y@jc, as.integer(x))
@@ -1990,6 +2102,7 @@ setMethod("add_months", signature(y = "Column", x = "numeric"),
#' @name date_add
#' @export
#' @examples \dontrun{date_add(df$d, 1)}
+#' @note date_add since 1.5.0
setMethod("date_add", signature(y = "Column", x = "numeric"),
function(y, x) {
jc <- callJStatic("org.apache.spark.sql.functions", "date_add", y@jc, as.integer(x))
@@ -2005,6 +2118,7 @@ setMethod("date_add", signature(y = "Column", x = "numeric"),
#' @name date_sub
#' @export
#' @examples \dontrun{date_sub(df$d, 1)}
+#' @note date_sub since 1.5.0
setMethod("date_sub", signature(y = "Column", x = "numeric"),
function(y, x) {
jc <- callJStatic("org.apache.spark.sql.functions", "date_sub", y@jc, as.integer(x))
@@ -2026,6 +2140,7 @@ setMethod("date_sub", signature(y = "Column", x = "numeric"),
#' @name format_number
#' @export
#' @examples \dontrun{format_number(df$n, 4)}
+#' @note format_number since 1.5.0
setMethod("format_number", signature(y = "Column", x = "numeric"),
function(y, x) {
jc <- callJStatic("org.apache.spark.sql.functions",
@@ -2046,6 +2161,7 @@ setMethod("format_number", signature(y = "Column", x = "numeric"),
#' @name sha2
#' @export
#' @examples \dontrun{sha2(df$c, 256)}
+#' @note sha2 since 1.5.0
setMethod("sha2", signature(y = "Column", x = "numeric"),
function(y, x) {
jc <- callJStatic("org.apache.spark.sql.functions", "sha2", y@jc, as.integer(x))
@@ -2062,6 +2178,7 @@ setMethod("sha2", signature(y = "Column", x = "numeric"),
#' @name shiftLeft
#' @export
#' @examples \dontrun{shiftLeft(df$c, 1)}
+#' @note shiftLeft since 1.5.0
setMethod("shiftLeft", signature(y = "Column", x = "numeric"),
function(y, x) {
jc <- callJStatic("org.apache.spark.sql.functions",
@@ -2080,6 +2197,7 @@ setMethod("shiftLeft", signature(y = "Column", x = "numeric"),
#' @name shiftRight
#' @export
#' @examples \dontrun{shiftRight(df$c, 1)}
+#' @note shiftRight since 1.5.0
setMethod("shiftRight", signature(y = "Column", x = "numeric"),
function(y, x) {
jc <- callJStatic("org.apache.spark.sql.functions",
@@ -2098,6 +2216,7 @@ setMethod("shiftRight", signature(y = "Column", x = "numeric"),
#' @name shiftRightUnsigned
#' @export
#' @examples \dontrun{shiftRightUnsigned(df$c, 1)}
+#' @note shiftRightUnsigned since 1.5.0
setMethod("shiftRightUnsigned", signature(y = "Column", x = "numeric"),
function(y, x) {
jc <- callJStatic("org.apache.spark.sql.functions",
@@ -2116,6 +2235,7 @@ setMethod("shiftRightUnsigned", signature(y = "Column", x = "numeric"),
#' @name concat_ws
#' @export
#' @examples \dontrun{concat_ws('-', df$s, df$d)}
+#' @note concat_ws since 1.5.0
setMethod("concat_ws", signature(sep = "character", x = "Column"),
function(sep, x, ...) {
jcols <- lapply(list(x, ...), function(x) { x@jc })
@@ -2132,6 +2252,7 @@ setMethod("concat_ws", signature(sep = "character", x = "Column"),
#' @name conv
#' @export
#' @examples \dontrun{conv(df$n, 2, 16)}
+#' @note conv since 1.5.0
setMethod("conv", signature(x = "Column", fromBase = "numeric", toBase = "numeric"),
function(x, fromBase, toBase) {
fromBase <- as.integer(fromBase)
@@ -2152,6 +2273,7 @@ setMethod("conv", signature(x = "Column", fromBase = "numeric", toBase = "numeri
#' @name expr
#' @export
#' @examples \dontrun{expr('length(name)')}
+#' @note expr since 1.5.0
setMethod("expr", signature(x = "character"),
function(x) {
jc <- callJStatic("org.apache.spark.sql.functions", "expr", x)
@@ -2167,6 +2289,7 @@ setMethod("expr", signature(x = "character"),
#' @name format_string
#' @export
#' @examples \dontrun{format_string('%d %s', df$a, df$b)}
+#' @note format_string since 1.5.0
setMethod("format_string", signature(format = "character", x = "Column"),
function(format, x, ...) {
jcols <- lapply(list(x, ...), function(arg) { arg@jc })
@@ -2191,6 +2314,7 @@ setMethod("format_string", signature(format = "character", x = "Column"),
#'from_unixtime(df$t)
#'from_unixtime(df$t, 'yyyy/MM/dd HH')
#'}
+#' @note from_unixtime since 1.5.0
setMethod("from_unixtime", signature(x = "Column"),
function(x, format = "yyyy-MM-dd HH:mm:ss") {
jc <- callJStatic("org.apache.spark.sql.functions",
@@ -2236,6 +2360,7 @@ setMethod("from_unixtime", signature(x = "Column"),
#' # Thirty-second windows every 10 seconds, e.g. 09:00:00-09:00:30, 09:00:10-09:00:40, ...
#' window(df$time, "30 seconds", "10 seconds")
#'}
+#' @note window since 2.0.0
setMethod("window", signature(x = "Column"),
function(x, windowDuration, slideDuration = NULL, startTime = NULL) {
stopifnot(is.character(windowDuration))
@@ -2273,6 +2398,7 @@ setMethod("window", signature(x = "Column"),
#' @name locate
#' @export
#' @examples \dontrun{locate('b', df$c, 1)}
+#' @note locate since 1.5.0
setMethod("locate", signature(substr = "character", str = "Column"),
function(substr, str, pos = 1) {
jc <- callJStatic("org.apache.spark.sql.functions",
@@ -2290,6 +2416,7 @@ setMethod("locate", signature(substr = "character", str = "Column"),
#' @name lpad
#' @export
#' @examples \dontrun{lpad(df$c, 6, '#')}
+#' @note lpad since 1.5.0
setMethod("lpad", signature(x = "Column", len = "numeric", pad = "character"),
function(x, len, pad) {
jc <- callJStatic("org.apache.spark.sql.functions",
@@ -2307,6 +2434,7 @@ setMethod("lpad", signature(x = "Column", len = "numeric", pad = "character"),
#' @name rand
#' @export
#' @examples \dontrun{rand()}
+#' @note rand since 1.5.0
setMethod("rand", signature(seed = "missing"),
function(seed) {
jc <- callJStatic("org.apache.spark.sql.functions", "rand")
@@ -2316,6 +2444,7 @@ setMethod("rand", signature(seed = "missing"),
#' @rdname rand
#' @name rand
#' @export
+#' @note rand(numeric) since 1.5.0
setMethod("rand", signature(seed = "numeric"),
function(seed) {
jc <- callJStatic("org.apache.spark.sql.functions", "rand", as.integer(seed))
@@ -2331,6 +2460,7 @@ setMethod("rand", signature(seed = "numeric"),
#' @name randn
#' @export
#' @examples \dontrun{randn()}
+#' @note randn since 1.5.0
setMethod("randn", signature(seed = "missing"),
function(seed) {
jc <- callJStatic("org.apache.spark.sql.functions", "randn")
@@ -2340,6 +2470,7 @@ setMethod("randn", signature(seed = "missing"),
#' @rdname randn
#' @name randn
#' @export
+#' @note randn(numeric) since 1.5.0
setMethod("randn", signature(seed = "numeric"),
function(seed) {
jc <- callJStatic("org.apache.spark.sql.functions", "randn", as.integer(seed))
@@ -2355,6 +2486,7 @@ setMethod("randn", signature(seed = "numeric"),
#' @name regexp_extract
#' @export
#' @examples \dontrun{regexp_extract(df$c, '(\d+)-(\d+)', 1)}
+#' @note regexp_extract since 1.5.0
setMethod("regexp_extract",
signature(x = "Column", pattern = "character", idx = "numeric"),
function(x, pattern, idx) {
@@ -2373,6 +2505,7 @@ setMethod("regexp_extract",
#' @name regexp_replace
#' @export
#' @examples \dontrun{regexp_replace(df$c, '(\\d+)', '--')}
+#' @note regexp_replace since 1.5.0
setMethod("regexp_replace",
signature(x = "Column", pattern = "character", replacement = "character"),
function(x, pattern, replacement) {
@@ -2391,6 +2524,7 @@ setMethod("regexp_replace",
#' @name rpad
#' @export
#' @examples \dontrun{rpad(df$c, 6, '#')}
+#' @note rpad since 1.5.0
setMethod("rpad", signature(x = "Column", len = "numeric", pad = "character"),
function(x, len, pad) {
jc <- callJStatic("org.apache.spark.sql.functions",
@@ -2415,6 +2549,7 @@ setMethod("rpad", signature(x = "Column", len = "numeric", pad = "character"),
#'substring_index(df$c, '.', 2)
#'substring_index(df$c, '.', -1)
#'}
+#' @note substring_index since 1.5.0
setMethod("substring_index",
signature(x = "Column", delim = "character", count = "numeric"),
function(x, delim, count) {
@@ -2436,6 +2571,7 @@ setMethod("substring_index",
#' @name translate
#' @export
#' @examples \dontrun{translate(df$c, 'rnlt', '123')}
+#' @note translate since 1.5.0
setMethod("translate",
signature(x = "Column", matchingString = "character", replaceString = "character"),
function(x, matchingString, replaceString) {
@@ -2458,6 +2594,7 @@ setMethod("translate",
#'unix_timestamp(df$t)
#'unix_timestamp(df$t, 'yyyy-MM-dd HH')
#'}
+#' @note unix_timestamp since 1.5.0
setMethod("unix_timestamp", signature(x = "missing", format = "missing"),
function(x, format) {
jc <- callJStatic("org.apache.spark.sql.functions", "unix_timestamp")
@@ -2467,6 +2604,7 @@ setMethod("unix_timestamp", signature(x = "missing", format = "missing"),
#' @rdname unix_timestamp
#' @name unix_timestamp
#' @export
+#' @note unix_timestamp(Column) since 1.5.0
setMethod("unix_timestamp", signature(x = "Column", format = "missing"),
function(x, format) {
jc <- callJStatic("org.apache.spark.sql.functions", "unix_timestamp", x@jc)
@@ -2476,6 +2614,7 @@ setMethod("unix_timestamp", signature(x = "Column", format = "missing"),
#' @rdname unix_timestamp
#' @name unix_timestamp
#' @export
+#' @note unix_timestamp(Column, character) since 1.5.0
setMethod("unix_timestamp", signature(x = "Column", format = "character"),
function(x, format = "yyyy-MM-dd HH:mm:ss") {
jc <- callJStatic("org.apache.spark.sql.functions", "unix_timestamp", x@jc, format)
@@ -2492,6 +2631,7 @@ setMethod("unix_timestamp", signature(x = "Column", format = "character"),
#' @seealso \link{ifelse}
#' @export
#' @examples \dontrun{when(df$age == 2, df$age + 1)}
+#' @note when since 1.5.0
setMethod("when", signature(condition = "Column", value = "ANY"),
function(condition, value) {
condition <- condition@jc
@@ -2514,6 +2654,7 @@ setMethod("when", signature(condition = "Column", value = "ANY"),
#' ifelse(df$a > 1 & df$b > 2, 0, 1)
#' ifelse(df$a > 1, df$a, 1)
#' }
+#' @note ifelse since 1.5.0
setMethod("ifelse",
signature(test = "Column", yes = "ANY", no = "ANY"),
function(test, yes, no) {
@@ -2544,6 +2685,7 @@ setMethod("ifelse",
#' @family window_funcs
#' @export
#' @examples \dontrun{cume_dist()}
+#' @note cume_dist since 1.6.0
setMethod("cume_dist",
signature(x = "missing"),
function() {
@@ -2566,6 +2708,7 @@ setMethod("cume_dist",
#' @family window_funcs
#' @export
#' @examples \dontrun{dense_rank()}
+#' @note dense_rank since 1.6.0
setMethod("dense_rank",
signature(x = "missing"),
function() {
@@ -2586,6 +2729,7 @@ setMethod("dense_rank",
#' @family window_funcs
#' @export
#' @examples \dontrun{lag(df$c)}
+#' @note lag since 1.6.0
setMethod("lag",
signature(x = "characterOrColumn"),
function(x, offset, defaultValue = NULL) {
@@ -2613,6 +2757,7 @@ setMethod("lag",
#' @family window_funcs
#' @export
#' @examples \dontrun{lead(df$c)}
+#' @note lead since 1.6.0
setMethod("lead",
signature(x = "characterOrColumn", offset = "numeric", defaultValue = "ANY"),
function(x, offset, defaultValue = NULL) {
@@ -2640,6 +2785,7 @@ setMethod("lead",
#' @family window_funcs
#' @export
#' @examples \dontrun{ntile(1)}
+#' @note ntile since 1.6.0
setMethod("ntile",
signature(x = "numeric"),
function(x) {
@@ -2662,6 +2808,7 @@ setMethod("ntile",
#' @family window_funcs
#' @export
#' @examples \dontrun{percent_rank()}
+#' @note percent_rank since 1.6.0
setMethod("percent_rank",
signature(x = "missing"),
function() {
@@ -2685,6 +2832,7 @@ setMethod("percent_rank",
#' @family window_funcs
#' @export
#' @examples \dontrun{rank()}
+#' @note rank since 1.6.0
setMethod("rank",
signature(x = "missing"),
function() {
@@ -2710,6 +2858,7 @@ setMethod("rank",
#' @family window_funcs
#' @export
#' @examples \dontrun{row_number()}
+#' @note row_number since 1.6.0
setMethod("row_number",
signature(x = "missing"),
function() {
@@ -2730,6 +2879,7 @@ setMethod("row_number",
#' @family collection_funcs
#' @export
#' @examples \dontrun{array_contains(df$c, 1)}
+#' @note array_contains since 1.6.0
setMethod("array_contains",
signature(x = "Column", value = "ANY"),
function(x, value) {
@@ -2746,6 +2896,7 @@ setMethod("array_contains",
#' @family collection_funcs
#' @export
#' @examples \dontrun{explode(df$c)}
+#' @note explode since 1.5.0
setMethod("explode",
signature(x = "Column"),
function(x) {
@@ -2762,6 +2913,7 @@ setMethod("explode",
#' @family collection_funcs
#' @export
#' @examples \dontrun{size(df$c)}
+#' @note size since 1.5.0
setMethod("size",
signature(x = "Column"),
function(x) {
@@ -2787,6 +2939,7 @@ setMethod("size",
#' sort_array(df$c)
#' sort_array(df$c, FALSE)
#' }
+#' @note sort_array since 1.6.0
setMethod("sort_array",
signature(x = "Column"),
function(x, asc = TRUE) {
diff --git a/R/pkg/R/group.R b/R/pkg/R/group.R
index 65b9e841be..51e151623c 100644
--- a/R/pkg/R/group.R
+++ b/R/pkg/R/group.R
@@ -31,6 +31,7 @@ setOldClass("jobj")
#'
#' @param sgd A Java object reference to the backing Scala GroupedData
#' @export
+#' @note GroupedData since 1.4.0
setClass("GroupedData",
slots = list(sgd = "jobj"))
@@ -46,6 +47,7 @@ groupedData <- function(sgd) {
#' @rdname show
+#' @note show(GroupedData) since 1.4.0
setMethod("show", "GroupedData",
function(object) {
cat("GroupedData\n")
@@ -64,6 +66,7 @@ setMethod("show", "GroupedData",
#' \dontrun{
#' count(groupBy(df, "name"))
#' }
+#' @note count since 1.4.0
setMethod("count",
signature(x = "GroupedData"),
function(x) {
@@ -90,6 +93,7 @@ setMethod("count",
#' df3 <- agg(df, ageSum = sum(df$age)) # Creates a new column named ageSum
#' df4 <- summarize(df, ageSum = max(df$age))
#' }
+#' @note agg since 1.4.0
setMethod("agg",
signature(x = "GroupedData"),
function(x, ...) {
@@ -117,6 +121,7 @@ setMethod("agg",
#' @rdname summarize
#' @name summarize
+#' @note summarize since 1.4.0
setMethod("summarize",
signature(x = "GroupedData"),
function(x, ...) {
@@ -190,6 +195,7 @@ createMethods()
#' 3 3 3.0
#' 1 1 1.5
#' }
+#' @note gapply(GroupedData) since 2.0.0
setMethod("gapply",
signature(x = "GroupedData"),
function(x, func, schema) {
diff --git a/R/pkg/R/jobj.R b/R/pkg/R/jobj.R
index 5b9142bbd3..4905e1fe5c 100644
--- a/R/pkg/R/jobj.R
+++ b/R/pkg/R/jobj.R
@@ -71,6 +71,7 @@ jobj <- function(objId) {
#'
#' @param x The JVM object reference
#' @param ... further arguments passed to or from other methods
+#' @note print.jobj since 1.4.0
print.jobj <- function(x, ...) {
name <- getClassName.jobj(x)
cat("Java ref type", name, "id", x$id, "\n", sep = " ")
diff --git a/R/pkg/R/mllib.R b/R/pkg/R/mllib.R
index 2127daee18..d6ff2aa22d 100644
--- a/R/pkg/R/mllib.R
+++ b/R/pkg/R/mllib.R
@@ -29,24 +29,28 @@
#'
#' @param jobj a Java object reference to the backing Scala GeneralizedLinearRegressionWrapper
#' @export
+#' @note GeneralizedLinearRegressionModel since 2.0.0
setClass("GeneralizedLinearRegressionModel", representation(jobj = "jobj"))
#' S4 class that represents a NaiveBayesModel
#'
#' @param jobj a Java object reference to the backing Scala NaiveBayesWrapper
#' @export
+#' @note NaiveBayesModel since 2.0.0
setClass("NaiveBayesModel", representation(jobj = "jobj"))
#' S4 class that represents a AFTSurvivalRegressionModel
#'
#' @param jobj a Java object reference to the backing Scala AFTSurvivalRegressionWrapper
#' @export
+#' @note AFTSurvivalRegressionModel since 2.0.0
setClass("AFTSurvivalRegressionModel", representation(jobj = "jobj"))
#' S4 class that represents a KMeansModel
#'
#' @param jobj a Java object reference to the backing Scala KMeansModel
#' @export
+#' @note KMeansModel since 2.0.0
setClass("KMeansModel", representation(jobj = "jobj"))
#' Fits a generalized linear model
@@ -73,6 +77,7 @@ setClass("KMeansModel", representation(jobj = "jobj"))
#' model <- spark.glm(df, Sepal_Length ~ Sepal_Width, family="gaussian")
#' summary(model)
#' }
+#' @note spark.glm since 2.0.0
setMethod(
"spark.glm",
signature(data = "SparkDataFrame", formula = "formula"),
@@ -120,6 +125,7 @@ setMethod(
#' model <- glm(Sepal_Length ~ Sepal_Width, df, family="gaussian")
#' summary(model)
#' }
+#' @note glm since 1.5.0
setMethod("glm", signature(formula = "formula", family = "ANY", data = "SparkDataFrame"),
function(formula, family = gaussian, data, epsilon = 1e-06, maxit = 25) {
spark.glm(data, formula, family, epsilon, maxit)
@@ -138,6 +144,7 @@ setMethod("glm", signature(formula = "formula", family = "ANY", data = "SparkDat
#' model <- glm(y ~ x, trainingData)
#' summary(model)
#' }
+#' @note summary(GeneralizedLinearRegressionModel) since 2.0.0
setMethod("summary", signature(object = "GeneralizedLinearRegressionModel"),
function(object, ...) {
jobj <- object@jobj
@@ -173,6 +180,7 @@ setMethod("summary", signature(object = "GeneralizedLinearRegressionModel"),
#' @rdname print
#' @name print.summary.GeneralizedLinearRegressionModel
#' @export
+#' @note print.summary.GeneralizedLinearRegressionModel since 2.0.0
print.summary.GeneralizedLinearRegressionModel <- function(x, ...) {
if (x$is.loaded) {
cat("\nSaved-loaded model does not support output 'Deviance Residuals'.\n")
@@ -215,6 +223,7 @@ print.summary.GeneralizedLinearRegressionModel <- function(x, ...) {
#' predicted <- predict(model, testData)
#' showDF(predicted)
#' }
+#' @note predict(GeneralizedLinearRegressionModel) since 1.5.0
setMethod("predict", signature(object = "GeneralizedLinearRegressionModel"),
function(object, newData) {
return(dataFrame(callJMethod(object@jobj, "transform", newData@sdf)))
@@ -236,6 +245,7 @@ setMethod("predict", signature(object = "GeneralizedLinearRegressionModel"),
#' predicted <- predict(model, testData)
#' showDF(predicted)
#'}
+#' @note predict(NaiveBayesModel) since 2.0.0
setMethod("predict", signature(object = "NaiveBayesModel"),
function(object, newData) {
return(dataFrame(callJMethod(object@jobj, "transform", newData@sdf)))
@@ -256,6 +266,7 @@ setMethod("predict", signature(object = "NaiveBayesModel"),
#' model <- spark.naiveBayes(trainingData, y ~ x)
#' summary(model)
#'}
+#' @note summary(NaiveBayesModel) since 2.0.0
setMethod("summary", signature(object = "NaiveBayesModel"),
function(object, ...) {
jobj <- object@jobj
@@ -289,6 +300,7 @@ setMethod("summary", signature(object = "NaiveBayesModel"),
#' \dontrun{
#' model <- spark.kmeans(data, ~ ., k=2, initMode="random")
#' }
+#' @note spark.kmeans since 2.0.0
setMethod("spark.kmeans", signature(data = "SparkDataFrame", formula = "formula"),
function(data, formula, k, maxIter = 10, initMode = c("random", "k-means||")) {
formula <- paste(deparse(formula), collapse = "")
@@ -313,6 +325,7 @@ setMethod("spark.kmeans", signature(data = "SparkDataFrame", formula = "formula"
#' fitted.model <- fitted(model)
#' showDF(fitted.model)
#'}
+#' @note fitted since 2.0.0
setMethod("fitted", signature(object = "KMeansModel"),
function(object, method = c("centers", "classes"), ...) {
method <- match.arg(method)
@@ -339,6 +352,7 @@ setMethod("fitted", signature(object = "KMeansModel"),
#' model <- spark.kmeans(trainingData, ~ ., 2)
#' summary(model)
#' }
+#' @note summary(KMeansModel) since 2.0.0
setMethod("summary", signature(object = "KMeansModel"),
function(object, ...) {
jobj <- object@jobj
@@ -374,6 +388,7 @@ setMethod("summary", signature(object = "KMeansModel"),
#' predicted <- predict(model, testData)
#' showDF(predicted)
#' }
+#' @note predict(KMeansModel) since 2.0.0
setMethod("predict", signature(object = "KMeansModel"),
function(object, newData) {
return(dataFrame(callJMethod(object@jobj, "transform", newData@sdf)))
@@ -396,6 +411,7 @@ setMethod("predict", signature(object = "KMeansModel"),
#' df <- createDataFrame(infert)
#' model <- spark.naiveBayes(df, education ~ ., laplace = 0)
#'}
+#' @note spark.naiveBayes since 2.0.0
setMethod("spark.naiveBayes", signature(data = "SparkDataFrame", formula = "formula"),
function(data, formula, laplace = 0, ...) {
formula <- paste(deparse(formula), collapse = "")
@@ -423,6 +439,7 @@ setMethod("spark.naiveBayes", signature(data = "SparkDataFrame", formula = "form
#' path <- "path/to/model"
#' write.ml(model, path)
#' }
+#' @note write.ml(NaiveBayesModel, character) since 2.0.0
setMethod("write.ml", signature(object = "NaiveBayesModel", path = "character"),
function(object, path, overwrite = FALSE) {
writer <- callJMethod(object@jobj, "write")
@@ -450,6 +467,7 @@ setMethod("write.ml", signature(object = "NaiveBayesModel", path = "character"),
#' path <- "path/to/model"
#' write.ml(model, path)
#' }
+#' @note write.ml(AFTSurvivalRegressionModel, character) since 2.0.0
setMethod("write.ml", signature(object = "AFTSurvivalRegressionModel", path = "character"),
function(object, path, overwrite = FALSE) {
writer <- callJMethod(object@jobj, "write")
@@ -477,6 +495,7 @@ setMethod("write.ml", signature(object = "AFTSurvivalRegressionModel", path = "c
#' path <- "path/to/model"
#' write.ml(model, path)
#' }
+#' @note write.ml(GeneralizedLinearRegressionModel, character) since 2.0.0
setMethod("write.ml", signature(object = "GeneralizedLinearRegressionModel", path = "character"),
function(object, path, overwrite = FALSE) {
writer <- callJMethod(object@jobj, "write")
@@ -504,6 +523,7 @@ setMethod("write.ml", signature(object = "GeneralizedLinearRegressionModel", pat
#' path <- "path/to/model"
#' write.ml(model, path)
#' }
+#' @note write.ml(KMeansModel, character) since 2.0.0
setMethod("write.ml", signature(object = "KMeansModel", path = "character"),
function(object, path, overwrite = FALSE) {
writer <- callJMethod(object@jobj, "write")
@@ -525,6 +545,7 @@ setMethod("write.ml", signature(object = "KMeansModel", path = "character"),
#' path <- "path/to/model"
#' model <- read.ml(path)
#' }
+#' @note read.ml since 2.0.0
read.ml <- function(path) {
path <- suppressWarnings(normalizePath(path))
jobj <- callJStatic("org.apache.spark.ml.r.RWrappers", "load", path)
@@ -558,6 +579,7 @@ read.ml <- function(path) {
#' df <- createDataFrame(ovarian)
#' model <- spark.survreg(df, Surv(futime, fustat) ~ ecog_ps + rx)
#' }
+#' @note spark.survreg since 2.0.0
setMethod("spark.survreg", signature(data = "SparkDataFrame", formula = "formula"),
function(data, formula, ...) {
formula <- paste(deparse(formula), collapse = "")
@@ -581,6 +603,7 @@ setMethod("spark.survreg", signature(data = "SparkDataFrame", formula = "formula
#' model <- spark.survreg(trainingData, Surv(futime, fustat) ~ ecog_ps + rx)
#' summary(model)
#' }
+#' @note summary(AFTSurvivalRegressionModel) since 2.0.0
setMethod("summary", signature(object = "AFTSurvivalRegressionModel"),
function(object, ...) {
jobj <- object@jobj
@@ -608,6 +631,7 @@ setMethod("summary", signature(object = "AFTSurvivalRegressionModel"),
#' predicted <- predict(model, testData)
#' showDF(predicted)
#' }
+#' @note predict(AFTSurvivalRegressionModel) since 2.0.0
setMethod("predict", signature(object = "AFTSurvivalRegressionModel"),
function(object, newData) {
return(dataFrame(callJMethod(object@jobj, "transform", newData@sdf)))
diff --git a/R/pkg/R/schema.R b/R/pkg/R/schema.R
index 89a2cfac3c..fb23c780ad 100644
--- a/R/pkg/R/schema.R
+++ b/R/pkg/R/schema.R
@@ -35,6 +35,7 @@
#' function(key, x) { y <- data.frame(key, mean(x$b), stringsAsFactors = FALSE) },
#' schema)
#' }
+#' @note structType since 1.4.0
structType <- function(x, ...) {
UseMethod("structType", x)
}
@@ -67,6 +68,7 @@ structType.structField <- function(x, ...) {
#'
#' @param x A StructType object
#' @param ... further arguments passed to or from other methods
+#' @note print.structType since 1.4.0
print.structType <- function(x, ...) {
cat("StructType\n",
sapply(x$fields(),
@@ -98,7 +100,7 @@ print.structType <- function(x, ...) {
#' function(key, x) { y <- data.frame(key, mean(x$b), stringsAsFactors = FALSE) },
#' schema)
#' }
-
+#' @note structField since 1.4.0
structField <- function(x, ...) {
UseMethod("structField", x)
}
@@ -202,6 +204,7 @@ structField.character <- function(x, type, nullable = TRUE) {
#'
#' @param x A StructField object
#' @param ... further arguments passed to or from other methods
+#' @note print.structField since 1.4.0
print.structField <- function(x, ...) {
cat("StructField(name = \"", x$name(),
"\", type = \"", x$dataType.toString(),
diff --git a/R/pkg/R/sparkR.R b/R/pkg/R/sparkR.R
index d05660ccf6..94d0e63c8a 100644
--- a/R/pkg/R/sparkR.R
+++ b/R/pkg/R/sparkR.R
@@ -31,6 +31,7 @@ connExists <- function(env) {
#' @rdname sparkR.session.stop
#' @name sparkR.stop
#' @export
+#' @note sparkR.stop since 1.4.0
sparkR.stop <- function() {
sparkR.session.stop()
}
@@ -41,7 +42,7 @@ sparkR.stop <- function() {
#' @rdname sparkR.session.stop
#' @name sparkR.session.stop
#' @export
-#' @note since 2.0.0
+#' @note sparkR.session.stop since 2.0.0
sparkR.session.stop <- function() {
env <- .sparkREnv
if (exists(".sparkRCon", envir = env)) {
@@ -112,7 +113,7 @@ sparkR.session.stop <- function() {
#' c("one.jar", "two.jar", "three.jar"),
#' c("com.databricks:spark-avro_2.10:2.0.1"))
#'}
-
+#' @note sparkR.init since 1.4.0
sparkR.init <- function(
master = "",
appName = "SparkR",
@@ -265,7 +266,7 @@ sparkR.sparkContext <- function(
#' sc <- sparkR.init()
#' sqlContext <- sparkRSQL.init(sc)
#'}
-
+#' @note sparkRSQL.init since 1.4.0
sparkRSQL.init <- function(jsc = NULL) {
.Deprecated("sparkR.session")
@@ -293,7 +294,7 @@ sparkRSQL.init <- function(jsc = NULL) {
#' sc <- sparkR.init()
#' sqlContext <- sparkRHive.init(sc)
#'}
-
+#' @note sparkRHive.init since 1.4.0
sparkRHive.init <- function(jsc = NULL) {
.Deprecated("sparkR.session")
@@ -334,8 +335,7 @@ sparkRHive.init <- function(jsc = NULL) {
#' c("com.databricks:spark-avro_2.10:2.0.1"))
#' sparkR.session(spark.master = "yarn-client", spark.executor.memory = "4g")
#'}
-#' @note since 2.0.0
-
+#' @note sparkR.session since 2.0.0
sparkR.session <- function(
master = "",
appName = "SparkR",
@@ -399,7 +399,7 @@ sparkR.session <- function(
#' sc <- sparkR.init()
#' setJobGroup(sc, "myJobGroup", "My job group description", TRUE)
#'}
-
+#' @note setJobGroup since 1.5.0
setJobGroup <- function(sc, groupId, description, interruptOnCancel) {
callJMethod(sc, "setJobGroup", groupId, description, interruptOnCancel)
}
@@ -412,7 +412,7 @@ setJobGroup <- function(sc, groupId, description, interruptOnCancel) {
#' sc <- sparkR.init()
#' clearJobGroup(sc)
#'}
-
+#' @note clearJobGroup since 1.5.0
clearJobGroup <- function(sc) {
callJMethod(sc, "clearJobGroup")
}
@@ -426,7 +426,7 @@ clearJobGroup <- function(sc) {
#' sc <- sparkR.init()
#' cancelJobGroup(sc, "myJobGroup")
#'}
-
+#' @note cancelJobGroup since 1.5.0
cancelJobGroup <- function(sc, groupId) {
callJMethod(sc, "cancelJobGroup", groupId)
}
diff --git a/R/pkg/R/stats.R b/R/pkg/R/stats.R
index 6b53517873..e92b9e3d84 100644
--- a/R/pkg/R/stats.R
+++ b/R/pkg/R/stats.R
@@ -40,6 +40,7 @@ setOldClass("jobj")
#' df <- read.json("/path/to/file.json")
#' ct <- crosstab(df, "title", "gender")
#' }
+#' @note crosstab since 1.5.0
setMethod("crosstab",
signature(x = "SparkDataFrame", col1 = "character", col2 = "character"),
function(x, col1, col2) {
@@ -65,6 +66,7 @@ setMethod("crosstab",
#' df <- read.json("/path/to/file.json")
#' cov <- cov(df, "title", "gender")
#' }
+#' @note cov since 1.6.0
setMethod("cov",
signature(x = "SparkDataFrame"),
function(x, col1, col2) {
@@ -95,6 +97,7 @@ setMethod("cov",
#' corr <- corr(df, "title", "gender")
#' corr <- corr(df, "title", "gender", method = "pearson")
#' }
+#' @note corr since 1.6.0
setMethod("corr",
signature(x = "SparkDataFrame"),
function(x, col1, col2, method = "pearson") {
@@ -123,6 +126,7 @@ setMethod("corr",
#' df <- read.json("/path/to/file.json")
#' fi = freqItems(df, c("title", "gender"))
#' }
+#' @note freqItems since 1.6.0
setMethod("freqItems", signature(x = "SparkDataFrame", cols = "character"),
function(x, cols, support = 0.01) {
statFunctions <- callJMethod(x@sdf, "stat")
@@ -160,6 +164,7 @@ setMethod("freqItems", signature(x = "SparkDataFrame", cols = "character"),
#' df <- read.json("/path/to/file.json")
#' quantiles <- approxQuantile(df, "key", c(0.5, 0.8), 0.0)
#' }
+#' @note approxQuantile since 2.0.0
setMethod("approxQuantile",
signature(x = "SparkDataFrame", col = "character",
probabilities = "numeric", relativeError = "numeric"),
@@ -188,6 +193,7 @@ setMethod("approxQuantile",
#' df <- read.json("/path/to/file.json")
#' sample <- sampleBy(df, "key", fractions, 36)
#' }
+#' @note sampleBy since 1.6.0
setMethod("sampleBy",
signature(x = "SparkDataFrame", col = "character",
fractions = "list", seed = "numeric"),
diff --git a/R/pkg/R/utils.R b/R/pkg/R/utils.R
index aafb34472f..d5c062d3bc 100644
--- a/R/pkg/R/utils.R
+++ b/R/pkg/R/utils.R
@@ -115,6 +115,7 @@ isRDD <- function(name, env) {
#' hashCode(1.0) # 1072693248
#' hashCode("1") # 49
#'}
+#' @note hashCode since 1.4.0
hashCode <- function(key) {
if (class(key) == "integer") {
as.integer(key[[1]])
diff --git a/R/pkg/R/window.R b/R/pkg/R/window.R
index 7ecf70abc6..e4bc933b9a 100644
--- a/R/pkg/R/window.R
+++ b/R/pkg/R/window.R
@@ -32,6 +32,7 @@
#' ws <- window.partitionBy(df$key1, df$key2)
#' df1 <- select(df, over(lead("value", 1), ws))
#' }
+#' @note window.partitionBy(character) since 2.0.0
setMethod("window.partitionBy",
signature(col = "character"),
function(col, ...) {
@@ -45,6 +46,7 @@ setMethod("window.partitionBy",
#' @rdname window.partitionBy
#' @name window.partitionBy
#' @export
+#' @note window.partitionBy(Column) since 2.0.0
setMethod("window.partitionBy",
signature(col = "Column"),
function(col, ...) {
@@ -72,6 +74,7 @@ setMethod("window.partitionBy",
#' ws <- window.orderBy(df$key1, df$key2)
#' df1 <- select(df, over(lead("value", 1), ws))
#' }
+#' @note window.orderBy(character) since 2.0.0
setMethod("window.orderBy",
signature(col = "character"),
function(col, ...) {
@@ -85,6 +88,7 @@ setMethod("window.orderBy",
#' @rdname window.orderBy
#' @name window.orderBy
#' @export
+#' @note window.orderBy(Column) since 2.0.0
setMethod("window.orderBy",
signature(col = "Column"),
function(col, ...) {