aboutsummaryrefslogtreecommitdiff
path: root/repl
diff options
context:
space:
mode:
Diffstat (limited to 'repl')
-rw-r--r--repl/scala-2.10/src/test/scala/org/apache/spark/repl/ReplSuite.scala10
-rw-r--r--repl/scala-2.11/src/test/scala/org/apache/spark/repl/ReplSuite.scala8
2 files changed, 9 insertions, 9 deletions
diff --git a/repl/scala-2.10/src/test/scala/org/apache/spark/repl/ReplSuite.scala b/repl/scala-2.10/src/test/scala/org/apache/spark/repl/ReplSuite.scala
index f150fec7db..5674dcd669 100644
--- a/repl/scala-2.10/src/test/scala/org/apache/spark/repl/ReplSuite.scala
+++ b/repl/scala-2.10/src/test/scala/org/apache/spark/repl/ReplSuite.scala
@@ -211,7 +211,7 @@ class ReplSuite extends SparkFunSuite {
}
test("local-cluster mode") {
- val output = runInterpreter("local-cluster[1,1,512]",
+ val output = runInterpreter("local-cluster[1,1,1024]",
"""
|var v = 7
|def getV() = v
@@ -233,7 +233,7 @@ class ReplSuite extends SparkFunSuite {
}
test("SPARK-1199 two instances of same class don't type check.") {
- val output = runInterpreter("local-cluster[1,1,512]",
+ val output = runInterpreter("local-cluster[1,1,1024]",
"""
|case class Sum(exp: String, exp2: String)
|val a = Sum("A", "B")
@@ -256,7 +256,7 @@ class ReplSuite extends SparkFunSuite {
test("SPARK-2576 importing SQLContext.implicits._") {
// We need to use local-cluster to test this case.
- val output = runInterpreter("local-cluster[1,1,512]",
+ val output = runInterpreter("local-cluster[1,1,1024]",
"""
|val sqlContext = new org.apache.spark.sql.SQLContext(sc)
|import sqlContext.implicits._
@@ -325,9 +325,9 @@ class ReplSuite extends SparkFunSuite {
assertDoesNotContain("Exception", output)
assertContains("ret: Array[Foo] = Array(Foo(1),", output)
}
-
+
test("collecting objects of class defined in repl - shuffling") {
- val output = runInterpreter("local-cluster[1,1,512]",
+ val output = runInterpreter("local-cluster[1,1,1024]",
"""
|case class Foo(i: Int)
|val list = List((1, Foo(1)), (1, Foo(2)))
diff --git a/repl/scala-2.11/src/test/scala/org/apache/spark/repl/ReplSuite.scala b/repl/scala-2.11/src/test/scala/org/apache/spark/repl/ReplSuite.scala
index e1cee97de3..bf8997998e 100644
--- a/repl/scala-2.11/src/test/scala/org/apache/spark/repl/ReplSuite.scala
+++ b/repl/scala-2.11/src/test/scala/org/apache/spark/repl/ReplSuite.scala
@@ -209,7 +209,7 @@ class ReplSuite extends SparkFunSuite {
}
test("local-cluster mode") {
- val output = runInterpreter("local-cluster[1,1,512]",
+ val output = runInterpreter("local-cluster[1,1,1024]",
"""
|var v = 7
|def getV() = v
@@ -231,7 +231,7 @@ class ReplSuite extends SparkFunSuite {
}
test("SPARK-1199 two instances of same class don't type check.") {
- val output = runInterpreter("local-cluster[1,1,512]",
+ val output = runInterpreter("local-cluster[1,1,1024]",
"""
|case class Sum(exp: String, exp2: String)
|val a = Sum("A", "B")
@@ -254,7 +254,7 @@ class ReplSuite extends SparkFunSuite {
test("SPARK-2576 importing SQLContext.createDataFrame.") {
// We need to use local-cluster to test this case.
- val output = runInterpreter("local-cluster[1,1,512]",
+ val output = runInterpreter("local-cluster[1,1,1024]",
"""
|val sqlContext = new org.apache.spark.sql.SQLContext(sc)
|import sqlContext.implicits._
@@ -314,7 +314,7 @@ class ReplSuite extends SparkFunSuite {
}
test("collecting objects of class defined in repl - shuffling") {
- val output = runInterpreter("local-cluster[1,1,512]",
+ val output = runInterpreter("local-cluster[1,1,1024]",
"""
|case class Foo(i: Int)
|val list = List((1, Foo(1)), (1, Foo(2)))