diff options
author | Sameer Agarwal <sameer@databricks.com> | 2016-04-25 18:22:06 -0700 |
---|---|---|
committer | Reynold Xin <rxin@databricks.com> | 2016-04-25 18:22:06 -0700 |
commit | c71c6853fcecd3b41de0e8273329ea83a9779320 (patch) | |
tree | 4102e00f2db78b75f0b206bf2399b4382cc48858 | |
parent | cfa64882fc0728d7becf55b8a424926e4ca93887 (diff) | |
download | spark-c71c6853fcecd3b41de0e8273329ea83a9779320.tar.gz spark-c71c6853fcecd3b41de0e8273329ea83a9779320.tar.bz2 spark-c71c6853fcecd3b41de0e8273329ea83a9779320.zip |
[SPARK-14870][SQL][FOLLOW-UP] Move decimalDataWithNulls in DataFrameAggregateSuite
## What changes were proposed in this pull request?
Minor followup to https://github.com/apache/spark/pull/12651
## How was this patch tested?
Test-only change
Author: Sameer Agarwal <sameer@databricks.com>
Closes #12674 from sameeragarwal/tpcds-fix-2.
-rw-r--r-- | sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala | 9 | ||||
-rw-r--r-- | sql/core/src/test/scala/org/apache/spark/sql/test/SQLTestData.scala | 15 |
2 files changed, 9 insertions, 15 deletions
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala index 0fcfb97d2b..2f685c5f9c 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala @@ -21,6 +21,7 @@ import org.apache.spark.sql.expressions.Window import org.apache.spark.sql.functions._ import org.apache.spark.sql.internal.SQLConf import org.apache.spark.sql.test.SharedSQLContext +import org.apache.spark.sql.test.SQLTestData.DecimalData import org.apache.spark.sql.types.DecimalType case class Fact(date: Int, hour: Int, minute: Int, room_name: String, temp: Double) @@ -69,6 +70,14 @@ class DataFrameAggregateSuite extends QueryTest with SharedSQLContext { Row(new java.math.BigDecimal(3.0), new java.math.BigDecimal(3.0))) ) + val decimalDataWithNulls = sqlContext.sparkContext.parallelize( + DecimalData(1, 1) :: + DecimalData(1, null) :: + DecimalData(2, 1) :: + DecimalData(2, null) :: + DecimalData(3, 1) :: + DecimalData(3, 2) :: + DecimalData(null, 2) :: Nil).toDF() checkAnswer( decimalDataWithNulls.groupBy("a").agg(sum("b")), Seq(Row(new java.math.BigDecimal(1.0), new java.math.BigDecimal(1.0)), diff --git a/sql/core/src/test/scala/org/apache/spark/sql/test/SQLTestData.scala b/sql/core/src/test/scala/org/apache/spark/sql/test/SQLTestData.scala index c5f25fa1df..7fa6760b71 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/test/SQLTestData.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/test/SQLTestData.scala @@ -103,19 +103,6 @@ private[sql] trait SQLTestData { self => df } - protected lazy val decimalDataWithNulls: DataFrame = { - val df = sqlContext.sparkContext.parallelize( - DecimalDataWithNulls(1, 1) :: - DecimalDataWithNulls(1, null) :: - DecimalDataWithNulls(2, 1) :: - DecimalDataWithNulls(2, null) :: - DecimalDataWithNulls(3, 1) :: - DecimalDataWithNulls(3, 2) :: - DecimalDataWithNulls(null, 2) :: Nil).toDF() - df.registerTempTable("decimalDataWithNulls") - df - } - protected lazy val binaryData: DataFrame = { val df = sqlContext.sparkContext.parallelize( BinaryData("12".getBytes(StandardCharsets.UTF_8), 1) :: @@ -280,7 +267,6 @@ private[sql] trait SQLTestData { self => negativeData largeAndSmallInts decimalData - decimalDataWithNulls binaryData upperCaseData lowerCaseData @@ -310,7 +296,6 @@ private[sql] object SQLTestData { case class TestData3(a: Int, b: Option[Int]) case class LargeAndSmallInts(a: Int, b: Int) case class DecimalData(a: BigDecimal, b: BigDecimal) - case class DecimalDataWithNulls(a: BigDecimal, b: BigDecimal) case class BinaryData(a: Array[Byte], b: Int) case class UpperCaseData(N: Int, L: String) case class LowerCaseData(n: Int, l: String) |