From c71c6853fcecd3b41de0e8273329ea83a9779320 Mon Sep 17 00:00:00 2001 From: Sameer Agarwal Date: Mon, 25 Apr 2016 18:22:06 -0700 Subject: [SPARK-14870][SQL][FOLLOW-UP] Move decimalDataWithNulls in DataFrameAggregateSuite ## What changes were proposed in this pull request? Minor followup to https://github.com/apache/spark/pull/12651 ## How was this patch tested? Test-only change Author: Sameer Agarwal Closes #12674 from sameeragarwal/tpcds-fix-2. --- .../org/apache/spark/sql/DataFrameAggregateSuite.scala | 9 +++++++++ .../scala/org/apache/spark/sql/test/SQLTestData.scala | 15 --------------- 2 files changed, 9 insertions(+), 15 deletions(-) (limited to 'sql') diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala index 0fcfb97d2b..2f685c5f9c 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala @@ -21,6 +21,7 @@ import org.apache.spark.sql.expressions.Window import org.apache.spark.sql.functions._ import org.apache.spark.sql.internal.SQLConf import org.apache.spark.sql.test.SharedSQLContext +import org.apache.spark.sql.test.SQLTestData.DecimalData import org.apache.spark.sql.types.DecimalType case class Fact(date: Int, hour: Int, minute: Int, room_name: String, temp: Double) @@ -69,6 +70,14 @@ class DataFrameAggregateSuite extends QueryTest with SharedSQLContext { Row(new java.math.BigDecimal(3.0), new java.math.BigDecimal(3.0))) ) + val decimalDataWithNulls = sqlContext.sparkContext.parallelize( + DecimalData(1, 1) :: + DecimalData(1, null) :: + DecimalData(2, 1) :: + DecimalData(2, null) :: + DecimalData(3, 1) :: + DecimalData(3, 2) :: + DecimalData(null, 2) :: Nil).toDF() checkAnswer( decimalDataWithNulls.groupBy("a").agg(sum("b")), Seq(Row(new java.math.BigDecimal(1.0), new java.math.BigDecimal(1.0)), diff --git a/sql/core/src/test/scala/org/apache/spark/sql/test/SQLTestData.scala b/sql/core/src/test/scala/org/apache/spark/sql/test/SQLTestData.scala index c5f25fa1df..7fa6760b71 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/test/SQLTestData.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/test/SQLTestData.scala @@ -103,19 +103,6 @@ private[sql] trait SQLTestData { self => df } - protected lazy val decimalDataWithNulls: DataFrame = { - val df = sqlContext.sparkContext.parallelize( - DecimalDataWithNulls(1, 1) :: - DecimalDataWithNulls(1, null) :: - DecimalDataWithNulls(2, 1) :: - DecimalDataWithNulls(2, null) :: - DecimalDataWithNulls(3, 1) :: - DecimalDataWithNulls(3, 2) :: - DecimalDataWithNulls(null, 2) :: Nil).toDF() - df.registerTempTable("decimalDataWithNulls") - df - } - protected lazy val binaryData: DataFrame = { val df = sqlContext.sparkContext.parallelize( BinaryData("12".getBytes(StandardCharsets.UTF_8), 1) :: @@ -280,7 +267,6 @@ private[sql] trait SQLTestData { self => negativeData largeAndSmallInts decimalData - decimalDataWithNulls binaryData upperCaseData lowerCaseData @@ -310,7 +296,6 @@ private[sql] object SQLTestData { case class TestData3(a: Int, b: Option[Int]) case class LargeAndSmallInts(a: Int, b: Int) case class DecimalData(a: BigDecimal, b: BigDecimal) - case class DecimalDataWithNulls(a: BigDecimal, b: BigDecimal) case class BinaryData(a: Array[Byte], b: Int) case class UpperCaseData(N: Int, L: String) case class LowerCaseData(n: Int, l: String) -- cgit v1.2.3