From b634901bb28070ac5d9a24a9bc7b7640472a54e2 Mon Sep 17 00:00:00 2001 From: Reynold Xin Date: Mon, 4 Jan 2016 21:05:27 -0800 Subject: [SPARK-12600][SQL] follow up: add range check for DecimalType This addresses davies' code review feedback in https://github.com/apache/spark/pull/10559 Author: Reynold Xin Closes #10586 from rxin/remove-deprecated-sql-followup. --- .../main/scala/org/apache/spark/sql/types/DecimalType.scala | 10 ++++++++++ 1 file changed, 10 insertions(+) diff --git a/sql/catalyst/src/main/scala/org/apache/spark/sql/types/DecimalType.scala b/sql/catalyst/src/main/scala/org/apache/spark/sql/types/DecimalType.scala index fdae2e82a0..cf5322125b 100644 --- a/sql/catalyst/src/main/scala/org/apache/spark/sql/types/DecimalType.scala +++ b/sql/catalyst/src/main/scala/org/apache/spark/sql/types/DecimalType.scala @@ -20,6 +20,7 @@ package org.apache.spark.sql.types import scala.reflect.runtime.universe.typeTag import org.apache.spark.annotation.DeveloperApi +import org.apache.spark.sql.AnalysisException import org.apache.spark.sql.catalyst.ScalaReflectionLock import org.apache.spark.sql.catalyst.expressions.Expression @@ -39,6 +40,15 @@ import org.apache.spark.sql.catalyst.expressions.Expression @DeveloperApi case class DecimalType(precision: Int, scale: Int) extends FractionalType { + if (scale > precision) { + throw new AnalysisException( + s"Decimal scale ($scale) cannot be greater than precision ($precision).") + } + + if (precision > DecimalType.MAX_PRECISION) { + throw new AnalysisException(s"DecimalType can only support precision up to 38") + } + // default constructor for Java def this(precision: Int) = this(precision, 0) def this() = this(10) -- cgit v1.2.3