aboutsummaryrefslogtreecommitdiff
path: root/sql/core
diff options
context:
space:
mode:
authorMike Timper <mike@aurorafeint.com>2014-10-09 14:02:27 -0700
committerMichael Armbrust <michael@databricks.com>2014-10-09 14:02:27 -0700
commitec4d40e48186af18e25517e0474020720645f583 (patch)
tree1d12e416c2b2c017c8a2861f7c7ec88970844a47 /sql/core
parente7edb723d22869f228b838fd242bf8e6fe73ee19 (diff)
downloadspark-ec4d40e48186af18e25517e0474020720645f583.tar.gz
spark-ec4d40e48186af18e25517e0474020720645f583.tar.bz2
spark-ec4d40e48186af18e25517e0474020720645f583.zip
[SPARK-3853][SQL] JSON Schema support for Timestamp fields
In JSONRDD.scala, add 'case TimestampType' in the enforceCorrectType function and a toTimestamp function. Author: Mike Timper <mike@aurorafeint.com> Closes #2720 from mtimper/master and squashes the following commits: 9386ab8 [Mike Timper] Fix and tests for SPARK-3853
Diffstat (limited to 'sql/core')
-rw-r--r--sql/core/src/main/scala/org/apache/spark/sql/json/JsonRDD.scala10
-rw-r--r--sql/core/src/test/scala/org/apache/spark/sql/json/JsonSuite.scala8
2 files changed, 18 insertions, 0 deletions
diff --git a/sql/core/src/main/scala/org/apache/spark/sql/json/JsonRDD.scala b/sql/core/src/main/scala/org/apache/spark/sql/json/JsonRDD.scala
index 0f27fd13e7..fbc2965e61 100644
--- a/sql/core/src/main/scala/org/apache/spark/sql/json/JsonRDD.scala
+++ b/sql/core/src/main/scala/org/apache/spark/sql/json/JsonRDD.scala
@@ -20,6 +20,7 @@ package org.apache.spark.sql.json
import scala.collection.Map
import scala.collection.convert.Wrappers.{JMapWrapper, JListWrapper}
import scala.math.BigDecimal
+import java.sql.Timestamp
import com.fasterxml.jackson.databind.ObjectMapper
@@ -361,6 +362,14 @@ private[sql] object JsonRDD extends Logging {
}
}
+ private def toTimestamp(value: Any): Timestamp = {
+ value match {
+ case value: java.lang.Integer => new Timestamp(value.asInstanceOf[Int].toLong)
+ case value: java.lang.Long => new Timestamp(value)
+ case value: java.lang.String => Timestamp.valueOf(value)
+ }
+ }
+
private[json] def enforceCorrectType(value: Any, desiredType: DataType): Any ={
if (value == null) {
null
@@ -377,6 +386,7 @@ private[sql] object JsonRDD extends Logging {
case ArrayType(elementType, _) =>
value.asInstanceOf[Seq[Any]].map(enforceCorrectType(_, elementType))
case struct: StructType => asRow(value.asInstanceOf[Map[String, Any]], struct)
+ case TimestampType => toTimestamp(value)
}
}
}
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/json/JsonSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/json/JsonSuite.scala
index 685e788207..3cfcb2b1aa 100644
--- a/sql/core/src/test/scala/org/apache/spark/sql/json/JsonSuite.scala
+++ b/sql/core/src/test/scala/org/apache/spark/sql/json/JsonSuite.scala
@@ -23,6 +23,8 @@ import org.apache.spark.sql.json.JsonRDD.{enforceCorrectType, compatibleType}
import org.apache.spark.sql.QueryTest
import org.apache.spark.sql.test.TestSQLContext._
+import java.sql.Timestamp
+
class JsonSuite extends QueryTest {
import TestJsonData._
TestJsonData
@@ -50,6 +52,12 @@ class JsonSuite extends QueryTest {
val doubleNumber: Double = 1.7976931348623157E308d
checkTypePromotion(doubleNumber.toDouble, enforceCorrectType(doubleNumber, DoubleType))
checkTypePromotion(BigDecimal(doubleNumber), enforceCorrectType(doubleNumber, DecimalType))
+
+ checkTypePromotion(new Timestamp(intNumber), enforceCorrectType(intNumber, TimestampType))
+ checkTypePromotion(new Timestamp(intNumber.toLong),
+ enforceCorrectType(intNumber.toLong, TimestampType))
+ val strDate = "2014-09-30 12:34:56"
+ checkTypePromotion(Timestamp.valueOf(strDate), enforceCorrectType(strDate, TimestampType))
}
test("Get compatible type") {