aboutsummaryrefslogtreecommitdiff
path: root/sql/core/src/test/resources/sql-tests/results
diff options
context:
space:
mode:
authorDilip Biswal <dbiswal@us.ibm.com>2017-04-04 09:53:05 +0900
committerTakuya UESHIN <ueshin@databricks.com>2017-04-04 09:53:05 +0900
commit3bfb639cb7352aec572ef6686d3471bd78748ffa (patch)
tree9d2054aecd8b3356a4ec7a38543def25a3ee380e /sql/core/src/test/resources/sql-tests/results
parente7877fd4728ed41e440d7c4d8b6b02bd0d9e873e (diff)
downloadspark-3bfb639cb7352aec572ef6686d3471bd78748ffa.tar.gz
spark-3bfb639cb7352aec572ef6686d3471bd78748ffa.tar.bz2
spark-3bfb639cb7352aec572ef6686d3471bd78748ffa.zip
[SPARK-10364][SQL] Support Parquet logical type TIMESTAMP_MILLIS
## What changes were proposed in this pull request? **Description** from JIRA The TimestampType in Spark SQL is of microsecond precision. Ideally, we should convert Spark SQL timestamp values into Parquet TIMESTAMP_MICROS. But unfortunately parquet-mr hasn't supported it yet. For the read path, we should be able to read TIMESTAMP_MILLIS Parquet values and pad a 0 microsecond part to read values. For the write path, currently we are writing timestamps as INT96, similar to Impala and Hive. One alternative is that, we can have a separate SQL option to let users be able to write Spark SQL timestamp values as TIMESTAMP_MILLIS. Of course, in this way the microsecond part will be truncated. ## How was this patch tested? Added new tests in ParquetQuerySuite and ParquetIOSuite Author: Dilip Biswal <dbiswal@us.ibm.com> Closes #15332 from dilipbiswal/parquet-time-millis.
Diffstat (limited to 'sql/core/src/test/resources/sql-tests/results')
0 files changed, 0 insertions, 0 deletions