aboutsummaryrefslogtreecommitdiff
path: root/sql
diff options
context:
space:
mode:
authorMichael Armbrust <michael@databricks.com>2015-05-18 14:04:04 -0700
committerAndrew Or <andrew@databricks.com>2015-05-18 14:04:04 -0700
commitfcf90b75ccf222bd2f1939addc3f8f052d2bd3ff (patch)
tree08024609e8886e999984e13b8a43e81e586ced29 /sql
parent0b6f503d5337a8387c37cc2c8e544f67c68f7dad (diff)
downloadspark-fcf90b75ccf222bd2f1939addc3f8f052d2bd3ff.tar.gz
spark-fcf90b75ccf222bd2f1939addc3f8f052d2bd3ff.tar.bz2
spark-fcf90b75ccf222bd2f1939addc3f8f052d2bd3ff.zip
[HOTFIX] Fix ORC build break
Fix break caused by merging #6225 and #6194. Author: Michael Armbrust <michael@databricks.com> Closes #6244 from marmbrus/fixOrcBuildBreak and squashes the following commits: b10e47b [Michael Armbrust] [HOTFIX] Fix ORC Build break
Diffstat (limited to 'sql')
-rw-r--r--sql/hive/src/main/scala/org/apache/spark/sql/hive/orc/OrcRelation.scala11
1 files changed, 6 insertions, 5 deletions
diff --git a/sql/hive/src/main/scala/org/apache/spark/sql/hive/orc/OrcRelation.scala b/sql/hive/src/main/scala/org/apache/spark/sql/hive/orc/OrcRelation.scala
index 9708199f07..e10d3a0b68 100644
--- a/sql/hive/src/main/scala/org/apache/spark/sql/hive/orc/OrcRelation.scala
+++ b/sql/hive/src/main/scala/org/apache/spark/sql/hive/orc/OrcRelation.scala
@@ -20,7 +20,7 @@ package org.apache.spark.sql.hive.orc
import java.util.{Objects, Properties}
import org.apache.hadoop.conf.Configuration
-import org.apache.hadoop.fs.Path
+import org.apache.hadoop.fs.{FileStatus, Path}
import org.apache.hadoop.hive.conf.HiveConf.ConfVars
import org.apache.hadoop.hive.ql.io.orc.{OrcInputFormat, OrcOutputFormat, OrcSerde, OrcSplit}
import org.apache.hadoop.hive.serde2.objectinspector.StructObjectInspector
@@ -171,9 +171,10 @@ private[sql] case class OrcRelation(
maybePartitionSpec)
}
- override def buildScan(requiredColumns: Array[String],
+ override def buildScan(
+ requiredColumns: Array[String],
filters: Array[Filter],
- inputPaths: Array[String]): RDD[Row] = {
+ inputPaths: Array[FileStatus]): RDD[Row] = {
val output = StructType(requiredColumns.map(dataSchema(_))).toAttributes
OrcTableScan(output, this, filters, inputPaths).execute()
}
@@ -194,7 +195,7 @@ private[orc] case class OrcTableScan(
attributes: Seq[Attribute],
@transient relation: OrcRelation,
filters: Array[Filter],
- inputPaths: Array[String])
+ inputPaths: Array[FileStatus])
extends Logging
with HiveInspectors {
@@ -256,7 +257,7 @@ private[orc] case class OrcTableScan(
addColumnIds(attributes, relation, conf)
if (inputPaths.nonEmpty) {
- FileInputFormat.setInputPaths(job, inputPaths.map(new Path(_)): _*)
+ FileInputFormat.setInputPaths(job, inputPaths.map(_.getPath): _*)
}
val inputFormatClass =