aboutsummaryrefslogtreecommitdiff
path: root/project/SparkBuild.scala
diff options
context:
space:
mode:
authorSean Owen <sowen@cloudera.com>2014-03-02 14:27:53 -0800
committerPatrick Wendell <pwendell@gmail.com>2014-03-02 14:27:53 -0800
commitfd31adbf27d824f00f62646e13c23f632d1b77d3 (patch)
treec0a03717075b62488f5720a5c49fe4d8cdd3fba6 /project/SparkBuild.scala
parent353ac6b4fa8f0740a7faabc1309022420c7b52fc (diff)
downloadspark-fd31adbf27d824f00f62646e13c23f632d1b77d3.tar.gz
spark-fd31adbf27d824f00f62646e13c23f632d1b77d3.tar.bz2
spark-fd31adbf27d824f00f62646e13c23f632d1b77d3.zip
SPARK-1084.2 (resubmitted)
(Ported from https://github.com/apache/incubator-spark/pull/650 ) This adds one more change though, to fix the scala version warning introduced by json4s recently. Author: Sean Owen <sowen@cloudera.com> Closes #32 from srowen/SPARK-1084.2 and squashes the following commits: 9240abd [Sean Owen] Avoid scala version conflict in scalap induced by json4s dependency 1561cec [Sean Owen] Remove "exclude *" dependencies that are causing Maven warnings, and that are apparently unneeded anyway
Diffstat (limited to 'project/SparkBuild.scala')
-rw-r--r--project/SparkBuild.scala20
1 files changed, 9 insertions, 11 deletions
diff --git a/project/SparkBuild.scala b/project/SparkBuild.scala
index 399886ca1b..d45f6773fa 100644
--- a/project/SparkBuild.scala
+++ b/project/SparkBuild.scala
@@ -243,13 +243,11 @@ object SparkBuild extends Build {
val slf4jVersion = "1.7.5"
- val excludeCglib = ExclusionRule(organization = "org.sonatype.sisu.inject")
- val excludeJackson = ExclusionRule(organization = "org.codehaus.jackson")
val excludeNetty = ExclusionRule(organization = "org.jboss.netty")
val excludeAsm = ExclusionRule(organization = "asm")
- val excludeSnappy = ExclusionRule(organization = "org.xerial.snappy")
val excludeCommonsLogging = ExclusionRule(organization = "commons-logging")
val excludeSLF4J = ExclusionRule(organization = "org.slf4j")
+ val excludeScalap = ExclusionRule(organization = "org.scala-lang", artifact = "scalap")
def coreSettings = sharedSettings ++ Seq(
name := "spark-core",
@@ -273,13 +271,13 @@ object SparkBuild extends Build {
"org.spark-project.akka" %% "akka-remote" % "2.2.3-shaded-protobuf" excludeAll(excludeNetty),
"org.spark-project.akka" %% "akka-slf4j" % "2.2.3-shaded-protobuf" excludeAll(excludeNetty),
"org.spark-project.akka" %% "akka-testkit" % "2.2.3-shaded-protobuf" % "test",
- "org.json4s" %% "json4s-jackson" % "3.2.6",
+ "org.json4s" %% "json4s-jackson" % "3.2.6" excludeAll(excludeScalap),
"it.unimi.dsi" % "fastutil" % "6.4.4",
"colt" % "colt" % "1.2.0",
"org.apache.mesos" % "mesos" % "0.13.0",
"net.java.dev.jets3t" % "jets3t" % "0.7.1" excludeAll(excludeCommonsLogging),
"org.apache.derby" % "derby" % "10.4.2.0" % "test",
- "org.apache.hadoop" % hadoopClient % hadoopVersion excludeAll(excludeJackson, excludeNetty, excludeAsm, excludeCglib, excludeCommonsLogging, excludeSLF4J),
+ "org.apache.hadoop" % hadoopClient % hadoopVersion excludeAll(excludeNetty, excludeAsm, excludeCommonsLogging, excludeSLF4J),
"org.apache.curator" % "curator-recipes" % "2.4.0" excludeAll(excludeNetty),
"com.codahale.metrics" % "metrics-core" % "3.0.0",
"com.codahale.metrics" % "metrics-jvm" % "3.0.0",
@@ -316,7 +314,7 @@ object SparkBuild extends Build {
exclude("io.netty", "netty")
exclude("jline","jline")
exclude("org.apache.cassandra.deps", "avro")
- excludeAll(excludeSnappy, excludeCglib, excludeSLF4J)
+ excludeAll(excludeSLF4J)
)
) ++ assemblySettings ++ extraAssemblySettings
@@ -379,10 +377,10 @@ object SparkBuild extends Build {
def yarnEnabledSettings = Seq(
libraryDependencies ++= Seq(
// Exclude rule required for all ?
- "org.apache.hadoop" % hadoopClient % hadoopVersion excludeAll(excludeJackson, excludeNetty, excludeAsm, excludeCglib),
- "org.apache.hadoop" % "hadoop-yarn-api" % hadoopVersion excludeAll(excludeJackson, excludeNetty, excludeAsm, excludeCglib),
- "org.apache.hadoop" % "hadoop-yarn-common" % hadoopVersion excludeAll(excludeJackson, excludeNetty, excludeAsm, excludeCglib),
- "org.apache.hadoop" % "hadoop-yarn-client" % hadoopVersion excludeAll(excludeJackson, excludeNetty, excludeAsm, excludeCglib)
+ "org.apache.hadoop" % hadoopClient % hadoopVersion excludeAll(excludeNetty, excludeAsm),
+ "org.apache.hadoop" % "hadoop-yarn-api" % hadoopVersion excludeAll(excludeNetty, excludeAsm),
+ "org.apache.hadoop" % "hadoop-yarn-common" % hadoopVersion excludeAll(excludeNetty, excludeAsm),
+ "org.apache.hadoop" % "hadoop-yarn-client" % hadoopVersion excludeAll(excludeNetty, excludeAsm)
)
)
@@ -428,7 +426,7 @@ object SparkBuild extends Build {
def flumeSettings() = sharedSettings ++ Seq(
name := "spark-streaming-flume",
libraryDependencies ++= Seq(
- "org.apache.flume" % "flume-ng-sdk" % "1.2.0" % "compile" excludeAll(excludeNetty, excludeSnappy)
+ "org.apache.flume" % "flume-ng-sdk" % "1.2.0" % "compile" excludeAll(excludeNetty)
)
)