diff options
author | Shivaram Venkataraman <shivaram@eecs.berkeley.edu> | 2013-10-09 04:24:34 -0700 |
---|---|---|
committer | Shivaram Venkataraman <shivaram@eecs.berkeley.edu> | 2013-10-09 04:24:34 -0700 |
commit | 484166d520e239d0da9501c3863435a59e235242 (patch) | |
tree | c3fe73af94ad28f31a7b6170116454fb2089a546 /project/SparkBuild.scala | |
parent | a106ed8b97e707b36818c11d1d7211fa28636178 (diff) | |
download | spark-484166d520e239d0da9501c3863435a59e235242.tar.gz spark-484166d520e239d0da9501c3863435a59e235242.tar.bz2 spark-484166d520e239d0da9501c3863435a59e235242.zip |
Add new SBT target for dependency assembly
Diffstat (limited to 'project/SparkBuild.scala')
-rw-r--r-- | project/SparkBuild.scala | 8 |
1 files changed, 7 insertions, 1 deletions
diff --git a/project/SparkBuild.scala b/project/SparkBuild.scala index d038a4f479..6d6a8ef992 100644 --- a/project/SparkBuild.scala +++ b/project/SparkBuild.scala @@ -56,6 +56,8 @@ object SparkBuild extends Build { lazy val assemblyProj = Project("assembly", file("assembly"), settings = assemblyProjSettings) .dependsOn(core, bagel, mllib, repl, streaming) dependsOn(maybeYarn: _*) + lazy val spark = TaskKey[Unit]("spark", "Build assembly of dependencies and spark packages") + // A configuration to set an alternative publishLocalConfiguration lazy val MavenCompile = config("m2r") extend(Compile) lazy val publishLocalBoth = TaskKey[Unit]("publish-local", "publish local for m2 and ivy") @@ -73,6 +75,8 @@ object SparkBuild extends Build { lazy val allProjects = Seq[ProjectReference]( core, repl, examples, bagel, streaming, mllib, tools, assemblyProj) ++ maybeYarnRef + lazy val packageProjects = Seq[ProjectReference](core, repl, bagel, streaming, mllib, tools) ++ maybeYarnRef + def sharedSettings = Defaults.defaultSettings ++ Seq( organization := "org.apache.spark", version := "0.8.0-SNAPSHOT", @@ -288,7 +292,9 @@ object SparkBuild extends Build { def assemblyProjSettings = sharedSettings ++ Seq( name := "spark-assembly", - jarName in assembly <<= version map { v => "spark-assembly-" + v + "-hadoop" + hadoopVersion + ".jar" } + spark in Compile <<= (packageProjects.map(packageBin in Compile in _) ++ Seq(packageDependency in Compile)).dependOn, + jarName in assembly <<= version map { v => "spark-assembly-" + v + "-hadoop" + hadoopVersion + ".jar" }, + jarName in packageDependency <<= version map { v => "spark-assembly-" + v + "-hadoop" + hadoopVersion + "-deps.jar" } ) ++ assemblySettings ++ extraAssemblySettings def extraAssemblySettings() = Seq( |