aboutsummaryrefslogtreecommitdiff
path: root/project
diff options
context:
space:
mode:
Diffstat (limited to 'project')
-rw-r--r--project/MimaBuild.scala6
-rw-r--r--project/MimaExcludes.scala10
-rw-r--r--project/SparkBuild.scala6
3 files changed, 17 insertions, 5 deletions
diff --git a/project/MimaBuild.scala b/project/MimaBuild.scala
index d919b18e09..f0cbf4e57b 100644
--- a/project/MimaBuild.scala
+++ b/project/MimaBuild.scala
@@ -30,7 +30,7 @@ object MimaBuild {
def excludeMember(fullName: String) = Seq(
ProblemFilters.exclude[MissingMethodProblem](fullName),
- // Sometimes excluded methods have default arguments and
+ // Sometimes excluded methods have default arguments and
// they are translated into public methods/fields($default$) in generated
// bytecode. It is not possible to exhaustively list everything.
// But this should be okay.
@@ -91,9 +91,9 @@ object MimaBuild {
def mimaSettings(sparkHome: File, projectRef: ProjectRef) = {
val organization = "org.apache.spark"
- val previousSparkVersion = "1.1.0"
+ val previousSparkVersion = "1.2.0"
val fullId = "spark-" + projectRef.project + "_2.10"
- mimaDefaultSettings ++
+ mimaDefaultSettings ++
Seq(previousArtifact := Some(organization % fullId % previousSparkVersion),
binaryIssueFilters ++= ignoredABIProblems(sparkHome, version.value))
}
diff --git a/project/MimaExcludes.scala b/project/MimaExcludes.scala
index 8a2a865867..6ab2d1b04c 100644
--- a/project/MimaExcludes.scala
+++ b/project/MimaExcludes.scala
@@ -33,6 +33,16 @@ import com.typesafe.tools.mima.core._
object MimaExcludes {
def excludes(version: String) =
version match {
+ case v if v.startsWith("1.3") =>
+ Seq(
+ MimaBuild.excludeSparkPackage("deploy"),
+ MimaBuild.excludeSparkPackage("graphx"),
+ // These are needed if checking against the sbt build, since they are part of
+ // the maven-generated artifacts in the 1.2 build.
+ MimaBuild.excludeSparkPackage("unused"),
+ ProblemFilters.exclude[MissingClassProblem]("com.google.common.base.Optional")
+ )
+
case v if v.startsWith("1.2") =>
Seq(
MimaBuild.excludeSparkPackage("deploy"),
diff --git a/project/SparkBuild.scala b/project/SparkBuild.scala
index c1879ce4ba..efc48e7f67 100644
--- a/project/SparkBuild.scala
+++ b/project/SparkBuild.scala
@@ -210,12 +210,14 @@ object OldDeps {
def versionArtifact(id: String): Option[sbt.ModuleID] = {
val fullId = id + "_2.10"
- Some("org.apache.spark" % fullId % "1.1.0")
+ Some("org.apache.spark" % fullId % "1.2.0")
}
def oldDepsSettings() = Defaults.coreDefaultSettings ++ Seq(
name := "old-deps",
scalaVersion := "2.10.4",
+ // TODO: remove this as soon as 1.2.0 is published on Maven central.
+ resolvers += "spark-staging-1038" at "https://repository.apache.org/content/repositories/orgapachespark-1038/",
retrieveManaged := true,
retrievePattern := "[type]s/[artifact](-[revision])(-[classifier]).[ext]",
libraryDependencies := Seq("spark-streaming-mqtt", "spark-streaming-zeromq",
@@ -385,7 +387,7 @@ object TestSettings {
javaOptions in Test ++= "-Xmx3g -XX:PermSize=128M -XX:MaxNewSize=256m -XX:MaxPermSize=1g"
.split(" ").toSeq,
// This places test scope jars on the classpath of executors during tests.
- javaOptions in Test +=
+ javaOptions in Test +=
"-Dspark.executor.extraClassPath=" + (fullClasspath in Test).value.files.
map(_.getAbsolutePath).mkString(":").stripSuffix(":"),
javaOptions += "-Xmx3g",