diff options
author | Y.CORP.YAHOO.COM\tgraves <tgraves@thatenemy-lm.champ.corp.yahoo.com> | 2013-09-03 10:50:21 -0500 |
---|---|---|
committer | Y.CORP.YAHOO.COM\tgraves <tgraves@thatenemy-lm.champ.corp.yahoo.com> | 2013-09-03 10:50:21 -0500 |
commit | c8cc27611002f4310bc386df858059700a3fec9d (patch) | |
tree | eba0dc6b9136c2e44ffcf769ed18e06a693b7bd8 /core | |
parent | 547fc4a4120a2bb3a6aeca66dd0cd568534ba65a (diff) | |
download | spark-c8cc27611002f4310bc386df858059700a3fec9d.tar.gz spark-c8cc27611002f4310bc386df858059700a3fec9d.tar.bz2 spark-c8cc27611002f4310bc386df858059700a3fec9d.zip |
Review comment changes and update to org.apache packaging
Diffstat (limited to 'core')
6 files changed, 20 insertions, 32 deletions
diff --git a/core/src/main/scala/org/apache/spark/SparkContext.scala b/core/src/main/scala/org/apache/spark/SparkContext.scala index f2641851cb..89318712a5 100644 --- a/core/src/main/scala/org/apache/spark/SparkContext.scala +++ b/core/src/main/scala/org/apache/spark/SparkContext.scala @@ -196,7 +196,7 @@ class SparkContext( case "yarn-standalone" => val scheduler = try { - val clazz = Class.forName("spark.scheduler.cluster.YarnClusterScheduler") + val clazz = Class.forName("org.apache.spark.scheduler.cluster.YarnClusterScheduler") val cons = clazz.getConstructor(classOf[SparkContext]) cons.newInstance(this).asInstanceOf[ClusterScheduler] } catch { diff --git a/core/src/main/scala/org/apache/spark/SparkEnv.scala b/core/src/main/scala/org/apache/spark/SparkEnv.scala index 478e5a0aaf..29968c273c 100644 --- a/core/src/main/scala/org/apache/spark/SparkEnv.scala +++ b/core/src/main/scala/org/apache/spark/SparkEnv.scala @@ -62,7 +62,7 @@ class SparkEnv ( val yarnMode = java.lang.Boolean.valueOf(System.getProperty("SPARK_YARN_MODE", System.getenv("SPARK_YARN_MODE"))) if(yarnMode) { try { - Class.forName("spark.deploy.yarn.YarnSparkHadoopUtil").newInstance.asInstanceOf[SparkHadoopUtil] + Class.forName("org.apache.spark.deploy.yarn.YarnSparkHadoopUtil").newInstance.asInstanceOf[SparkHadoopUtil] } catch { case th: Throwable => throw new SparkException("Unable to load YARN support", th) } diff --git a/core/src/main/scala/org/apache/spark/scheduler/cluster/SparkDeploySchedulerBackend.scala b/core/src/main/scala/org/apache/spark/scheduler/cluster/SparkDeploySchedulerBackend.scala index 3f8a96fc47..9a2cf20de7 100644 --- a/core/src/main/scala/org/apache/spark/scheduler/cluster/SparkDeploySchedulerBackend.scala +++ b/core/src/main/scala/org/apache/spark/scheduler/cluster/SparkDeploySchedulerBackend.scala @@ -50,7 +50,7 @@ private[spark] class SparkDeploySchedulerBackend( "org.apache.spark.executor.StandaloneExecutorBackend", args, sc.executorEnvs) val sparkHome = sc.getSparkHome().getOrElse(null) val appDesc = new ApplicationDescription(appName, maxCores, executorMemory, command, sparkHome, - sc.ui.appHttpUIAddress) + "http://" + sc.ui.appUIAddress) client = new Client(sc.env.actorSystem, master, appDesc, this) client.start() diff --git a/core/src/main/scala/org/apache/spark/ui/SparkUI.scala b/core/src/main/scala/org/apache/spark/ui/SparkUI.scala index 5d54420e72..48eb096063 100644 --- a/core/src/main/scala/org/apache/spark/ui/SparkUI.scala +++ b/core/src/main/scala/org/apache/spark/ui/SparkUI.scala @@ -79,7 +79,6 @@ private[spark] class SparkUI(sc: SparkContext) extends Logging { server.foreach(_.stop()) } - private[spark] def appHttpUIAddress = "http://" + appUIAddress private[spark] def appUIAddress = host + ":" + boundPort.getOrElse("-1") } diff --git a/core/src/main/scala/org/apache/spark/ui/UIUtils.scala b/core/src/main/scala/org/apache/spark/ui/UIUtils.scala index 9190b8ac2c..d3af75630e 100644 --- a/core/src/main/scala/org/apache/spark/ui/UIUtils.scala +++ b/core/src/main/scala/org/apache/spark/ui/UIUtils.scala @@ -33,49 +33,38 @@ private[spark] object UIUtils { return uiRoot + resource } - private[spark] val storageStr = addBaseUri("/storage") - private[spark] val stagesStr = addBaseUri("/stages") - private[spark] val envStr = addBaseUri("/environment") - private[spark] val executorsStr = addBaseUri("/executors") - private[spark] val bootstrapMinCssStr = addBaseUri("/static/bootstrap.min.css") - private[spark] val webuiCssStr = addBaseUri("/static/webui.css") - private[spark] val sortTableStr = addBaseUri("/static/sorttable.js") - private[spark] val sparkLogoHdStr = addBaseUri("/static/spark-logo-77x50px-hd.png") - private[spark] val sparkLogoStr = addBaseUri("/static/spark_logo.png") - - /** Returns a spark page with correctly formatted headers */ def headerSparkPage(content: => Seq[Node], sc: SparkContext, title: String, page: Page.Value) : Seq[Node] = { val jobs = page match { - case Stages => <li class="active"><a href={stagesStr}>Stages</a></li> - case _ => <li><a href="/stages">Stages</a></li> + case Stages => <li class="active"><a href={"%s/stages".format(UIUtils.addBaseUri())}>Stages</a></li> + case _ => <li><a href={"%s/stages".format(UIUtils.addBaseUri())}>Stages</a></li> } val storage = page match { - case Storage => <li class="active"><a href={storageStr}>Storage</a></li> - case _ => <li><a href={storageStr}>Storage</a></li> + case Storage => <li class="active"><a href={"%s/storage".format(UIUtils.addBaseUri())}>Storage</a></li> + case _ => <li><a href={"%s/storage".format(UIUtils.addBaseUri())}>Storage</a></li> } val environment = page match { - case Environment => <li class="active"><a href={envStr}>Environment</a></li> - case _ => <li><a href={envStr}>Environment</a></li> + case Environment => <li class="active"><a href={"%s/environment".format(UIUtils.addBaseUri())}>Environment</a></li> + case _ => <li><a href={"%s/environment".format(UIUtils.addBaseUri())}>Environment</a></li> } val executors = page match { - case Executors => <li class="active"><a href={executorsStr}>Executors</a></li> - case _ => <li><a href={executorsStr}>Executors</a></li> + case Executors => <li class="active"><a href={"%s/executors".format(UIUtils.addBaseUri())}>Executors</a></li> + case _ => <li><a href={"%s/executors".format(UIUtils.addBaseUri())}>Executors</a></li> } <html> <head> <meta http-equiv="Content-type" content="text/html; charset=utf-8" /> - <link rel="stylesheet" href={bootstrapMinCssStr} type="text/css" /> - <link rel="stylesheet" href={webuiCssStr} type="text/css" /> - <script src={sortTableStr}></script> + <link rel="stylesheet" href={"%s/static/bootstrap.min.css".format(UIUtils.addBaseUri())} type="text/css" /> + <link rel="stylesheet" href={"%s/static/webui.css".format(UIUtils.addBaseUri())} type="text/css" /> + <script src={"%s/static/sorttable.js".format(UIUtils.addBaseUri())} ></script> <title>{sc.appName} - {title}</title> </head> <body> <div class="navbar navbar-static-top"> <div class="navbar-inner"> - <a href="/" class="brand"><img src={sparkLogoHdStr} /></a> + <a href={"%s/".format(UIUtils.addBaseUri())} class="brand"><img src={"%s/static/spark-logo-77x50px-hd.png".format(UIUtils.addBaseUri())} /></a> <ul class="nav"> {jobs} {storage} @@ -105,9 +94,9 @@ private[spark] object UIUtils { <html> <head> <meta http-equiv="Content-type" content="text/html; charset=utf-8" /> - <link rel="stylesheet" href={bootstrapMinCssStr} type="text/css" /> - <link rel="stylesheet" href="/static/webui.css" type="text/css" /> - <script src={sortTableStr}></script> + <link rel="stylesheet" href={"%s/static/bootstrap.min.css".format(UIUtils.addBaseUri())} type="text/css" /> + <link rel="stylesheet" href={"%s/static/webui.css".format(UIUtils.addBaseUri())} type="text/css" /> + <script src={"%s/static/sorttable.js".format(UIUtils.addBaseUri())} ></script> <title>{title}</title> </head> <body> @@ -115,7 +104,7 @@ private[spark] object UIUtils { <div class="row-fluid"> <div class="span12"> <h3 style="vertical-align: middle; display: inline-block;"> - <img src="/static/spark-logo-77x50px-hd.png" style="margin-right: 15px;" /> + <img src={"%s/static/spark-logo-77x50px-hd.png".format(UIUtils.addBaseUri())} style="margin-right: 15px;" /> {title} </h3> </div> diff --git a/core/src/main/scala/org/apache/spark/util/Utils.scala b/core/src/main/scala/org/apache/spark/util/Utils.scala index bb47fc0a2c..468800b2bd 100644 --- a/core/src/main/scala/org/apache/spark/util/Utils.scala +++ b/core/src/main/scala/org/apache/spark/util/Utils.scala @@ -613,7 +613,7 @@ private[spark] object Utils extends Logging { * A regular expression to match classes of the "core" Spark API that we want to skip when * finding the call site of a method. */ - private val SPARK_CLASS_REGEX = """^spark(\.api\.java)?(\.rdd)?\.[A-Z]""".r + private val SPARK_CLASS_REGEX = """^org\.apache\.spark(\.api\.java)?(\.util)?(\.rdd)?\.[A-Z]""".r private[spark] class CallSiteInfo(val lastSparkMethod: String, val firstUserFile: String, val firstUserLine: Int, val firstUserClass: String) |