diff options
author | Mingfei <mingfei.shi@intel.com> | 2013-06-08 15:45:47 +0800 |
---|---|---|
committer | Mingfei <mingfei.shi@intel.com> | 2013-06-08 15:45:47 +0800 |
commit | 4fd86e0e10149ad1803831a308a056c7105cbe67 (patch) | |
tree | b44c320a669cf53eaa7b91937e0de619febc0c49 | |
parent | 362f0f93acf38aefa872158fe9cc9d51a43d81a1 (diff) | |
download | spark-4fd86e0e10149ad1803831a308a056c7105cbe67.tar.gz spark-4fd86e0e10149ad1803831a308a056c7105cbe67.tar.bz2 spark-4fd86e0e10149ad1803831a308a056c7105cbe67.zip |
delete test code for joblogger in SparkContext
-rw-r--r-- | core/src/main/scala/spark/SparkContext.scala | 3 |
1 files changed, 1 insertions, 2 deletions
diff --git a/core/src/main/scala/spark/SparkContext.scala b/core/src/main/scala/spark/SparkContext.scala index b67a2066c8..70a9d7698c 100644 --- a/core/src/main/scala/spark/SparkContext.scala +++ b/core/src/main/scala/spark/SparkContext.scala @@ -48,7 +48,6 @@ import spark.scheduler.local.LocalScheduler import spark.scheduler.mesos.{CoarseMesosSchedulerBackend, MesosSchedulerBackend} import spark.storage.{BlockManagerUI, StorageStatus, StorageUtils, RDDInfo} import spark.util.{MetadataCleaner, TimeStampedHashMap} -import spark.scheduler.JobLogger /** * Main entry point for Spark functionality. A SparkContext represents the connection to a Spark @@ -510,7 +509,7 @@ class SparkContext( def addSparkListener(listener: SparkListener) { dagScheduler.sparkListeners += listener } - addSparkListener(new JobLogger) + /** * Return a map from the slave to the max memory available for caching and the remaining * memory available for caching. |