diff options
author | Kay Ousterhout <kayousterhout@gmail.com> | 2013-12-24 14:18:39 -0800 |
---|---|---|
committer | Kay Ousterhout <kayousterhout@gmail.com> | 2013-12-24 14:18:39 -0800 |
commit | 1efe3adf560d207f9106ffd4e15934e422adb636 (patch) | |
tree | 713dc7354844451da7c3b20145e1e86c8793ead6 /core/src/main/scala/org | |
parent | b7bfae1afecad0ae79d5d040d2e02e390c272efb (diff) | |
download | spark-1efe3adf560d207f9106ffd4e15934e422adb636.tar.gz spark-1efe3adf560d207f9106ffd4e15934e422adb636.tar.bz2 spark-1efe3adf560d207f9106ffd4e15934e422adb636.zip |
Responded to Reynold's style comments
Diffstat (limited to 'core/src/main/scala/org')
3 files changed, 7 insertions, 6 deletions
diff --git a/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala b/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala index 7409168f7b..dbac6b96ac 100644 --- a/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala +++ b/core/src/main/scala/org/apache/spark/scheduler/TaskSchedulerImpl.scala @@ -46,9 +46,10 @@ import org.apache.spark.scheduler.SchedulingMode.SchedulingMode * we are holding a lock on ourselves. */ private[spark] class TaskSchedulerImpl( - val sc: SparkContext, - val maxTaskFailures : Int = System.getProperty("spark.task.maxFailures", "4").toInt, - isLocal: Boolean = false) extends TaskScheduler with Logging { + val sc: SparkContext, + val maxTaskFailures : Int = System.getProperty("spark.task.maxFailures", "4").toInt, + isLocal: Boolean = false) + extends TaskScheduler with Logging { // How often to check for speculative tasks val SPECULATION_INTERVAL = System.getProperty("spark.speculation.interval", "100").toLong diff --git a/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala b/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala index aa3fb0b35a..c676e73e03 100644 --- a/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala +++ b/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala @@ -17,6 +17,7 @@ package org.apache.spark.scheduler +import java.io.NotSerializableException import java.util.Arrays import scala.collection.mutable.ArrayBuffer @@ -28,8 +29,7 @@ import scala.math.min import org.apache.spark.{ExceptionFailure, FetchFailed, Logging, Resubmitted, SparkEnv, Success, TaskEndReason, TaskKilled, TaskResultLost, TaskState} import org.apache.spark.TaskState.TaskState -import org.apache.spark.util.{SystemClock, Clock} -import java.io.NotSerializableException +import org.apache.spark.util.{Clock, SystemClock} /** diff --git a/core/src/main/scala/org/apache/spark/scheduler/local/LocalBackend.scala b/core/src/main/scala/org/apache/spark/scheduler/local/LocalBackend.scala index 69c1c04843..4edc6a0d3f 100644 --- a/core/src/main/scala/org/apache/spark/scheduler/local/LocalBackend.scala +++ b/core/src/main/scala/org/apache/spark/scheduler/local/LocalBackend.scala @@ -93,7 +93,7 @@ private[spark] class LocalBackend(scheduler: TaskSchedulerImpl, val totalCores: } override def reviveOffers() { - localActor ! ReviveOffers + localActor ! ReviveOffers } override def defaultParallelism() = totalCores |