aboutsummaryrefslogtreecommitdiff
path: root/core
diff options
context:
space:
mode:
authorKay Ousterhout <kayousterhout@gmail.com>2015-10-22 11:39:06 -0700
committerKay Ousterhout <kayousterhout@gmail.com>2015-10-22 11:39:06 -0700
commit3535b91ddc9fd05b613a121e09263b0f378bd5fa (patch)
treef6c0247ce50670a513c47d91884172ff32a2222b /core
parent7bb6d31cff279776f90744407291682774cfe1c2 (diff)
downloadspark-3535b91ddc9fd05b613a121e09263b0f378bd5fa.tar.gz
spark-3535b91ddc9fd05b613a121e09263b0f378bd5fa.tar.bz2
spark-3535b91ddc9fd05b613a121e09263b0f378bd5fa.zip
[SPARK-11163] Remove unnecessary addPendingTask calls.
This commit removes unnecessary calls to addPendingTask in TaskSetManager.executorLost. These calls are unnecessary: for tasks that are still pending and haven't been launched, they're still in all of the correct pending lists, so calling addPendingTask has no effect. For tasks that are currently running (which may still be in the pending lists, depending on how they were scheduled), we call addPendingTask in handleFailedTask, so the calls at the beginning of executorLost are redundant. I think these calls are left over from when we re-computed the locality levels in addPendingTask; now that we call recomputeLocality separately, I don't think these are necessary. Now that those calls are removed, the readding parameter in addPendingTask is no longer necessary, so this commit also removes that parameter. markhamstra can you take a look at this? cc vanzin Author: Kay Ousterhout <kayousterhout@gmail.com> Closes #9154 from kayousterhout/SPARK-11163.
Diffstat (limited to 'core')
-rw-r--r--core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala27
1 files changed, 5 insertions, 22 deletions
diff --git a/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala b/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala
index c02597c436..987800d3d1 100644
--- a/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala
+++ b/core/src/main/scala/org/apache/spark/scheduler/TaskSetManager.scala
@@ -177,14 +177,11 @@ private[spark] class TaskSetManager(
var emittedTaskSizeWarning = false
- /**
- * Add a task to all the pending-task lists that it should be on. If readding is set, we are
- * re-adding the task so only include it in each list if it's not already there.
- */
- private def addPendingTask(index: Int, readding: Boolean = false) {
- // Utility method that adds `index` to a list only if readding=false or it's not already there
+ /** Add a task to all the pending-task lists that it should be on. */
+ private def addPendingTask(index: Int) {
+ // Utility method that adds `index` to a list only if it's not already there
def addTo(list: ArrayBuffer[Int]) {
- if (!readding || !list.contains(index)) {
+ if (!list.contains(index)) {
list += index
}
}
@@ -219,9 +216,7 @@ private[spark] class TaskSetManager(
addTo(pendingTasksWithNoPrefs)
}
- if (!readding) {
- allPendingTasks += index // No point scanning this whole list to find the old task there
- }
+ allPendingTasks += index // No point scanning this whole list to find the old task there
}
/**
@@ -783,18 +778,6 @@ private[spark] class TaskSetManager(
/** Called by TaskScheduler when an executor is lost so we can re-enqueue our tasks */
override def executorLost(execId: String, host: String, reason: ExecutorLossReason) {
- logInfo("Re-queueing tasks for " + execId + " from TaskSet " + taskSet.id)
-
- // Re-enqueue pending tasks for this host based on the status of the cluster. Note
- // that it's okay if we add a task to the same queue twice (if it had multiple preferred
- // locations), because dequeueTaskFromList will skip already-running tasks.
- for (index <- getPendingTasksForExecutor(execId)) {
- addPendingTask(index, readding = true)
- }
- for (index <- getPendingTasksForHost(host)) {
- addPendingTask(index, readding = true)
- }
-
// Re-enqueue any tasks that ran on the failed executor if this is a shuffle map stage,
// and we are not using an external shuffle server which could serve the shuffle outputs.
// The reason is the next stage wouldn't be able to fetch the data from this dead executor