aboutsummaryrefslogtreecommitdiff
path: root/project
diff options
context:
space:
mode:
authorEric Liang <ekl@databricks.com>2017-03-23 23:30:40 -0700
committerKay Ousterhout <kayousterhout@gmail.com>2017-03-23 23:30:44 -0700
commit8e558041aa0c41ba9fb2ce242daaf6d6ed4d85b7 (patch)
tree6c35820390b647b95394e2d478f12756f1d3d608 /project
parent19596c28b6ef6e7abe0cfccfd2269c2fddf1fdee (diff)
downloadspark-8e558041aa0c41ba9fb2ce242daaf6d6ed4d85b7.tar.gz
spark-8e558041aa0c41ba9fb2ce242daaf6d6ed4d85b7.tar.bz2
spark-8e558041aa0c41ba9fb2ce242daaf6d6ed4d85b7.zip
[SPARK-19820][CORE] Add interface to kill tasks w/ a reason
This commit adds a killTaskAttempt method to SparkContext, to allow users to kill tasks so that they can be re-scheduled elsewhere. This also refactors the task kill path to allow specifying a reason for the task kill. The reason is propagated opaquely through events, and will show up in the UI automatically as `(N killed: $reason)` and `TaskKilled: $reason`. Without this change, there is no way to provide the user feedback through the UI. Currently used reasons are "stage cancelled", "another attempt succeeded", and "killed via SparkContext.killTask". The user can also specify a custom reason through `SparkContext.killTask`. cc rxin In the stage overview UI the reasons are summarized: ![1](https://cloud.githubusercontent.com/assets/14922/23929209/a83b2862-08e1-11e7-8b3e-ae1967bbe2e5.png) Within the stage UI you can see individual task kill reasons: ![2](https://cloud.githubusercontent.com/assets/14922/23929200/9a798692-08e1-11e7-8697-72b27ad8a287.png) Existing tests, tried killing some stages in the UI and verified the messages are as expected. Author: Eric Liang <ekl@databricks.com> Author: Eric Liang <ekl@google.com> Closes #17166 from ericl/kill-reason.
Diffstat (limited to 'project')
-rw-r--r--project/MimaExcludes.scala13
1 files changed, 13 insertions, 0 deletions
diff --git a/project/MimaExcludes.scala b/project/MimaExcludes.scala
index 9925a8ba72..8ce9367c9b 100644
--- a/project/MimaExcludes.scala
+++ b/project/MimaExcludes.scala
@@ -66,6 +66,19 @@ object MimaExcludes {
// [SPARK-17161] Removing Python-friendly constructors not needed
ProblemFilters.exclude[DirectMissingMethodProblem]("org.apache.spark.ml.classification.OneVsRestModel.this"),
+ // [SPARK-19820] Allow reason to be specified to task kill
+ ProblemFilters.exclude[MissingTypesProblem]("org.apache.spark.TaskKilled$"),
+ ProblemFilters.exclude[DirectMissingMethodProblem]("org.apache.spark.TaskKilled.productElement"),
+ ProblemFilters.exclude[DirectMissingMethodProblem]("org.apache.spark.TaskKilled.productArity"),
+ ProblemFilters.exclude[DirectMissingMethodProblem]("org.apache.spark.TaskKilled.canEqual"),
+ ProblemFilters.exclude[DirectMissingMethodProblem]("org.apache.spark.TaskKilled.productIterator"),
+ ProblemFilters.exclude[DirectMissingMethodProblem]("org.apache.spark.TaskKilled.countTowardsTaskFailures"),
+ ProblemFilters.exclude[DirectMissingMethodProblem]("org.apache.spark.TaskKilled.productPrefix"),
+ ProblemFilters.exclude[DirectMissingMethodProblem]("org.apache.spark.TaskKilled.toErrorString"),
+ ProblemFilters.exclude[FinalMethodProblem]("org.apache.spark.TaskKilled.toString"),
+ ProblemFilters.exclude[ReversedMissingMethodProblem]("org.apache.spark.TaskContext.killTaskIfInterrupted"),
+ ProblemFilters.exclude[ReversedMissingMethodProblem]("org.apache.spark.TaskContext.getKillReason"),
+
// [SPARK-19876] Add one time trigger, and improve Trigger APIs
ProblemFilters.exclude[IncompatibleTemplateDefProblem]("org.apache.spark.sql.streaming.Trigger"),
ProblemFilters.exclude[MissingTypesProblem]("org.apache.spark.sql.streaming.ProcessingTime")