diff options
author | Matei Zaharia <matei@eecs.berkeley.edu> | 2013-09-08 00:41:18 -0400 |
---|---|---|
committer | Matei Zaharia <matei@eecs.berkeley.edu> | 2013-09-08 00:29:11 -0700 |
commit | f261d2a60fe9c0ec81c7a93a24fd79062c31f7ae (patch) | |
tree | 76b1b22ca6a947c4ba649cc6b85763ce5e8578b0 /docs/monitoring.md | |
parent | 651a96adf7b53085bd810e153f8eabf52eed1994 (diff) | |
download | spark-f261d2a60fe9c0ec81c7a93a24fd79062c31f7ae.tar.gz spark-f261d2a60fe9c0ec81c7a93a24fd79062c31f7ae.tar.bz2 spark-f261d2a60fe9c0ec81c7a93a24fd79062c31f7ae.zip |
Added cluster overview doc, made logo higher-resolution, and added more
details on monitoring
Diffstat (limited to 'docs/monitoring.md')
-rw-r--r-- | docs/monitoring.md | 30 |
1 files changed, 21 insertions, 9 deletions
diff --git a/docs/monitoring.md b/docs/monitoring.md index 0ec987107c..e9832e0466 100644 --- a/docs/monitoring.md +++ b/docs/monitoring.md @@ -3,19 +3,30 @@ layout: global title: Monitoring and Instrumentation --- -There are several ways to monitor the progress of Spark jobs. +There are several ways to monitor Spark applications. # Web Interfaces -When a SparkContext is initialized, it launches a web server (by default at port 3030) which -displays useful information. This includes a list of active and completed scheduler stages, -a summary of RDD blocks and partitions, and environmental information. If multiple SparkContexts -are running on the same host, they will bind to succesive ports beginning with 3030 (3031, 3032, -etc). -Spark's Standlone Mode scheduler also has its own -[web interface](spark-standalone.html#monitoring-and-logging). +Every SparkContext launches a web UI, by default on port 3030, that +displays useful information about the application. This includes: + +* A list of scheduler stages and tasks +* A summary of RDD sizes and memory usage +* Information about the running executors +* Environmental information. + +You can access this interface by simply opening `http://<driver-node>:3030` in a web browser. +If multiple SparkContexts are running on the same host, they will bind to succesive ports +beginning with 3030 (3031, 3032, etc). + +Spark's Standlone Mode cluster manager also has its own +[web UI](spark-standalone.html#monitoring-and-logging). + +Note that in both of these UIs, the tables are sortable by clicking their headers, +making it easy to identify slow tasks, data skew, etc. + +# Metrics -# Spark Metrics Spark has a configurable metrics system based on the [Coda Hale Metrics Library](http://metrics.codahale.com/). This allows users to report Spark metrics to a variety of sinks including HTTP, JMX, and CSV @@ -35,6 +46,7 @@ The syntax of the metrics configuration file is defined in an example configurat `$SPARK_HOME/conf/metrics.conf.template`. # Advanced Instrumentation + Several external tools can be used to help profile the performance of Spark jobs: * Cluster-wide monitoring tools, such as [Ganglia](http://ganglia.sourceforge.net/), can provide |