aboutsummaryrefslogtreecommitdiff
path: root/docs/tuning.md
diff options
context:
space:
mode:
authorStephen Haberman <stephen@exigencecorp.com>2013-02-16 00:29:11 -0600
committerStephen Haberman <stephen@exigencecorp.com>2013-02-16 00:29:11 -0600
commit6cd68c31cbebb1e3c6b35026f067a3c82ce9fdfb (patch)
treeb44f937da2b8011b32cebe499169bba0d12acc24 /docs/tuning.md
parent680f42e6cd1ee8593136323a539dc5117b165377 (diff)
downloadspark-6cd68c31cbebb1e3c6b35026f067a3c82ce9fdfb.tar.gz
spark-6cd68c31cbebb1e3c6b35026f067a3c82ce9fdfb.tar.bz2
spark-6cd68c31cbebb1e3c6b35026f067a3c82ce9fdfb.zip
Update default.parallelism docs, have StandaloneSchedulerBackend use it.
Only brand new RDDs (e.g. parallelize and makeRDD) now use default parallelism, everything else uses their largest parent's partitioner or partition size.
Diffstat (limited to 'docs/tuning.md')
-rw-r--r--docs/tuning.md8
1 files changed, 4 insertions, 4 deletions
diff --git a/docs/tuning.md b/docs/tuning.md
index 9aaa53cd65..e9b4d6717c 100644
--- a/docs/tuning.md
+++ b/docs/tuning.md
@@ -213,10 +213,10 @@ but at a high level, managing how frequently full GC takes place can help in red
Clusters will not be fully utilized unless you set the level of parallelism for each operation high
enough. Spark automatically sets the number of "map" tasks to run on each file according to its size
-(though you can control it through optional parameters to `SparkContext.textFile`, etc), but for
-distributed "reduce" operations, such as `groupByKey` and `reduceByKey`, it uses a default value of 8.
-You can pass the level of parallelism as a second argument (see the
-[`spark.PairRDDFunctions`](api/core/index.html#spark.PairRDDFunctions) documentation),
+(though you can control it through optional parameters to `SparkContext.textFile`, etc), and for
+distributed "reduce" operations, such as `groupByKey` and `reduceByKey`, it uses the largest
+parent RDD's number of partitions. You can pass the level of parallelism as a second argument
+(see the [`spark.PairRDDFunctions`](api/core/index.html#spark.PairRDDFunctions) documentation),
or set the system property `spark.default.parallelism` to change the default.
In general, we recommend 2-3 tasks per CPU core in your cluster.