| Commit message (Collapse) | Author | Age | Files | Lines |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| |
(Continued from old repo, prior discussion at https://github.com/apache/incubator-spark/pull/615)
This patch cements our deprecation of the SPARK_MEM environment variable by replacing it with three more specialized variables:
SPARK_DAEMON_MEMORY, SPARK_EXECUTOR_MEMORY, and SPARK_DRIVER_MEMORY
The creation of the latter two variables means that we can safely set driver/job memory without accidentally setting the executor memory. Neither is public.
SPARK_EXECUTOR_MEMORY is only used by the Mesos scheduler (and set within SparkContext). The proper way of configuring executor memory is through the "spark.executor.memory" property.
SPARK_DRIVER_MEMORY is the new way of specifying the amount of memory run by jobs launched by spark-class, without possibly affecting executor memory.
Other memory considerations:
- The repl's memory can be set through the "--drivermem" command-line option, which really just sets SPARK_DRIVER_MEMORY.
- run-example doesn't use spark-class, so the only way to modify examples' memory is actually an unusual use of SPARK_JAVA_OPTS (which is normally overriden in all cases by spark-class).
This patch also fixes a lurking bug where spark-shell misused spark-class (the first argument is supposed to be the main class name, not java options), as well as a bug in the Windows spark-class2.cmd. I have not yet tested this patch on either Windows or Mesos, however.
Author: Aaron Davidson <aaron@databricks.com>
Closes #99 from aarondav/sparkmem and squashes the following commits:
9df4c68 [Aaron Davidson] SPARK-929: Fully deprecate usage of SPARK_MEM
|
|
|
|
|
|
|
|
|
|
| |
The default value of "spark.storage.memoryFraction" has been changed from 0.66 to 0.6 . So it should be 60% of the memory to cache while 40% used for task execution.
Author: Chen Chao <crazyjvm@gmail.com>
Closes #66 from CrazyJvm/master and squashes the following commits:
0f84d86 [Chen Chao] update proportion of memory
|
|
|
|
|
|
|
|
| |
Author: Andrew Ash <andrew@andrewash.com>
Closes #647 from ash211/doc-tuning and squashes the following commits:
b87de0a [Andrew Ash] Include reference to twitter/chill in tuning docs
|
|
|
| |
remove "-XX:+UseCompressedStrings" option from tuning guide since jdk7 no longer supports this.
|
| |
|
|
|
| |
Clarify when serializer is used based on recent user@ mailing list discussion.
|
|
|
| |
The documentation here is inconsistent with the coded default and other documentation.
|
|
|
|
| |
Also fix a couple HTML/Markdown issues in other files.
|
|
|
|
|
|
| |
* RDD, *RDDFunctions -> org.apache.spark.rdd
* Utils, ClosureCleaner, SizeEstimator -> org.apache.spark.util
* JavaSerializer, KryoSerializer -> org.apache.spark.serializer
|
| |
|
|
|
|
|
|
| |
Conflicts:
core/src/main/scala/spark/SparkContext.scala
|
|
|
| |
Make the example more compilable
|
|\
| |
| |
| |
| |
| |
| | |
Conflicts:
core/src/main/scala/spark/RDD.scala
core/src/main/scala/spark/scheduler/cluster/StandaloneSchedulerBackend.scala
core/src/test/scala/spark/ShuffleSuite.scala
|
| | |
|
|/
|
|
|
|
| |
Only brand new RDDs (e.g. parallelize and makeRDD) now use default
parallelism, everything else uses their largest parent's partitioner
or partition size.
|
| |
|
|
|
|
|
|
|
|
| |
- Edited quick start and tuning guide to simplify them a little
- Simplified top menu bar
- Made private a SparkContext constructor parameter that was left as
public
- Various small fixes
|
|
|
|
|
|
|
|
|
| |
throughout the docs: SPARK_VERSION, SCALA_VERSION, and MESOS_VERSION.
To use them, e.g. use {{site.SPARK_VERSION}}.
Also removes uses of {{HOME_PATH}} which were being resolved to ""
by the templating system anyway.
|
|
|
|
|
|
|
|
| |
1. Slight change in organization
2. Added pre-requisites
3. Made a new section about determining memory footprint
of an RDD
4. Other small changes
|
| |
|
|
|