From b547f69bdb5f4a6d5f471a2d998c2df6fb2a9347 Mon Sep 17 00:00:00 2001 From: Matei Zaharia <matei@databricks.com> Date: Sat, 26 Jul 2014 22:44:17 -0700 Subject: [PATCH] SPARK-2680: Lower spark.shuffle.memoryFraction to 0.2 by default Author: Matei Zaharia <matei@databricks.com> Closes #1593 from mateiz/spark-2680 and squashes the following commits: 3c949c4 [Matei Zaharia] Lower spark.shuffle.memoryFraction to 0.2 by default --- .../apache/spark/util/collection/ExternalAppendOnlyMap.scala | 2 +- docs/configuration.md | 2 +- 2 files changed, 2 insertions(+), 2 deletions(-) diff --git a/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala b/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala index be8f6529f7..c22bb8d9c6 100644 --- a/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala +++ b/core/src/main/scala/org/apache/spark/util/collection/ExternalAppendOnlyMap.scala @@ -74,7 +74,7 @@ class ExternalAppendOnlyMap[K, V, C]( // Collective memory threshold shared across all running tasks private val maxMemoryThreshold = { - val memoryFraction = sparkConf.getDouble("spark.shuffle.memoryFraction", 0.3) + val memoryFraction = sparkConf.getDouble("spark.shuffle.memoryFraction", 0.2) val safetyFraction = sparkConf.getDouble("spark.shuffle.safetyFraction", 0.8) (Runtime.getRuntime.maxMemory * memoryFraction * safetyFraction).toLong } diff --git a/docs/configuration.md b/docs/configuration.md index 4e4b78153a..46e3dd914b 100644 --- a/docs/configuration.md +++ b/docs/configuration.md @@ -239,7 +239,7 @@ Apart from these, the following properties are also available, and may be useful </tr> <tr> <td><code>spark.shuffle.memoryFraction</code></td> - <td>0.3</td> + <td>0.2</td> <td> Fraction of Java heap to use for aggregation and cogroups during shuffles, if <code>spark.shuffle.spill</code> is true. At any given time, the collective size of -- GitLab