From ccf48388cd8d8f4e1777be77ab30032d1b15c8c6 Mon Sep 17 00:00:00 2001 From: Matei Zaharia <matei@eecs.berkeley.edu> Date: Thu, 14 Jul 2011 17:37:04 -0400 Subject: [PATCH] Lowered default number of splits for files --- core/src/main/scala/spark/SparkContext.scala | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/core/src/main/scala/spark/SparkContext.scala b/core/src/main/scala/spark/SparkContext.scala index 74e6ff1fec..18bd5c8817 100644 --- a/core/src/main/scala/spark/SparkContext.scala +++ b/core/src/main/scala/spark/SparkContext.scala @@ -257,7 +257,7 @@ extends Logging { def defaultParallelism: Int = scheduler.defaultParallelism // Default min number of splits for Hadoop RDDs when not given by user - def defaultMinSplits: Int = Math.min(defaultParallelism, 4) + def defaultMinSplits: Int = Math.min(defaultParallelism, 2) private var nextShuffleId = new AtomicInteger(0) -- GitLab