diff --git a/core/src/main/scala/spark/SparkContext.scala b/core/src/main/scala/spark/SparkContext.scala index 74e6ff1fec14404a652d17ce31c5036afd7d7e5d..18bd5c8817e0342bff2a63ee157443fdf06f919e 100644 --- a/core/src/main/scala/spark/SparkContext.scala +++ b/core/src/main/scala/spark/SparkContext.scala @@ -257,7 +257,7 @@ extends Logging { def defaultParallelism: Int = scheduler.defaultParallelism // Default min number of splits for Hadoop RDDs when not given by user - def defaultMinSplits: Int = Math.min(defaultParallelism, 4) + def defaultMinSplits: Int = Math.min(defaultParallelism, 2) private var nextShuffleId = new AtomicInteger(0)