diff --git a/core/src/main/scala/spark/SparkContext.scala b/core/src/main/scala/spark/SparkContext.scala
index 74e6ff1fec14404a652d17ce31c5036afd7d7e5d..18bd5c8817e0342bff2a63ee157443fdf06f919e 100644
--- a/core/src/main/scala/spark/SparkContext.scala
+++ b/core/src/main/scala/spark/SparkContext.scala
@@ -257,7 +257,7 @@ extends Logging {
   def defaultParallelism: Int = scheduler.defaultParallelism
 
   // Default min number of splits for Hadoop RDDs when not given by user
-  def defaultMinSplits: Int = Math.min(defaultParallelism, 4)
+  def defaultMinSplits: Int = Math.min(defaultParallelism, 2)
 
   private var nextShuffleId = new AtomicInteger(0)