diff --git a/core/src/main/scala/spark/CoalescedRDD.scala b/core/src/main/scala/spark/CoalescedRDD.scala
index ac0627a4c2758147991f099d9fd14655784d0a89..f1ae346a445a89a5e3d26b3d062fcf4e795652ac 100644
--- a/core/src/main/scala/spark/CoalescedRDD.scala
+++ b/core/src/main/scala/spark/CoalescedRDD.scala
@@ -4,7 +4,7 @@ private class CoalescedRDDSplit(val index: Int, val parents: Array[Split]) exten
 
 /**
  * Coalesce the partitions of a parent RDD (`prev`) into fewer partitions, so that each partition of
- * this RDD computes zero or more of the parent ones. Will produce exactly `maxPartitions` if the
+ * this RDD computes one or more of the parent ones. Will produce exactly `maxPartitions` if the
  * parent had more than this many partitions, or fewer if the parent had fewer.
  *
  * This transformation is useful when an RDD with many partitions gets filtered into a smaller one,
diff --git a/core/src/test/scala/spark/RDDSuite.scala b/core/src/test/scala/spark/RDDSuite.scala
index 961d05bc82512cd158be44c137c4268d9256c540..a1fe63beaf44155d27fac45e2ccd05aec6f4fd7c 100644
--- a/core/src/test/scala/spark/RDDSuite.scala
+++ b/core/src/test/scala/spark/RDDSuite.scala
@@ -79,7 +79,7 @@ class RDDSuite extends FunSuite with BeforeAndAfter {
     val coalesced1 = new CoalescedRDD(data, 2)
     assert(coalesced1.collect().toList === (1 to 10).toList)
     assert(coalesced1.glom().collect().map(_.toList).toList ===
-           List(List(1, 2, 3, 4, 5), List(6, 7, 8, 9, 10)))
+      List(List(1, 2, 3, 4, 5), List(6, 7, 8, 9, 10)))
 
     // Check that the narrow dependency is also specified correctly
     assert(coalesced1.dependencies.head.getParents(0).toList === List(0, 1, 2, 3, 4))