diff --git a/core/src/main/scala/spark/CoalescedRDD.scala b/core/src/main/scala/spark/CoalescedRDD.scala index ac0627a4c2758147991f099d9fd14655784d0a89..f1ae346a445a89a5e3d26b3d062fcf4e795652ac 100644 --- a/core/src/main/scala/spark/CoalescedRDD.scala +++ b/core/src/main/scala/spark/CoalescedRDD.scala @@ -4,7 +4,7 @@ private class CoalescedRDDSplit(val index: Int, val parents: Array[Split]) exten /** * Coalesce the partitions of a parent RDD (`prev`) into fewer partitions, so that each partition of - * this RDD computes zero or more of the parent ones. Will produce exactly `maxPartitions` if the + * this RDD computes one or more of the parent ones. Will produce exactly `maxPartitions` if the * parent had more than this many partitions, or fewer if the parent had fewer. * * This transformation is useful when an RDD with many partitions gets filtered into a smaller one, diff --git a/core/src/test/scala/spark/RDDSuite.scala b/core/src/test/scala/spark/RDDSuite.scala index 961d05bc82512cd158be44c137c4268d9256c540..a1fe63beaf44155d27fac45e2ccd05aec6f4fd7c 100644 --- a/core/src/test/scala/spark/RDDSuite.scala +++ b/core/src/test/scala/spark/RDDSuite.scala @@ -79,7 +79,7 @@ class RDDSuite extends FunSuite with BeforeAndAfter { val coalesced1 = new CoalescedRDD(data, 2) assert(coalesced1.collect().toList === (1 to 10).toList) assert(coalesced1.glom().collect().map(_.toList).toList === - List(List(1, 2, 3, 4, 5), List(6, 7, 8, 9, 10))) + List(List(1, 2, 3, 4, 5), List(6, 7, 8, 9, 10))) // Check that the narrow dependency is also specified correctly assert(coalesced1.dependencies.head.getParents(0).toList === List(0, 1, 2, 3, 4))