aboutsummaryrefslogtreecommitdiff
path: root/core/src/test/scala/org/apache/spark/rdd/SortingSuite.scala
diff options
context:
space:
mode:
Diffstat (limited to 'core/src/test/scala/org/apache/spark/rdd/SortingSuite.scala')
-rw-r--r--core/src/test/scala/org/apache/spark/rdd/SortingSuite.scala125
1 files changed, 125 insertions, 0 deletions
diff --git a/core/src/test/scala/org/apache/spark/rdd/SortingSuite.scala b/core/src/test/scala/org/apache/spark/rdd/SortingSuite.scala
new file mode 100644
index 0000000000..2f7bd370fc
--- /dev/null
+++ b/core/src/test/scala/org/apache/spark/rdd/SortingSuite.scala
@@ -0,0 +1,125 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one or more
+ * contributor license agreements. See the NOTICE file distributed with
+ * this work for additional information regarding copyright ownership.
+ * The ASF licenses this file to You under the Apache License, Version 2.0
+ * (the "License"); you may not use this file except in compliance with
+ * the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+
+package org.apache.spark.rdd
+
+import org.scalatest.FunSuite
+import org.scalatest.BeforeAndAfter
+import org.scalatest.matchers.ShouldMatchers
+
+import org.apache.spark.{Logging, SharedSparkContext}
+import org.apache.spark.SparkContext._
+
+class SortingSuite extends FunSuite with SharedSparkContext with ShouldMatchers with Logging {
+
+ test("sortByKey") {
+ val pairs = sc.parallelize(Array((1, 0), (2, 0), (0, 0), (3, 0)), 2)
+ assert(pairs.sortByKey().collect() === Array((0,0), (1,0), (2,0), (3,0)))
+ }
+
+ test("large array") {
+ val rand = new scala.util.Random()
+ val pairArr = Array.fill(1000) { (rand.nextInt(), rand.nextInt()) }
+ val pairs = sc.parallelize(pairArr, 2)
+ val sorted = pairs.sortByKey()
+ assert(sorted.partitions.size === 2)
+ assert(sorted.collect() === pairArr.sortBy(_._1))
+ }
+
+ test("large array with one split") {
+ val rand = new scala.util.Random()
+ val pairArr = Array.fill(1000) { (rand.nextInt(), rand.nextInt()) }
+ val pairs = sc.parallelize(pairArr, 2)
+ val sorted = pairs.sortByKey(true, 1)
+ assert(sorted.partitions.size === 1)
+ assert(sorted.collect() === pairArr.sortBy(_._1))
+ }
+
+ test("large array with many partitions") {
+ val rand = new scala.util.Random()
+ val pairArr = Array.fill(1000) { (rand.nextInt(), rand.nextInt()) }
+ val pairs = sc.parallelize(pairArr, 2)
+ val sorted = pairs.sortByKey(true, 20)
+ assert(sorted.partitions.size === 20)
+ assert(sorted.collect() === pairArr.sortBy(_._1))
+ }
+
+ test("sort descending") {
+ val rand = new scala.util.Random()
+ val pairArr = Array.fill(1000) { (rand.nextInt(), rand.nextInt()) }
+ val pairs = sc.parallelize(pairArr, 2)
+ assert(pairs.sortByKey(false).collect() === pairArr.sortWith((x, y) => x._1 > y._1))
+ }
+
+ test("sort descending with one split") {
+ val rand = new scala.util.Random()
+ val pairArr = Array.fill(1000) { (rand.nextInt(), rand.nextInt()) }
+ val pairs = sc.parallelize(pairArr, 1)
+ assert(pairs.sortByKey(false, 1).collect() === pairArr.sortWith((x, y) => x._1 > y._1))
+ }
+
+ test("sort descending with many partitions") {
+ val rand = new scala.util.Random()
+ val pairArr = Array.fill(1000) { (rand.nextInt(), rand.nextInt()) }
+ val pairs = sc.parallelize(pairArr, 2)
+ assert(pairs.sortByKey(false, 20).collect() === pairArr.sortWith((x, y) => x._1 > y._1))
+ }
+
+ test("more partitions than elements") {
+ val rand = new scala.util.Random()
+ val pairArr = Array.fill(10) { (rand.nextInt(), rand.nextInt()) }
+ val pairs = sc.parallelize(pairArr, 30)
+ assert(pairs.sortByKey().collect() === pairArr.sortBy(_._1))
+ }
+
+ test("empty RDD") {
+ val pairArr = new Array[(Int, Int)](0)
+ val pairs = sc.parallelize(pairArr, 2)
+ assert(pairs.sortByKey().collect() === pairArr.sortBy(_._1))
+ }
+
+ test("partition balancing") {
+ val pairArr = (1 to 1000).map(x => (x, x)).toArray
+ val sorted = sc.parallelize(pairArr, 4).sortByKey()
+ assert(sorted.collect() === pairArr.sortBy(_._1))
+ val partitions = sorted.collectPartitions()
+ logInfo("Partition lengths: " + partitions.map(_.length).mkString(", "))
+ partitions(0).length should be > 180
+ partitions(1).length should be > 180
+ partitions(2).length should be > 180
+ partitions(3).length should be > 180
+ partitions(0).last should be < partitions(1).head
+ partitions(1).last should be < partitions(2).head
+ partitions(2).last should be < partitions(3).head
+ }
+
+ test("partition balancing for descending sort") {
+ val pairArr = (1 to 1000).map(x => (x, x)).toArray
+ val sorted = sc.parallelize(pairArr, 4).sortByKey(false)
+ assert(sorted.collect() === pairArr.sortBy(_._1).reverse)
+ val partitions = sorted.collectPartitions()
+ logInfo("partition lengths: " + partitions.map(_.length).mkString(", "))
+ partitions(0).length should be > 180
+ partitions(1).length should be > 180
+ partitions(2).length should be > 180
+ partitions(3).length should be > 180
+ partitions(0).last should be > partitions(1).head
+ partitions(1).last should be > partitions(2).head
+ partitions(2).last should be > partitions(3).head
+ }
+}
+