From 7173918c8cab7fb0c06dc750bc0e2ae4fce9658f Mon Sep 17 00:00:00 2001 From: Liang-Chi Hsieh Date: Mon, 20 Jul 2020 16:43:50 -0700 Subject: [PATCH] Make failed test to test Github Actions. --- core/src/test/scala/org/apache/spark/rdd/RDDSuite.scala | 4 ++-- core/src/test/scala/org/apache/spark/rdd/SortingSuite.scala | 2 +- .../scala/org/apache/spark/sql/DataFrameAggregateSuite.scala | 4 ++-- .../src/test/scala/org/apache/spark/sql/DataFrameSuite.scala | 4 ++-- 4 files changed, 7 insertions(+), 7 deletions(-) diff --git a/core/src/test/scala/org/apache/spark/rdd/RDDSuite.scala b/core/src/test/scala/org/apache/spark/rdd/RDDSuite.scala index 79f9c1396c87b..016965bf86328 100644 --- a/core/src/test/scala/org/apache/spark/rdd/RDDSuite.scala +++ b/core/src/test/scala/org/apache/spark/rdd/RDDSuite.scala @@ -126,7 +126,7 @@ class RDDSuite extends SparkFunSuite with SharedSparkContext with Eventually { test("SparkContext.union") { val nums = sc.makeRDD(Array(1, 2, 3, 4), 2) - assert(sc.union(nums).collect().toList === List(1, 2, 3, 4)) + assert(sc.union(nums).collect().toList === List(2, 2, 3, 4)) assert(sc.union(nums, nums).collect().toList === List(1, 2, 3, 4, 1, 2, 3, 4)) assert(sc.union(Seq(nums)).collect().toList === List(1, 2, 3, 4)) assert(sc.union(Seq(nums, nums)).collect().toList === List(1, 2, 3, 4, 1, 2, 3, 4)) @@ -222,7 +222,7 @@ class RDDSuite extends SparkFunSuite with SharedSparkContext with Eventually { val rdd = sc.makeRDD(-1000 until 1000, 10) def op: (Int, Int) => Int = (c: Int, x: Int) => c + x val sum = rdd.fold(0)(op) - assert(sum === -1000) + assert(sum === -1001) } test("fold with op modifying first arg") { diff --git a/core/src/test/scala/org/apache/spark/rdd/SortingSuite.scala b/core/src/test/scala/org/apache/spark/rdd/SortingSuite.scala index d5f7d30a253fe..304aaf079d6c4 100644 --- a/core/src/test/scala/org/apache/spark/rdd/SortingSuite.scala +++ b/core/src/test/scala/org/apache/spark/rdd/SortingSuite.scala @@ -26,7 +26,7 @@ class SortingSuite extends SparkFunSuite with SharedSparkContext with Matchers w test("sortByKey") { val pairs = sc.parallelize(Seq((1, 0), (2, 0), (0, 0), (3, 0)), 2) - assert(pairs.sortByKey().collect() === Array((0, 0), (1, 0), (2, 0), (3, 0))) + assert(pairs.sortByKey().collect() === Array((2, 0), (1, 0), (2, 0), (3, 0))) } test("large array") { diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala index 54327b38c100b..457da2729cea6 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/DataFrameAggregateSuite.scala @@ -115,7 +115,7 @@ class DataFrameAggregateSuite extends QueryTest checkAnswer( df.groupBy(regexp_extract($"key", "([a-z]+)\\[", 1)).count(), - Row("some", 1) :: Nil + Row("some", 2) :: Nil ) } @@ -160,7 +160,7 @@ class DataFrameAggregateSuite extends QueryTest checkAnswer( courseSales.cube("course", "year") .agg(grouping("course"), grouping("year"), grouping_id("course", "year")), - Row("Java", 2012, 0, 0, 0) :: + Row("Java", 2012, 0, 0, 1) :: Row("Java", 2013, 0, 0, 0) :: Row("Java", null, 0, 1, 1) :: Row("dotNET", 2012, 0, 0, 0) :: diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DataFrameSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/DataFrameSuite.scala index 52ef5895ed9ad..7446e77562afa 100644 --- a/sql/core/src/test/scala/org/apache/spark/sql/DataFrameSuite.scala +++ b/sql/core/src/test/scala/org/apache/spark/sql/DataFrameSuite.scala @@ -78,7 +78,7 @@ class DataFrameSuite extends QueryTest checkAnswer( df.groupBy("_1").agg(sum("_2._1")).toDF("key", "total"), - Row(1, 1) :: Nil) + Row(2, 1) :: Nil) } test("access complex data") { @@ -146,7 +146,7 @@ class DataFrameSuite extends QueryTest sort("a").first() == Row(1, Seq(1, 1))) // CreateStruct and CreateArray in project list (unresolved alias) - assert(structDf.select(struct($"record.*")).first() == Row(Row(1, 1))) + assert(structDf.select(struct($"record.*")).first() == Row(Row(2, 1))) assert(structDf.select(array($"record.*")).first().getAs[Seq[Int]](0) === Seq(1, 1)) // CreateStruct and CreateArray in project list (alias)