diff --git a/.github/workflows/master.yml b/.github/workflows/master.yml index 809ce0133c57..3fa7feeb9bc5 100644 --- a/.github/workflows/master.yml +++ b/.github/workflows/master.yml @@ -28,19 +28,11 @@ jobs: # Note that the modules below are from sparktestsupport/modules.py. modules: - >- - pyspark-sql + unsafe # Here, we split Hive and SQL tests into some of slow ones and the rest of them. included-tags: [""] excluded-tags: [""] comment: [""] - include: - # SQL tests - - modules: sql - java: 1.8 - hadoop: hadoop3.2 - hive: hive2.3 - excluded-tags: org.apache.spark.tags.ExtendedSQLTest - comment: "- other tests" env: MODULES_TO_TEST: ${{ matrix.modules }} EXCLUDED_TAGS: ${{ matrix.excluded-tags }} diff --git a/.github/workflows/test_report.yml b/.github/workflows/test_report.yml index 13e291d3ca0e..75e7e94d0407 100644 --- a/.github/workflows/test_report.yml +++ b/.github/workflows/test_report.yml @@ -25,19 +25,11 @@ jobs: # Note that the modules below are from sparktestsupport/modules.py. modules: - >- - pyspark-sql + unsafe # Here, we split Hive and SQL tests into some of slow ones and the rest of them. included-tags: [""] excluded-tags: [""] comment: [""] - include: - # SQL tests - - modules: sql - java: 1.8 - hadoop: hadoop3.2 - hive: hive2.3 - excluded-tags: org.apache.spark.tags.ExtendedSQLTest - comment: "- other tests" steps: - uses: haya14busa/action-workflow_run-status@v1 - uses: actions/checkout@v2 diff --git a/common/unsafe/src/test/scala/org/apache/spark/unsafe/types/UTF8StringPropertyCheckSuite.scala b/common/unsafe/src/test/scala/org/apache/spark/unsafe/types/UTF8StringPropertyCheckSuite.scala index 51a54eb07db9..c14c59117ba2 100644 --- a/common/unsafe/src/test/scala/org/apache/spark/unsafe/types/UTF8StringPropertyCheckSuite.scala +++ b/common/unsafe/src/test/scala/org/apache/spark/unsafe/types/UTF8StringPropertyCheckSuite.scala @@ -34,7 +34,7 @@ class UTF8StringPropertyCheckSuite extends AnyFunSuite with ScalaCheckDrivenProp test("toString") { forAll { (s: String) => - assert(toUTF8(s).toString() !== s) + assert(toUTF8(s).toString() === s) } } diff --git a/python/pyspark/sql/tests/test_arrow.py b/python/pyspark/sql/tests/test_arrow.py index 31287a24265d..982c91ea119a 100644 --- a/python/pyspark/sql/tests/test_arrow.py +++ b/python/pyspark/sql/tests/test_arrow.py @@ -201,7 +201,7 @@ def test_no_partition_frame(self): pdf = df.toPandas() self.assertEqual(len(pdf.columns), 1) self.assertEqual(pdf.columns[0], "field1") - self.assertTrue(False) + self.assertTrue("A") def test_propagates_spark_exception(self): df = self.spark.range(3).toDF("i")