This is an automated email from the ASF dual-hosted git repository.

beliefer pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
     new d8dbb6674d0 [SPARK-45514][SQL][MLLIB] Replace 
`scala.runtime.Tuple3Zipped` to `scala.collection.LazyZip3`
d8dbb6674d0 is described below

commit d8dbb6674d0f3087f072dc08263f86e02a263693
Author: Jiaan Geng <[email protected]>
AuthorDate: Mon Oct 16 10:00:26 2023 +0800

    [SPARK-45514][SQL][MLLIB] Replace `scala.runtime.Tuple3Zipped` to 
`scala.collection.LazyZip3`
    
    ### What changes were proposed in this pull request?
    Since scala 2.13.0, `scala.runtime.Tuple3Zipped` marked as deprecated and 
`scala.collection.LazyZip3` recommended.
    
    ### Why are the changes needed?
    Replace `scala.runtime.Tuple3Zipped` to `scala.collection.LazyZip3`.
    
    ### Does this PR introduce _any_ user-facing change?
    'No'.
    
    ### How was this patch tested?
    Exists test cases.
    
    ### Was this patch authored or co-authored using generative AI tooling?
    'No'.
    
    Closes #43363 from beliefer/SPARK-45514.
    
    Authored-by: Jiaan Geng <[email protected]>
    Signed-off-by: Jiaan Geng <[email protected]>
---
 .../spark/mllib/feature/ElementwiseProductSuite.scala    |  2 +-
 .../org/apache/spark/mllib/feature/NormalizerSuite.scala |  6 +++---
 .../apache/spark/mllib/feature/StandardScalerSuite.scala | 16 ++++++++--------
 .../spark/sql/execution/arrow/ArrowConvertersSuite.scala |  2 +-
 .../spark/sql/hive/execution/HiveComparisonTest.scala    |  2 +-
 5 files changed, 14 insertions(+), 14 deletions(-)

diff --git 
a/mllib/src/test/scala/org/apache/spark/mllib/feature/ElementwiseProductSuite.scala
 
b/mllib/src/test/scala/org/apache/spark/mllib/feature/ElementwiseProductSuite.scala
index 9eca2d682d6..a764ac8551f 100644
--- 
a/mllib/src/test/scala/org/apache/spark/mllib/feature/ElementwiseProductSuite.scala
+++ 
b/mllib/src/test/scala/org/apache/spark/mllib/feature/ElementwiseProductSuite.scala
@@ -48,7 +48,7 @@ class ElementwiseProductSuite extends SparkFunSuite with 
MLlibTestSparkContext {
     val data2 = sparseData.map(transformer.transform)
     val data2RDD = transformer.transform(dataRDD)
 
-    assert((sparseData, data2, data2RDD.collect()).zipped.forall {
+    assert(sparseData.lazyZip(data2).lazyZip(data2RDD.collect()).forall {
       case (v1: DenseVector, v2: DenseVector, v3: DenseVector) => true
       case (v1: SparseVector, v2: SparseVector, v3: SparseVector) => true
       case _ => false
diff --git 
a/mllib/src/test/scala/org/apache/spark/mllib/feature/NormalizerSuite.scala 
b/mllib/src/test/scala/org/apache/spark/mllib/feature/NormalizerSuite.scala
index 71ce26360b8..77eade64db9 100644
--- a/mllib/src/test/scala/org/apache/spark/mllib/feature/NormalizerSuite.scala
+++ b/mllib/src/test/scala/org/apache/spark/mllib/feature/NormalizerSuite.scala
@@ -43,7 +43,7 @@ class NormalizerSuite extends SparkFunSuite with 
MLlibTestSparkContext {
     val data1 = data.map(l1Normalizer.transform)
     val data1RDD = l1Normalizer.transform(dataRDD)
 
-    assert((data, data1, data1RDD.collect()).zipped.forall {
+    assert(data.lazyZip(data1).lazyZip(data1RDD.collect()).forall {
       case (v1: DenseVector, v2: DenseVector, v3: DenseVector) => true
       case (v1: SparseVector, v2: SparseVector, v3: SparseVector) => true
       case _ => false
@@ -70,7 +70,7 @@ class NormalizerSuite extends SparkFunSuite with 
MLlibTestSparkContext {
     val data2 = data.map(l2Normalizer.transform)
     val data2RDD = l2Normalizer.transform(dataRDD)
 
-    assert((data, data2, data2RDD.collect()).zipped.forall {
+    assert(data.lazyZip(data2).lazyZip(data2RDD.collect()).forall {
       case (v1: DenseVector, v2: DenseVector, v3: DenseVector) => true
       case (v1: SparseVector, v2: SparseVector, v3: SparseVector) => true
       case _ => false
@@ -97,7 +97,7 @@ class NormalizerSuite extends SparkFunSuite with 
MLlibTestSparkContext {
     val dataInf = data.map(lInfNormalizer.transform)
     val dataInfRDD = lInfNormalizer.transform(dataRDD)
 
-    assert((data, dataInf, dataInfRDD.collect()).zipped.forall {
+    assert(data.lazyZip(dataInf).lazyZip(dataInfRDD.collect()).forall {
       case (v1: DenseVector, v2: DenseVector, v3: DenseVector) => true
       case (v1: SparseVector, v2: SparseVector, v3: SparseVector) => true
       case _ => false
diff --git 
a/mllib/src/test/scala/org/apache/spark/mllib/feature/StandardScalerSuite.scala 
b/mllib/src/test/scala/org/apache/spark/mllib/feature/StandardScalerSuite.scala
index a2c72de4231..2fad944344b 100644
--- 
a/mllib/src/test/scala/org/apache/spark/mllib/feature/StandardScalerSuite.scala
+++ 
b/mllib/src/test/scala/org/apache/spark/mllib/feature/StandardScalerSuite.scala
@@ -87,19 +87,19 @@ class StandardScalerSuite extends SparkFunSuite with 
MLlibTestSparkContext {
     val summary2 = computeSummary(data2RDD)
     val summary3 = computeSummary(data3RDD)
 
-    assert((denseData, data1, data1RDD.collect()).zipped.forall {
+    assert(denseData.lazyZip(data1).lazyZip(data1RDD.collect()).forall {
       case (v1: DenseVector, v2: DenseVector, v3: DenseVector) => true
       case (v1: SparseVector, v2: SparseVector, v3: SparseVector) => true
       case _ => false
     }, "The vector type should be preserved after standardization.")
 
-    assert((denseData, data2, data2RDD.collect()).zipped.forall {
+    assert(denseData.lazyZip(data2).lazyZip(data2RDD.collect()).forall {
       case (v1: DenseVector, v2: DenseVector, v3: DenseVector) => true
       case (v1: SparseVector, v2: SparseVector, v3: SparseVector) => true
       case _ => false
     }, "The vector type should be preserved after standardization.")
 
-    assert((denseData, data3, data3RDD.collect()).zipped.forall {
+    assert(denseData.lazyZip(data3).lazyZip(data3RDD.collect()).forall {
       case (v1: DenseVector, v2: DenseVector, v3: DenseVector) => true
       case (v1: SparseVector, v2: SparseVector, v3: SparseVector) => true
       case _ => false
@@ -151,19 +151,19 @@ class StandardScalerSuite extends SparkFunSuite with 
MLlibTestSparkContext {
     val summary2 = computeSummary(data2RDD)
     val summary3 = computeSummary(data3RDD)
 
-    assert((denseData, data1, data1RDD.collect()).zipped.forall {
+    assert(denseData.lazyZip(data1).lazyZip(data1RDD.collect()).forall {
       case (v1: DenseVector, v2: DenseVector, v3: DenseVector) => true
       case (v1: SparseVector, v2: SparseVector, v3: SparseVector) => true
       case _ => false
     }, "The vector type should be preserved after standardization.")
 
-    assert((denseData, data2, data2RDD.collect()).zipped.forall {
+    assert(denseData.lazyZip(data2).lazyZip(data2RDD.collect()).forall {
       case (v1: DenseVector, v2: DenseVector, v3: DenseVector) => true
       case (v1: SparseVector, v2: SparseVector, v3: SparseVector) => true
       case _ => false
     }, "The vector type should be preserved after standardization.")
 
-    assert((denseData, data3, data3RDD.collect()).zipped.forall {
+    assert(denseData.lazyZip(data3).lazyZip(data3RDD.collect()).forall {
       case (v1: DenseVector, v2: DenseVector, v3: DenseVector) => true
       case (v1: SparseVector, v2: SparseVector, v3: SparseVector) => true
       case _ => false
@@ -219,7 +219,7 @@ class StandardScalerSuite extends SparkFunSuite with 
MLlibTestSparkContext {
     val summary2 = computeSummary(data2RDD)
     val summary3 = computeSummary(data3RDD)
 
-    assert((sparseData, data2, data2RDD.collect()).zipped.forall {
+    assert(sparseData.lazyZip(data2).lazyZip(data2RDD.collect()).forall {
       case (v1: DenseVector, v2: DenseVector, v3: DenseVector) => true
       case (v1: SparseVector, v2: SparseVector, v3: SparseVector) => true
       case _ => false
@@ -268,7 +268,7 @@ class StandardScalerSuite extends SparkFunSuite with 
MLlibTestSparkContext {
     val summary2 = computeSummary(data2RDD)
     val summary3 = computeSummary(data3RDD)
 
-    assert((sparseData, data2, data2RDD.collect()).zipped.forall {
+    assert(sparseData.lazyZip(data2).lazyZip(data2RDD.collect()).forall {
       case (v1: DenseVector, v2: DenseVector, v3: DenseVector) => true
       case (v1: SparseVector, v2: SparseVector, v3: SparseVector) => true
       case _ => false
diff --git 
a/sql/core/src/test/scala/org/apache/spark/sql/execution/arrow/ArrowConvertersSuite.scala
 
b/sql/core/src/test/scala/org/apache/spark/sql/execution/arrow/ArrowConvertersSuite.scala
index 4a947f18532..4b95b1f634b 100644
--- 
a/sql/core/src/test/scala/org/apache/spark/sql/execution/arrow/ArrowConvertersSuite.scala
+++ 
b/sql/core/src/test/scala/org/apache/spark/sql/execution/arrow/ArrowConvertersSuite.scala
@@ -550,7 +550,7 @@ class ArrowConvertersSuite extends SharedSparkSession {
     val upperCase = Seq("A", "B", "C")
     val lowerCase = Seq("a", "b", "c")
     val nullStr = Seq("ab", "CDE", null)
-    val df = (upperCase, lowerCase, nullStr).zipped.toList
+    val df = upperCase.lazyZip(lowerCase).lazyZip(nullStr).toList
       .toDF("upper_case", "lower_case", "null_str")
 
     collectAndValidate(df, json, "stringData.json")
diff --git 
a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveComparisonTest.scala
 
b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveComparisonTest.scala
index 67b51f93478..f0feccb4f49 100644
--- 
a/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveComparisonTest.scala
+++ 
b/sql/hive/src/test/scala/org/apache/spark/sql/hive/execution/HiveComparisonTest.scala
@@ -364,7 +364,7 @@ abstract class HiveComparisonTest extends SparkFunSuite 
with BeforeAndAfterAll {
           }
         }
 
-        (queryList, hiveResults, catalystResults).zipped.foreach {
+        queryList.lazyZip(hiveResults).lazyZip(catalystResults).foreach {
           case (query, hive, (hiveQuery, catalyst)) =>
             // Check that the results match unless its an EXPLAIN query.
             val preparedHive = prepareAnswer(hiveQuery, hive)


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to