This is an automated email from the ASF dual-hosted git repository.

gurwls223 pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git


The following commit(s) were added to refs/heads/master by this push:
     new d148e9be24f4 [SPARK-52877][PYTHON][FOLLOW-UP] Use columns instead of 
itercolumns in RecordBatch
d148e9be24f4 is described below

commit d148e9be24f49b79bf85aa46e48ae7e71bda2f13
Author: Hyukjin Kwon <[email protected]>
AuthorDate: Fri Jul 25 14:37:52 2025 +0900

    [SPARK-52877][PYTHON][FOLLOW-UP] Use columns instead of itercolumns in 
RecordBatch
    
    ### What changes were proposed in this pull request?
    
    This PR proposes to use `columns` instead of `itercolumns` in RecordBatch, 
which does not exist in the old version of PyArrow.
    
    ### Why are the changes needed?
    
    To recover the build 
https://github.com/apache/spark/actions/runs/16507806777/job/46682838114
    This is just a temporary workaround.
    
    ### Does this PR introduce _any_ user-facing change?
    
    No, test-only.
    
    ### How was this patch tested?
    
    Manually.
    
    ### Was this patch authored or co-authored using generative AI tooling?
    
    No,
    
    Closes #51661 from HyukjinKwon/SPARK-52877.
    
    Authored-by: Hyukjin Kwon <[email protected]>
    Signed-off-by: Hyukjin Kwon <[email protected]>
---
 python/pyspark/sql/pandas/serializers.py | 2 +-
 1 file changed, 1 insertion(+), 1 deletion(-)

diff --git a/python/pyspark/sql/pandas/serializers.py 
b/python/pyspark/sql/pandas/serializers.py
index 8367dca0228c..b6f9282b3a2f 100644
--- a/python/pyspark/sql/pandas/serializers.py
+++ b/python/pyspark/sql/pandas/serializers.py
@@ -793,7 +793,7 @@ class 
ArrowBatchUDFSerializer(ArrowStreamArrowUDFSerializer):
         for batch in super().load_stream(stream):
             columns = [
                 [conv(v) for v in column.to_pylist()] if conv is not None else 
column.to_pylist()
-                for column, conv in zip(batch.itercolumns(), converters)
+                for column, conv in zip(batch.columns, converters)
             ]
             if len(columns) == 0:
                 yield [[pyspark._NoValue] * batch.num_rows]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to