This is an automated email from the ASF dual-hosted git repository. gurwls223 pushed a commit to branch master in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/master by this push: new cd9af61ae11e [MINOR][DOCS] Add migration note for mapInPandas and mapInArrow validation cd9af61ae11e is described below commit cd9af61ae11e4828c096a597a319315d25f68ec2 Author: Haoyu Weng <wengh...@gmail.com> AuthorDate: Mon Apr 28 17:07:30 2025 +0900 [MINOR][DOCS] Add migration note for mapInPandas and mapInArrow validation ### What changes were proposed in this pull request? ### Why are the changes needed? ### Does this PR introduce _any_ user-facing change? ### How was this patch tested? ### Was this patch authored or co-authored using generative AI tooling? Closes #50722 from wengh/arrow_validation_guide. Authored-by: Haoyu Weng <wengh...@gmail.com> Signed-off-by: Hyukjin Kwon <gurwls...@apache.org> --- docs/sql-migration-guide.md | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/docs/sql-migration-guide.md b/docs/sql-migration-guide.md index 24b5851ce9be..edc0a61d8f1d 100644 --- a/docs/sql-migration-guide.md +++ b/docs/sql-migration-guide.md @@ -22,6 +22,10 @@ license: | * Table of contents {:toc} +## Upgrading from Spark SQL 4.0 to 4.1 + +- Since Spark 4.1, `mapInPandas` and `mapInArrow` enforces strict validation of the result against the schema. The column names must exactly match and types must match with compatible nullability. To restore the previous behavior, set `spark.sql.execution.arrow.pyspark.validateSchema.enabled` to `false`. + ## Upgrading from Spark SQL 3.5 to 4.0 - Since Spark 4.0, `spark.sql.ansi.enabled` is on by default. To restore the previous behavior, set `spark.sql.ansi.enabled` to `false` or `SPARK_ANSI_SQL_MODE` to `false`. --------------------------------------------------------------------- To unsubscribe, e-mail: commits-unsubscr...@spark.apache.org For additional commands, e-mail: commits-h...@spark.apache.org