This is an automated email from the ASF dual-hosted git repository.
dongjoon pushed a commit to branch master
in repository https://gitbox.apache.org/repos/asf/spark.git
The following commit(s) were added to refs/heads/master by this push:
new 3d395a6b874b [SPARK-46774][SQL][AVRO] Use
mapreduce.output.fileoutputformat.compress instead of deprecated
mapred.output.compress in Avro write jobs
3d395a6b874b is described below
commit 3d395a6b874bfd4609323551773c61d42fb60b8a
Author: Kent Yao <[email protected]>
AuthorDate: Fri Jan 19 10:22:13 2024 -0800
[SPARK-46774][SQL][AVRO] Use mapreduce.output.fileoutputformat.compress
instead of deprecated mapred.output.compress in Avro write jobs
### What changes were proposed in this pull request?
According to
[DeprecatedProperties](https://hadoop.apache.org/docs/r3.3.6/hadoop-project-dist/hadoop-common/DeprecatedProperties.html),
`mapred.output.compress` is deprecated. So in this PR, we use
`mapreduce.output.fileoutputformat.compress` .
### Why are the changes needed?
remove usage of deprecated Hadoop configurations
### Does this PR introduce _any_ user-facing change?
no
### How was this patch tested?
I tested locally by verifying the compressed output files before and after
this change.
### Was this patch authored or co-authored using generative AI tooling?
no
Closes #44799 from yaooqinn/SPARK-46774.
Authored-by: Kent Yao <[email protected]>
Signed-off-by: Dongjoon Hyun <[email protected]>
---
.../avro/src/main/scala/org/apache/spark/sql/avro/AvroUtils.scala | 4 ++--
1 file changed, 2 insertions(+), 2 deletions(-)
diff --git
a/connector/avro/src/main/scala/org/apache/spark/sql/avro/AvroUtils.scala
b/connector/avro/src/main/scala/org/apache/spark/sql/avro/AvroUtils.scala
index f0b70f09aa55..05562c913b19 100644
--- a/connector/avro/src/main/scala/org/apache/spark/sql/avro/AvroUtils.scala
+++ b/connector/avro/src/main/scala/org/apache/spark/sql/avro/AvroUtils.scala
@@ -107,9 +107,9 @@ private[sql] object AvroUtils extends Logging {
val jobConf = job.getConfiguration
AvroCompressionCodec.fromString(codecName) match {
case UNCOMPRESSED =>
- jobConf.setBoolean("mapred.output.compress", false)
+ jobConf.setBoolean("mapreduce.output.fileoutputformat.compress",
false)
case compressed =>
- jobConf.setBoolean("mapred.output.compress", true)
+ jobConf.setBoolean("mapreduce.output.fileoutputformat.compress",
true)
jobConf.set(AvroJob.CONF_OUTPUT_CODEC, compressed.getCodecName)
if (compressed.getSupportCompressionLevel) {
val level =
sqlConf.getConfString(s"spark.sql.avro.$codecName.level",
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]