rdblue commented on code in PR #11839: URL: https://github.com/apache/iceberg/pull/11839#discussion_r1925832536
########## flink/v1.20/flink/src/test/java/org/apache/iceberg/flink/data/TestFlinkParquetReader.java: ########## @@ -236,4 +457,44 @@ protected void writeAndValidate(Schema schema) throws IOException { RandomGenericData.generateFallbackRecords(schema, NUM_RECORDS, 21124, NUM_RECORDS / 20), schema); } + + @Override + protected void writeAndValidate(Schema writeSchema, Schema expectedSchema) throws IOException { + assumeThat( + TypeUtil.find( + writeSchema, + type -> type.isMapType() && type.asMapType().keyType() != Types.StringType.get())) + .as("Parquet Avro cannot write non-string map keys") + .isNull(); Review Comment: Why is this using parquet-avro to read when the version of `writeAndValidate` just above does not have this restriction? There are assertions to validate `Record` with `RowData` so I think this should use the same approach. That would also allow you to share the implementation of both `writeAndValidate` versions, as was done in the other cases. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org For additional commands, e-mail: issues-h...@iceberg.apache.org