lichaohao opened a new issue, #6330: URL: https://github.com/apache/iceberg/issues/6330
### Query engine iceberg:1.0.0 spark:3.2.0 flink:1.13.2 catalog:hive-catalog ### Question iceberg:1.0.0 spark:3.2.0 flink:1.13.2 catalog:hive-catalog source table:mysql cdc table: mysql_cdc_source sink table:iceberg table: my_iceberg_sink ==> primary key id ,format-version=2, write.upsert.enabled=true execute sql: (checkpoint 1min) upsert into my_iceberg_sink select * from mysql_cdc_source; ps: mysql exists insert and update operation when the job is running somet time, i want to rewrite the iceberg data file into bigger one, 【spark execute】 (call "hive_prod.system.rewrite_data_files(my_iceberg_sink)") get the following exception message: can not commit,found new position delete for replaced data file: GenericDataFile....hdfs://xxxxdata.parquet all above that what should i do can execute "call hive_prod.system.rewrite_data_files(my_iceberg_sink) " correctly? Thank you for your answer! -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org For additional commands, e-mail: issues-h...@iceberg.apache.org