Tonylin1998 opened a new issue, #9956: URL: https://github.com/apache/iceberg/issues/9956
### Query engine spark ### Question I was using Iceberg with PySpark, and using JDBC catalog, and set warehouse to GCS I creata a table using `date` as partition key. I write some data into table, and decide to delete date=20240220, so I ``` spark.sql(f"DELETE FROM {iceberg_table} WHERE date = '20241220'") spark.sql(f"CALL {catalog_name}.system.expire_snapshots('{iceberg_table}')") ``` I find that the parquet file under `date=20240220` is deleted, but the folder `date=20240220` still remain Also the same, when I drop the table using ``` spark.sql(f"DROP TABLE {iceberg_table} PURGE") ``` the data will be deleted, but all the partition folders will still remain These behavior cause many empty folders in my gcs, I wonder if there is any way I can do in iceberg to prevent this from happening? -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org For additional commands, e-mail: issues-h...@iceberg.apache.org