Tonylin1998 opened a new issue, #9956:
URL: https://github.com/apache/iceberg/issues/9956

   ### Query engine
   
   spark
   
   ### Question
   
   I was using Iceberg with PySpark, and using JDBC catalog, and set warehouse 
to GCS
   
   I creata a table using `date` as partition key.
   I write some data into table, and decide to delete date=20240220, so I
   ```
   spark.sql(f"DELETE FROM {iceberg_table} WHERE date = '20241220'")
   spark.sql(f"CALL {catalog_name}.system.expire_snapshots('{iceberg_table}')")
   ```
   I find that the parquet file under `date=20240220` is deleted, but the 
folder `date=20240220` still remain
   
   Also the same, when I drop the table using
   ```
   spark.sql(f"DROP TABLE {iceberg_table} PURGE")
   ```
   the data will be deleted, but all the partition folders will still remain
   These behavior cause many empty folders in my gcs, I wonder if there is any 
way I can do in iceberg to prevent this from happening?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org
For additional commands, e-mail: issues-h...@iceberg.apache.org

Reply via email to