github-actions[bot] closed issue #8862: Distributed execution of
DeleteReachableFilesSparkAction
URL: https://github.com/apache/iceberg/issues/8862
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to th
github-actions[bot] commented on issue #8862:
URL: https://github.com/apache/iceberg/issues/8862#issuecomment-2401026987
This issue has been closed because it has not received any activity in the
last 14 days since being marked as 'stale'
--
This is an automated message from the Apache Gi
github-actions[bot] commented on issue #8862:
URL: https://github.com/apache/iceberg/issues/8862#issuecomment-2367034243
This issue has been automatically marked as stale because it has been open
for 180 days with no activity. It will be closed in next 14 days if no further
activity occurs.
RussellSpitzer commented on issue #8862:
URL: https://github.com/apache/iceberg/issues/8862#issuecomment-1769209631
Yep, in our experience we have also seen a single driver is sufficient to
issue a very large number of deletes.
--
This is an automated message from the Apache Git Service.
tmnd1991 commented on issue #8862:
URL: https://github.com/apache/iceberg/issues/8862#issuecomment-1767681093
Hi @RussellSpitzer , yes I mean having the delete operations distributed. I
guess it’s difficult because you don’t know how many spark executor cores you
might have in any given mom
RussellSpitzer commented on issue #8862:
URL: https://github.com/apache/iceberg/issues/8862#issuecomment-1767364918
It is distributed in file discovery. Do you mean have the deletes
distributed? Previously we didn't want to do this because it's very difficult
to control parallelism of delet