liurenjie1024 commented on issue #172:
URL: https://github.com/apache/iceberg-rust/issues/172#issuecomment-2496515292

   > I _think_ that should work, the DataFusion wrapper can just hook the 
iceberg metadata operations into via that StorageProvider trait, and then use 
the DataFusion machinery directly for the actual file I/O. I presume this would 
be possible, or does this library expect to also perform the file I/O?
   
   I think that depends. This library provides several functionality for 
reading:
   1. Planning data files to read for table scan. This involves access to 
metadata files, and it use `FileIO` to do that.
   2. A builtin reader to convert planned table scan tasks to arrow record 
batch stream. In this case, we also use `FileIO` to access data files.
   
   For integration with query engines like datafusion, datafusion only need to 
use planning file api, and perform scanning by itself.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


---------------------------------------------------------------------
To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org
For additional commands, e-mail: issues-h...@iceberg.apache.org

Reply via email to