sdd commented on code in PR #982:
URL: https://github.com/apache/iceberg-rust/pull/982#discussion_r2056910313
##########
crates/iceberg/src/arrow/delete_file_manager.rs:
##########
@@ -47,47 +60,533 @@ impl DeleteFileManager for CachingDeleteFileManager {
))
}
}
+// Equality deletes may apply to more than one DataFile in a scan, and so
+// the same equality delete file may be present in more than one invocation of
+// DeleteFileManager::load_deletes in the same scan. We want to deduplicate
these
+// to avoid having to load them twice, so we immediately store cloneable
futures in the
+// state that can be awaited upon to get te EQ deletes. That way we can check
to see if
+// a load of each Eq delete file is already in progress and avoid starting
another one.
+#[derive(Debug, Clone)]
+struct EqDelFuture {
+ result: OnceLock<Predicate>,
+}
+
+impl EqDelFuture {
+ pub fn new() -> (oneshot::Sender<Predicate>, Self) {
+ let (tx, rx) = oneshot::channel();
+ let result = OnceLock::new();
+
+ crate::runtime::spawn({
+ let result = result.clone();
+ async move { result.set(rx.await.unwrap()) }
+ });
+
+ (tx, Self { result })
+ }
+}
+
+impl Future for EqDelFuture {
+ type Output = Predicate;
+
+ fn poll(self: Pin<&mut Self>, _cx: &mut Context<'_>) -> Poll<Self::Output>
{
+ match self.result.get() {
+ None => Poll::Pending,
+ Some(predicate) => Poll::Ready(predicate.clone()),
+ }
+ }
+}
+
+#[derive(Debug, Default)]
+struct DeleteFileManagerState {
+ // delete vectors and positional deletes get merged when loaded into a
single delete vector
+ // per data file
+ delete_vectors: HashMap<String, Arc<RwLock<DeleteVector>>>,
+
+ // equality delete files are parsed into unbound `Predicate`s. We store
them here as
+ // cloneable futures (see note below)
+ equality_deletes: HashMap<String, EqDelFuture>,
+}
+
+type StateRef = Arc<RwLock<DeleteFileManagerState>>;
+
+// Intermediate context during processing of a delete file task.
+enum DeleteFileContext {
+ // TODO: Delete Vector loader from Puffin files
+ InProgEqDel(EqDelFuture),
+ PosDels(ArrowRecordBatchStream),
+ FreshEqDel {
+ batch_stream: ArrowRecordBatchStream,
+ sender: oneshot::Sender<Predicate>,
+ },
+}
+
+// Final result of the processing of a delete file task before
+// results are fully merged into the DeleteFileManager's state
+enum ParsedDeleteFileContext {
+ InProgEqDel(EqDelFuture),
+ DelVecs(HashMap<String, DeleteVector>),
+ EqDel,
+}
#[allow(unused_variables)]
impl CachingDeleteFileManager {
- pub fn new(file_io: FileIO, concurrency_limit_data_files: usize) ->
CachingDeleteFileManager {
- Self {
+ pub(crate) fn new(file_io: FileIO, concurrency_limit_data_files: usize) ->
Self {
+ CachingDeleteFileManager {
file_io,
concurrency_limit_data_files,
+ state: Arc::new(Default::default()),
}
}
+ /// Load the deletes for all the specified tasks
+ ///
+ /// Returned future completes once all loading has finished.
+ ///
+ /// * Create a single stream of all delete file tasks irrespective of
type,
+ /// so that we can respect the combined concurrency limit
+ /// * We then process each in two phases: load and parse.
+ /// * for positional deletes the load phase instantiates an
ArrowRecordBatchStream to
+ /// stream the file contents out
+ /// * for eq deletes, we first check if the EQ delete is already loaded
or being loaded by
+ /// another concurrently processing data file scan task. If it is, we
return a future
+ /// for the pre-existing task from the load phase. If not, we create
such a future
+ /// and store it in the state to prevent other data file tasks from
starting to load
+ /// the same equality delete file, and return a record batch stream
from the load phase
+ /// as per the other delete file types - only this time it is
accompanied by a one-shot
+ /// channel sender that we will eventually use to resolve the shared
future that we stored
+ /// in the state.
+ /// * When this gets updated to add support for delete vectors, the load
phase will return
+ /// a PuffinReader for them.
+ /// * The parse phase parses each record batch stream according to its
associated data type.
+ /// The result of this is a map of data file paths to delete vectors
for the positional
+ /// delete tasks (and in future for the delete vector tasks). For
equality delete
+ /// file tasks, this results in an unbound Predicate.
+ /// * The unbound Predicates resulting from equality deletes are sent to
their associated oneshot
+ /// channel to store them in the right place in the delete file
managers state.
+ /// * The results of all of these futures are awaited on in parallel with
the specified
+ /// level of concurrency and collected into a vec. We then combine
all of the delete
+ /// vector maps that resulted from any positional delete or delete
vector files into a
+ /// single map and persist it in the state.
+ ///
+ ///
+ /// Conceptually, the data flow is like this:
+ /// ```none
+ /// FileScanTaskDeleteFile
+ /// |
+ /// Already-loading EQ Delete |
Everything Else
+ ///
+---------------------------------------------------+
+ /// |
|
+ /// [get existing future] [load
recordbatch stream / puffin]
+ /// DeleteFileContext::InProgEqDel
DeleteFileContext
+ /// |
|
+ /// |
|
+ /// |
+-----------------------------+--------------------------+
+ /// | Pos Del Del Vec (Not
yet Implemented) EQ Del
+ /// | |
| |
+ /// | [parse pos del stream] [parse
del vec puffin] [parse eq del]
+ /// | HashMap<String, RoaringTreeMap> HashMap<String,
RoaringTreeMap> (Predicate, Sender)
+ /// | |
| |
+ /// | |
| [persist to state]
+ /// | |
| ()
+ /// | |
| |
+ /// |
+-----------------------------+--------------------------+
+ /// |
|
+ /// |
[buffer unordered]
+ /// |
|
+ /// |
[combine del vectors]
+ /// |
HashMap<String, RoaringTreeMap>
+ /// |
|
+ /// | [persist
del vectors to state]
+ /// |
()
+ /// |
|
+ ///
+-------------------------+-------------------------+
+ /// |
+ /// [join!]
+ /// ```
pub(crate) async fn load_deletes(
&self,
- delete_file_entries: Vec<FileScanTaskDeleteFile>,
+ delete_file_entries: &[FileScanTaskDeleteFile],
) -> Result<()> {
- // TODO
+ let stream_items = delete_file_entries
+ .iter()
+ .map(|t| (t.clone(), self.file_io.clone(), self.state.clone()))
+ .collect::<Vec<_>>();
+ // NOTE: removing the collect and just passing the iterator to
futures::stream:iter
+ // results in an error 'implementation of `std::ops::FnOnce` is not
general enough'
- if !delete_file_entries.is_empty() {
- Err(Error::new(
- ErrorKind::FeatureUnsupported,
- "Reading delete files is not yet supported",
- ))
- } else {
- Ok(())
+ let task_stream = futures::stream::iter(stream_items.into_iter());
+
+ let results: Vec<ParsedDeleteFileContext> = task_stream
+ .map(move |(task, file_io, state_ref)| async {
+ Self::load_file_for_task(task, file_io, state_ref).await
+ })
+ .map(move |ctx| Ok(async {
Self::parse_file_content_for_task(ctx.await?).await }))
+ .try_buffer_unordered(self.concurrency_limit_data_files)
+ .try_collect::<Vec<_>>()
+ .await?;
+
+ // wait for all in-progress EQ deletes from other tasks
+ let _ = join_all(results.iter().filter_map(|i| {
+ if let ParsedDeleteFileContext::InProgEqDel(fut) = i {
+ Some(fut.clone())
+ } else {
+ None
+ }
+ }))
+ .await;
+
+ let merged_delete_vectors = results
+ .into_iter()
+ .fold(HashMap::default(), Self::merge_delete_vectors);
+
+ self.state.write().unwrap().delete_vectors = merged_delete_vectors;
+
+ Ok(())
+ }
+
+ async fn load_file_for_task(
+ task: FileScanTaskDeleteFile,
+ file_io: FileIO,
+ state: StateRef,
+ ) -> Result<DeleteFileContext> {
+ match task.file_type {
+ DataContentType::PositionDeletes => Ok(DeleteFileContext::PosDels(
+ Self::parquet_to_batch_stream(&task.file_path, file_io).await?,
+ )),
+
+ DataContentType::EqualityDeletes => {
+ let sender = {
+ let mut state = state.write().unwrap();
+ if let Some(existing) =
state.equality_deletes.get(&task.file_path) {
+ return
Ok(DeleteFileContext::InProgEqDel(existing.clone()));
+ }
+
+ let (sender, fut) = EqDelFuture::new();
+
+ state
+ .equality_deletes
+ .insert(task.file_path.to_string(), fut);
+
+ sender
+ };
+
+ Ok(DeleteFileContext::FreshEqDel {
+ batch_stream:
Self::parquet_to_batch_stream(&task.file_path, file_io).await?,
+ sender,
+ })
+ }
+
+ DataContentType::Data => Err(Error::new(
+ ErrorKind::Unexpected,
+ "tasks with files of type Data not expected here",
+ )),
}
}
- pub(crate) fn build_delete_predicate(
+ async fn parse_file_content_for_task(
+ ctx: DeleteFileContext,
+ ) -> Result<ParsedDeleteFileContext> {
+ match ctx {
+ DeleteFileContext::InProgEqDel(fut) =>
Ok(ParsedDeleteFileContext::InProgEqDel(fut)),
+ DeleteFileContext::PosDels(batch_stream) => {
+ let del_vecs =
+
Self::parse_positional_deletes_record_batch_stream(batch_stream).await?;
+ Ok(ParsedDeleteFileContext::DelVecs(del_vecs))
+ }
+ DeleteFileContext::FreshEqDel {
+ sender,
+ batch_stream,
+ } => {
+ let predicate =
+
Self::parse_equality_deletes_record_batch_stream(batch_stream).await?;
+
+ sender
+ .send(predicate)
+ .map_err(|err| {
+ Error::new(
+ ErrorKind::Unexpected,
+ "Could not send eq delete predicate to state",
+ )
+ })
+ .map(|_| ParsedDeleteFileContext::EqDel)
+ }
+ }
+ }
+
+ fn merge_delete_vectors(
+ mut merged_delete_vectors: HashMap<String, Arc<RwLock<DeleteVector>>>,
+ item: ParsedDeleteFileContext,
+ ) -> HashMap<String, Arc<RwLock<DeleteVector>>> {
+ if let ParsedDeleteFileContext::DelVecs(del_vecs) = item {
+ del_vecs.into_iter().for_each(|(key, val)| {
+ let entry = merged_delete_vectors.entry(key).or_default();
+ {
+ let mut inner = entry.write().unwrap();
+ (*inner).intersect_assign(&val);
+ }
+ });
+ }
+
+ merged_delete_vectors
+ }
+
+ /// Loads a RecordBatchStream for a given datafile.
+ async fn parquet_to_batch_stream(
+ data_file_path: &str,
+ file_io: FileIO,
+ ) -> Result<ArrowRecordBatchStream> {
+ /*
+ Essentially a super-cut-down ArrowReader. We can't use ArrowReader
directly
+ as that introduces a circular dependency.
+ */
+ let record_batch_stream =
ArrowReader::create_parquet_record_batch_stream_builder(
+ data_file_path,
+ file_io.clone(),
+ false,
+ )
+ .await?
+ .build()?
+ .map_err(|e| Error::new(ErrorKind::Unexpected, format!("{}", e)));
+
+ Ok(Box::pin(record_batch_stream) as ArrowRecordBatchStream)
+ }
+
+ /// Parses a record batch stream coming from positional delete files
+ ///
+ /// Returns a map of data file path to a delete vector
+ async fn parse_positional_deletes_record_batch_stream(
+ stream: ArrowRecordBatchStream,
+ ) -> Result<HashMap<String, DeleteVector>> {
+ // TODO
+
+ Ok(HashMap::default())
+ }
+
+ /// Parses record batch streams from individual equality delete files
+ ///
+ /// Returns an unbound Predicate for each batch stream
+ async fn parse_equality_deletes_record_batch_stream(
+ streams: ArrowRecordBatchStream,
+ ) -> Result<Predicate> {
+ // TODO
+
+ Ok(AlwaysTrue)
+ }
+
+ /// Builds eq delete predicate for the provided task.
+ ///
+ /// Must await on load_deletes before calling this.
+ pub(crate) async fn build_delete_predicate_for_task(
&self,
- snapshot_schema: SchemaRef,
+ file_scan_task: &FileScanTask,
) -> Result<Option<BoundPredicate>> {
- // TODO
+ // * Filter the task's deletes into just the Equality deletes
+ // * Retrieve the unbound predicate for each from
self.state.equality_deletes
+ // * Logical-AND them all together to get a single combined `Predicate`
+ // * Bind the predicate to the task's schema to get a `BoundPredicate`
+
+ let mut combined_predicate = AlwaysTrue;
+ for delete in &file_scan_task.deletes {
+ if !is_equality_delete(delete) {
+ continue;
+ }
+
+ let predicate = {
+ let state = self.state.read().unwrap();
- Ok(None)
+ let Some(predicate) =
state.equality_deletes.get(&delete.file_path) else {
+ return Err(Error::new(
+ ErrorKind::Unexpected,
+ format!(
+ "Missing predicate for equality delete file '{}'",
+ delete.file_path
+ ),
+ ));
+ };
+
+ predicate.clone()
+ };
+
+ combined_predicate = combined_predicate.and(predicate.await);
Review Comment:
I've updated https://github.com/apache/iceberg-rust/pull/1017 to make use of
the equality_ids passed through the file scan tasks to ensure that only columns
whose field ids are present in a given delete file's eq_ids are included in the
predicate:
https://github.com/apache/iceberg-rust/pull/1017/commits/f62ac3f98942707eeb7a06410a52a9a5eea951ac#diff-7d31974b51961466c4be7da66f05f93972ac55f35b7f78252aa6df3a6d087ea4R440
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]