amogh-jahagirdar commented on code in PR #13400:
URL: https://github.com/apache/iceberg/pull/13400#discussion_r2282712017


##########
core/src/main/java/org/apache/iceberg/RestTableScan.java:
##########
@@ -0,0 +1,256 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *   http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing,
+ * software distributed under the License is distributed on an
+ * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+ * KIND, either express or implied.  See the License for the
+ * specific language governing permissions and limitations
+ * under the License.
+ */
+package org.apache.iceberg;
+
+import java.util.List;
+import java.util.Map;
+import java.util.function.Supplier;
+import java.util.stream.Collectors;
+import org.apache.iceberg.catalog.TableIdentifier;
+import org.apache.iceberg.io.CloseableIterable;
+import org.apache.iceberg.relocated.com.google.common.collect.Lists;
+import org.apache.iceberg.rest.ErrorHandlers;
+import org.apache.iceberg.rest.ParserContext;
+import org.apache.iceberg.rest.PlanStatus;
+import org.apache.iceberg.rest.RESTClient;
+import org.apache.iceberg.rest.ResourcePaths;
+import org.apache.iceberg.rest.requests.PlanTableScanRequest;
+import org.apache.iceberg.rest.responses.FetchPlanningResultResponse;
+import org.apache.iceberg.rest.responses.PlanTableScanResponse;
+import org.apache.iceberg.types.Types;
+import org.apache.iceberg.util.ParallelIterable;
+
+public class RestTableScan extends DataTableScan {
+  private final RESTClient client;
+  private final String path;
+  private final Supplier<Map<String, String>> headers;
+  private final TableOperations operations;
+  private final Table table;
+  private final ResourcePaths resourcePaths;
+  private final TableIdentifier tableIdentifier;
+
+  // TODO revisit if this property should be configurable
+  private static final int FETCH_PLANNING_SLEEP_DURATION_MS = 1000;
+
+  public RestTableScan(
+      Table table,
+      Schema schema,
+      TableScanContext context,
+      RESTClient client,
+      String path,
+      Supplier<Map<String, String>> headers,
+      TableOperations operations,
+      TableIdentifier tableIdentifier,
+      ResourcePaths resourcePaths) {
+    super(table, schema, context);
+    this.table = table;
+    this.client = client;
+    this.headers = headers;
+    this.path = path;
+    this.operations = operations;
+    this.tableIdentifier = tableIdentifier;
+    this.resourcePaths = resourcePaths;
+  }
+
+  @Override
+  protected TableScan newRefinedScan(
+      Table refinedTable, Schema refinedSchema, TableScanContext 
refinedContext) {
+    return new RestTableScan(
+        refinedTable,
+        refinedSchema,
+        refinedContext,
+        client,
+        path,
+        headers,
+        operations,
+        tableIdentifier,
+        resourcePaths);
+  }
+
+  @Override
+  public CloseableIterable<FileScanTask> planFiles() {
+    List<String> selectedColumns =
+        
schema().columns().stream().map(Types.NestedField::name).collect(Collectors.toList());
+
+    List<String> statsFields = null;
+    if (columnsToKeepStats() != null) {
+      statsFields =
+          columnsToKeepStats().stream()
+              .map(columnId -> schema().findColumnName(columnId))
+              .collect(Collectors.toList());
+    }
+
+    Long startSnapshotId = context().fromSnapshotId();
+    Long endSnapshotId = context().toSnapshotId();
+    Long snapshotId = snapshotId();

Review Comment:
   I'd just inline these where they're actually used for improved readability



##########
core/src/main/java/org/apache/iceberg/ScanTasksIterable.java:
##########
@@ -0,0 +1,217 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *   http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing,
+ * software distributed under the License is distributed on an
+ * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+ * KIND, either express or implied.  See the License for the
+ * specific language governing permissions and limitations
+ * under the License.
+ */
+package org.apache.iceberg;
+
+import java.io.IOException;
+import java.util.List;
+import java.util.Map;
+import java.util.concurrent.ExecutorService;
+import java.util.function.Supplier;
+import org.apache.iceberg.catalog.TableIdentifier;
+import org.apache.iceberg.io.CloseableIterable;
+import org.apache.iceberg.io.CloseableIterator;
+import org.apache.iceberg.relocated.com.google.common.collect.Lists;
+import org.apache.iceberg.rest.ErrorHandlers;
+import org.apache.iceberg.rest.ParserContext;
+import org.apache.iceberg.rest.RESTClient;
+import org.apache.iceberg.rest.ResourcePaths;
+import org.apache.iceberg.rest.requests.FetchScanTasksRequest;
+import org.apache.iceberg.rest.responses.FetchScanTasksResponse;
+import org.apache.iceberg.util.ParallelIterable;
+
+public class ScanTasksIterable implements CloseableIterable<FileScanTask> {
+  private final RESTClient client;
+  private final ResourcePaths resourcePaths;
+  private final TableIdentifier tableIdentifier;
+  private final Supplier<Map<String, String>> headers;
+  // parallelizing on this where a planTask produces a list of file scan 
tasks, as
+  //  well more planTasks.
+  private final String planTask;
+  private final List<FileScanTask> fileScanTasks;
+  private final ExecutorService executorService;
+  private final Map<Integer, PartitionSpec> specsById;
+  private final boolean caseSensitive;
+
+  public ScanTasksIterable(
+      String planTask,
+      RESTClient client,
+      ResourcePaths resourcePaths,
+      TableIdentifier tableIdentifier,
+      Supplier<Map<String, String>> headers,
+      ExecutorService executorService,
+      Map<Integer, PartitionSpec> specsById,
+      boolean caseSensitive) {
+    this.planTask = planTask;
+    this.fileScanTasks = null;
+    this.client = client;
+    this.resourcePaths = resourcePaths;
+    this.tableIdentifier = tableIdentifier;
+    this.headers = headers;
+    this.executorService = executorService;
+    this.specsById = specsById;
+    this.caseSensitive = caseSensitive;
+  }
+
+  public ScanTasksIterable(
+      List<FileScanTask> fileScanTasks,
+      RESTClient client,
+      ResourcePaths resourcePaths,
+      TableIdentifier tableIdentifier,
+      Supplier<Map<String, String>> headers,
+      ExecutorService executorService,
+      Map<Integer, PartitionSpec> specsById,
+      boolean caseSensitive) {
+    this.planTask = null;
+    this.fileScanTasks = fileScanTasks;
+    this.client = client;
+    this.resourcePaths = resourcePaths;
+    this.tableIdentifier = tableIdentifier;
+    this.headers = headers;
+    this.executorService = executorService;
+    this.specsById = specsById;
+    this.caseSensitive = caseSensitive;
+  }
+
+  @Override
+  public CloseableIterator<FileScanTask> iterator() {
+    return new ScanTasksIterator(
+        planTask,
+        fileScanTasks,
+        client,
+        resourcePaths,
+        tableIdentifier,
+        headers,
+        executorService,
+        specsById,
+        caseSensitive);
+  }
+
+  @Override
+  public void close() throws IOException {}
+
+  private static class ScanTasksIterator implements 
CloseableIterator<FileScanTask> {
+    private final RESTClient client;
+    private final ResourcePaths resourcePaths;
+    private final TableIdentifier tableIdentifier;
+    private final Supplier<Map<String, String>> headers;
+    private String planTask;
+    private final List<FileScanTask> fileScanTasks;
+    private final ExecutorService executorService;
+    private final Map<Integer, PartitionSpec> specsById;
+    private final boolean caseSensitive;
+
+    ScanTasksIterator(
+        String planTask,
+        List<FileScanTask> fileScanTasks,
+        RESTClient client,
+        ResourcePaths resourcePaths,
+        TableIdentifier tableIdentifier,
+        Supplier<Map<String, String>> headers,
+        ExecutorService executorService,
+        Map<Integer, PartitionSpec> specsById,
+        boolean caseSensitive) {
+      this.client = client;
+      this.resourcePaths = resourcePaths;
+      this.tableIdentifier = tableIdentifier;
+      this.headers = headers;
+      this.planTask = planTask;
+      this.fileScanTasks = fileScanTasks != null ? fileScanTasks : 
Lists.newArrayList();
+      this.executorService = executorService;
+      this.specsById = specsById;
+      this.caseSensitive = caseSensitive;
+    }
+
+    @Override
+    public boolean hasNext() {
+      if (!fileScanTasks.isEmpty()) {
+        // Have file scan tasks so continue to consume
+        return true;
+      }
+      // Out of file scan tasks, so need to now fetch more from each planTask
+      // Service can send back more planTasks which acts as pagination
+      if (planTask != null) {
+        fetchScanTasks(planTask);
+        planTask = null;
+        // Make another hasNext() call, as more fileScanTasks have been fetched
+        return hasNext();
+      }
+      // we have no file scan tasks left to consume
+      // so means we are finished
+      return false;
+    }
+
+    @Override
+    public FileScanTask next() {
+      return fileScanTasks.remove(0);
+    }
+
+    private void fetchScanTasks(String withPlanTask) {
+      FetchScanTasksRequest fetchScanTasksRequest = new 
FetchScanTasksRequest(withPlanTask);
+      // we need injectable values here
+      ParserContext parserContext =
+          ParserContext.builder()
+              .add("specsById", specsById)
+              .add("caseSensitive", caseSensitive)
+              .build();
+      FetchScanTasksResponse response =
+          client.post(
+              resourcePaths.fetchScanTasks(tableIdentifier),
+              fetchScanTasksRequest,
+              FetchScanTasksResponse.class,
+              headers.get(),
+              ErrorHandlers.defaultErrorHandler(),
+              stringStringMap -> {},
+              parserContext);
+      if (response.fileScanTasks() != null) {
+        fileScanTasks.addAll(response.fileScanTasks());
+      }
+
+      if (response.planTasks() != null) {
+        // this is the case where a plan task returned an additional plan 
task, so ensure that this
+        // result is added to top level fileScanTasks list.
+        // confirmed working with catalog test
+        // 
#testPlanTableScanAndFetchScanTasksWithCompletedStatusAndNestedPlanTasks
+        Iterable<FileScanTask> fileScanTasksFromPlanTasks =
+            getScanTasksIterable(response.planTasks());
+        fileScanTasksFromPlanTasks.forEach(fileScanTasks::add);
+      }
+    }
+
+    public CloseableIterable<FileScanTask> getScanTasksIterable(List<String> 
planTasks) {
+      List<ScanTasksIterable> iterableOfScanTaskIterables = 
Lists.newArrayList();
+      for (String withPlanTask : planTasks) {
+        ScanTasksIterable iterable =
+            new ScanTasksIterable(
+                withPlanTask,
+                client,
+                resourcePaths,
+                tableIdentifier,
+                headers,
+                executorService,
+                specsById,
+                caseSensitive);
+        iterableOfScanTaskIterables.add(iterable);
+      }
+      return new ParallelIterable<>(iterableOfScanTaskIterables, 
executorService);
+    }
+
+    @Override
+    public void close() throws IOException {}

Review Comment:
   Shouldn't `close` cancel any ongoing work? 



##########
core/src/main/java/org/apache/iceberg/RestTableScan.java:
##########
@@ -0,0 +1,256 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *   http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing,
+ * software distributed under the License is distributed on an
+ * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+ * KIND, either express or implied.  See the License for the
+ * specific language governing permissions and limitations
+ * under the License.
+ */
+package org.apache.iceberg;
+
+import java.util.List;
+import java.util.Map;
+import java.util.function.Supplier;
+import java.util.stream.Collectors;
+import org.apache.iceberg.catalog.TableIdentifier;
+import org.apache.iceberg.io.CloseableIterable;
+import org.apache.iceberg.relocated.com.google.common.collect.Lists;
+import org.apache.iceberg.rest.ErrorHandlers;
+import org.apache.iceberg.rest.ParserContext;
+import org.apache.iceberg.rest.PlanStatus;
+import org.apache.iceberg.rest.RESTClient;
+import org.apache.iceberg.rest.ResourcePaths;
+import org.apache.iceberg.rest.requests.PlanTableScanRequest;
+import org.apache.iceberg.rest.responses.FetchPlanningResultResponse;
+import org.apache.iceberg.rest.responses.PlanTableScanResponse;
+import org.apache.iceberg.types.Types;
+import org.apache.iceberg.util.ParallelIterable;
+
+public class RestTableScan extends DataTableScan {
+  private final RESTClient client;
+  private final String path;
+  private final Supplier<Map<String, String>> headers;
+  private final TableOperations operations;
+  private final Table table;
+  private final ResourcePaths resourcePaths;
+  private final TableIdentifier tableIdentifier;
+
+  // TODO revisit if this property should be configurable
+  private static final int FETCH_PLANNING_SLEEP_DURATION_MS = 1000;
+
+  public RestTableScan(
+      Table table,
+      Schema schema,
+      TableScanContext context,
+      RESTClient client,
+      String path,
+      Supplier<Map<String, String>> headers,
+      TableOperations operations,
+      TableIdentifier tableIdentifier,
+      ResourcePaths resourcePaths) {
+    super(table, schema, context);
+    this.table = table;
+    this.client = client;
+    this.headers = headers;
+    this.path = path;
+    this.operations = operations;
+    this.tableIdentifier = tableIdentifier;
+    this.resourcePaths = resourcePaths;
+  }
+
+  @Override
+  protected TableScan newRefinedScan(
+      Table refinedTable, Schema refinedSchema, TableScanContext 
refinedContext) {
+    return new RestTableScan(
+        refinedTable,
+        refinedSchema,
+        refinedContext,
+        client,
+        path,
+        headers,
+        operations,
+        tableIdentifier,
+        resourcePaths);
+  }
+
+  @Override
+  public CloseableIterable<FileScanTask> planFiles() {
+    List<String> selectedColumns =
+        
schema().columns().stream().map(Types.NestedField::name).collect(Collectors.toList());
+
+    List<String> statsFields = null;
+    if (columnsToKeepStats() != null) {
+      statsFields =
+          columnsToKeepStats().stream()
+              .map(columnId -> schema().findColumnName(columnId))
+              .collect(Collectors.toList());
+    }
+
+    Long startSnapshotId = context().fromSnapshotId();
+    Long endSnapshotId = context().toSnapshotId();
+    Long snapshotId = snapshotId();
+
+    PlanTableScanRequest.Builder planTableScanRequestBuilder =
+        new PlanTableScanRequest.Builder()
+            .withSelect(selectedColumns)
+            .withFilter(filter())
+            .withCaseSensitive(isCaseSensitive())
+            .withStatsFields(statsFields);
+
+    if (startSnapshotId != null && endSnapshotId != null) {
+      planTableScanRequestBuilder
+          .withStartSnapshotId(startSnapshotId)
+          .withEndSnapshotId(endSnapshotId)
+          .withUseSnapshotSchema(true);
+
+    } else if (snapshotId != null) {
+      boolean useSnapShotSchema = snapshotId != 
table.currentSnapshot().snapshotId();
+      planTableScanRequestBuilder
+          .withSnapshotId(snapshotId)
+          .withUseSnapshotSchema(useSnapShotSchema);

Review Comment:
   I don't follow, in this case shouldn't `useSnapshotSchema` always be true? 
When time travelling to a specific snapshot  we always use the snapshot schema



##########
core/src/main/java/org/apache/iceberg/RestTableScan.java:
##########
@@ -0,0 +1,256 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *   http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing,
+ * software distributed under the License is distributed on an
+ * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+ * KIND, either express or implied.  See the License for the
+ * specific language governing permissions and limitations
+ * under the License.
+ */
+package org.apache.iceberg;
+
+import java.util.List;
+import java.util.Map;
+import java.util.function.Supplier;
+import java.util.stream.Collectors;
+import org.apache.iceberg.catalog.TableIdentifier;
+import org.apache.iceberg.io.CloseableIterable;
+import org.apache.iceberg.relocated.com.google.common.collect.Lists;
+import org.apache.iceberg.rest.ErrorHandlers;
+import org.apache.iceberg.rest.ParserContext;
+import org.apache.iceberg.rest.PlanStatus;
+import org.apache.iceberg.rest.RESTClient;
+import org.apache.iceberg.rest.ResourcePaths;
+import org.apache.iceberg.rest.requests.PlanTableScanRequest;
+import org.apache.iceberg.rest.responses.FetchPlanningResultResponse;
+import org.apache.iceberg.rest.responses.PlanTableScanResponse;
+import org.apache.iceberg.types.Types;
+import org.apache.iceberg.util.ParallelIterable;
+
+public class RestTableScan extends DataTableScan {

Review Comment:
   I feel like this should reside in the REST module? 



##########
core/src/main/java/org/apache/iceberg/ScanTasksIterable.java:
##########
@@ -0,0 +1,217 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *   http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing,
+ * software distributed under the License is distributed on an
+ * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+ * KIND, either express or implied.  See the License for the
+ * specific language governing permissions and limitations
+ * under the License.
+ */
+package org.apache.iceberg;
+
+import java.io.IOException;
+import java.util.List;
+import java.util.Map;
+import java.util.concurrent.ExecutorService;
+import java.util.function.Supplier;
+import org.apache.iceberg.catalog.TableIdentifier;
+import org.apache.iceberg.io.CloseableIterable;
+import org.apache.iceberg.io.CloseableIterator;
+import org.apache.iceberg.relocated.com.google.common.collect.Lists;
+import org.apache.iceberg.rest.ErrorHandlers;
+import org.apache.iceberg.rest.ParserContext;
+import org.apache.iceberg.rest.RESTClient;
+import org.apache.iceberg.rest.ResourcePaths;
+import org.apache.iceberg.rest.requests.FetchScanTasksRequest;
+import org.apache.iceberg.rest.responses.FetchScanTasksResponse;
+import org.apache.iceberg.util.ParallelIterable;
+
+public class ScanTasksIterable implements CloseableIterable<FileScanTask> {

Review Comment:
   Same as above, I feel like this should reside in the REST module



##########
core/src/main/java/org/apache/iceberg/RestTableScan.java:
##########
@@ -0,0 +1,256 @@
+/*
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements.  See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership.  The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License.  You may obtain a copy of the License at
+ *
+ *   http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing,
+ * software distributed under the License is distributed on an
+ * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
+ * KIND, either express or implied.  See the License for the
+ * specific language governing permissions and limitations
+ * under the License.
+ */
+package org.apache.iceberg;
+
+import java.util.List;
+import java.util.Map;
+import java.util.function.Supplier;
+import java.util.stream.Collectors;
+import org.apache.iceberg.catalog.TableIdentifier;
+import org.apache.iceberg.io.CloseableIterable;
+import org.apache.iceberg.relocated.com.google.common.collect.Lists;
+import org.apache.iceberg.rest.ErrorHandlers;
+import org.apache.iceberg.rest.ParserContext;
+import org.apache.iceberg.rest.PlanStatus;
+import org.apache.iceberg.rest.RESTClient;
+import org.apache.iceberg.rest.ResourcePaths;
+import org.apache.iceberg.rest.requests.PlanTableScanRequest;
+import org.apache.iceberg.rest.responses.FetchPlanningResultResponse;
+import org.apache.iceberg.rest.responses.PlanTableScanResponse;
+import org.apache.iceberg.types.Types;
+import org.apache.iceberg.util.ParallelIterable;
+
+public class RestTableScan extends DataTableScan {
+  private final RESTClient client;
+  private final String path;
+  private final Supplier<Map<String, String>> headers;
+  private final TableOperations operations;
+  private final Table table;
+  private final ResourcePaths resourcePaths;
+  private final TableIdentifier tableIdentifier;
+
+  // TODO revisit if this property should be configurable
+  private static final int FETCH_PLANNING_SLEEP_DURATION_MS = 1000;
+
+  public RestTableScan(
+      Table table,
+      Schema schema,
+      TableScanContext context,
+      RESTClient client,
+      String path,
+      Supplier<Map<String, String>> headers,
+      TableOperations operations,
+      TableIdentifier tableIdentifier,
+      ResourcePaths resourcePaths) {
+    super(table, schema, context);
+    this.table = table;
+    this.client = client;
+    this.headers = headers;
+    this.path = path;
+    this.operations = operations;
+    this.tableIdentifier = tableIdentifier;
+    this.resourcePaths = resourcePaths;
+  }
+
+  @Override
+  protected TableScan newRefinedScan(
+      Table refinedTable, Schema refinedSchema, TableScanContext 
refinedContext) {
+    return new RestTableScan(
+        refinedTable,
+        refinedSchema,
+        refinedContext,
+        client,
+        path,
+        headers,
+        operations,
+        tableIdentifier,
+        resourcePaths);
+  }
+
+  @Override
+  public CloseableIterable<FileScanTask> planFiles() {
+    List<String> selectedColumns =
+        
schema().columns().stream().map(Types.NestedField::name).collect(Collectors.toList());
+
+    List<String> statsFields = null;
+    if (columnsToKeepStats() != null) {
+      statsFields =
+          columnsToKeepStats().stream()
+              .map(columnId -> schema().findColumnName(columnId))
+              .collect(Collectors.toList());
+    }
+
+    Long startSnapshotId = context().fromSnapshotId();
+    Long endSnapshotId = context().toSnapshotId();
+    Long snapshotId = snapshotId();
+
+    PlanTableScanRequest.Builder planTableScanRequestBuilder =
+        new PlanTableScanRequest.Builder()
+            .withSelect(selectedColumns)
+            .withFilter(filter())
+            .withCaseSensitive(isCaseSensitive())
+            .withStatsFields(statsFields);
+
+    if (startSnapshotId != null && endSnapshotId != null) {
+      planTableScanRequestBuilder
+          .withStartSnapshotId(startSnapshotId)
+          .withEndSnapshotId(endSnapshotId)
+          .withUseSnapshotSchema(true);
+
+    } else if (snapshotId != null) {
+      boolean useSnapShotSchema = snapshotId != 
table.currentSnapshot().snapshotId();
+      planTableScanRequestBuilder
+          .withSnapshotId(snapshotId)
+          .withUseSnapshotSchema(useSnapShotSchema);
+
+    } else {
+      
planTableScanRequestBuilder.withSnapshotId(table().currentSnapshot().snapshotId());
+    }
+
+    return planTableScan(planTableScanRequestBuilder.build());
+  }
+
+  private CloseableIterable<FileScanTask> planTableScan(PlanTableScanRequest 
planTableScanRequest) {
+    // we need to inject specById map here and also the caseSensitive
+    ParserContext context =
+        ParserContext.builder()
+            .add("specsById", table.specs())
+            .add("caseSensitive", context().caseSensitive())
+            .build();
+    PlanTableScanResponse response =
+        client.post(
+            resourcePaths.planTableScan(tableIdentifier),
+            planTableScanRequest,
+            PlanTableScanResponse.class,
+            headers.get(),
+            ErrorHandlers.defaultErrorHandler(),
+            stringStringMap -> {},
+            context);
+
+    PlanStatus planStatus = response.planStatus();
+    switch (planStatus) {
+      case COMPLETED:
+        // List<FileScanTask> fileScanTasks = 
bindFileScanTasksWithSpec(response.fileScanTasks());
+        return getScanTasksIterable(response.planTasks(), 
response.fileScanTasks());
+      case SUBMITTED:
+        return fetchPlanningResult(response.planId());
+      case FAILED:
+        throw new RuntimeException(
+            "Received \"failed\" status from service when planning a table 
scan");
+      default:

Review Comment:
   Can it not also be a cancelled status here as well? 
https://github.com/apache/iceberg/blob/main/open-api/rest-catalog-open-api.yaml#L3443



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: [email protected]

For queries about this service, please contact Infrastructure at:
[email protected]


---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to