advancedxy commented on PR #7733: URL: https://github.com/apache/iceberg/pull/7733#issuecomment-1567894428
> > ``` > > set spark.datasource.iceberg.locality=false; -- works only for DataFrame > > set spark.sql.iceberg.locality.enabled=false; -- works both DataFrame and Spark SQL > > ``` > > The division is not true, the key point here is where the table is loaded through TableProvider or CatalogPlugin. > > 1. table loaded through TableProvider examples > > * DataFrame cases > > ``` > spark.read.format("iceberg").xxx > > df.write.format("iceberg").xxx > ``` > > * SQL cases > > ``` > create table t_iceberg (...) using iceberg; > select ... from t_iceberg; > insert into t_iceberg select ...; > ``` > > 2. table loaded through CatalogPlugin > > Assume `iceberg` catalog is pre-setup properly > > * DataFrame cases > > ``` > spark.table("iceberg.db.tbl")... > df.writeTo.("iceberg.db.tbl")... > ``` > > * SQL cases > > ``` > select ... from iceberg.db.tbl; > insert into iceberg.db.tbl select ...; > ``` Thanks for the detail explanation. By `DataFrame` I mean `spark.read.format` and `df.write.format` cases, the ` create table t_iceberg (...) using iceberg` doesn't occur to me, but it's indeed loaded by `TableProvider`. Anyway you can get my point is that tables loaded from `TableProvider` and from `CatalogPlugin` don't have unified configuration settings. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org For additional commands, e-mail: issues-h...@iceberg.apache.org