liurenjie1024 opened a new issue, #13873:
URL: https://github.com/apache/iceberg/issues/13873
### Apache Iceberg version
1.6.1
### Query engine
Spark
### Please describe the bug 🐞
1. Run `mkdir warehouse/default` to create dir
2. Download attachment for table data
3. Run `tar -xzf iceberg.tar.gz -C warehouse/default` to put table data into
dir
4. Start spark shell
```
/home/ubuntu/Apps/spark-3.5.4-bin-hadoop3/bin/spark-sql \
--packages org.apache.iceberg:iceberg-spark-runtime-3.5_2.12:1.9.2\
--conf
spark.sql.catalog.spark_catalog=org.apache.iceberg.spark.SparkCatalog \
--conf spark.sql.catalog.spark_catalog.type=hadoop \
--conf spark.sql.catalog.spark_catalog.warehouse=Workspace/warehouse
```
5. Run count(*) it shows 705 rows.
```
spark-sql (default)> select count(*) from tmp_table_gw0_374721098_0;
705
Time taken: 2.726 seconds, Fetched 1 row(s)
```
6. Run ```select * from tmp_table_gw0_374721098_0``` it reports fetching
rows not 750 rows.
```
spark-sql (default)> select * from tmp_table_gw0_374721098_0;
...
Time taken: 0.403 seconds, Fetched 723 row(s)
```
The oddest part is that the difference only happens when *the first time*
you run `select *` after `select count(*)`, after that when you run `select *`
everything is back to normal.
I tested againts 1.6.1, 1.7.2, and 1.9.2, buth all failed.
[iceberg.tar.gz](https://github.com/user-attachments/files/21889282/iceberg.tar.gz)
### Willingness to contribute
- [ ] I can contribute a fix for this bug independently
- [ ] I would be willing to contribute a fix for this bug with guidance from
the Iceberg community
- [x] I cannot contribute a fix for this bug at this time
--
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
To unsubscribe, e-mail: [email protected]
For queries about this service, please contact Infrastructure at:
[email protected]
---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]