hussein-awala commented on PR #9902: URL: https://github.com/apache/iceberg/pull/9902#issuecomment-2005106161
> In that case I would probably just add a new test class where reading and writing is done through Spark. I think the purpose of TestSparkReaderWithBloomFilter was to actually only read through Spark Yes I can do that, but some logics will be duplicated (data writer for avro and spark, random data generator, ...), knowing that replacing avro writer with spark df write reduced the code size (`+435 −742` despite the addition of a new test), and simplified it. If you check the logic of the existing test, it didn't change, I just updated the extraction of row values to do the assertion. WDYT? If you still consider it an unnecessary change, I can create a new class for the new test. -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@iceberg.apache.org For additional commands, e-mail: issues-h...@iceberg.apache.org