aylonsk commented on PR #14932: URL: https://github.com/apache/lucene/pull/14932#issuecomment-3156552514
Thank you for your suggestions @kaivalnp, I have pushed these changes to the PR. @jpountz I ran the knnPerfTest on the baseline VarInt vs candidate GroupVarInt implementations. These tests was run with fairly standard hyperparameters, and for each test, the median results of 3 runs was taken (a PR that will hopefully be approved in LuceneUtil). Looking at the results, it seems that removing the top-level abstraction from the reader did not visibly affect the performance improvement, which is good. ``` VarInt Median Results: recall latency(ms) netCPU avgCpuCount nDoc topK fanout maxConn beamWidth quantized index(s) index_docs/s num_segments index_size(MB) vec_disk(MB) vec_RAM(MB) indexType 0.938 1.411 1.392 0.988 50000 100 50 64 250 no 0.00 Infinity 8 22.30 19.073 19.073 HNSW ``` ``` GroupVarInt Median Results: recall latency(ms) netCPU avgCpuCount nDoc topK fanout maxConn beamWidth quantized index(s) index_docs/s num_segments index_size(MB) vec_disk(MB) vec_RAM(MB) indexType 0.937 1.329 1.311 0.986 50000 100 50 64 250 no 0.00 Infinity 8 20.55 19.073 19.073 HNSW ``` Median Latency Improvement: ~5.81% -- This is an automated message from the Apache Git Service. To respond to the message, please log on to GitHub and use the URL above to go to the specific comment. To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org For queries about this service, please contact Infrastructure at: us...@infra.apache.org --------------------------------------------------------------------- To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org For additional commands, e-mail: issues-h...@lucene.apache.org