[GitHub] [lucene] jbellis closed pull request #12303: Address HNSW Searcher performance regression

2023-05-21 Thread via GitHub


jbellis closed pull request #12303: Address HNSW Searcher performance regression
URL: https://github.com/apache/lucene/pull/12303


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene] gsmiller commented on issue #12317: Option for disabling term dictionary compression

2023-05-21 Thread via GitHub


gsmiller commented on issue #12317:
URL: https://github.com/apache/lucene/issues/12317#issuecomment-1556202535

   I'm no expert in this area of our codec, but I'm curious to understand the 
issue a bit better. From the flame chart you provide, it looks like you're 
primarily looking at an indexing-related performance issue. Is that correct? 
When you disabled the patch, did you notice query-time performance changes? 
Compression isn't only useful for saving disk space; it's useful for keeping 
index pages hot in the OS cache and getting better data locality, which 
translates to better query-time performance. If your customer is running a 16gb 
heap on a machine with 32gb of memory, I wonder if they're benefitting much 
from the page cache? Is the heap squeezing out available memory for the page 
cache?


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene] rmuir commented on pull request #12311: Integrate the Incubating Panama Vector API

2023-05-21 Thread via GitHub


rmuir commented on PR #12311:
URL: https://github.com/apache/lucene/pull/12311#issuecomment-1556224675

   i made the benchmarks easier to run with something like this:
   ```
   git clone https://github.com/rmuir/vectorbench
   cd vectorbench
   mvn verify
   java -jar target/vectorbench.jar
   ```
   
   I can confirm everything works on aarch64 and i am experimenting with the 
unrolling. definitely if we remove the unrolling it gets way slower.
   
   so i tried to unroll again (4x instead of 2x), it is only a slight 
improvement in performance on my skylake:
   ```
   Benchmark (size)   Mode  Cnt   Score   Error   
Units
   DotProductBenchmark.dotProductNew   1024  thrpt5   9.997 ± 0.999  
ops/us
   DotProductBenchmark.dotProductNewNew1024  thrpt5  11.285 ± 0.161  
ops/us
   DotProductBenchmark.dotProductOld   1024  thrpt5   2.024 ± 0.028  
ops/us
   ```
   
   But on the aarch64 mac additionall unrolling basically doubles throughput 
again (7.785 -> 14.912). I am experimenting more with this.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene] ChrisHegarty commented on pull request #12311: Integrate the Incubating Panama Vector API

2023-05-21 Thread via GitHub


ChrisHegarty commented on PR #12311:
URL: https://github.com/apache/lucene/pull/12311#issuecomment-1556278132

   I didn't get an anywhere with Luceneutil yet! :-(   (I haven't been able to 
run it successfully, getting OOM errors )


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene] rmuir commented on pull request #12311: Integrate the Incubating Panama Vector API

2023-05-21 Thread via GitHub


rmuir commented on PR #12311:
URL: https://github.com/apache/lucene/pull/12311#issuecomment-1556284488

   thanks for sanity checking! i'm still working on the repo and making 
improvements. would be super-curious if you could 'git pull' and try 
-psize=1024 on your avx512 machine. hopefully it looks better now there, we 
were being inefficient.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene] rmuir commented on pull request #12311: Integrate the Incubating Panama Vector API

2023-05-21 Thread via GitHub


rmuir commented on PR #12311:
URL: https://github.com/apache/lucene/pull/12311#issuecomment-1556285923

   With latest commits to that vectorbench I see this on my m1:
   ```
   Benchmark (size)   Mode  Cnt   Score   Error   
Units
   DotProductBenchmark.dotProductNew   1024  thrpt5   9.463 ± 0.004  
ops/us
   DotProductBenchmark.dotProductNewNew1024  thrpt5  16.106 ± 0.048  
ops/us
   DotProductBenchmark.dotProductOld   1024  thrpt5   3.828 ± 0.003  
ops/us
   ```


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene] rmuir commented on pull request #12311: Integrate the Incubating Panama Vector API

2023-05-21 Thread via GitHub


rmuir commented on PR #12311:
URL: https://github.com/apache/lucene/pull/12311#issuecomment-1556287294

   thanks, glad it fixes the problem. i am running it across all the sizes we 
test and seeing how it looks on both my machines.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene] ChrisHegarty commented on pull request #12311: Integrate the Incubating Panama Vector API

2023-05-21 Thread via GitHub


ChrisHegarty commented on PR #12311:
URL: https://github.com/apache/lucene/pull/12311#issuecomment-1556288076

   >  we were being inefficient.
   
   If I understand this correctly, the inefficiency was too many reduceLances, 
right? You replaced it with addition of the accumulators before reducing. 
Sounds good.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene] rmuir commented on pull request #12311: Integrate the Incubating Panama Vector API

2023-05-21 Thread via GitHub


rmuir commented on PR #12311:
URL: https://github.com/apache/lucene/pull/12311#issuecomment-1556288827

   yes, i think we have to imagine it as a scalar operation that gets slower as 
vector size increases. i looked into it and read this answer and changed the 
code: 
https://stackoverflow.com/questions/6996764/fastest-way-to-do-horizontal-sse-vector-sum-or-other-reduction/35270026#35270026


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene] rmuir commented on pull request #12311: Integrate the Incubating Panama Vector API

2023-05-21 Thread via GitHub


rmuir commented on PR #12311:
URL: https://github.com/apache/lucene/pull/12311#issuecomment-1556297297

   i pushed one more commit to improve for "unaligned" vectors. the way to 
think about it, with unrolling, we do 64-at-a-time on avx512. 
   
   So it isn't good to do worst-case 63 scalar computations just because user 
had 1023 dimensions or something like that. better to be bounded to 15. It 
makes things more well-rounded and prevents seeing slowdowns for sizes such as 
702 in the test.


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene] rmuir commented on pull request #12311: Integrate the Incubating Panama Vector API

2023-05-21 Thread via GitHub


rmuir commented on PR #12311:
URL: https://github.com/apache/lucene/pull/12311#issuecomment-1556297355

   aarch64: 
   ```
   Benchmark (size)   Mode  CntScore   Error   
Units
   DotProductBenchmark.dotProductNew  1  thrpt5  322.255 ± 0.496  
ops/us
   DotProductBenchmark.dotProductNew  4  thrpt5  247.637 ± 1.027  
ops/us
   DotProductBenchmark.dotProductNew  6  thrpt5  222.331 ± 0.156  
ops/us
   DotProductBenchmark.dotProductNew  8  thrpt5  184.218 ± 0.074  
ops/us
   DotProductBenchmark.dotProductNew 13  thrpt5  128.962 ± 0.017  
ops/us
   DotProductBenchmark.dotProductNew 16  thrpt5  165.321 ± 0.057  
ops/us
   DotProductBenchmark.dotProductNew 25  thrpt5  132.347 ± 0.435  
ops/us
   DotProductBenchmark.dotProductNew 32  thrpt5  145.280 ± 5.810  
ops/us
   DotProductBenchmark.dotProductNew 64  thrpt5  111.581 ± 9.717  
ops/us
   DotProductBenchmark.dotProductNew100  thrpt5   81.562 ± 0.310  
ops/us
   DotProductBenchmark.dotProductNew128  thrpt5   80.488 ± 0.255  
ops/us
   DotProductBenchmark.dotProductNew207  thrpt5   41.804 ± 0.360  
ops/us
   DotProductBenchmark.dotProductNew256  thrpt5   40.422 ± 0.090  
ops/us
   DotProductBenchmark.dotProductNew300  thrpt5   33.164 ± 0.131  
ops/us
   DotProductBenchmark.dotProductNew512  thrpt5   21.133 ± 0.020  
ops/us
   DotProductBenchmark.dotProductNew702  thrpt5   13.383 ± 0.010  
ops/us
   DotProductBenchmark.dotProductNew   1024  thrpt59.402 ± 0.198  
ops/us
   DotProductBenchmark.dotProductNewNew   1  thrpt5  322.162 ± 1.042  
ops/us
   DotProductBenchmark.dotProductNewNew   4  thrpt5  247.393 ± 3.673  
ops/us
   DotProductBenchmark.dotProductNewNew   6  thrpt5  219.753 ± 2.000  
ops/us
   DotProductBenchmark.dotProductNewNew   8  thrpt5  189.224 ± 3.233  
ops/us
   DotProductBenchmark.dotProductNewNew  13  thrpt5  148.712 ± 6.845  
ops/us
   DotProductBenchmark.dotProductNewNew  16  thrpt5  169.608 ± 0.200  
ops/us
   DotProductBenchmark.dotProductNewNew  25  thrpt5  105.866 ± 0.506  
ops/us
   DotProductBenchmark.dotProductNewNew  32  thrpt5  146.394 ± 0.802  
ops/us
   DotProductBenchmark.dotProductNewNew  64  thrpt5  119.317 ± 0.385  
ops/us
   DotProductBenchmark.dotProductNewNew 100  thrpt5   84.921 ± 2.819  
ops/us
   DotProductBenchmark.dotProductNewNew 128  thrpt5   87.055 ± 0.473  
ops/us
   DotProductBenchmark.dotProductNewNew 207  thrpt5   51.933 ± 0.270  
ops/us
   DotProductBenchmark.dotProductNewNew 256  thrpt5   55.509 ± 3.926  
ops/us
   DotProductBenchmark.dotProductNewNew 300  thrpt5   26.955 ± 0.016  
ops/us
   DotProductBenchmark.dotProductNewNew 512  thrpt5   20.205 ± 0.076  
ops/us
   DotProductBenchmark.dotProductNewNew 702  thrpt5   20.980 ± 0.029  
ops/us
   DotProductBenchmark.dotProductNewNew1024  thrpt5   16.244 ± 0.172  
ops/us
   DotProductBenchmark.dotProductOld  1  thrpt5  347.324 ± 5.857  
ops/us
   DotProductBenchmark.dotProductOld  4  thrpt5  247.653 ± 0.285  
ops/us
   DotProductBenchmark.dotProductOld  6  thrpt5  262.647 ± 1.425  
ops/us
   DotProductBenchmark.dotProductOld  8  thrpt5  227.780 ± 0.477  
ops/us
   DotProductBenchmark.dotProductOld 13  thrpt5  154.562 ± 2.007  
ops/us
   DotProductBenchmark.dotProductOld 16  thrpt5  154.300 ± 0.460  
ops/us
   DotProductBenchmark.dotProductOld 25  thrpt5   92.981 ± 0.083  
ops/us
   DotProductBenchmark.dotProductOld 32  thrpt5   89.480 ± 0.211  
ops/us
   DotProductBenchmark.dotProductOld 64  thrpt5   50.378 ± 0.064  
ops/us
   DotProductBenchmark.dotProductOld100  thrpt5   38.107 ± 0.115  
ops/us
   DotProductBenchmark.dotProductOld128  thrpt5   27.906 ± 0.030  
ops/us
   DotProductBenchmark.dotProductOld207  thrpt5   21.094 ± 0.016  
ops/us
   DotProductBenchmark.dotProductOld256  thrpt5   14.728 ± 0.004  
ops/us
   DotProductBenchmark.dotProductOld300  thrpt5   14.860 ± 0.025  
ops/us
   DotProductBenchmark.dotProductOld512  thrpt57.496 ± 0.041  
ops/us
   DotProductBenchmark.dotProductOld702  thrpt56.369 ± 0.015  
ops/us
   DotProductBenchmark.dotProductOld   1024  thrpt53.826 ± 0.002  
ops/us
   ```


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To

[GitHub] [lucene] rmuir commented on pull request #12311: Integrate the Incubating Panama Vector API

2023-05-21 Thread via GitHub


rmuir commented on PR #12311:
URL: https://github.com/apache/lucene/pull/12311#issuecomment-1556297561

   skylake:
   ```
   Benchmark (size)   Mode  CntScore   Error   
Units
   DotProductBenchmark.dotProductNew  1  thrpt5  153.702 ± 2.576  
ops/us
   DotProductBenchmark.dotProductNew  4  thrpt5   95.861 ± 1.403  
ops/us
   DotProductBenchmark.dotProductNew  6  thrpt5   93.582 ± 1.640  
ops/us
   DotProductBenchmark.dotProductNew  8  thrpt5   81.923 ± 1.045  
ops/us
   DotProductBenchmark.dotProductNew 13  thrpt5   66.178 ± 0.789  
ops/us
   DotProductBenchmark.dotProductNew 16  thrpt5   62.173 ± 1.191  
ops/us
   DotProductBenchmark.dotProductNew 25  thrpt5   40.726 ± 0.455  
ops/us
   DotProductBenchmark.dotProductNew 32  thrpt5   59.063 ± 6.797  
ops/us
   DotProductBenchmark.dotProductNew 64  thrpt5   51.108 ± 1.368  
ops/us
   DotProductBenchmark.dotProductNew100  thrpt5   35.460 ± 0.310  
ops/us
   DotProductBenchmark.dotProductNew128  thrpt5   39.522 ± 0.356  
ops/us
   DotProductBenchmark.dotProductNew207  thrpt5   21.369 ± 0.182  
ops/us
   DotProductBenchmark.dotProductNew256  thrpt5   26.010 ± 0.112  
ops/us
   DotProductBenchmark.dotProductNew300  thrpt5   19.118 ± 0.389  
ops/us
   DotProductBenchmark.dotProductNew512  thrpt5   17.368 ± 0.755  
ops/us
   DotProductBenchmark.dotProductNew702  thrpt5   11.338 ± 0.143  
ops/us
   DotProductBenchmark.dotProductNew   1024  thrpt5   10.073 ± 0.113  
ops/us
   DotProductBenchmark.dotProductNewNew   1  thrpt5  152.223 ± 0.854  
ops/us
   DotProductBenchmark.dotProductNewNew   4  thrpt5  114.786 ± 1.555  
ops/us
   DotProductBenchmark.dotProductNewNew   6  thrpt5   91.451 ± 0.874  
ops/us
   DotProductBenchmark.dotProductNewNew   8  thrpt5   81.767 ± 0.345  
ops/us
   DotProductBenchmark.dotProductNewNew  13  thrpt5   67.915 ± 0.889  
ops/us
   DotProductBenchmark.dotProductNewNew  16  thrpt5   64.509 ± 1.064  
ops/us
   DotProductBenchmark.dotProductNewNew  25  thrpt5   53.764 ± 1.037  
ops/us
   DotProductBenchmark.dotProductNewNew  32  thrpt5   62.759 ± 0.942  
ops/us
   DotProductBenchmark.dotProductNewNew  64  thrpt5   55.151 ± 0.396  
ops/us
   DotProductBenchmark.dotProductNewNew 100  thrpt5   37.558 ± 0.996  
ops/us
   DotProductBenchmark.dotProductNewNew 128  thrpt5   46.005 ± 0.733  
ops/us
   DotProductBenchmark.dotProductNewNew 207  thrpt5   26.135 ± 0.780  
ops/us
   DotProductBenchmark.dotProductNewNew 256  thrpt5   30.208 ± 0.115  
ops/us
   DotProductBenchmark.dotProductNewNew 300  thrpt5   22.830 ± 1.903  
ops/us
   DotProductBenchmark.dotProductNewNew 512  thrpt5   17.916 ± 0.216  
ops/us
   DotProductBenchmark.dotProductNewNew 702  thrpt5   12.854 ± 1.727  
ops/us
   DotProductBenchmark.dotProductNewNew1024  thrpt5   11.620 ± 0.291  
ops/us
   DotProductBenchmark.dotProductOld  1  thrpt5  162.477 ± 3.116  
ops/us
   DotProductBenchmark.dotProductOld  4  thrpt5  120.188 ± 2.748  
ops/us
   DotProductBenchmark.dotProductOld  6  thrpt5  120.427 ± 1.619  
ops/us
   DotProductBenchmark.dotProductOld  8  thrpt5   98.704 ± 2.279  
ops/us
   DotProductBenchmark.dotProductOld 13  thrpt5   76.331 ± 1.940  
ops/us
   DotProductBenchmark.dotProductOld 16  thrpt5   67.417 ± 1.456  
ops/us
   DotProductBenchmark.dotProductOld 25  thrpt5   47.443 ± 0.513  
ops/us
   DotProductBenchmark.dotProductOld 32  thrpt5   43.270 ± 4.112  
ops/us
   DotProductBenchmark.dotProductOld 64  thrpt5   26.506 ± 0.826  
ops/us
   DotProductBenchmark.dotProductOld100  thrpt5   16.793 ± 0.163  
ops/us
   DotProductBenchmark.dotProductOld128  thrpt5   14.332 ± 0.207  
ops/us
   DotProductBenchmark.dotProductOld207  thrpt59.032 ± 0.155  
ops/us
   DotProductBenchmark.dotProductOld256  thrpt57.853 ± 0.115  
ops/us
   DotProductBenchmark.dotProductOld300  thrpt56.331 ± 0.025  
ops/us
   DotProductBenchmark.dotProductOld512  thrpt54.027 ± 0.023  
ops/us
   DotProductBenchmark.dotProductOld702  thrpt52.762 ± 0.041  
ops/us
   DotProductBenchmark.dotProductOld   1024  thrpt52.003 ± 0.020  
ops/us
   ```


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To 

[GitHub] [lucene] rmuir commented on pull request #12311: Integrate the Incubating Panama Vector API

2023-05-21 Thread via GitHub


rmuir commented on PR #12311:
URL: https://github.com/apache/lucene/pull/12311#issuecomment-1556403183

   I pushed a new benchmark to https://github.com/rmuir/vectorbench for the 
binary dot product. 
   
   Basically this has to act like:
   ```
   int sum = 0;
   for (...) {
 short product = (short) (a[i] * b[i]);
 sum += (int) product;
   }
   ```
   
   So it is tricky to do with totally generic implementation (just using 
SPECIES_PREFERRED). for avx256, it means you read byte vector of length 32 and 
then work each half as short (2 short vectors of length 16), and then the same 
thing again for each half as int (4 int vectors of length 8). This generic 
approach only gives me a 2x speedup which is a little disappointing.
   
   but this is a stupid approach if you have 256-bit vectors. You can just use 
ByteVector.SPECIES_64, ShortVector.SPECIES_128, and IntVector.SPECIES_256 and 
the whole thing is much faster. 
   
   on my skylake (has avx 256 and gets the optimized 256-bit impl)
   ```
   Benchmark(size)   Mode  CntScore   Error 
  Units
   BinaryDotProductBenchmark.dotProductNew   1  thrpt5  159.476 ± 8.177 
 ops/us
   BinaryDotProductBenchmark.dotProductNew 128  thrpt5   41.759 ± 0.267 
 ops/us
   BinaryDotProductBenchmark.dotProductNew 207  thrpt5   25.094 ± 0.107 
 ops/us
   BinaryDotProductBenchmark.dotProductNew 256  thrpt5   24.841 ± 0.124 
 ops/us
   BinaryDotProductBenchmark.dotProductNew 300  thrpt5   19.624 ± 0.891 
 ops/us
   BinaryDotProductBenchmark.dotProductNew 512  thrpt5   13.763 ± 0.171 
 ops/us
   BinaryDotProductBenchmark.dotProductNew 702  thrpt59.792 ± 0.388 
 ops/us
   BinaryDotProductBenchmark.dotProductNew1024  thrpt56.878 ± 0.834 
 ops/us
   BinaryDotProductBenchmark.dotProductOld   1  thrpt5  160.423 ± 6.845 
 ops/us
   BinaryDotProductBenchmark.dotProductOld 128  thrpt5   13.300 ± 0.159 
 ops/us
   BinaryDotProductBenchmark.dotProductOld 207  thrpt58.678 ± 0.293 
 ops/us
   BinaryDotProductBenchmark.dotProductOld 256  thrpt56.892 ± 0.331 
 ops/us
   BinaryDotProductBenchmark.dotProductOld 300  thrpt56.008 ± 0.438 
 ops/us
   BinaryDotProductBenchmark.dotProductOld 512  thrpt53.613 ± 0.192 
 ops/us
   BinaryDotProductBenchmark.dotProductOld 702  thrpt52.710 ± 0.167 
 ops/us
   BinaryDotProductBenchmark.dotProductOld1024  thrpt51.825 ± 0.125 
 ops/us
   ```


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene] rmuir commented on pull request #12311: Integrate the Incubating Panama Vector API

2023-05-21 Thread via GitHub


rmuir commented on PR #12311:
URL: https://github.com/apache/lucene/pull/12311#issuecomment-1556405150

   and here's the results on my aarch64 mac, which has only 128-bit vectors and 
gets that disappointing generic impl:
   ```
   Benchmark(size)   Mode  CntScore   Error 
  Units
   BinaryDotProductBenchmark.dotProductNew   1  thrpt5  334.839 ± 0.368 
 ops/us
   BinaryDotProductBenchmark.dotProductNew 128  thrpt5   34.097 ± 0.026 
 ops/us
   BinaryDotProductBenchmark.dotProductNew 207  thrpt5   22.045 ± 0.063 
 ops/us
   BinaryDotProductBenchmark.dotProductNew 256  thrpt5   18.782 ± 0.603 
 ops/us
   BinaryDotProductBenchmark.dotProductNew 300  thrpt5   15.932 ± 0.049 
 ops/us
   BinaryDotProductBenchmark.dotProductNew 512  thrpt59.985 ± 0.010 
 ops/us
   BinaryDotProductBenchmark.dotProductNew 702  thrpt57.322 ± 0.001 
 ops/us
   BinaryDotProductBenchmark.dotProductNew1024  thrpt55.146 ± 0.002 
 ops/us
   BinaryDotProductBenchmark.dotProductOld   1  thrpt5  343.722 ± 0.640 
 ops/us
   BinaryDotProductBenchmark.dotProductOld 128  thrpt5   24.981 ± 0.102 
 ops/us
   BinaryDotProductBenchmark.dotProductOld 207  thrpt5   14.944 ± 0.128 
 ops/us
   BinaryDotProductBenchmark.dotProductOld 256  thrpt5   12.541 ± 0.006 
 ops/us
   BinaryDotProductBenchmark.dotProductOld 300  thrpt5   10.663 ± 0.005 
 ops/us
   BinaryDotProductBenchmark.dotProductOld 512  thrpt56.198 ± 0.019 
 ops/us
   BinaryDotProductBenchmark.dotProductOld 702  thrpt54.531 ± 0.062 
 ops/us
   BinaryDotProductBenchmark.dotProductOld1024  thrpt53.108 ± 0.003 
 ops/us
   ```
   
   I'd be curious how we could implement this with better performance, 
especially if we can just have a single generic impl like the float one. I feel 
like I must be doing it wrong :)


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene] tang-hi opened a new pull request, #12322: NeighborQueue set incomplemete false when call clear

2023-05-21 Thread via GitHub


tang-hi opened a new pull request, #12322:
URL: https://github.com/apache/lucene/pull/12322

   ### Description
   solve the bug that @msokolov mentioned in 
[PR](https://github.com/apache/lucene/pull/12255#issuecomment-1553088549)
   


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene] zhaih merged pull request #12257: Add multi-thread searchability to OnHeapHnswGraph

2023-05-21 Thread via GitHub


zhaih merged PR #12257:
URL: https://github.com/apache/lucene/pull/12257


-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org



[GitHub] [lucene] zhaih commented on a diff in pull request #12246: Set word2vec getSynonyms method synchronized

2023-05-21 Thread via GitHub


zhaih commented on code in PR #12246:
URL: https://github.com/apache/lucene/pull/12246#discussion_r1199956128


##
lucene/analysis/common/src/java/org/apache/lucene/analysis/synonym/word2vec/Word2VecSynonymProvider.java:
##
@@ -42,6 +42,7 @@ public class Word2VecSynonymProvider {
   private static final VectorSimilarityFunction SIMILARITY_FUNCTION =
   VectorSimilarityFunction.DOT_PRODUCT;
   private static final VectorEncoding VECTOR_ENCODING = VectorEncoding.FLOAT32;
+  private static final int NO_LIMIT_ON_VISITED_NODES = Integer.MAX_VALUE;
   private final Word2VecModel word2VecModel;
   private final HnswGraph hnswGraph;

Review Comment:
   I merged https://github.com/apache/lucene/pull/12257, but I think we still 
need to make a change here to declare it as `OnHeapHnswGraph` to make it work.



-- 
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.

To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org

For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


-
To unsubscribe, e-mail: issues-unsubscr...@lucene.apache.org
For additional commands, e-mail: issues-h...@lucene.apache.org