You have a 12G heap for a 200MB index? Can you just try changing Xmx to, like, 1g ?
On Tue, Oct 6, 2020 at 7:43 AM Karol Grzyb <grz...@gmail.com> wrote: > > Hi, > > I'm involved in investigation of issue that involves huge GC overhead > that happens during performance tests on Solr Nodes. Solr version is > 6.1. Last test were done on staging env, and we run into problems for > <100 requests/second. > > The size of the index itself is ~200MB ~ 50K docs > Index has small updates every 15min. > > > > Queries involve sorting and faceting. > > I've gathered some heap dumps, I can see from them that most of heap > memory is retained because of object of following classes: > > -org.apache.lucene.search.grouping.term.TermSecondPassGroupingCollector > (>4G, 91% of heap) > -org.apache.lucene.search.grouping.AbstractSecondPassGroupingCollector$SearchGroupDocs > -org.apache.lucene.search.FieldValueHitQueue$MultiComparatorsFieldValueHitQueue > -org.apache.lucene.search.TopFieldCollector$SimpleFieldCollector > (>3.7G 76% of heap) > > > > Based on information above is there anything generic that can been > looked at as source of potential improvement without diving deeply > into schema and queries (which may be very difficlut to change at this > moment)? I don't see docvalues being enabled - could this help, as if > I get the docs correctly, it's specifically helpful when there are > many sorts/grouping/facets? Or I > > Additionaly I see, that many threads are blocked on LRUCache.get, > should I recomend switching to FastLRUCache? > > Also, I wonder if -Xmx12288m for java heap is not too much for 16G > memory? I see some (~5/s) page faults in Dynatrace during the biggest > traffic. > > Thank you very much for any help, > Kind regards, > Karol