This is misleading and not particularly good advice. Solr 8 does NOT contain G1. G1GC is a feature of the JVM. We’ve been using it with Java 8 and Solr 6.6.2 for a few years.
A test with eighty documents doesn’t test anything. Try a million documents to get Solr memory usage warmed up. GC_TUNE has been in the solr.in.sh file for a long time. Here are the settings we use with Java 8. We have about 120 hosts running Solr in six prod clusters. SOLR_HEAP=8g # Use G1 GC -- wunder 2017-01-23 # Settings from https://wiki.apache.org/solr/ShawnHeisey GC_TUNE=" \ -XX:+UseG1GC \ -XX:+ParallelRefProcEnabled \ -XX:G1HeapRegionSize=8m \ -XX:MaxGCPauseMillis=200 \ -XX:+UseLargePages \ -XX:+AggressiveOpts \ " wunder Walter Underwood wun...@wunderwood.org http://observer.wunderwood.org/ (my blog) > On Sep 1, 2020, at 8:39 AM, Joe Doupnik <j...@netlab1.net> wrote: > > Erick states this correctly. To give some numbers from my experiences, > here are two slides from my presentation about installing Solr > (https://netlab1.net/ <https://netlab1.net/>, locate item "Solr/Lucene Search > Service"): >> <hbifonfjanlomngl.png> > >> <phnahkoblmojphjo.png> > > Thus we see a) experiments are the key, just as Erick says, and b) the > choice of garbage collection algorithm plays a major role. > In my setup I assigned SOLR_HEAP to be 2048m, SOLR_OPTS has -Xss1024k, > plus stock GC_TUNE values. Your "memorage" may vary. > Thanks, > Joe D. > > On 01/09/2020 15:33, Erick Erickson wrote: >> You want to run with the smallest heap you can due to Lucene’s use of >> MMapDirectory, >> see the excellent: >> >> https://blog.thetaphi.de/2012/07/use-lucenes-mmapdirectory-on-64bit.html >> <https://blog.thetaphi.de/2012/07/use-lucenes-mmapdirectory-on-64bit.html> >> >> There’s also little reason to have different Xms and Xmx values, that just >> means you’ll >> eventually move a bunch of memory around as the heap expands, I usually set >> them both >> to the same value. >> >> How to determine what “the smallest heap you can” is? Unfortunately there’s >> no good way >> outside of stress-testing your application with less and less memory until >> you have problems, >> then add some extra… >> >> Best, >> Erick >> >>> On Sep 1, 2020, at 10:27 AM, Dominique Bejean <dominique.bej...@eolya.fr> >>> <mailto:dominique.bej...@eolya.fr> wrote: >>> >>> Hi, >>> >>> As all Java applications the Heap memory is regularly cleaned by the >>> garbage collector (some young items moved to the old generation heap zone >>> and unused old items removed from the old generation heap zone). This >>> causes heap usage to continuously grow and reduce. >>> >>> Regards >>> >>> Dominique >>> >>> >>> >>> >>> Le mar. 1 sept. 2020 à 13:50, yaswanth kumar <yaswanth...@gmail.com> >>> <mailto:yaswanth...@gmail.com> a >>> écrit : >>> >>>> Can someone make me understand on how the value % on the column Heap is >>>> calculated. >>>> >>>> I did created a new solr cloud with 3 solr nodes and one zookeeper, its >>>> not yet live neither interms of indexing or searching, but I do see some >>>> spikes in the HEAP column against nodes when I refresh the page multiple >>>> times. Its like almost going to 95% (sometimes) and then coming down to 50% >>>> >>>> Solr version: 8.2 >>>> Zookeeper: 3.4 >>>> >>>> JVM size configured in solr.in.sh is min of 1GB to max of 10GB (actually >>>> RAM size on the node is 16GB) >>>> >>>> Basically need to understand if I need to worry about this heap % which >>>> was quite altering before making it live? or is that quite normal, because >>>> this is new UI change on solr cloud is kind of new to us as we used to have >>>> solr 5 version before and this UI component doesn't exists then. >>>> >>>> -- >>>> Thanks & Regards, >>>> Yaswanth Kumar Konathala. >>>> yaswanth...@gmail.com <mailto:yaswanth...@gmail.com> >>>> >>>> Sent from my iPhone >