This is misleading and not particularly good advice.

Solr 8 does NOT contain G1. G1GC is a feature of the JVM. We’ve been using
it with Java 8 and Solr 6.6.2 for a few years.

A test with eighty documents doesn’t test anything. Try a million documents to
get Solr memory usage warmed up.

GC_TUNE has been in the solr.in.sh file for a long time. Here are the settings
we use with Java 8. We have about 120 hosts running Solr in six prod clusters.

SOLR_HEAP=8g
# Use G1 GC  -- wunder 2017-01-23
# Settings from https://wiki.apache.org/solr/ShawnHeisey
GC_TUNE=" \
-XX:+UseG1GC \
-XX:+ParallelRefProcEnabled \
-XX:G1HeapRegionSize=8m \
-XX:MaxGCPauseMillis=200 \
-XX:+UseLargePages \
-XX:+AggressiveOpts \
"

wunder
Walter Underwood
wun...@wunderwood.org
http://observer.wunderwood.org/  (my blog)

> On Sep 1, 2020, at 8:39 AM, Joe Doupnik <j...@netlab1.net> wrote:
> 
>     Erick states this correctly. To give some numbers from my experiences, 
> here are two slides from my presentation about installing Solr 
> (https://netlab1.net/ <https://netlab1.net/>, locate item "Solr/Lucene Search 
> Service"):
>> <hbifonfjanlomngl.png>
> 
>> <phnahkoblmojphjo.png>
> 
>     Thus we see a) experiments are the key, just as Erick says, and b) the 
> choice of garbage collection algorithm plays a major role.
>     In my setup I assigned SOLR_HEAP to be 2048m, SOLR_OPTS has -Xss1024k, 
> plus stock GC_TUNE values. Your "memorage" may vary.
>     Thanks,
>     Joe D.
> 
> On 01/09/2020 15:33, Erick Erickson wrote:
>> You want to run with the smallest heap you can due to Lucene’s use of 
>> MMapDirectory, 
>> see the excellent:
>> 
>> https://blog.thetaphi.de/2012/07/use-lucenes-mmapdirectory-on-64bit.html 
>> <https://blog.thetaphi.de/2012/07/use-lucenes-mmapdirectory-on-64bit.html>
>> 
>> There’s also little reason to have different Xms and Xmx values, that just 
>> means you’ll
>> eventually move a bunch of memory around as the heap expands, I usually set 
>> them both
>> to the same value.
>> 
>> How to determine what “the smallest heap you can” is? Unfortunately there’s 
>> no good way
>> outside of stress-testing your application with less and less memory until 
>> you have problems,
>> then add some extra…
>> 
>> Best,
>> Erick
>> 
>>> On Sep 1, 2020, at 10:27 AM, Dominique Bejean <dominique.bej...@eolya.fr> 
>>> <mailto:dominique.bej...@eolya.fr> wrote:
>>> 
>>> Hi,
>>> 
>>> As all Java applications the Heap memory is regularly cleaned by the
>>> garbage collector (some young items moved to the old generation heap zone
>>> and unused old items removed from the old generation heap zone). This
>>> causes heap usage to continuously grow and reduce.
>>> 
>>> Regards
>>> 
>>> Dominique
>>> 
>>> 
>>> 
>>> 
>>> Le mar. 1 sept. 2020 à 13:50, yaswanth kumar <yaswanth...@gmail.com> 
>>> <mailto:yaswanth...@gmail.com> a
>>> écrit :
>>> 
>>>> Can someone make me understand on how the value % on the column Heap is
>>>> calculated.
>>>> 
>>>> I did created a new solr cloud with 3 solr nodes and one zookeeper, its
>>>> not yet live neither interms of indexing or searching, but I do see some
>>>> spikes in the HEAP column against nodes when I refresh the page multiple
>>>> times. Its like almost going to 95% (sometimes) and then coming down to 50%
>>>> 
>>>> Solr version: 8.2
>>>> Zookeeper: 3.4
>>>> 
>>>> JVM size configured in solr.in.sh is min of 1GB to max of 10GB (actually
>>>> RAM size on the node is 16GB)
>>>> 
>>>> Basically need to understand if I need to worry about this heap % which
>>>> was quite altering before making it live? or is that quite normal, because
>>>> this is new UI change on solr cloud is kind of new to us as we used to have
>>>> solr 5 version before and this UI component doesn't exists then.
>>>> 
>>>> --
>>>> Thanks & Regards,
>>>> Yaswanth Kumar Konathala.
>>>> yaswanth...@gmail.com <mailto:yaswanth...@gmail.com>
>>>> 
>>>> Sent from my iPhone
> 

Reply via email to