For readability, I’d use -Xmx12G instead of -XX:MaxHeapSize=12884901888. Also, I always use a start size the same as the max size, since servers will eventually grow to the max size. So:
-Xmx12G -Xms12G wunder Walter Underwood wun...@wunderwood.org http://observer.wunderwood.org/ (my blog) > On Apr 11, 2018, at 6:29 AM, Sujay Bawaskar <sujaybawas...@gmail.com> wrote: > > What is directory factory defined in solrconfig.xml? Your JVM heap should > be tuned up with respect to that. > How solr is being use, is it more updates and less query or less updates > more queries? > What is OOM error? Is it frequent GC or Error 12? > > On Wed, Apr 11, 2018 at 6:05 PM, Adam Harrison-Fuller < > aharrison-ful...@mintel.com> wrote: > >> Hey Jesus, >> >> Thanks for the suggestions. The Solr nodes have 4 CPUs assigned to them. >> >> Cheers! >> Adam >> >> On 11 April 2018 at 11:22, Jesus Olivan <jesus.oli...@letgo.com> wrote: >> >>> Hi Adam, >>> >>> IMHO you could try increasing heap to 20 Gb (with 46 Gb of physical RAM, >>> your JVM can afford more RAM without threading penalties due to outside >>> heap RAM lacks. >>> >>> Another good one would be to increase -XX:CMSInitiatingOccupancyFraction >>> =50 >>> to 75. I think that CMS collector works better when Old generation space >> is >>> more populated. >>> >>> I usually use to set Survivor spaces to lesser size. If you want to try >>> SurvivorRatio to 6, i think performance would be improved. >>> >>> Another good practice for me would be to set an static NewSize instead >>> of -XX:NewRatio=3. >>> You could try to set -XX:NewSize=7000m and -XX:MaxNewSize=7000Mb (one >> third >>> of total heap space is recommended). >>> >>> Finally, my best results after a deep JVM I+D related to Solr, came >>> removing ScavengeBeforeRemark flag and applying this new one: + >>> ParGCCardsPerStrideChunk. >>> >>> However, It would be a good one to set ParallelGCThreads and >>> *ConcGCThreads *to their optimal value, and we need you system CPU number >>> to know it. Can you provide this data, please? >>> >>> Regards >>> >>> >>> 2018-04-11 12:01 GMT+02:00 Adam Harrison-Fuller < >>> aharrison-ful...@mintel.com >>>> : >>> >>>> Hey all, >>>> >>>> I was wondering if I could get some JVM/GC tuning advice to resolve an >>>> issue that we are experiencing. >>>> >>>> Full disclaimer, I am in no way a JVM/Solr expert so any advice you can >>>> render would be greatly appreciated. >>>> >>>> Our Solr cloud nodes are having issues throwing OOM exceptions under >>> load. >>>> This issue has only started manifesting itself over the last few months >>>> during which time the only change I can discern is an increase in index >>>> size. They are running Solr 5.5.2 on OpenJDK version "1.8.0_101". The >>>> index is currently 58G and the server has 46G of physical RAM and runs >>>> nothing other than the Solr node. >>>> >>>> The JVM is invoked with the following JVM options: >>>> -XX:CMSInitiatingOccupancyFraction=50 -XX:CMSMaxAbortablePrecleanTime= >>> 6000 >>>> -XX:+CMSParallelRemarkEnabled -XX:+CMSScavengeBeforeRemark >>>> -XX:ConcGCThreads=4 -XX:InitialHeapSize=12884901888 >>> -XX:+ManagementServer >>>> -XX:MaxHeapSize=12884901888 -XX:MaxTenuringThreshold=8 >>>> -XX:NewRatio=3 -XX:OldPLABSize=16 >>>> -XX:OnOutOfMemoryError=/opt/solr/bin/oom_solr.sh 30000 >>>> /data/gnpd/solr/logs >>>> -XX:ParallelGCThreads=4 >>>> -XX:+ParallelRefProcEnabled -XX:PretenureSizeThreshold=67108864 >>>> -XX:+PrintGC -XX:+PrintGCApplicationStoppedTime -XX:+PrintGCDateStamps >>>> -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintHeapAtGC >>>> -XX:+PrintTenuringDistribution -XX:SurvivorRatio=4 >>>> -XX:TargetSurvivorRatio=90 >>>> -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseCompressedClassPointers >>>> -XX:+UseCompressedOops -XX:+UseConcMarkSweepGC -XX:+UseParNewGC >>>> >>>> These values were decided upon serveral years by a colleague based upon >>>> some suggestions from this mailing group with an index size ~25G. >>>> >>>> I have imported the GC logs into GCViewer and attached a link to a >>>> screenshot showing the lead up to a OOM crash. Interestingly the young >>>> generation space is almost empty before the repeated GC's and >> subsequent >>>> crash. >>>> https://imgur.com/a/Wtlez >>>> >>>> I was considering slowly increasing the amount of heap available to the >>> JVM >>>> slowly until the crashes, any other suggestions? I'm looking at trying >>> to >>>> get the nodes stable without having issues with the GC taking forever >> to >>>> run. >>>> >>>> Additional information can be provided on request. >>>> >>>> Cheers! >>>> Adam >>>> >>>> -- >>>> >>>> Mintel Group Ltd | 11 Pilgrim Street | London | EC4V 6RN >>>> Registered in >>>> England: Number 1475918. | VAT Number: GB 232 9342 72 >>>> >>>> Contact details for >>>> our other offices can be found at http://www.mintel.com/office- >> locations >>>> <http://www.mintel.com/office-locations>. >>>> >>>> This email and any attachments >>>> may include content that is confidential, privileged >>>> or otherwise >>>> protected under applicable law. Unauthorised disclosure, copying, >>>> distribution >>>> or use of the contents is prohibited and may be unlawful. If >>>> you have received this email in error, >>>> including without appropriate >>>> authorisation, then please reply to the sender about the error >>>> and delete >>>> this email and any attachments. >>>> >>>> >>> >> >> -- >> >> Mintel Group Ltd | 11 Pilgrim Street | London | EC4V 6RN >> Registered in >> England: Number 1475918. | VAT Number: GB 232 9342 72 >> >> Contact details for >> our other offices can be found at http://www.mintel.com/office-locations >> <http://www.mintel.com/office-locations>. >> >> This email and any attachments >> may include content that is confidential, privileged >> or otherwise >> protected under applicable law. Unauthorised disclosure, copying, >> distribution >> or use of the contents is prohibited and may be unlawful. If >> you have received this email in error, >> including without appropriate >> authorisation, then please reply to the sender about the error >> and delete >> this email and any attachments. >> >> > > > -- > Thanks, > Sujay P Bawaskar > M:+91-77091 53669