Initially we were getting the warning message as ulimit is low i.e. 1024 so we changed it to 65000 Using ulimit -u 65000.
Then the error was failed to reserve shared memory error =1 Because of this we removed -xx : +uselargepages Now in console log it is showing Could not find or load main class \ And solr is not starting up On Mon, 20 Jan, 2020, 7:20 AM Walter Underwood, <wun...@wunderwood.org> wrote: > What message do you get that means the heap space is full? > > Java will always use all of the heap, either as live data or > not-yet-collected garbage. > > wunder > Walter Underwood > wun...@wunderwood.org > http://observer.wunderwood.org/ (my blog) > > > On Jan 19, 2020, at 5:47 PM, Rajdeep Sahoo <rajdeepsahoo2...@gmail.com> > wrote: > > > > Hi, > > Currently there is no request or indexing is happening. > > It's just start up > > And during that time heap is getting full. > > Index size is approx 1 g. > > > > > > On Mon, 20 Jan, 2020, 7:01 AM Walter Underwood, <wun...@wunderwood.org> > > wrote: > > > >> A new garbage collector won’t fix it, but it might help a bit. > >> > >> Requesting 200 facet fields and having 50-60 of them with results is a > >> huge amount of work for Solr. A typical faceting implementation might > have > >> three to five facets. Your requests will be at least 10X to 20X slower. > >> > >> Check the CPU during one request. It should use nearly 100% of a single > >> CPU. If it a lot lower than 100%, you have another bottleneck. That > might > >> be insufficient heap or accessing disk during query requests (not enough > >> RAM). If it is near 100%, the only thing you can do is get a faster CPU. > >> > >> One other question, how frequently is the index updated? > >> > >> wunder > >> Walter Underwood > >> wun...@wunderwood.org > >> http://observer.wunderwood.org/ (my blog) > >> > >>> On Jan 19, 2020, at 4:49 PM, Rajdeep Sahoo <rajdeepsahoo2...@gmail.com > > > >> wrote: > >>> > >>> Hi, > >>> Still facing the same issue... > >>> Anything else that we need to check. > >>> > >>> > >>> On Mon, 20 Jan, 2020, 4:07 AM Walter Underwood, <wun...@wunderwood.org > > > >>> wrote: > >>> > >>>> With Java 1.8, I would use the G1 garbage collector. We’ve been > running > >>>> that combination in prod for three years with no problems. > >>>> > >>>> SOLR_HEAP=8g > >>>> # Use G1 GC -- wunder 2017-01-23 > >>>> # Settings from https://wiki.apache.org/solr/ShawnHeisey > >>>> GC_TUNE=" \ > >>>> -XX:+UseG1GC \ > >>>> -XX:+ParallelRefProcEnabled \ > >>>> -XX:G1HeapRegionSize=8m \ > >>>> -XX:MaxGCPauseMillis=200 \ > >>>> -XX:+UseLargePages \ > >>>> -XX:+AggressiveOpts \ > >>>> “ > >>>> > >>>> wunder > >>>> Walter Underwood > >>>> wun...@wunderwood.org > >>>> http://observer.wunderwood.org/ (my blog) > >>>> > >>>>> On Jan 19, 2020, at 9:25 AM, Rajdeep Sahoo < > rajdeepsahoo2...@gmail.com > >>> > >>>> wrote: > >>>>> > >>>>> We are using solr 7.7 . Ram size is 24 gb and allocated space is 12 > gb. > >>>> We > >>>>> have completed indexing after starting the server suddenly heap space > >> is > >>>>> getting full. > >>>>> Added gc params , still not working and jdk version is 1.8 . > >>>>> Please find the below gc params > >>>>> -XX:NewRatio=2 > >>>>> -XX:SurvivorRatio=3 > >>>>> -XX:TargetSurvivorRatio=90 \ > >>>>> -XX:MaxTenuringThreshold=8 \ > >>>>> -XX:+UseConcMarkSweepGC \ > >>>>> -XX:+CMSScavengeBeforeRemark \ > >>>>> -XX:ConcGCThreads=4 -XX:ParallelGCThreads=4 \ > >>>>> -XX:PretenureSizeThreshold=512m \ > >>>>> -XX:CMSFullGCsBeforeCompaction=1 \ > >>>>> -XX:+UseCMSInitiatingOccupancyOnly \ > >>>>> -XX:CMSInitiatingOccupancyFraction=70 \ > >>>>> -XX:CMSMaxAbortablePrecleanTime=6000 \ > >>>>> -XX:+CMSParallelRemarkEnabled > >>>>> -XX:+ParallelRefProcEnabled > >>>>> -XX:+UseLargePages \ > >>>>> -XX:+AggressiveOpts \ > >>>> > >>>> > >> > >> > >