Initially we were getting the warning message as  ulimit is low i.e. 1024
so we changed it to 65000
Using ulimit -u 65000.

Then the error was failed to reserve shared memory error =1
 Because of this we removed
   -xx : +uselargepages

Now in console log it is showing
Could not find or load main class \

And solr is not starting up



On Mon, 20 Jan, 2020, 7:20 AM Walter Underwood, <wun...@wunderwood.org>
wrote:

> What message do you get that means the heap space is full?
>
> Java will always use all of the heap, either as live data or
> not-yet-collected garbage.
>
> wunder
> Walter Underwood
> wun...@wunderwood.org
> http://observer.wunderwood.org/  (my blog)
>
> > On Jan 19, 2020, at 5:47 PM, Rajdeep Sahoo <rajdeepsahoo2...@gmail.com>
> wrote:
> >
> > Hi,
> > Currently there is no request or indexing is happening.
> >  It's just start up
> > And during that time heap is getting full.
> > Index size is approx 1 g.
> >
> >
> > On Mon, 20 Jan, 2020, 7:01 AM Walter Underwood, <wun...@wunderwood.org>
> > wrote:
> >
> >> A new garbage collector won’t fix it, but it might help a bit.
> >>
> >> Requesting 200 facet fields and having 50-60 of them with results is a
> >> huge amount of work for Solr. A typical faceting implementation might
> have
> >> three to five facets. Your requests will be at least 10X to 20X slower.
> >>
> >> Check the CPU during one request. It should use nearly 100% of a single
> >> CPU. If it a lot lower than 100%, you have another bottleneck. That
> might
> >> be insufficient heap or accessing disk during query requests (not enough
> >> RAM). If it is near 100%, the only thing you can do is get a faster CPU.
> >>
> >> One other question, how frequently is the index updated?
> >>
> >> wunder
> >> Walter Underwood
> >> wun...@wunderwood.org
> >> http://observer.wunderwood.org/  (my blog)
> >>
> >>> On Jan 19, 2020, at 4:49 PM, Rajdeep Sahoo <rajdeepsahoo2...@gmail.com
> >
> >> wrote:
> >>>
> >>> Hi,
> >>> Still facing the same issue...
> >>> Anything else that we need to check.
> >>>
> >>>
> >>> On Mon, 20 Jan, 2020, 4:07 AM Walter Underwood, <wun...@wunderwood.org
> >
> >>> wrote:
> >>>
> >>>> With Java 1.8, I would use the G1 garbage collector. We’ve been
> running
> >>>> that combination in prod for three years with no problems.
> >>>>
> >>>> SOLR_HEAP=8g
> >>>> # Use G1 GC  -- wunder 2017-01-23
> >>>> # Settings from https://wiki.apache.org/solr/ShawnHeisey
> >>>> GC_TUNE=" \
> >>>> -XX:+UseG1GC \
> >>>> -XX:+ParallelRefProcEnabled \
> >>>> -XX:G1HeapRegionSize=8m \
> >>>> -XX:MaxGCPauseMillis=200 \
> >>>> -XX:+UseLargePages \
> >>>> -XX:+AggressiveOpts \
> >>>> “
> >>>>
> >>>> wunder
> >>>> Walter Underwood
> >>>> wun...@wunderwood.org
> >>>> http://observer.wunderwood.org/  (my blog)
> >>>>
> >>>>> On Jan 19, 2020, at 9:25 AM, Rajdeep Sahoo <
> rajdeepsahoo2...@gmail.com
> >>>
> >>>> wrote:
> >>>>>
> >>>>> We are using solr 7.7 . Ram size is 24 gb and allocated space is 12
> gb.
> >>>> We
> >>>>> have completed indexing after starting the server suddenly heap space
> >> is
> >>>>> getting full.
> >>>>> Added gc params  , still not working and jdk version is 1.8 .
> >>>>> Please find the below gc  params
> >>>>> -XX:NewRatio=2
> >>>>> -XX:SurvivorRatio=3
> >>>>> -XX:TargetSurvivorRatio=90 \
> >>>>> -XX:MaxTenuringThreshold=8 \
> >>>>> -XX:+UseConcMarkSweepGC \
> >>>>> -XX:+CMSScavengeBeforeRemark \
> >>>>> -XX:ConcGCThreads=4 -XX:ParallelGCThreads=4 \
> >>>>> -XX:PretenureSizeThreshold=512m \
> >>>>> -XX:CMSFullGCsBeforeCompaction=1 \
> >>>>> -XX:+UseCMSInitiatingOccupancyOnly \
> >>>>> -XX:CMSInitiatingOccupancyFraction=70 \
> >>>>> -XX:CMSMaxAbortablePrecleanTime=6000 \
> >>>>> -XX:+CMSParallelRemarkEnabled
> >>>>> -XX:+ParallelRefProcEnabled
> >>>>> -XX:+UseLargePages \
> >>>>> -XX:+AggressiveOpts \
> >>>>
> >>>>
> >>
> >>
>
>

Reply via email to