We have allocated 16 gb of heap space out of 24 g. There are 3 solr cores here, for one core when the no of documents are getting increased i.e. around 4.5 lakhs,then this scenario is happening.
On Sun, 2 Feb, 2020, 9:02 PM Erick Erickson, <erickerick...@gmail.com> wrote: > Allocate more heap and possibly add more RAM. > > What are you expectations? You can't continue to > add documents to your Solr instance without regard to > how much heap you’ve allocated. You’ve put over 4x > the number of docs on the node. There’s no magic here. > You can’t continue to add docs to a Solr instance without > increasing the heap at some point. > > And as far as I know, you’ve never told us how much heap yo > _are_ allocating. The default for Java processes is 512M, which > is quite small. so perhaps it’s a simple matter of starting Solr > with the -XmX parameter set to something larger. > > Best, > Erick > > > On Feb 2, 2020, at 10:19 AM, Rajdeep Sahoo <rajdeepsahoo2...@gmail.com> > wrote: > > > > What can we do in this scenario as the solr master node is going down and > > the indexing is failing. > > Please provide some workaround for this issue. > > > > On Sat, 1 Feb, 2020, 11:51 PM Walter Underwood, <wun...@wunderwood.org> > > wrote: > > > >> What message do you get about the heap space. > >> > >> It is completely normal for Java to use all of heap before running a > major > >> GC. That > >> is how the JVM works. > >> > >> wunder > >> Walter Underwood > >> wun...@wunderwood.org > >> http://observer.wunderwood.org/ (my blog) > >> > >>> On Feb 1, 2020, at 6:35 AM, Rajdeep Sahoo <rajdeepsahoo2...@gmail.com> > >> wrote: > >>> > >>> Please reply anyone > >>> > >>> On Fri, 31 Jan, 2020, 11:37 PM Rajdeep Sahoo, < > >> rajdeepsahoo2...@gmail.com> > >>> wrote: > >>> > >>>> This is happening when the no of indexed document count is increasing. > >>>> With 1 million docs it's working fine but when it's crossing 4.5 > >>>> million it's heap space is getting full. > >>>> > >>>> > >>>> On Wed, 22 Jan, 2020, 7:05 PM Michael Gibney, < > >> mich...@michaelgibney.net> > >>>> wrote: > >>>> > >>>>> Rajdeep, you say that "suddenly" heap space is getting full ... does > >>>>> this mean that some variant of this configuration was working for you > >>>>> at some point, or just that the failure happens quickly? > >>>>> > >>>>> If heap space and faceting are indeed the bottleneck, you might make > >>>>> sure that you have docValues enabled for your facet field fieldTypes, > >>>>> and perhaps set uninvertible=false. > >>>>> > >>>>> I'm not seeing where large numbers of facets initially came from in > >>>>> this thread? But on that topic this is perhaps relevant, regarding > the > >>>>> potential utility of a facet cache: > >>>>> https://issues.apache.org/jira/browse/SOLR-13807 > >>>>> > >>>>> Michael > >>>>> > >>>>> On Wed, Jan 22, 2020 at 7:16 AM Toke Eskildsen <t...@kb.dk> wrote: > >>>>>> > >>>>>> On Sun, 2020-01-19 at 21:19 -0500, Mehai, Lotfi wrote: > >>>>>>> I had a similar issue with a large number of facets. There is no > way > >>>>>>> (At least I know) your can get an acceptable response time from > >>>>>>> search engine with high number of facets. > >>>>>> > >>>>>> Just for the record then it is doable under specific circumstances > >>>>>> (static single-shard index, only String fields, Solr 4 with patch, > >>>>>> fixed list of facet fields): > >>>>>> https://sbdevel.wordpress.com/2013/03/20/over-9000-facet-fields/ > >>>>>> > >>>>>> More usable for the current case would be to play with facet.threads > >>>>>> and throw hardware with many CPU-cores after the problem. > >>>>>> > >>>>>> - Toke Eskildsen, Royal Danish Library > >>>>>> > >>>>>> > >>>>> > >>>> > >> > >> > >