Thank Markus,

So I will go ahead with 7.3.1 release.

On Tue, May 1, 2018 at 9:41 PM Markus Jelsma <markus.jel...@openindex.io>
wrote:

> Mạnh, Shalin,
>
> I tried to reproduce it locally but i failed, it is not just a stream of
> queries and frequent updates/commits. We will temporarily abuse a
> production machine to run 7.3 and a control  machine on 7.2 to rule some
> things out.
>
> We have plenty custom plugins, so when i can reproduce it again, we can
> rule stuff out and hopefully get back at you guys!
>
> Thanks,
> Markus
>
> -----Original message-----
> > From:Đạt Cao Mạnh <caomanhdat...@gmail.com>
> > Sent: Monday 30th April 2018 4:07
> > To: solr-user@lucene.apache.org
> > Subject: Re: 7.3 appears to leak
> >
> > Hi Markus,
> >
> > I tried indexing documents and query documents with queries and filter
> > question, but can't not find any leak problems. Can you give us more
> > information about the leak?
> >
> > Thanks!
> >
> > On Fri, Apr 27, 2018 at 5:11 PM Shalin Shekhar Mangar <
> > shalinman...@gmail.com> wrote:
> >
> > > Hi Markus,
> > >
> > > Can you give an idea of what your filter queries look like? Any custom
> > > plugins or things we should be aware of? Simple indexing artificial
> docs,
> > > querying and committing doesn't seem to reproduce the issue for me.
> > >
> > > On Thu, Apr 26, 2018 at 10:13 PM, Markus Jelsma <
> > > markus.jel...@openindex.io>
> > > wrote:
> > >
> > > > Hello,
> > > >
> > > > We just finished upgrading our three separate clusters from 7.2.1 to
> 7.3,
> > > > which went fine, except for our main text search collection, it
> appears
> > > to
> > > > leak memory on commit!
> > > >
> > > > After initial upgrade we saw the cluster slowly starting to run out
> of
> > > > memory within about an hour and a half. We increased heap in case 7.3
> > > just
> > > > requires more of it, but the heap consumption graph is still growing
> on
> > > > each commit. Heap space cannot be reclaimed by forcing the garbage
> > > > collector to run, everything just piles up in the OldGen. Running
> with
> > > this
> > > > slightly larger heap, the first nodes will run out of memory in
> about two
> > > > and a half hours after cluster restart.
> > > >
> > > > The heap eating cluster is a 2shard/3replica system on separate
> nodes.
> > > > Each replica is about 50 GB in size and about 8.5 million documents.
> On
> > > > 7.2.1 it ran fine with just a 2 GB heap. With 7.3 and 2.5 GB heap, it
> > > will
> > > > take just a little longer for it to run out of memory.
> > > >
> > > > I inspected reports shown by the sampler of VisualVM and spotted one
> > > > peculiarity, the number of instances of SortedIntDocSet kept growing
> on
> > > > each commit by about the same amount as the number of cached filter
> > > > queries. But this doesn't happen on the logs cluster, SortedIntDocSet
> > > > instances are neatly collected there. The number of instances also
> > > accounts
> > > > for the number of commits since start up times the cache sizes
> > > >
> > > > Our other two clusters don't have this problem, one of them receives
> very
> > > > few commits per day, but the other receives data all the time, it
> logs
> > > user
> > > > interactions so a large amount of data is coming in all the time. I
> > > cannot
> > > > reproduce it locally by indexing data and committing all the time,
> the
> > > peak
> > > > usage in OldGen stays about the same. But, i can reproduce it locally
> > > when
> > > > i introduce queries, and filter queries while indexing pieces of
> data and
> > > > committing it.
> > > >
> > > > So, what is the problem? I dug in the CHANGES.txt of both Lucene and
> > > Solr,
> > > > but nothing really caught my attention. Does anyone here have an idea
> > > where
> > > > to look?
> > > >
> > > > Many thanks,
> > > > Markus
> > > >
> > >
> > >
> > >
> > > --
> > > Regards,
> > > Shalin Shekhar Mangar.
> > >
> >
>

Reply via email to