Hi Jean-Sebastien,

One thing you didn't mention is whether as you are increasing(I assume)
cache sizes you actually see performance improve?  If not, then maybe there
is no value increasing cache sizes.

I assume you changed only one cache at a time? Were you able to get any one
of them to the point where there were no evictions without things breaking?

What are your queries like, can you share a few examples?

Otis
--
Performance Monitoring * Log Analytics * Search Analytics
Solr & Elasticsearch Support * http://sematext.com/


On Mon, Jun 2, 2014 at 11:09 AM, Jean-Sebastien Vachon <
jean-sebastien.vac...@wantedanalytics.com> wrote:

> Thanks for your quick response.
>
> Our JVM is configured with a heap of 8GB. So we are pretty close of the
> "optimal" configuration you are mentioning. The only other programs running
> is Zookeeper (which has its own storage device) and a proprietary API (with
> a heap of 1GB) we have on top of Solr to server our customer`s requests.
>
> I will look into the filterCache to see if we can better use it.
>
> Thanks for your help
>
> > -----Original Message-----
> > From: Shawn Heisey [mailto:s...@elyograg.org]
> > Sent: June-02-14 10:48 AM
> > To: solr-user@lucene.apache.org
> > Subject: Re: Strange behaviour when tuning the caches
> >
> > On 6/2/2014 8:24 AM, Jean-Sebastien Vachon wrote:
> > > We have yet to determine where the exact breaking point is.
> > >
> > > The two patterns we are seeing are:
> > >
> > > -          less cache (around 20-30% hit/ratio), poor performance but
> > > overall good stability
> >
> > When caches are too small, a low hit ratio is expected.  Increasing them
> is a
> > good idea, but only increase them a little bit at a time.  The
> filterCache in
> > particular should not be increased dramatically, especially the
> > autowarmCount value.  Filters can take a very long time to execute, so a
> high
> > autowarmCount can result in commits taking forever.
> >
> > Each filter entry can take up a lot of heap memory -- in terms of bytes,
> it is
> > the number of documents in the core divided by 8.  This means that if the
> > core has 10 million documents, each filter entry (for JUST that
> > core) will take over a megabyte of RAM.
> >
> > > -          more cache (over 90% hit/ratio), improved performance but
> > > almost no stability. In that case, we start seeing messages such as
> > > "No shards hosting shard X" or "cancelElection did not find election
> > > node to remove"
> >
> > This would not be a direct result of increasing the cache size, unless
> perhaps
> > you've increased them so they are *REALLY* big and you're running out of
> > RAM for the heap or OS disk cache.
> >
> > > Anyone, has any advice on what could cause this? I am beginning to
> > > suspect the JVM version, is there any minimal requirements regarding
> > > the JVM?
> >
> > Oracle Java 7 is recommended for all releases, and required for Solr
> 4.8.  You
> > just need to stay away from 7u40, 7u45, and 7u51 because of bugs in Java
> > itself.  Right now, the latest release is recommended, which is 7u60.
>  The
> > 7u21 release that you are running should be perfectly fine.
> >
> > With six 9.4GB cores per node, you'll achieve the best performance if you
> > have about 60GB of RAM left over for the OS disk cache to use -- the
> size of
> > your index data on disk.  You did mention that you have 92GB of RAM per
> > node, but you have not said how big your Java heap is, or whether there
> is
> > other software on the machine that may be eating up RAM for its heap or
> > data.
> >
> > http://wiki.apache.org/solr/SolrPerformanceProblems
> >
> > Thanks,
> > Shawn
> >
> > -----
> > Aucun virus trouvé dans ce message.
> > Analyse effectuée par AVG - www.avg.fr
> > Version: 2014.0.4570 / Base de données virale: 3950/7571 - Date:
> > 27/05/2014
>

Reply via email to