i wonder how do you manage 200 instances
On 20 Oct 2011, at 09:21, Sujatha Arun wrote:
Yes 200 Individual Solr Instances not solr cores.
We get an avg response time of below 1 sec.
The number of documents is not many most of the isntances ,some of
the
instnaces have about 5 lac documents on average.
Regards
Sujahta
On Thu, Oct 20, 2011 at 3:35 AM, Jaeger, Jay - DOT >wrote:
200 instances of what? The Solr application with lucene, etc. per
usual?
Solr cores? ???
Either way, 200 seems to be very very very many: unusually so. Why
so
many?
If you have 200 instances of Solr in a 20 GB JVM, that would only
be 100MB
per Solr instance.
If you have 200 instances of Solr all accessing the same physical
disk, the
results are not likely to be satisfactory - the disk head will go
nuts
trying to handle all of the requests.
JRJ
-Original Message-
From: Sujatha Arun [mailto:suja.a...@gmail.com]
Sent: Wednesday, October 19, 2011 12:25 AM
To: solr-user@lucene.apache.org; Otis Gospodnetic
Subject: Re: OS Cache - Solr
Thanks ,Otis,
This is our Solr Cache Allocation.We have the same Cache
allocation for
all
our *200+ instances* in the single Server.Is this too high?
*Query Result Cache*:LRU Cache(maxSize=16384, initialSize=4096,
autowarmCount=1024, )
*Document Cache *:LRU Cache(maxSize=16384, initialSize=16384)
*Filter Cache* LRU Cache(maxSize=16384, initialSize=4096,
autowarmCount=4096, )
Regards
Sujatha
On Wed, Oct 19, 2011 at 4:05 AM, Otis Gospodnetic <
otis_gospodne...@yahoo.com> wrote:
Maybe your Solr Document cache is big and that's consuming a big
part of
that JVM heap?
If you want to be able to run with a smaller heap, consider making
your
caches smaller.
Otis
Sematext :: http://sematext.com/ :: Solr - Lucene - Nutch
Lucene ecosystem search :: http://search-lucene.com/
From: Sujatha Arun
To: solr-user@lucene.apache.org
Sent: Tuesday, October 18, 2011 12:53 AM
Subject: Re: OS Cache - Solr
Hello Jan,
Thanks for your response and clarification.
We are monitoring the JVM cache utilization and we are currently
using
about
18 GB of the 20 GB assigned to JVM. Out total index size being
abt 14GB
Regards
Sujatha
On Tue, Oct 18, 2011 at 1:19 AM, Jan Høydahl t.com>
wrote:
Hi Sujatha,
Are you sure you need 20Gb for Tomcat? Have you profiled using
JConsole
or
similar? Try with 15Gb and see how it goes. The reason why this is
beneficial is that you WANT your OS to have available memory for
disk
caching. If you have 17Gb free after starting Solr, your OS will
be
able
to
cache all index files in memory and you get very high search
performance.
With your current settings, there is only 12Gb free for both
caching
the
index and for your MySql activities. Chances are that when you
backup
MySql, the cached part of your Solr index gets flushed from disk
caches
and
need to be re-cached later.
How to interpret memory stats vary between OSes, and seing 163Mb
free
may
simply mean that your OS has used most RAM for various caches and
paging,
but will flush it once an application asks for more memory. Have
you
seen
http://wiki.apache.org/solr/SolrPerformanceFactors ?
You should also slim down your index maximally by setting
stored=false
and
indexed=false wherever possible. I would also upgrade to a more
current
Solr
version.
--
Jan Høydahl, search solution architect
Cominvent AS - www.cominvent.com
Solr Training - www.solrtraining.com
On 17. okt. 2011, at 19:51, Sujatha Arun wrote:
Hello
I am trying to understand the OS cache utilization of Solr .Our
server
has
several solr instances on a server .The total combined Index
size of
all
instances is abt 14 Gb and the size of the maximum single Index
is
abt
2.5
GB .
Our Server has Quad processor with 32 GB RAM .Out of which 20
GB has
been
assigned to JVM. We are running solr1.3 on tomcat 5.5 and
Java 1.6
Our current Statistics indicate that solr uses 18-19 GB of 20 GB
RAM
assigned to JVM .However the Free physical seems to remain
constant
as
below.
Free physical memory = 163 Mb
Total physical memory = 32,232 Mb,
The server also serves as a backup server for Mysql where the
application
DB
is backed up and restored .During this activity we see that lot
of
queries
that nearly take even 10+ minutes to execute .But other wise
maximum query time is less than 1-2 secs
The physical memory that is free seems to be constant . Why is
this
constant
and how this will be used between the Mysql backup and solr
while
backup activity is happening How much free physical memory
should
be
available to OS given out stats.?
Any pointers would be helpful.
Regards
Sujatha