Hi Dimitry, I pointed to the wiki page to enable JMX, then you can use any tool that visualizes JMX stuff like Zabbix. See http://www.lucidimagination.com/blog/2011/10/02/monitoring-apache-solr-and-lucidworks-with-zabbix/
On Wed, Dec 7, 2011 at 11:49 AM, Dmitry Kan <dmitry....@gmail.com> wrote: > The culprit seems to be the merger (frontend) SOLR. Talking to one shard > directly takes substantially less time (1-2 sec). > > On Wed, Dec 7, 2011 at 4:10 PM, Dmitry Kan <dmitry....@gmail.com> wrote: > > > Tomás: thanks. The page you gave didn't mention cache specifically, is > > there more documentation on this specifically? I have used solrmeter > tool, > > it draws the cache diagrams, is there a similar tool, but which would use > > jmx directly and present the cache usage in runtime? > > > > pravesh: > > I have increased the size of filterCache, but the search hasn't become > any > > faster, taking almost 9 sec on avg :( > > > > name: search > > class: org.apache.solr.handler.component.SearchHandler > > version: $Revision: 1052938 $ > > description: Search using components: > > > org.apache.solr.handler.component.QueryComponent,org.apache.solr.handler.component.FacetComponent,org.apache.solr.handler.component.MoreLikeThisComponent,org.apache.solr.handler.component.HighlightComponent,org.apache.solr.handler.component.StatsComponent,org.apache.solr.handler.component.DebugComponent, > > > > stats: handlerStart : 1323255147351 > > requests : 100 > > errors : 3 > > timeouts : 0 > > totalTime : 885438 > > avgTimePerRequest : 8854.38 > > avgRequestsPerSecond : 0.008789442 > > > > the stats (copying fieldValueCache as well here, to show term > statistics): > > > > name: fieldValueCache > > class: org.apache.solr.search.FastLRUCache > > version: 1.0 > > description: Concurrent LRU Cache(maxSize=10000, initialSize=10, > > minSize=9000, acceptableSize=9500, cleanupThread=false) > > stats: lookups : 79 > > hits : 77 > > hitratio : 0.97 > > inserts : 1 > > evictions : 0 > > size : 1 > > warmupTime : 0 > > cumulative_lookups : 79 > > cumulative_hits : 77 > > cumulative_hitratio : 0.97 > > cumulative_inserts : 1 > > cumulative_evictions : 0 > > item_shingleContent_trigram : > > > {field=shingleContent_trigram,memSize=326924381,tindexSize=4765394,time=215426,phase1=213868,nTerms=14827061,bigTerms=35,termInstances=114359167,uses=78} > > name: filterCache > > class: org.apache.solr.search.FastLRUCache > > version: 1.0 > > description: Concurrent LRU Cache(maxSize=153600, initialSize=4096, > > minSize=138240, acceptableSize=145920, cleanupThread=false) > > stats: lookups : 1082854 > > hits : 940370 > > hitratio : 0.86 > > inserts : 142486 > > evictions : 0 > > size : 142486 > > warmupTime : 0 > > cumulative_lookups : 1082854 > > cumulative_hits : 940370 > > cumulative_hitratio : 0.86 > > cumulative_inserts : 142486 > > cumulative_evictions : 0 > > > > > > index size: 3,25 GB > > > > Does anyone have some pointers to where to look at and optimize for query > > time? > > > > > > 2011/12/7 Tomás Fernández Löbbe <tomasflo...@gmail.com> > > > >> Hi Dimitry, cache information is exposed via JMX, so you should be able > to > >> monitor that information with any JMX tool. See > >> http://wiki.apache.org/solr/SolrJmx > >> > >> On Wed, Dec 7, 2011 at 6:19 AM, Dmitry Kan <dmitry....@gmail.com> > wrote: > >> > >> > Yes, we do require that much. > >> > Ok, thanks, I will try increasing the maxsize. > >> > > >> > On Wed, Dec 7, 2011 at 10:56 AM, pravesh <suyalprav...@yahoo.com> > >> wrote: > >> > > >> > > >>facet.limit=500000 > >> > > your facet.limit seems too high. Do you actually require this much? > >> > > > >> > > Since there a lot of evictions from filtercache, so, increase the > >> maxsize > >> > > value to your acceptable limit. > >> > > > >> > > Regards > >> > > Pravesh > >> > > > >> > > -- > >> > > View this message in context: > >> > > > >> > > >> > http://lucene.472066.n3.nabble.com/cache-monitoring-tools-tp3566645p3566811.html > >> > > Sent from the Solr - User mailing list archive at Nabble.com. > >> > > > >> > > >> > > >> > > >> > -- > >> > Regards, > >> > > >> > Dmitry Kan > >> > > >> > > > > > > > > -- > > Regards, > > > > Dmitry Kan > > > > > > -- > Regards, > > Dmitry Kan >