Hi Dmitry, I had everything on one Solr Instance before, but this got to heavy and I had the same issue here, that the 1st facet.query was really slow.
When querying the facet: - facet.limit = 100 Cache settings are like this: <filterCache class="solr.FastLRUCache" size="16384" initialSize="4096" autowarmCount="4096"/> <queryResultCache class="solr.LRUCache" size="512" initialSize="512" autowarmCount="0"/> <documentCache class="solr.LRUCache" size="512" initialSize="512" autowarmCount="0"/> How big was your index? Did it fit into the RAM which you gave the Solr instance? Thanks On Tue, Jan 17, 2012 at 1:56 PM, Dmitry Kan <dmitry....@gmail.com> wrote: > I had a similar problem for a similar task. And in my case merging the > results from two shards turned out to be a culprit. If you can logically > store your data just in one shard, your faceting should become faster. Size > wise it should not be a problem for SOLR. > > Also, you didn't say anything about the facet.limit value, cache > parameters, usage of filter queries. Some of these can be interconnected. > > Dmitry > > On Tue, Jan 17, 2012 at 2:49 PM, Daniel Bruegge < > daniel.brue...@googlemail.com> wrote: > > > Hi, > > > > I have 2 Solr-shards. One is filled with approx. 25mio documents (local > > index 6GB), the other with 10mio documents (2.7GB size). > > I am trying to create some kind of 'word cloud' to see the frequency of > > words for a *text_general *field. > > For this I am currently using a facet over this field and I am also > > restricting the documents by using some other filters in the query. > > > > The performance is really bad for the first call and then pretty fast for > > the following calls. > > > > The maximum Java heap size is 3G for each shard. Both shards are running > on > > the same physical server which has 12G RAM. > > > > Question: Should I reduce the documents in one shard, so that the index > is > > equal or less the Java Heap size for this shard? Or is > > there another method to avoid this slow calls? > > > > Thank you > > > > Daniel > > > > > > -- > Regards, > > Dmitry Kan >