Yes, why are you doing this? A suggester is designed to have a smaller set of 
terms than the entire index.

I would never expect a 130 million term suggester to work. I’m astonished that 
it works with 50 million terms.

We typically have about 50 thousand terms in a suggester.

Also, you haven’t said which kind of suggester you have configured. Some of 
them are in memory.

wunder
Walter Underwood
wun...@wunderwood.org
http://observer.wunderwood.org/  (my blog)

> On Jun 4, 2018, at 9:09 AM, Erick Erickson <erickerick...@gmail.com> wrote:
> 
> bq. I have 130 million documents and each document has unique document id. I
> want to build suggester on document id.
> 
> Why do it this way? I'm supposing you want to have someone start
> typing in the doc ID
> then do autocomplete on it. For such a simple operation, it would be
> far easier and
> pretty certainly fast enough to just use the Terms component and specify
> terms.prefix. See:
> https://lucene.apache.org/solr/guide/6_6/the-terms-component.html
> 
> This would not require any build step, would be as up-to-date as your
> last commit,
> would not consume the additional resources a suggester would work if
> you shard.....
> 
> Best,
> Erick
> 
> On Mon, Jun 4, 2018 at 4:23 AM, Yogendra Kumar Soni
> <yogendra.ku...@dolcera.com> wrote:
>> I sent  log of node to which i  sent the request. need to check other nodes
>> log
>>>> In SolrCloud an investigation does not isolate to a single Solr log : you
>>>> see a timeout, i would recommend to check both the nodes involved.
>> 
>> 
>> monitored from admin UI, could not find any clue at the time of failure.
>> 
>>>> Are you monitoring the memory utilisation for your Solr nodes ?
>> 
>> 
>>>> When you say : " heap usage is around 10 GB - 12 GB per node.", do you
>> refer
>>    to the effective usage by the Solr JVM or the allocated heap ?
>> 
>> 
>> heap usage varies from 5 gb to 12 gb . Initially it was 5 gb then increased
>> to 12 gb gradually and decreasing to 5 gb again. (may be because of garbage
>> collection)
>> 10-12 GB maximum  heap uses, allocated is 50 GB.
>> 
>>>> Are Garbage Collection cycles behaving correctly ?
>>>> When a timeout occurs, something bad happened in the communication between
>>>> the Solr nodes.
>> 
>> Need to  analyze GC pause. Any suggestion how i can monitor resource usage
>> and GC pause effectively.
>>>> It could be network, but in your case it may be some Stop World situation
>>>> caused by GC.
>> 
>> On Mon, Jun 4, 2018 at 3:27 PM, Alessandro Benedetti <a.benede...@sease.io>
>> wrote:
>> 
>>> Hi Yogendra,
>>> you mentioned you are using SolrCloud.
>>> In SolrCloud an investigation does not isolate to a single Solr log : you
>>> see a timeout, i would recommend to check both the nodes involved.
>>> 
>>> When you say : " heap usage is around 10 GB - 12 GB per node.", do you
>>> refer
>>> to the effective usage by the Solr JVM or the allocated heap ?
>>> Are you monitoring the memory utilisation for your Solr nodes ?
>>> Are Garbage Collection cycles behaving correctly ?
>>> When a timeout occurs, something bad happened in the communication between
>>> the Solr nodes.
>>> It could be network, but in your case it may be some Stop World situation
>>> caused by GC.
>>> 
>>> 
>>> 
>>> 
>>> -----
>>> ---------------
>>> Alessandro Benedetti
>>> Search Consultant, R&D Software Engineer, Director
>>> Sease Ltd. - www.sease.io
>>> --
>>> Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html
>>> 
>> 
>> 
>> 
>> --
>> *Thanks and Regards,*
>> *Yogendra Kumar Soni*

Reply via email to