Yes, why are you doing this? A suggester is designed to have a smaller set of terms than the entire index.
I would never expect a 130 million term suggester to work. I’m astonished that it works with 50 million terms. We typically have about 50 thousand terms in a suggester. Also, you haven’t said which kind of suggester you have configured. Some of them are in memory. wunder Walter Underwood wun...@wunderwood.org http://observer.wunderwood.org/ (my blog) > On Jun 4, 2018, at 9:09 AM, Erick Erickson <erickerick...@gmail.com> wrote: > > bq. I have 130 million documents and each document has unique document id. I > want to build suggester on document id. > > Why do it this way? I'm supposing you want to have someone start > typing in the doc ID > then do autocomplete on it. For such a simple operation, it would be > far easier and > pretty certainly fast enough to just use the Terms component and specify > terms.prefix. See: > https://lucene.apache.org/solr/guide/6_6/the-terms-component.html > > This would not require any build step, would be as up-to-date as your > last commit, > would not consume the additional resources a suggester would work if > you shard..... > > Best, > Erick > > On Mon, Jun 4, 2018 at 4:23 AM, Yogendra Kumar Soni > <yogendra.ku...@dolcera.com> wrote: >> I sent log of node to which i sent the request. need to check other nodes >> log >>>> In SolrCloud an investigation does not isolate to a single Solr log : you >>>> see a timeout, i would recommend to check both the nodes involved. >> >> >> monitored from admin UI, could not find any clue at the time of failure. >> >>>> Are you monitoring the memory utilisation for your Solr nodes ? >> >> >>>> When you say : " heap usage is around 10 GB - 12 GB per node.", do you >> refer >> to the effective usage by the Solr JVM or the allocated heap ? >> >> >> heap usage varies from 5 gb to 12 gb . Initially it was 5 gb then increased >> to 12 gb gradually and decreasing to 5 gb again. (may be because of garbage >> collection) >> 10-12 GB maximum heap uses, allocated is 50 GB. >> >>>> Are Garbage Collection cycles behaving correctly ? >>>> When a timeout occurs, something bad happened in the communication between >>>> the Solr nodes. >> >> Need to analyze GC pause. Any suggestion how i can monitor resource usage >> and GC pause effectively. >>>> It could be network, but in your case it may be some Stop World situation >>>> caused by GC. >> >> On Mon, Jun 4, 2018 at 3:27 PM, Alessandro Benedetti <a.benede...@sease.io> >> wrote: >> >>> Hi Yogendra, >>> you mentioned you are using SolrCloud. >>> In SolrCloud an investigation does not isolate to a single Solr log : you >>> see a timeout, i would recommend to check both the nodes involved. >>> >>> When you say : " heap usage is around 10 GB - 12 GB per node.", do you >>> refer >>> to the effective usage by the Solr JVM or the allocated heap ? >>> Are you monitoring the memory utilisation for your Solr nodes ? >>> Are Garbage Collection cycles behaving correctly ? >>> When a timeout occurs, something bad happened in the communication between >>> the Solr nodes. >>> It could be network, but in your case it may be some Stop World situation >>> caused by GC. >>> >>> >>> >>> >>> ----- >>> --------------- >>> Alessandro Benedetti >>> Search Consultant, R&D Software Engineer, Director >>> Sease Ltd. - www.sease.io >>> -- >>> Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html >>> >> >> >> >> -- >> *Thanks and Regards,* >> *Yogendra Kumar Soni*