bq. I have 130 million documents and each document has unique document id. I want to build suggester on document id.
Why do it this way? I'm supposing you want to have someone start typing in the doc ID then do autocomplete on it. For such a simple operation, it would be far easier and pretty certainly fast enough to just use the Terms component and specify terms.prefix. See: https://lucene.apache.org/solr/guide/6_6/the-terms-component.html This would not require any build step, would be as up-to-date as your last commit, would not consume the additional resources a suggester would work if you shard..... Best, Erick On Mon, Jun 4, 2018 at 4:23 AM, Yogendra Kumar Soni <yogendra.ku...@dolcera.com> wrote: > I sent log of node to which i sent the request. need to check other nodes > log >>>In SolrCloud an investigation does not isolate to a single Solr log : you >>>see a timeout, i would recommend to check both the nodes involved. > > > monitored from admin UI, could not find any clue at the time of failure. > >>>Are you monitoring the memory utilisation for your Solr nodes ? > > >>>When you say : " heap usage is around 10 GB - 12 GB per node.", do you > refer > to the effective usage by the Solr JVM or the allocated heap ? > > > heap usage varies from 5 gb to 12 gb . Initially it was 5 gb then increased > to 12 gb gradually and decreasing to 5 gb again. (may be because of garbage > collection) > 10-12 GB maximum heap uses, allocated is 50 GB. > >>>Are Garbage Collection cycles behaving correctly ? >>>When a timeout occurs, something bad happened in the communication between >>>the Solr nodes. > > Need to analyze GC pause. Any suggestion how i can monitor resource usage > and GC pause effectively. >>>It could be network, but in your case it may be some Stop World situation >>>caused by GC. > > On Mon, Jun 4, 2018 at 3:27 PM, Alessandro Benedetti <a.benede...@sease.io> > wrote: > >> Hi Yogendra, >> you mentioned you are using SolrCloud. >> In SolrCloud an investigation does not isolate to a single Solr log : you >> see a timeout, i would recommend to check both the nodes involved. >> >> When you say : " heap usage is around 10 GB - 12 GB per node.", do you >> refer >> to the effective usage by the Solr JVM or the allocated heap ? >> Are you monitoring the memory utilisation for your Solr nodes ? >> Are Garbage Collection cycles behaving correctly ? >> When a timeout occurs, something bad happened in the communication between >> the Solr nodes. >> It could be network, but in your case it may be some Stop World situation >> caused by GC. >> >> >> >> >> ----- >> --------------- >> Alessandro Benedetti >> Search Consultant, R&D Software Engineer, Director >> Sease Ltd. - www.sease.io >> -- >> Sent from: http://lucene.472066.n3.nabble.com/Solr-User-f472068.html >> > > > > -- > *Thanks and Regards,* > *Yogendra Kumar Soni*