When you're doing hard commits, is it with openSeacher = true or false? It should probably be false...
Here's a rundown of the soft/hard commit consequences: http://searchhub.org/2013/08/23/understanding-transaction-logs-softcommit-and-commit-in-sorlcloud/ I suspect (but, of course, can't prove) that you're over-committing and hitting segment merges without meaning to... FWIW, Erick On Wed, Jan 22, 2014 at 1:46 PM, Software Dev <static.void....@gmail.com> wrote: > A suggestion would be to hard commit much less often, ie every 10 > minutes, and see if there is a change. > > - Will try this > > How much system RAM ? JVM Heap ? Enough space in RAM for system disk cache ? > > - We have 18G of ram 12 dedicated to Solr but as of right now the total > index size is only 5GB > > Ah, and what about network IO ? Could that be a limiting factor ? > > - What is the size of your documents ? A few KB, MB, ... ? > > Under 1MB > > - Again, total index size is only 5GB so I dont know if this would be a > problem > > > > > > > On Wed, Jan 22, 2014 at 12:26 AM, Andre Bois-Crettez > <andre.b...@kelkoo.com>wrote: > >> 1 node having more load should be the leader (because of the extra work >> of receiving and distributing updates, but my experiences show only a >> bit more CPU usage, and no difference in disk IO). >> >> A suggestion would be to hard commit much less often, ie every 10 >> minutes, and see if there is a change. >> How much system RAM ? JVM Heap ? Enough space in RAM for system disk cache >> ? >> What is the size of your documents ? A few KB, MB, ... ? >> Ah, and what about network IO ? Could that be a limiting factor ? >> >> >> André >> >> >> On 2014-01-21 23:40, Software Dev wrote: >> >>> Any other suggestions? >>> >>> >>> On Mon, Jan 20, 2014 at 2:49 PM, Software Dev <static.void....@gmail.com> >>> wrote: >>> >>> 4.6.0 >>>> >>>> >>>> On Mon, Jan 20, 2014 at 2:47 PM, Mark Miller <markrmil...@gmail.com >>>> >wrote: >>>> >>>> What version are you running? >>>>> >>>>> - Mark >>>>> >>>>> On Jan 20, 2014, at 5:43 PM, Software Dev <static.void....@gmail.com> >>>>> wrote: >>>>> >>>>> We also noticed that disk IO shoots up to 100% on 1 of the nodes. Do >>>>>> all >>>>>> updates get sent to one machine or something? >>>>>> >>>>>> >>>>>> On Mon, Jan 20, 2014 at 2:42 PM, Software Dev < >>>>>> >>>>> static.void....@gmail.com>wrote: >>>>> >>>>>> We commit have a soft commit every 5 seconds and hard commit every 30. >>>>>>> >>>>>> As >>>>> >>>>>> far as docs/second it would guess around 200/sec which doesn't seem >>>>>>> >>>>>> that >>>>> >>>>>> high. >>>>>>> >>>>>>> >>>>>>> On Mon, Jan 20, 2014 at 2:26 PM, Erick Erickson < >>>>>>> >>>>>> erickerick...@gmail.com>wrote: >>>>> >>>>>> Questions: How often do you commit your updates? What is your >>>>>>>> indexing rate in docs/second? >>>>>>>> >>>>>>>> In a SolrCloud setup, you should be using a CloudSolrServer. If the >>>>>>>> server is having trouble keeping up with updates, switching to CUSS >>>>>>>> probably wouldn't help. >>>>>>>> >>>>>>>> So I suspect there's something not optimal about your setup that's >>>>>>>> the culprit. >>>>>>>> >>>>>>>> Best, >>>>>>>> Erick >>>>>>>> >>>>>>>> On Mon, Jan 20, 2014 at 4:00 PM, Software Dev < >>>>>>>> >>>>>>> static.void....@gmail.com> >>>>> >>>>>> wrote: >>>>>>>> >>>>>>>>> We are testing our shiny new Solr Cloud architecture but we are >>>>>>>>> experiencing some issues when doing bulk indexing. >>>>>>>>> >>>>>>>>> We have 5 solr cloud machines running and 3 indexing machines >>>>>>>>> >>>>>>>> (separate >>>>> >>>>>> from the cloud servers). The indexing machines pull off ids from a >>>>>>>>> >>>>>>>> queue >>>>> >>>>>> then they index and ship over a document via a CloudSolrServer. It >>>>>>>>> >>>>>>>> appears >>>>>>>> >>>>>>>>> that the indexers are too fast because the load (particularly disk >>>>>>>>> >>>>>>>> io) >>>>> >>>>>> on >>>>>>>> >>>>>>>>> the solr cloud machines spikes through the roof making the entire >>>>>>>>> >>>>>>>> cluster >>>>>>>> >>>>>>>>> unusable. It's kind of odd because the total index size is not even >>>>>>>>> large..ie, < 10GB. Are there any optimization/enhancements I could >>>>>>>>> >>>>>>>> try >>>>> >>>>>> to >>>>>>>> >>>>>>>>> help alleviate these problems? >>>>>>>>> >>>>>>>>> I should note that for the above collection we have only have 1 >>>>>>>>> shard >>>>>>>>> >>>>>>>> thats >>>>>>>> >>>>>>>>> replicated across all machines so all machines have the full index. >>>>>>>>> >>>>>>>>> Would we benefit from switching to a ConcurrentUpdateSolrServer >>>>>>>>> where >>>>>>>>> >>>>>>>> all >>>>>>>> >>>>>>>>> updates get sent to 1 machine and 1 machine only? We could then >>>>>>>>> >>>>>>>> remove >>>>> >>>>>> this >>>>>>>> >>>>>>>>> machine from our cluster than that handles user requests. >>>>>>>>> >>>>>>>>> Thanks for any input. >>>>>>>>> >>>>>>>> >>>>>>> >>>>> >>> -- >>> André Bois-Crettez >>> >>> Software Architect >>> Search Developer >>> http://www.kelkoo.com/ >>> >> >> Kelkoo SAS >> Société par Actions Simplifiée >> Au capital de € 4.168.964,30 >> Siège social : 8, rue du Sentier 75002 Paris >> 425 093 069 RCS Paris >> >> Ce message et les pièces jointes sont confidentiels et établis à >> l'attention exclusive de leurs destinataires. Si vous n'êtes pas le >> destinataire de ce message, merci de le détruire et d'en avertir >> l'expéditeur. >>