It actually is Beta that we're working with. /Martin
On Mon, Aug 27, 2012 at 10:38 PM, Martin Koch <m...@issuu.com> wrote: > (I'm working with Raghav on this): We've got several parallel workers that > add documents in batches of 16 through pysolr, and using commitWithin at 60 > seconds when the commit causes solr to freeze; if the commit is only 5 > seconds, then everything seems to work fine. In both cases, throughput is > around 500 documents / second. > > We can certainly give it a try with the Beta. > > Thanks, > /Martin > > > On Mon, Aug 27, 2012 at 7:30 PM, Mark Miller <markrmil...@gmail.com>wrote: > >> How are you adding the docs? In batch, streaming, a doc at a time? >> >> Any chance you can try with the Beta? >> >> On Mon, Aug 27, 2012 at 9:35 AM, Raghav Karol <r...@issuu.com> wrote: >> > Hello *, >> > >> > We are using SolrClould 4.0 - Alpha and have a 4 machine setup. >> > >> > Machine 1 - 16 Solr cores - Shard 1 - 16 >> > Machine 2 - 16 Solr cores - Shard 17 - 32 >> > Machine 3 - 16 Solr cores - Replica 1 - 16 >> > Machine 4 - 16 Solr cores - Replice 17 - 32 >> > >> > Index at 500 docs/sec and committing every 60 seconds, i.e., 30,000 >> Documents causes Solr to freeze. There is nothing in the logs to indicate >> errors or replication activity - Solr just appeared to freeze. >> > >> > Increasing the commit frequency we observed that commits of at most >> 2,500 docs worked fine. >> > >> > Are we using SolrCloud and replications incorrectly? >> > >> > -- >> > Raghav >> > >> > >> > >> >> >> >> -- >> - Mark >> >> http://www.lucidimagination.com >> > >