Hi, I used solr 5.2.1 version. It is fast, I think. But again, I am stuck on concurrent searching and threading. I changed <maxWarmingSearchers>*2*</maxWarmingSearchers> to <maxWarmingSearchers>*100*</maxWarmingSearchers>. And apply simultaneous searching using 100 workers. It works fast but not upto the mark.
It increases searching from 1.5 to 0.5 seconds. But If I run only single worker then searching time is 0.03 seconds, it is too fast but not possible with 100 workers simultaneously. As Shawn said - "Making 100 concurrent indexing requests at the same time as 100 concurrent queries will overwhelm *any* single Solr server". I got your point. But MongoDB can handle concurrent searching and indexing faster. Then why not solr? Sorry for this.. On Mon, Aug 10, 2015 at 2:39 AM Shawn Heisey <apa...@elyograg.org> wrote: > On 8/7/2015 1:15 PM, Nitin Solanki wrote: > > I wrote a python script for indexing and using > > urllib and urllib2 for indexing data via http.. > > There are a number of Solr python clients. Using a client makes your > code much easier to write and understand. > > https://wiki.apache.org/solr/SolPython > > I have no experience with any of these clients, but I can say that the > one encountered most often when Python developers come into the #solr > IRC channel is pysolr. Our wiki page says the last update for pysolr > happened in December of 2013, but I can see that the last version on > their web page is dated 2015-05-26. > > Making 100 concurrent indexing requests at the same time as 100 > concurrent queries will overwhelm *any* single Solr server. In a > previous message you said that you have 4 CPU cores. The load you're > trying to put on Solr will require at *LEAST* 200 threads. It may be > more than that. Any single system is going to have trouble with that. > A system with 4 cores will be *very* overloaded. > > Thanks, > Shawn > >