Hello,

1. It depends on your query types & data (complexity, featureset, paging) - geospatial could be something with calculation inside solr? 2. It depends massively on the document size & field-selection (load a hundred of 100MB documents can take some time) 3. It depends especially on your disc IO / Ram utilization - are these dedicated machines ? 4. It depends on how often you changing your documents (cache warm-ups!!!, disc IO)! 5. What is the bottleneck? CPU ? RAM ? Disc? You should be able to give some more information about this. 6. It depends on the amount of cores (more cores must not be better - CPU-caching, OS-management overhead...) 7. Force cache hit-rate - means: control the type of queries, cluster them and send them to A or B - to have a higher chance for a cache hit.

Maybe you can give some more details about the points I mentioned.

Ralf

On 07/16/2013 04:42 PM, adfel70 wrote:
Hi
I need to create a solr cluster that contains geospatial information and
provides the ability to perform a few hundreds queries per second, each
query should retrieve around 100k results.
The data is around 100k documents, around 300gb total.

I started with 2 shard cluster (replicationFactor 1) and a portion of the
data - 20 gb.

I run some load-tests and see that when 100 requests are sent in one second,
the average qTime is around 4 seconds, but the average total response time
(measuring from sending the request to solr untill getting a response )
reaches 20-25 seconds which is very bad.

Currently I load-balance myself between the 2 solr servers (each request is
sent to another server)

Any advice on  which resources do I need and how my solr cluster should look
like?
More shards? more replicas? another webserver?

Thanks.





--
View this message in context: 
http://lucene.472066.n3.nabble.com/Need-advice-on-performing-300-queries-per-second-on-solr-index-tp4078353.html
Sent from the Solr - User mailing list archive at Nabble.com.

Reply via email to