Norgorn [lsunnyd...@mail.ru] wrote:
> Collection contains about billion of documents.

So 3-400M documents per core. That is a challenge with frequent updates and 
facets, but with your simple queries it should be doable.

> At the end, I want to reach several seconds per search query (for not cached
> query =) ), so, please, give me some reference points.

The frustratingly true answer is
http://lucidworks.com/blog/sizing-hardware-in-the-abstract-why-we-dont-have-a-definitive-answer/

> How much (roughly) will I need RAM with and without SSDs?
> I know, it depends, but at least sommething, please.

Okay, something it is: We have a 256GB machine, running a SolrCloud with 17 
shards, each shard being about 900GB / 300M documents and put on a dedicated 
SSD. The machine currently has 160GB free for disk cache or about 1% of the 
total index size. For very simple unwarmed searches (just query on 1-3 terms, 
edismax over 6 fields with 2 phrase fields), median response time is < 200ms 
and nearly all response times are < 1 second. An extremely rough downscale with 
a factor 15 to approximate your 1TB index would leave 11GB for disk cache; 
divide it by 3 for your machines and it's 4GB disk cache/machine + whatever it 
takes to run your Solrs and the system itself.

BUT! All the shards are fully optimized and never updated, range filters can be 
tricky, multiple filters takes time, you have more documents/bytes than we have 
etc. 

> And HS means HelioSearch,

Ah. Of course. Although it helps with processing performance, it cannot do 
anything for your IO-problem,

How much memory is used for disk caching with your current setup?

- Toke

Reply via email to