Thanks, Erick. I should probably keep a tally of how many beers I owe
you ;-)


On 21/09/2020 14:50, Erick Erickson wrote:
> In a word, yes. G1GC still has spikes, and the larger the heap the more 
> likely you’ll be to encounter them. So having multiple JVMS rather than one 
> large JVM with a ginormous heap is still recommended.
> 
> I’ve seen some cases that used the Zing zero-pause product with very large 
> heaps, but they were forced into that by the project requirements.
> 
> That said, when Java has a ZCG option, I think we’re in uncharted territory. 
> I frankly don’t know what using very large heaps without having to worry 
> about GC pauses will mean for Solr. I suspect we’ll have to do something to 
> take advantage of that. For instance, could we support a topology where all 
> shards had at least one replica in the same JVM that didn’t make any HTTP 
> requests? Would that topology be common enough to support? Maybe extend “rack 
> aware” to be “JVM aware”? Etc.
> 
> One thing that does worry me is that it’ll be easier and easier to “just 
> throw more memory at it” rather than examine whether you’re choosing options 
> that minimize heap requirements. And Lucene has done a lot to move memory to 
> the OS rather than heap (e.g. docValues, MMapDirectory etc.).
> 
> Anyway, carry on as before for the nonce.
> 
> Best,
> Erick
> 
>> On Sep 21, 2020, at 6:06 AM, Bram Van Dam <bram.van...@intix.eu> wrote:
>>
>> Hey folks,
>>
>> I've always heard that it's preferred to have a SolrCloud setup with
>> many smaller instances under the CompressedOops limit in terms of
>> memory, instead of having larger instances with, say, 256GB worth of
>> heap space.
>>
>> Does this recommendation still hold true with newer garbage collectors?
>> G1 is pretty fast on large heaps. ZGC and Shenandoah promise even more
>> improvements.
>>
>> Thx,
>>
>> - Bram
> 

Reply via email to