On Sat, Nov 26, 2016 at 6:30 PM, Dikang Gu <dikan...@gmail.com> wrote:
> Hi Marcus,
>
> Do you have some stack trace to show that which function in the `
> getNextBackgroundTask` is most expensive?
>
> Yeah, I think having 15-20K sstables in L0 is very bad, in our heavy-write
> cluster, I try my best to reduce the impact of repair, and keep number of
> sstables in L0 < 100.
>
> Thanks
> Dikang.
>
> On Thu, Nov 24, 2016 at 12:53 PM, Nate McCall <zznat...@gmail.com> wrote:
>
>> > The reason is described here:
>> https://issues.apache.org/jira/browse/CASSANDRA-5371?
>> focusedCommentId=13621679&page=com.atlassian.jira.
>> plugin.system.issuetabpanels:comment-tabpanel#comment-13621679
>> >
>> > /Marcus
>>
>> "...a lot of the work you've done you will redo when you compact your now
>> bigger L0 sstable against L1."
>>
>> ^ Sylvain's hypothesis (next comment down) is actually something we see
>> occasionally in practice: having to re-write the contents of L1 too often
>> when large L0 SSTables are pulled in. Here is an example we took on a
>> system with pending compaction spikes that was seeing this specific issue
>> with four LCS-based tables:
>>
>> https://gist.github.com/zznate/d22812551fa7a527d4c0d931f107c950
>>
>> The significant part of this particular workload is a burst of heavy writes
>> from long-duration scheduled jobs.
>>
>
>
>
> --
> Dikang



-- 
Eric Evans
john.eric.ev...@gmail.com

Reply via email to