On Sat, Nov 26, 2016 at 6:30 PM, Dikang Gu <dikan...@gmail.com> wrote: > Hi Marcus, > > Do you have some stack trace to show that which function in the ` > getNextBackgroundTask` is most expensive? > > Yeah, I think having 15-20K sstables in L0 is very bad, in our heavy-write > cluster, I try my best to reduce the impact of repair, and keep number of > sstables in L0 < 100. > > Thanks > Dikang. > > On Thu, Nov 24, 2016 at 12:53 PM, Nate McCall <zznat...@gmail.com> wrote: > >> > The reason is described here: >> https://issues.apache.org/jira/browse/CASSANDRA-5371? >> focusedCommentId=13621679&page=com.atlassian.jira. >> plugin.system.issuetabpanels:comment-tabpanel#comment-13621679 >> > >> > /Marcus >> >> "...a lot of the work you've done you will redo when you compact your now >> bigger L0 sstable against L1." >> >> ^ Sylvain's hypothesis (next comment down) is actually something we see >> occasionally in practice: having to re-write the contents of L1 too often >> when large L0 SSTables are pulled in. Here is an example we took on a >> system with pending compaction spikes that was seeing this specific issue >> with four LCS-based tables: >> >> https://gist.github.com/zznate/d22812551fa7a527d4c0d931f107c950 >> >> The significant part of this particular workload is a burst of heavy writes >> from long-duration scheduled jobs. >> > > > > -- > Dikang
-- Eric Evans john.eric.ev...@gmail.com