>> Also i have read in Quinn and Keough 2002, design and analysis of
>> experiments for
>> biologists, that a variance component analysis should only be conducted
>> after a rejection
>> of the null hypothesis of no variance at that level.

Hmmm...

This does rather assume that 'no significant result' means 'near-zero variance 
contribution'.

These are not identical statements if the anova has low power; an 
'insignificant' term can conceal practically important sizes of effect. So if 
you have a smallish number of groups (say, ten or less) you might want to find 
out what that estimated between-group variance could have been before you throw 
it away. That's especially important if you're expecting to say something about 
standard errors or confidence intervals of fixed effects. 

I may well be biased, here, though. In the kinds of nested design I get 
involved in (often inter-laboratory or homogeneity studies in chemistry), there 
is nearly always a between-group  effect; the only question is its size. Under 
those circumstances, the null hypothesis is not a particularly compelling 
starting point.  I'd rather have a variance component estimate and know how 
vague it was than assume it wasn't there at all.

But if you have good power and a good reason for believing there's no case to 
answer, sure; assume zero unless proven otherwise.

S


*******************************************************************
This email and any attachments are confidential. Any use...{{dropped:8}}

______________________________________________
R-help@r-project.org mailing list
https://stat.ethz.ch/mailman/listinfo/r-help
PLEASE do read the posting guide http://www.R-project.org/posting-guide.html
and provide commented, minimal, self-contained, reproducible code.

Reply via email to