;>> proposed by Michael and Clifford is a good one, but the solution assumes
>>> that the standard deviation parameter is the same for all three models.
>>>
>>> You may want to consider the degree by which the standard deviation
>>> estimates differ for the three
the predicted
values at the ends of the 3 regression lines are significantly
different... But I'm not sure how to do the Johnson-Neyman procedure
in R, so I think testing for slope differences will suffice!
Thanks to any who may be able to help!
Doug Adams
__
wledging? Does my aov syntax seem appropriate in the first
place?
Thanks everyone very much for any help you can give,
Doug Adams
question 1 question 2 question 3
question 4
group 1
subject 1 # # #
Ah, that did it. Thank you!
-
Doug Adams
MStat Student
University of Utah
--
View this message in context:
http://n4.nabble.com/lmer-mcmcsamp-coda-HPDinterval-tp1457803p1459380.html
Sent from the R help mailing list archive at Nabble.com.
__
R
of my fixed effects:
A6post <- mcmcsamp(A6mlm, 5)
library(coda)
HPDinterval(A6post)
..but I got this message:
"no applicable method for 'HPDinterval' applied to an object of class
"merMCMC""
Should I be coercing A6post to another type, or am I missing other steps
al
u type
>>
>> lme(y~x, random=~1|subjetc)
>>
>> On lme4 library you type
>>
>> lmer(y~x+(1|subject))
>>
>> You mixed them.
>>
>> At your disposal.
>
> Which is what I tell my wife when I am standing by our sink.
>
>>
some students are
registered as juniors & others as seniors within the same school.)
So schools are random, division is fixed, and the student Score is the
outcome variable. This is what I've tried:
lmer(data=Age6m, Score ~ division + (1|school), random=~1 | school)
Am I on the right track?
7 matches
Mail list logo