Here is an update (final for now) with 10 models and 4 hierarchical
sequences. Comments, details, and a demonstration with the Machines data
are available in a new RPub[1]. I switched to a notation with numeric
covariates to reduce potential confusion about terms containing `1+f` and
`0+f`; `c1` and `c2` are contrasts defined for a factor with levels `A`,
`B`, and `C`.
## LMMs
### Prototype LMMs
max_LMM: m1 = y ~ 1 + c1 + c2 + (1 + c1 + c2 | subj)
prsm1_LMM: m2 = y ~ 1 + c1 + c2 + (1 | subj) + (0 + c1 + c2 | subj)
zcp_LMM: m3 = y ~ 1 + c1 + c2 + (1 + c1 + c2 || subj)
prsm2_LMM: m4 = y ~ 1 + c1 + c2 + (1 + c2 || subj)
min_LMM: m5 = y ~ 1 + c1 + c2 + (1 | subj)
Protoype refers to prsm1 and prsm2; there are also variations of them.
Prsm1 are models pruning correlation parameters; prsm2 are models pruning
variance components in the absence of correlation parameters. The order
reflects hierarchical decreasing model complexity.
### LMMs with interaction term
int_LMM: m6 = y ~ 1 + c1 + c2 + (1 | subj) + (1 | factor:subj)
min2_Lmm: m7 = y ~ 1 + c1 + c2 + (1 | factor:subj)
### LMMs with (0 + f | g) RE structures
maxL_MM_RE0: m8 = y ~ 1 + c1 + c2 + (0 + A + B + C | subj)
prsm1_LMM_RE0: m9 = y ~ 1 + c1 + c2 + (0 + A | subj) + (0 + B + C | subj)
zcp_LMM_RE0: m10 = y ~ 1 + c1 + c2 + (0 + A + B + C || subj)
## Hierarchical model sequences
Here are the sequences I am confident about.
(1) max_LMM -> prsm1_LMM -> zcp_LMM -> prsm2_LMM -> min1_LMM
(2) max_LMM -> int_LMM -> min1_LMM
(3) max_LMM -> int_LMM -> min2_LMM
(4) max_LMM_RE0 -> prsm1_LMM_RE0 -> zcp_LMM_RE0
So far, in my research, I have worked almost exclusively with Sequence (1)
and do not recall ever experiencing technical problems or inconsistencies
as long as there was no overparameterization. It has served me well in the
determination of parsimonious LMMs[2, 3].
For complex fixed-effect structures (i.e., for models with many factors or
with factors with many levels), the number of correlation parameters grows
very rapidly. If this goes together with a modest number of observations or
levels of the random factor, Sequences (2) and (3) might be a good place to
start to avoid convergence problems. Of course, if your hypotheses are
about correlation parameters, these LMMs will not get you very far.
Finally, Sequence (4) could be a default strategy if one is interested in
the fixed effects and if one does not want to spend much time in wondering
about the meaning of CPs. However, as correlations between levels of fixed
factors are can be very large (at least larger than correlations between
effects), LMMs in this sequence may be prone to convergence problems.
A few open questions -
(1) Rune Haubo proposed that min2_LMM (his fm2) is nested under zcp_LMM_RE0
(his fm3) and could serve as a baseline model, but I don't understand why
it is nested.
(2) Marteen Jung wonders about Rune Haubo's fm5 (with four VCs; there was
typo in my last post). The ten models above have at most 3 VCs (i.e., the
number of levels of the factor). I also don't see a good place for it in
the sequences. Am I missing something?
(3) Any suggestions for models between maxLMM and intLMM and for models
between intLMM and minLMM?
Best
Reinhold Kliegl
[1] http://rpubs.com/Reinhold/391828
[2] https://arxiv.org/abs/1506.04967
[3] https://www.sciencedirect.com/science/article/pii/S0749596X17300013
Post by Reinhold KlieglThere is an interpretable alternative to fm5 (actually there are many
...), called fm8 below, that avoids the redundancy between variance
components. The change is to switch from (1 |g) + (0 + f | g) = (1 | g) +
(0 + A + B + C | g) to 1 | g) + (0 + c1 + c2 |g ), where c1 and c2 are the
contrasts defined for f. (I have actually used such LMMs quite often.) With
this specification the difference to the maxLMM (fm6) is that the
correlation between intercept and contrasts is suppressed to zero. The
correlation parameters now refer to the correlations between effects of c1
and c2, not to the correlations between A, B, and C. Actually, this is but
one example of many LMMs one could slot into this position of the
hierarchical model sequences. At this level of model complexity one can
suppress various subsets of correlation parameters (as illustrated in Bates
et al. (2015)[1] and various vignettes of the RePsychLing package).
Post by Reinhold Klieglfm1 = y ~ 1 + f + (1 | g) # minimal LMM version 1
(min1LMM)
Post by Reinhold Klieglfm2 = y ~ 1 + f + (1 | f:g) # minimal LMM version 2
(min2LMM)
Post by Reinhold Klieglfm3 = y ~ 1 + f + (0 + f || g) # zcpLMM with 0 in RE
(zcpLMM_RE0)
Post by Reinhold Klieglfm4 = y ~ 1 + f + (1 | g) + (1 | f:g) # LMM w/ f x g interaction
(intLMM)
Post by Reinhold Klieglfm5 = y ~ 1 + f + (1 | g) + (0 + f | g) # N/A
fm6 = y ~ 1 + f + (1 + f | g) # maximal LMM (maxLMM)
fm7 = y ~ 1 + f + (1 + f || g) # zcpLMM with 1 in RE
(zcpLMM_RE1)
Post by Reinhold Klieglfm8 = y ~ 1 + f + (1 | g) + (0 + c1 + c2 | g) # parsimonious LMM (prsmLMM)
Hierarchical model sequences
(1) maxLMM_RE1 -> prsmLMM -> intLMM -> min1LMM # fm6 -> fm8 -> fm4
-> fm1
Post by Reinhold Kliegl(2) maxLMM_RE1 -> prsmLMM -> intLMM -> min2LMM # fm6 -> fm8 -> fm4
-> fm2
Post by Reinhold Kliegl(3) maxLMM_RE0 -> prsmLMM -> zcpLMM_RE0 -> min2LMM # fm6 -> fm8 -> fm3
-> fm2
Post by Reinhold Kliegl(4) maxLMM_RE1 -> prsmLMM -> zcpLMM_RE1 -> min1LMM # fm6 -> fm8 -> fm7
-> fm1 (new sequence)
Post by Reinhold Kliegl```
I will update the RPub in the next days.
[1] https://arxiv.org/pdf/1506.04967.pdf
Best regards,
Reinhold Kliegl
On Tue, May 22, 2018 at 11:00 AM, Maarten Jung <
Post by Maarten JungI see that fm2 is nested within fm3 and fm4.
But I have a hard time understanding fm3 and fm2 because, as Reinhold
Kiegl said, they specify the f:g interaction but without the g main effect.
Can someone provide an intuition for these models?
Post by Reinhold KlieglPost by Maarten JungAlso, it is not entirely clear to me what fm5 represents. It looks to
me, and again I am with Reinhold Kiegl , as if there were
over-parameterization going on.
Post by Reinhold KlieglPost by Maarten JungCheers,
Maarten
On Tue, May 22, 2018 at 9:45 AM, Reinhold Kliegl <
Post by Reinhold KlieglOk, I figured out the answer to the question about fm2.
fm2 is indeed a very nice baseline for fm3 and fm4. So I distinguish
between min1LMM and min2LMM.
(min1LMM)
(min2LMM)
Post by Reinhold KlieglPost by Maarten JungPost by Reinhold Klieglfm3 = y ~ 1 + f + (0 + f || g) # zcpLMM with 0 in RE (zcpLMM_RE0)
fm4 = y ~ 1 + f + (1 | g) + (1 | f:g) # LMM w/ f x g interaction (intLMM)
fm5 = y ~ 1 + f + (1 | g) + (0 + f | g) # N/A
fm6 = y ~ 1 + f + (1 + f | g) # maximal LMM (maxLMM)
fm7 = y ~ 1 + f + (1 + f || g) # zcpLMM with 1 in RE (zcpLMM_RE1)
(1) maxLMM_RE1 -> intLMM -> min1LMM # fm6 -> fm4 -> fm1
(2) maxLMM_RE1 -> intLMM -> min2LMM # fm6 -> fm4 -> fm2
(3) maxLMM_RE0 -> zcpLMM_RE0 -> min2LMM # fm6 -> fm3 -> fm2
(4) maxLMM_RE1 -> zcpLMM_RE1 -> min1LMM # fm6 -> fm7 -> fm1 (new sequence)
On Tue, May 22, 2018 at 12:21 AM, Reinhold Kliegl <
Post by Reinhold KlieglSorry, I am somewhat late to this conversation. I am responding to
this thread, because it fits my comment very well, but it was initially
triggered by a previous thread, especially Rune Haubo's post here [1]. So I
hope it is ok to continue here.
put up along with this post [2]. I start with a translation between Rune
Post by Reinhold KlieglPost by Maarten JungPost by Reinhold KlieglPost by Reinhold Klieglfm1 = y ~ 1 + f + (1 | g) # minimal LMM (minLMM)
fm3 = y ~ 1 + f + (0 + f || g) # zero-corr param LMM with 0 in
RE (zcpLMM_RE0)
interaction (intLMM),
Post by Reinhold KlieglPost by Maarten JungPost by Reinhold KlieglPost by Reinhold Klieglfm6 = y ~ 1 + f + (1 + f | g) # maximal LMM (maxLMM)
fm7 = y ~ 1 + f + (1 + f || g) # zero-corr param LMM with 1 in
RE (zcpLMM_RE1)
are in Rune Haubo's post; fm7 is new (added by me). I have not used fm2 and
fm5 so far (see below).
under zcpLMM. I had included this LRT in my older RPub cited in the thread,
but I stand corrected and agree with Rune Haubo that intLMM is not nested
under zcpLMM. For example, in the new RPub, I show that slightly modified
Machines data exhibit smaller deviance for intLMM than zcpLMM despite an
additional model parameter in the latter. Thanks for the critical reading.
my translation (right)
Post by Reinhold KlieglPost by Maarten JungPost by Reinhold KlieglPost by Reinhold Kliegl(1) fm6 -> fm5 -> fm4 -> fm1 # maxLMM_RE1 -> fm5 -> intLMM -> minLMM
(2) fm6 -> fm5 -> fm4 -> fm2 # maxLMM_RE1 -> fm5 -> intLMM -> fm2
(3) fm6 -> fm5 -> fm3 -> fm2 # maxLMM_RE1 -> fm5 -> zcpLMM_RE0 -> fm2
and here are sequences I came up with (left) augmented with
translation into RH's fm's.
Post by Reinhold KlieglPost by Maarten JungPost by Reinhold KlieglPost by Reinhold Kliegl(1) maxLMM_RE1 -> intLMM -> minLMM # fm6 -> fm4 -> fm1
(3) maxLMM_RE0 -> zcpLMM_RE0 # fm6 -> fm3
(4) maxLMM_RE1 -> zcpLMM_RE1 -> minLMM # fm6 -> fm7 -> fm1 (new sequence)
(III) I have questions about fm2 and fm5.
fm2: fm2 redefines the levels of the group factor (e.g., in the
cake data there are 45 groups in fm2 compared to 15 in the other models).
Why is fm2 nested under fm3 and fm6? Somehow it looks to me that you
include an f:g interaction without the g main effect (relative to fm4).
This looks like an interesting model; I would appreciate a bit more
conceptual support for its interpretation in the model hierarchy.
only 3 levels. So to me this looks like there is redundancy built into the
model. In support of this intuition, for the cake data, one of the VCs is
estimated with 0. However, in the Machine data the model was not
degenerate. So I am not sure. In other words, if the factor levels are A,
B, C, and the two contrasts are c1 and c2, I thought I can specify either
(1 + c1 + c2) or (0 + A + B + C). fm5 specifies (1 + A + B + C) which is
rank deficient in the fixed effect part, but not necessarily in the
random-effect term. What am I missing here?
https://stat.ethz.ch/pipermail/r-sig-mixed-models/2018q2/026775.html
Post by Reinhold KlieglPost by Maarten JungPost by Reinhold KlieglPost by Reinhold Kliegl[2] http://rpubs.com/Reinhold/391027
Best,
Reinhold Kliegl
On Thu, May 17, 2018 at 12:43 PM, Maarten Jung <
Post by Maarten JungDear list,
When one wants to specify a lmer model including variance components but no
correlation parameters for categorical predictors (factors) afaik one has
to convert the factors to numeric covariates or use lme4::dummy(). Until
recently I thought m2a (or equivalently m2b using the double-bar syntax)
would be the correct way to specify such a zero-correlation parameter model.
But in this thread [1] Rune Haubo Bojesen Christensen pointed out that this
model does not make sense to him. Instead he suggests m3 as an appropriate
model.
I think this is a *highly relevant difference* for everyone who uses
factors in lmer and therefore I'm bringing up this issue again. But maybe
I'm mistaken and just don't get what is quite obvious for more experienced
mixed modelers.
Please note that the question is on CrossValidated [2] but some consider it
as off-topic and I don't think there will be an answer any time soon.
How should one specify a lmm without correlation parameters for factors and
what are the differences between m2a and m3?
Is there a preferred model for model comparison with m4 (this model is also
discussed here [3])?
library("lme4")
data("Machines", package = "MEMSS")
d <- Machines
contrasts(d$Machine) # default coding: contr.sum
m1 <- lmer(score ~ Machine + (Machine | Worker), d)
c1 <- model.matrix(m1)[, 2]
c2 <- model.matrix(m1)[, 3]
m2a <- lmer(score ~ Machine + (1 | Worker) + (0 + c1 | Worker) + (0 + c2 |
Worker), d)
m2b <- lmer(score ~ Machine + (c1 + c2 || Worker), d)
VarCorr(m2a)
Groups Name Std.Dev.
Worker (Intercept) 5.24354
Worker.1 c1 2.58446
Worker.2 c2 3.71504
Residual 0.96256
m3 <- lmer(score ~ Machine + (1 | Worker) + (0 + dummy(Machine, "A") |
Worker) +
(0 + dummy(Machine, "B") |
Worker) +
(0 + dummy(Machine, "C") |
Worker), d)
VarCorr(m3)
Groups Name Std.Dev.
Worker (Intercept) 3.78595
Worker.1 dummy(Machine, "A") 1.94032
Worker.2 dummy(Machine, "B") 5.87402
Worker.3 dummy(Machine, "C") 2.84547
Residual 0.96158
m4 <- lmer(score ~ Machine + (1 | Worker) + (1 | Worker:Machine), d)
[1]
https://stat.ethz.ch/pipermail/r-sig-mixed-models/2018q2/026775.html
Post by Reinhold KlieglPost by Maarten JungPost by Reinhold KlieglPost by Reinhold KlieglPost by Maarten Jung[2] https://stats.stackexchange.com/q/345842/136579
[3] https://stats.stackexchange.com/q/304374/136579
Best regards,
Maarten
[[alternative HTML version deleted]]
_______________________________________________
https://stat.ethz.ch/mailman/listinfo/r-sig-mixed-models
[[alternative HTML version deleted]]