ABSTRACT
Over the last decade or two, multilevel structural equation modeling (ML-SEM) has become a prominent modeling approach in the social sciences because it allows researchers to correct for sampling and measurement errors and thus to estimate the effects of Level 2 (L2) constructs without bias. Because the latent variable modeling software Mplus uses maximum likelihood (ML) by default, many researchers in the social sciences have applied ML to obtain estimates of L2 regression coefficients. However, one drawback of ML is that covariance matrices of the predictor variables at L2 tend to be degenerate, and thus, estimates of L2 regression coefficients tend to be rather inaccurate when sample sizes are small. In this article, I show how an approach for stabilizing covariance matrices at L2 can be used to obtain more accurate estimates of L2 regression coefficients. A simulation study is conducted to compare the proposed approach with ML, and I illustrate its application with an example from organizational research.
Notes
1 Analysis of variance (ANOVA) estimates (e.g., Searle et al., Citation1992) can be derived as (2) (2) (3) (3) where N = Jn is the total number of observations.
2 Applying the ANOVA method yields (6) (6) (7) (7)
3 Kr is the number of indicators of the rth predictor variable.
4 The covariances of the measurement errors are assumed to be zero and thus need not to be estimated.
5 is a rather difficult to interpret. However, it may help to recognize that is close to the average number of persons per group when the data are nearly balanced (see Hox, Citation2010).
6 Note that this does not necessarily mean that degenerate matrices are completely useless in practice (see Savalei & Kolenikov, Citation2008, for the argument that degenerate matrices can be useful in the evaluation of model fit).
7 d must be chosen such that is an integer. Typically, d is much smaller than J (see Simulation Study). Once d has been selected, subsamples can be created by the following algorithm: First, divide (1, …, J) into subsets that do not overlap. For example, (d + 1, …, J), …, (1, …, J − d). Next, use these subsets to create the subsamples. More specifically, create the first subsample by maintaining groups d + 1, …, J, the second subsample by maintaining groups 1, …, d and 2d + 1, …, J, the third subsample by maintaining groups 1, …, 2d and 3d + 1, …, J, and so forth.
8 It should be noted that focusing on the smaller of the two between-group regression coefficients (βB, 2) instead would not have changed these findings.
9 Mplus runs a maximum of 500 iterations by default and uses a criterion of 0.001 to assess convergence (Muthén & Muthén, Citation2012, pp. 590–629).
10 In an additional analysis, I also investigated the statistical properties of the estimates of the within-group regression coefficient. With an absolute value of the estimated relative bias below 0.05, all approaches to estimating the ML-SEM were unbiased except the manifest model. They did not differ with respect to the relative RMSE, and the coverage rates were all close to the nominal 95% level.
11 Please find the equations that were used to compute these estimates as well as further details about their derivations in the Appendix.