655
Views
14
CrossRef citations to date
0
Altmetric
Articles

Model Selection and Model Averaging for Mixed-Effects Models with Crossed Random Effects for Subjects and Items

ORCID Icon, ORCID Icon & ORCID Icon
Pages 603-619 | Published online: 26 Feb 2021
 

Abstract

A good deal of experimental research is characterized by the presence of random effects on subjects and items. A standard modeling approach that includes such sources of variability is the mixed-effects models (MEMs) with crossed random effects. However, under-parameterizing or over-parameterizing the random structure of MEMs bias the estimations of the Standard Errors (SEs) of fixed effects. In this simulation study, we examined two different but complementary perspectives: model selection with likelihood-ratio tests, AIC, and BIC; and model averaging with Akaike weights. Results showed that true model selection was constant across the different strategies examined (including ML and REML estimators). However, sample size and variance of random slopes were found to explain true model selection and SE bias of fixed effects. No relevant differences in SE bias were found for model selection and model averaging. Sample size and variance of random slopes interacted with the estimator to explain SE bias. Only the within-subjects effect showed significant underestimation of SEs with smaller number of items and larger item random slopes. SE bias was higher for ML than REML, but the variability of SE bias was the opposite. Such variability can be translated into high rates of unacceptable bias in many replications.

Article information

Conflict of Interest Disclosures: Each author signed a form for disclosure of potential conflicts of interest. No authors reported any financial or other conflicts of interest in relation to the work described.

Ethical Principles: The authors affirm having followed professional ethical guidelines in preparing this work. These guidelines include obtaining informed consent from human participants, maintaining ethical treatment and respect for the rights of human or animal participants, and ensuring the privacy of participants and their data, such as ensuring that individual participants cannot be identified in reported results or from publicly available original or archival data.

Funding: This work was not supported.

Role of the Funders/Sponsors: None of the funders or sponsors of this research had any role in the design and conduct of the study; collection, management, analysis, and interpretation of data; preparation, review, or approval of the manuscript; or decision to submit the manuscript for publication.

Acknowledgments: The ideas and opinions expressed herein are those of the authors alone, and endorsement by the authors’ institutions is not intended and should not be inferred.

Notes

1 In some contexts, fixed and random effects of MEMs are called models for the mean and variance, respectively.

2 Note that random effects are relevant for different reasons. One of them is their relevance to estimate SEs and p-values of fixed effects. Other more common reason is they provide information about variance in the data. Not only the presence of relevant random effects can be useful, but can be the focus of hypothesis testing. This aspect is particularly relevant because the parameterization of random effects should be considered a confirmatory hypothesis testing itself (Barr, Citation2013).

3 Given that the effect size can be influenced by random slopes variances, two different versions of the effect size were tested in the present study. The first one was the simulated effect size (Table 1). The second one was an effect size that was corrected by the sampling variance considering the error term and the size of the random slopes of subjects and items. Willett (Citation1989) presents two different illustrations for longitudinal modeling, and Judd et al. (Citation2017) present a similar approach for experimental research. In the present study, no relevant differences were found between these two approaches when evaluating the influence of effect size in true model selection and SE bias. Thus, we report the results of the original effect size from Table 1.

4 We called it minimal model following Matuschek et al. (Citation2017).

5 We called it maximal model following Barr et al. (Citation2013).

6 We also analyzed the performance of AICc correcting for the sample size (number of subjects) and the number of observations (number of subjects multiplied by number of items), but no relevant differences in true model selection were obtained with AIC index. For the sake of brevity, we decided to focus on AIC and BIC indices.

Log in via your institution

Log in to Taylor & Francis Online

PDF download + Online access

  • 48 hours access to article PDF & online version
  • Article PDF can be downloaded
  • Article PDF can be printed
USD 53.00 Add to cart

Issue Purchase

  • 30 days online access to complete issue
  • Article PDFs can be downloaded
  • Article PDFs can be printed
USD 352.00 Add to cart

* Local tax will be added as applicable

Related Research

People also read lists articles that other readers of this article have read.

Recommended articles lists articles that we recommend and is powered by our AI driven recommendation engine.

Cited by lists all citing articles based on Crossref citations.
Articles with the Crossref icon will open in a new tab.