ABSTRACT
The aim of this study was to evaluate the inter-rater reliability of the Phillips-checklist, a proposed framework for the quality assessment of modeling studies. Six raters evaluated nine modeling studies from three different medical specialties. Intra-class correlation (ICC) and corresponding variance components were estimated from these studies. Raters were asked to comment on their experience with the framework. While overall the mean inter-rater reliability showed no significant rater-effect (ICC = 0.69, p = 0.064), there was – presumably as a result of a lower study variability – a significant rater effect for clopidogrel only (p < 0.001). The framework allowed a more structured methodological assessment but several items remained unclear. Regarding the quality assessment of modeling studies with the proposed framework, the rater variability is similar or even higher than variability because of studies or residual effects. Several scoring items can and should be improved to ease interpretation.
Acknowledgments
There was no funding provided for this work from any organization, company or institute. The authors have no financial interests to disclose directly or indirectly related to the research in the article.
Financial & competing interests disclosure
The authors have no relevant affiliations or financial involvement with any organization or entity with a financial interest in or financial conflict with the subject matter or materials discussed in the manuscript. This includes employment, consultancies, honoraria, stock ownership or options, expert testimony, grants or patents received or pending, or royalties.
Supplemental data
Supplemental data for this article can be accessed here.