3,097
Views
4
CrossRef citations to date
0
Altmetric
Editorials

Quantitative methods in interprofessional education research: some critical reflections and ideas to improving rigor

ORCID Icon &

Introduction

Although quantitative research on interprofessional education (IPE) is proliferating, the accelerating growth in this empirical work does not necessarily imply a stronger evidence base. The contrast between more empiricism and little-improved evidence is perplexing. On the one hand, quantitative research output has accelerated with little signs of slowing down. On the other hand, best-evidence syntheses have been limited to descriptive summaries, in part, due to the varietal methodologies across IPE studies (e.g. Reeves et al., Citation2016; Reeves, Palaganas, & Zierler, Citation2017). As Costanza (Citation2015) noted recently, “in many ways, poor research methodology is the reason… research to date fails to show the impact of IPE” (p.35). Moreover, the methodological under-achievement of IPE may have been borne by early untended calls to focus IPE’s empiricism (Reeves, Citation2010).

This editorial provides an overview of the common quantitative study designs for IPE and addresses some implications for their analytic approaches. Subsequently, the editorial turns toward the prevailing measurement paradigm employed in IPE​—classical test theory (CTT). Specifically, the editorial discusses key limitations of CTT for IPE research and suggests how the adoption of modern psychometrics in IPE research – namely, item response theory (IRT) can remedy CTT’s limitations using substantive examples from IPE’s empirical literature.

Growth of quantitative research

Although the global increased-incidence of IPE is noteworthy, the concomitant use of varietal study methods limits the valuations and substantive impact of this form of education on practice and patients (e.g. Harath et al., Citation2017). Inconsistent methods partly owe to scarce guidelines, which have only recently become available for IPE researchers (e.g., Institute of Medicine, Citation2015; Reeves, Boet, Zierler, & Kitto, Citation2015). For quantitative research to sustain the development of IPE, stronger evidence may be achieved by offering more specific guidelines. Drawing on critical reflections from extant empirical literature, representative cases are presented below for use by IPE researchers.

Some common designs

In the Institute of Medicine (Citation2015) report on IPE effectiveness, the conventional ‘gold-standard’ design of randomized-control trials (RCTs) is tempered to, “have limited utility for measuring the effects of IPE” (p.40). Because experimental designs are lesser-used in IPE studies, it is worth mentioning that the experiment’s “limitation” is likely due to classical inactive controls, which constricts the efficacy of IPE to absolute terms, such as ‘IPE’ vs. ‘no IPE’ (Brashers, Phillips, Malpass, & Owen, Citation2015). In contrast, active-controls (‘IPE comparators’) are suitable and appropriate for IPE evaluation in relative terms. For example, IPE experiments using ‘active controls’ may readily evaluate, comparative pedagogies (didactic lecture vs. problem-based learning), comparative measurements (single-Likert response vs. forced-choice format), or comparative contexts (classroom/acute-hospital vs. community-based settings).

For non-experimental (observational) designs, programs are typically evaluated by their usefulness for improving meaningful outcomes. For IPE researchers, different designs are optimal for different research questions and outcomes evaluated. Carpenter’s (Citation1995) early seminal IPE study, for example, utilized a before-after design to evaluate changes in students’ interprofessional attitudes. In another example, as IPE institutionalization leads to more formalized curricula, the value-added of controlled before-after designs (CBAs) vis-à-vis ‘procedural confounds’Footnote2 and ‘maturational threats’ can reasonably be expected to diminish for IPE research. Finally, the normative expectancies for IPE may approximate an ‘interrupted time-series’ (ITS) design with repeated IPE programs implemented to student populations over time. The changing proportion of IPE ‘elective’/‘compulsory’ offerings may inform the ‘usefulness’ of IPE reaction measures developed in the early stages of IPE institutionalization. More pragmatic for education administrators, the resource-intensiveness of ITS designs may limit their cost-benefit justification, particular for under-funded IPE programs. Of notable design-integration value, the ‘training transfer’ focus of long-term effectiveness of IPE (e.g., patient impact) is, recently, being directly abridged via community-based placements and regular IPE assessment (Kent, Martin, & Keating, Citation2016).

Modern measurement—practical benefits over classical approaches

Having overviewed some common IPE study designs, the next section below will turn to address some IPE measurement tools. The vast majority of existing IPE measures (e.g., Readiness for Interprofessional Learning Scale [RIPLS], Interprofessional Education Perceptions Scale [IEPS]) use CTT to construct and validate scales (Oates & Davidson, Citation2015). Modern measurement theories, such as IRT, offer several practical advantages over IPE’s current, CTT paradigm. We summarize a few, select advantages with illustrative examples to help substantiate researchers’ interpretation for the IPE field.

First, IRT allows for equating different measures that have been mapped onto a common metric (For example, within each level of Kirkpatrick’s modified-IPE outcomes framework).Footnote1 Also, because IRT achieves greater reliability with fewer items, IPE studies may use a common subset of items (i.e., core items), which facilitates administration with shorter evaluations and improves student response rates with lower time-on-test (i.e., reduced response burden). Both of these features of IRT confer an advantage for strengthening IPE comparability across studies and sites that use different IPE measures.

Second, IRT enables item-level examination of measurement bias to strengthen IPE generalizability. Related to this more detailed, item-level scope of measurement bias, IRT allows for the use of IPE instruments beyond their original validation sample (e.g., student vs. practitioners). Removing this sample-dependent barrier to measurement selection gives greater access to researchers for optimizing content-appropriate IPE evaluations.

Third, in terms of understanding IPE change-scores, classical approaches severely limits interpretation, because well-documented baseline differences (e.g., Kerry, Heimberg, & Schmutz, Citation2017) of students and professions confounds meaningful assessments of change for learner outcomes. A salient example of CTT’s limitation for assessing change in IPE outcomes comes from Kashner and colleagues’ (Citation2017) ten-year study of medical students’ preferences for IPE, where findings indicated a normative-increase in interprofessional preferences over time. Under a classical paradigm, this normative change would confer a decrease in before-after studies of IPE effectiveness. In contrast, IRT’s item-level examination would allow for revealing the – seemingly – ‘decreased effectiveness’ of IPE effectiveness as attributable to an item-level ceiling effect from normative changes of expectancies from IP training—Simply restated, when there’s less room to go up, there’s more chance of going down (Whereas CTT would suggest less IPE effectiveness, however, IRT would indicate a ceiling-bias and need for measurement revision).

Finally, related to IRT’s flexibility over baseline differences for IPE change-scores, IRT is also more flexible in combining scores from multiple sources (e.g., self-report/observer-rated), multiple item-formats (dichotomous ‘yes/no’ or checklists/Likert-type scales/forced-choice), and multiple content domains (reactions/attitudes/knowledge/behavior). This latter flexibility-advantage is imperative for IPE researchers wishing to evaluate the broader impact of IPE’s heterogeneous competency framework with a composite, global index.

Concluding comments

Analytically, very few IPE researchers have studied the identification of student individual-differences that impact on, for example, IPE receptivity, motivation for transfer, or candidacy for student-facilitator training. Identifying student characteristics can help IPE researchers strengthen the short-term effectiveness of IPE by targeting delivery to ‘in-need’ learners (Kerry et al., Citation2017). Accounting for relevant student characteristics would also help IPE researchers disentangle IPE’s long-term effectiveness, over the complex training pathway, for distal care practice and patient outcomes. Summarizing, IPE’s quantitative evidence may be strengthened by experimental comparisons of relative-IPE effectiveness (active-controls), a shift from classical to modern measurement (CTT→IRT), and more substantive analytics (individual differences). Meeting the complexities of IPE programs with such quantitative best-practice methodologies should, ultimately, sustain IPE’s future development and impact.

Disclosure statement

The authors report no conflicts of interest. The authors alone are responsible for the content and writing of the article.

Notes

1. Kirkpatrick’s training evaluation framework, consisting of four criteria (trainee reactions, learning, transfer/behavior change, and organization results) was adapted for IPE by Barr and colleagues (Citation2005). These authors expanded the original model to six outcome levels: (1) learner reactions; (2a) attitude/perception modification; (2b) knowledge/skill acquisition; (3) behavioral change; (4a) organizational change; and (4b) improvement to patient health/well-being.

2. Both terms refer to concepts known to weaken internal validity of non-experimental research study designs. Interested readers can learn by the following instructional link, http://www.creative-wisdom.com/teaching/WBI/threat.shtml.

References

  • Barr, H., Koppel, I., Reeves, S., Hammick, M., & Freeth, D. (2005). Effective interprofessional education: Assumption, argument and evidence. Oxford, UK: Blackwell Publishing.
  • Brashers, V., Phillips, E., Malpass, J., & Owen, J. (2015). Review: Measuring the impact of interprofessional education (IPE) on collaborative practice and patient outcomes. In Institute of Medicine (Ed.), Measuring the impact of interprofessional education on collaborative practice and patient outcomes (pp. 67–133). Washington, DC: The National Academies Press.
  • Carpenter, J. (1995). Doctors and nurses: Stereotypes and stereotype change in interprofessional education. Journal of Interprofessional Care, 9, 151–161.
  • Costanza, M. E. (2015). Measuring the impact of interprofessional education on collaborative practice and patient outcomes. Journal of Interprofessional Education & Practice, 1(2), 34–35.
  • Herath, C., Zhou, Y., Gan, Y., Nakandawire, N., Gong, Y., & Lu, Z. (2017). A comparative study of interprofessional education in global health care: A systematic review. Medicine, 96(38), e7336.
  • Institute of Medicine. (2015). Measuring the impact of interprofessional education on collaborative practice and patient outcomes. Washington, DC: The National Academies Press.
  • Kashner, T. M., Hettler, D. L., Zeiss, R. A., Aron, D. C., Bernett, D. S., Brannen, J. L., … Gilman, S. C. (2017). Has interprofessional education changed learning preferences? A national perspective. Health Services Research, 52(1), 268–290.
  • Kent, F., Martin, N., & Keating, J. L. (2016). Interprofessional student-led clinics: An innovative approach to the support of older people in the community. Journal of Interprofessional Care, 30(1), 123–128.
  • Kerry, M. J., Heimberg, E., & Schmutz, J. B. (2017). Investigating predictive power of agency and communion for interprofessional learner outcomes: A before–After study. Journal of Interprofessional Care, 31(6), 789–792.
  • Oates, M., & Davidson, M. (2015). A critical appraisal of instruments to measure outcomes of interprofessional education. Medical Education, 49(4), 386–398.
  • Reeves, S. (2010). Ideas for the development of the interprofessional field. Journal of Interprofessional Care, 24(3), 217–219.
  • Reeves, S., Boet, S., Zierler, B., & Kitto, S. (2015). Interprofessional Education and Practice Guide No. 3: Evaluating interprofessional education. Journal of Interprofessional Care, 29(4), 305–312.
  • Reeves, S., Fletcher, S., Barr, H., Birch, I., Boet, S., Davies, N., … Kitto, S. (2016). A BEME systematic review of the effects of interprofessional education: BEME Guide No. 39. Medical Teacher, 38(7), 656–668.
  • Reeves, S., Palaganas, J., & Zierler, B. (2017). An updated synthesis of review evidence of interprofessional education. Journal of Allied Health, 46(1), 56–61.
  • Steven, K., Howden, S., Mires, G., Rowe, I., Lafferty, N., Arnold, A., & Strath, A. (2017). Toward interprofessional learning and education: Mapping common outcomes for prequalifying healthcare professional programs in the United Kingdom. Medical Teacher, 39(7), 720–744.
  • West, S. G., Duan, N., Pequegnat, W., Gaist, P., Des Jarlais, D. C., Holtgrave, D., … Mullen, P. D. (2008). Alternatives to the randomized controlled trial. American Journal of Public Health, 98(8), 1359–1366.

Reprints and Corporate Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

To request a reprint or corporate permissions for this article, please click on the relevant link below:

Academic Permissions

Please note: Selecting permissions does not provide access to the full text of the article, please see our help page How do I view content?

Obtain permissions instantly via Rightslink by clicking on the button below:

If you are unable to obtain permissions via Rightslink, please complete and submit this Permissions form. For more information, please visit our Permissions help page.