4,010
Views
10
CrossRef citations to date
0
Altmetric
Trend Article

A developmental assessment of clinical reasoning in preclinical medical education

, , &
Article: 1591257 | Received 10 Dec 2018, Accepted 28 Feb 2019, Published online: 01 Apr 2019

ABSTRACT

Background: Clinical reasoning is an essential skill to be learned during medical education. A developmental framework for the assessment and measurement of this skill has not yet been described in the literature.

Objective: The authors describe the creation and pilot implementation of a rubric designed to assess the development of clinical reasoning skills in pre-clinical medical education.

Design: The multi-disciplinary course team used Backwards Design to develop course goals, objectives, and assessment for a new Clinical Reasoning Course. The team focused on behaviors that students were expected to demonstrate, identifying each as a ‘desired result’ element and aligning these with three levels of performance: emerging, acquiring, and mastering.

Results: The first draft of the rubric was reviewed and piloted by faculty using sample student entries; this provided feedback on ease of use and appropriateness. After the first semester, the course team evaluated whether the rubric distinguished between different levels of student performance in each competency. A systematic approach based on descriptive analysis of mid- and end of semester assessments of student performance revealed that from mid- to end-of-semester, over half the students received higher competency scores at semester end.

Conclusion: The assessment rubric allowed students in the early stages of clinical reasoning development to understand their trajectory and provided faculty a framework from which to give meaningful feedback. The multi-disciplinary background of the course team supported a systematic and robust course and assessment design process. The authors strongly encourage other colleges to support the use of collaborative and multi-disciplinary course teams.

Introduction

Clinical reasoning is a necessary skill for physicians, and teaching this skill is an important aspect of medical education. Integrating the instruction of clinical reasoning and assessment of this skill has proven to be challenging and variable, particularly within the pre-clinical phase of undergraduate medical education. While several strategies to assess clinical reasoning of learners have been utilized, a developmental approach to measuring and assessing the progression of these skills has not yet been described in medical education literature.

There is no single ‘gold standard’ for assessing clinical reasoning. Two approaches, patient management problems and computer-based exam projects, ask students to indicate what clinical examinations and diagnostic tests they would perform to narrow the differential diagnosis [Citation1]. These approaches revealed challenges and errors, including inter-rater reliability, variability in the predictability of performance scores in terms of tracking long-term development of clinical reasoning, and the observation that experienced physicians did not perform as well as medical students on some of the same cases or decision-making items [Citation1]. Other strategies shared in the literature include script concordance and Think Aloud protocols [Citation2Citation4]. The script concordance test is based on the understanding that experienced clinicians have sets of knowledge or scripts, that are used to understand situations and make decisions related to diagnosis and treatment [Citation2]. Learners are provided with a short clinical vignette and asked how an additional piece of new information affects their decisions regarding diagnosis, investigational study, or therapy [Citation2,Citation3]. These responses are then compared to responses by a panel of experts [Citation2,Citation3]. The Think Aloud method calls for individuals to verbalize how they are using the information to generate a solution to a problem [Citation4]. It has been suggested that if learners, particularly in graduate medical education, pause intermittently during clinical presentations to explicitly explain how information is affecting their decision-making, supervisors could assess their clinical reasoning and address errors in real time [Citation4].

Assessing clinical reasoning skills is difficult because it is not one well-defined skill, but a combination of skills that involves understanding the complexity of a problem and the ability to translate and synthesize existing knowledge and skills to new problems. Assessing variation in the development of this skill set poses challenges. For example, some students might arrive at a diagnosis or plan but cannot articulate how they did, while others might arrive at the correct answer and articulate an approach other than one typically taken. Any assessment process, then, should address both a medical student’s current performance of clinical reasoning and their learning progression toward developing this skill set that calls on the simultaneous exercise of medical knowledge, communication and physical examination skills, and critical and reflective thinking. In other words, assessments should be both context-specific and global – allowing for an assessment of performance over time [Citation1]. In discussing future directions in clinical reasoning assessment, Lang et al. described a milestones approach, a developmental assessment process, as a way to capture both diagnostic reasoning (evaluating decision-making points) and therapeutic reasoning that includes ethical and patient-specific information [Citation1].

In 2014, the University of Arizona College of Medicine – Tucson created a Clinical Reasoning course (CRC) that spans the first three semesters of the pre-clerkship curriculum. CRC evolved from Case-Based Instruction (CBI), a case-based activity that was integrated into each preclinical block course and had the primary goal of expanding students’ medical knowledge. CRC was created as a separate, longitudinal, small-group and case-based course that had the primary goal of developing students’ clinical reasoning skills. The change in course goals was paired with the adoption of an online platform called ThinkShareTM, where students explained in writing their thinking about the cases before coming to their small-group sessions. Students progressed through each case in a stepwise process, with opportunities to respond to new clinical or patient information disclosed at progressive junctures.

A challenge for the new course was to devise a system to assess student performance. We describe the development and implementation of a rubric designed to assess the development of clinical reasoning skills in pre-clinical medical education. This rubric reflects the developmental process and ‘milestones’ expected as a learner progresses through a clinical reasoning curriculum.

Materials and methods

The CRC team implemented a Backward Design process to create the new course goals, learning objectives, and student assessments [Citation5]. An important aspect of the CRC course design process was that the team included individuals from multiple disciplines with advanced degrees in the basic sciences, clinical medicine, education and public health and with expertise in instructional design, assessment, and course/curriculum evaluation. The course goals were developed by reviewing a crosswalk of the existing goals of CBI, the program level educational competencies, and the first level of the Accreditation Council for Graduate Medical Education (ACGME) residency milestones related to critical thinking skills of four specialties, Medicine, Family & Community Medicine, Pediatrics, and Obstetrics and Gynecology.

The crosswalk process enabled the team to identify similar expectations between the original CBI modules, the goals of the undergraduate medical education program and the expectations for performance in residency. Once the course goals were created, the team then identified the ‘acceptable evidence’ associated with the course goals by asking the questions ‘How will I know if students have achieved the desired results?’ and ‘What will I accept as evidence of student understanding and proficiency?’ The discussion of desired results and evidence for those results prompted the course team to discuss the possibility of improving the connection between the formative feedback and assessment in the CRC to the ACGME Milestones approach used in residency programs. The course team also relied significantly on work implemented at the University of New Mexico [Citation6].

The ACGME identified six core competencies for medical education: practice-based learning and improvement, patient care and procedural skills, system-based practice, medical knowledge, interpersonal and communication skills, and professionalism [Citation7]. Because CRC provided the best opportunity in our pre-clinical curriculum for faculty to observe the performance of individual students and provide feedback, CRC was used to assess the competencies of Medical Knowledge (in this case, clinical reasoning), Interpersonal and Communication Skills, Practice-based Learning and Improvement, and Professionalism.

The ACGME residency milestones were implemented in 2013. The milestones created a developmental framework for the assessment of professional skill development in residency training and provided more meaningful assessment and outcome measures. Using these milestones as a reference, the CRC team aimed to create a student performance assessment rubric that was developmental, clearly described the trajectory of performance, skills and behaviors we expected students to develop throughout the three semesters of the course, and included the future target for performance (connecting each behavior to related expectations in clerkship).

The course team focused on behaviors we expected students to be able to demonstrate, identifying each as a ‘desired result’ element and aligning these with four levels of performance: pre-emergent, emerging, acquiring, and mastering. Each level was defined by a descriptive anchor. Students received a mid-semester and an end-semester assessment (weighted 20% and 80% toward their final grade). As the students proceeded through the three semesters of the course, the threshold assessment rating necessary to pass the course increased.

The first iteration of the assessment rubric included 12 items in four competencies (). The full version of the assessment rubric with the descriptive anchors is provided in Appendix 1.

Table 1. First version of assessment rubric items.

Once the draft rubric was created the course team implemented an initial review process. The course team provided five faculty facilitators with written examples of student reasoning that demonstrated different performance levels. The team used multiple student case entries within ThinkShare from three students over a year of CBI. All case entries were from students who took the course in prior years and were blinded to the reviewer. The team used qualitative analysis to identify themes in post-case reflections that demonstrated students’ engagement in a range of analytical, evaluative and other forms of higher order thinking and metacognitive engagement as defined by Bloom’s Taxonomy Revised and Shraw and Dennison [Citation8]. Codes were derived to define each theme. Using Atlas.ti, we analyzed entries from one year of students and then selected eight post-case reflections for each of two cases that demonstrated the three levels of performance described in the new pilot rubric (emerging, acquiring, and mastering).

The CRC team asked the faculty reviewers to assess these student entries using the new assessment rubric. The faculty scored the entries, and provided feedback on timing, ease of use, and appropriateness. Their feedback was used by the CRC team to make adjustments to the rubric. The CRC team reviewed the assessment ratings between each reviewer and relative to the prior assessment of performance (which was blinded to the pilot faculty reviewers) and subsequently revised the assessment rubric based on faculty reviewer feedback.

The assessment rubric was used for the first time in the fall of 2015 when the course was implemented in its longitudinal format. Faculty facilitators for the CRC course were trained to use the assessment rubric by the Director of Faculty Development and the Manager for Student Assessment, including applying the rubric to the sample student entries from the pilot. The faculty facilitators for each student group assessed student performance at mid- and end of the semester using the new assessment rubric.

In addition to reviewing the descriptive data, the course team continued the iterative improvement process that began during the first year of implementation to gather faculty facilitator feedback beginning with a focus group. Based on feedback from the focus group, the CRC team members paired with two faculty facilitators to make additional revisions to the items and behavioral anchors in each competency area. This process led to a second version of the assessment rubric with nine items in four competencies (Appendix 2). This version of the rubric was then used for the assessment of students at every mid-semester and end of semester evaluation from that point forward.

Ethical approval was waived by the University of Arizona Institutional Review Board.

Results

After the first semester implementation of the assessment rubric, the CRC team identified a key question to evaluate the rubric: did the assessment rubric distinguish between different levels of student performance in each competency? In order to answer this question, the team implemented a systematic approach based on a descriptive analysis of the mid- and end of semester assessment ratings of student performance. Since the assessment was implemented with first-year students in their first semester of medical school, the team expected that ratings would increase from the mid- to the end of the semester. and present our descriptive data from the mid-and end of semester assessments for the 116 students enrolled in the course. The CRC team reviewed the descriptive data and interpreted the results as promising. At mid-semester assessment, the distribution of the assessment ratings was evenly split between 1 = emerging and 2 = acquiring for most individual items with very few ratings of 0 = pre-emergent or 3 = Mastery. At the end of the semester, this distribution shifted towards more students receiving ratings of 2 = acquiring for most of the items. This is reflected in the change in median scores for each competency. presents the results of a matched analysis using a Wilcoxon-signed rank test. The CRC team used this data to understand the trend for individual students matched scores. This data shows that from mid- to end-of-semester, over half the students (ranging from 53% to 62%) received a higher competency score at the end-of-the semester.

Table 2. Descriptive statistics mid- and end-semester assessment.

Table 3. Number of students with decrease, tied, or increased total rating score by competency.

Discussion

The CRC assessment rubric has undergone revisions over the first three years of the course based on feedback from students and faculty. Continued work to describe the developmental trajectory of clinical reasoning skills for pre-clinical medical students is necessary and will inform future iterations of this assessment. We believe that this rubric has allowed students who are in the early stages of their clinical reasoning development to visualize their trajectory and has provided faculty facilitators a framework from which to provide meaningful and concrete feedback. The expected development of these skills in early learners is well described by our assessment rubric.

CRC continues to be an integral component of our medical school curriculum though the leadership of the course has changed. The new course directors have continued to build on the developmental approach integrated in the rubric. We do recognize the need for further testing on the validity and reliability of our rubric. Structural and organizational limitations in managing a small group course, facilitated by faculty, for an entire cohort of students presented a logistical challenge to implementing additional reliability testing. The course did implement an online Health Sciences Reasoning Test (HSRT) with the first cohort of students with the plan to analyze the possible relationship between the rubric and student results on the HSRT. However, this analysis was not able to be conducted due to organizational constraints. These limitations highlight the importance of dedicating enough time and focus to the follow up reliability and validity analysis of any new tool [Citation9].

Another limitation relates to our findings of the mid- and end of semester ratings. It is possible that the differences in ratings over time could be due to several factors including changes in rater expectations or differences in rater training. In addition, as the weighting of mid-semester assessment was worth 20% of their final grade while the end of semester was worth 80%, students may have put forth more effort into their work at the end of the semester.

Future work and applications for our clinical reasoning rubric could be focused on use during clerkship rotations. We did include ‘Clerkship Level’ goals for each behavior which could be used to modify the rubric with higher level expectations for learners on their clinical rotations.

The course design process and the development and pilot process for the assessment rubric would not have been possible without the commitment of the multi-disciplinary CRC team members. Our experience emphasizes the need for Colleges of Medicine to provide structural support to the course design and implementation process. This includes valuing the time and effort dedicated to course design by individuals from multiple disciplines with advanced degrees in the basic sciences, clinical medicine, education and public health and with expertise in instructional design, assessment, and course/curriculum evaluation.

Acknowledgments

The authors wish to thank the members of the CRC team, past and present, with special thanks to John Bloom, MD; Susan Ellis, MA, EdS; Amber Hansen, MS, RD; Jennifer Cogan; Herman Gordon, PhD; David Bear, PhD.

Disclosure statement

No potential conflict of interest was reported by the authors.

References

  • Lang VJ, Schuwirth L, Durning SJ, et al. Assessment of clinical reasoning. In: Trowbridge RL, Rencic JJ, Durning SJ, editors. Teaching clinical reasoning, ACP teaching medicine series. Philadelphia (PA): Americal College of Physicians: VersaPress; 2015. pp. 117–154.
  • Charlin B, Roy L, Brailovsky C, et al. The script concordance test: a tool to assess the reflective clinician. Teach Learn Med. 2000;12(4):189–9.
  • Humbert AJ, Besinger B, Miech EJ. Assessing clinical reasoning skills in scenarios of uncertainty: convergent validity for a script concordance test in an emergency medicine clerkship and residency. Acad Emerg Med. 2011;18:627–634.
  • Pinnock R, Young L, Spence F, et al. Can think aloud be used to teach and assess clinical reasoning in graduate medical education? J Grad Med Educ. 2015;7(3):334–337.
  • Wiggins GP, McTighe J. Understanding by design, expanded. 2nd ed. Danvers (MA): Association for Supervision and Curriculum Development; 2005.
  • Mitchell S, Richter D, Eldredge J, et al. Providing feedback that matters: a model for evaluating students in a developmental problembased learning curriculum. Association of American Medical Colleges. Western Group on Educational Affairs (AAMC/WGEA) Conference. Asilomar, CA. March 30-April 3, 2012.
  • NEJM Knowledge+. Exploring the ACGME core competencies. [updated Jun 2016; cited 2018 Jul 15]. Available from: https://knowledgeplus.nejm.org/blog/exploring-acgme-core-competencies/
  • Shraw G, Dennison RS. Assessing metacognitive awareness. Contemp Educ Psychol. 1994;19:460–475.
  • Health Science Reasoning Test (HSRT). 2019. Available from: https://www.insightassessment.com/Products/Products-Summary/Critical-Thinking-Skills-Tests/Health-Sciences-Reasoning-Test-HSRT

Appendix 1

Milestones of Observable Behaviors for Clinical Reasoning Course

Appendix 2

☐ YES ☐ NO This student submitted ThinkShare entries for all steps of all cases.

☐ YES ☐ NO This student met with the facilitator for individual, in-person formative feedback.

Please provide written comments about this student’s performance during this assessment period: