119
Views
2
CrossRef citations to date
0
Altmetric
Perspectives

Applying a reflexive framework to evaluate a communication skills curriculum

Pages 587-592 | Published online: 14 Oct 2016

Abstract

After creating and delivering an educational curriculum, medical educators must ultimately evaluate the effectiveness of the implemented curriculum. Seasoned educators can benefit from using an established framework to help them structure a thorough, complete curricular evaluation; however, novice educators may have difficulty in transforming the concept of evaluation into a concrete process. The RUFDATA (Reasons and purpose, Uses, Focus, Data and evidence, Audience, Timing, and Agency) framework is one such paradigm. It is a well-recognized tool consisting of a reflexive framework that can guide medical educators to evaluate their own medical education curriculum. Just as important, it enables medical educators to reflect on the reasons behind the evaluation. This insight, in turn, can foster a spirit of evaluation, thus helping to ingrain it into the local educational culture. By using the evaluation of our communication skills curriculum as an example, this article describes how educators can apply the RUFDATA framework to evaluate their own curriculum.

Introduction

Once medical educators have developed and implemented a curriculum for their learners, they must evaluate the effectiveness of that curriculum to ensure if it is fulfilling its intended purpose.Citation1,Citation2 Here, we define curriculum evaluation as gathering information or data about an educational program to judge its merit.Citation3

However, a proper curricular evaluation can be difficult to conduct for faculty who lack experience or tools to methodically plan the evaluation.Citation4Citation6 A reflexive framework known as RUFDATA (Reasons and purpose, Uses, Focus, Data and evidence, Audience, Timing, and Agency) is a tool that can transform the concept of evaluation into a concrete process.Citation7

This article shows how medical educators can use the RUFDATA framework to plan and conduct an evaluation of their own curriculum. After providing a brief account of our communication skills curriculum, this article gives an overview of the RUFDATA framework and describes how to employ each component of the framework to conduct curricular evaluation, using the evaluation of our communication skills curriculum as an example.

Background

Teaching communication skills is an essential component of the curriculum in residency education.Citation8,Citation9 For our 2-year subspecialty residency program in respirology, we first performed an analysis with past cohorts of residents about their needs. Then, we developed and implemented a communication skills curriculum after reviewing the methods to teach communication skills in the literature. Our curriculum thus included several small group seminars that incorporated active learning,Citation10Citation13 reflective essay assignments,Citation14 and simulation exercisesCitation15Citation17 and covered topics such as writing proper consultation letters, communicating with patients and their families, collaborating with allied health professionals, and giving effective presentations.

RUFDATA framework

The RUFDATA framework by SaundersCitation7 asks 7 questions that educators should ponder when planning their curricular evaluation. Reflexive questioningCitation18 (ie, by asking open-ended questions to help the educator reflect on many possible options) helps start the evaluation process by addressing key aspects of the evaluation’s design. In other words, it helps educators decide what they want to get out of the evaluation and thus to plan and execute it accordingly.

Reasons

First, educators should ask themselves the reasons for evaluating the curriculum as the answer(s) to this question will be the answers to the rest of the questions in the RUFDATA framework. Broadly speaking, a curriculum is evaluated to demonstrate accountability to various stakeholders, to develop further knowledge about the curricular topic, and to improve the curriculum itself.Citation19

We specifically wanted to evaluate our communication skills curriculum to ensure that our residency program was meeting accreditation standards in this specific area, the first small step in evaluating all aspects of our residency curriculum. We also wanted to ensure that our residents were learning the requisite skills to work with allied health professionals and were communicating effectively with patients and their families. In addition, we sought to determine whether there was any effect on the residents’ communication skills during their teaching presentations and whether there were any unplanned learning experiences occurring, such as “curriculum drift”Citation20 and a “hidden curriculum”,Citation21Citation25 which failed to match the expectations of our planned curriculum.

Uses

After reflecting on the reasons behind the evaluation, educators should ask how the information gathered will be used. In other words, what action can and/or will be taken once the data are collected and the results of the evaluation are known? How the information will be used, in turn, informs the focus of the evaluation and the data one collects.

We planned to use the information for ongoing quality improvement of our program ie, to conduct formative evaluation.Citation26 The information gained from our evaluation would enable us to efficiently target areas for improvement in our teaching and assessment of communication skills for the residents. We also wanted to incorporate the evaluation results into our organizational databaseCitation27 to allow future residency program directors to see what worked (and did not work) for us. If insufficient resources such as money, time, and personnel were found to be the cause of any deficiencies in our curriculum, we planned to use our evidence to lobby for investment in these resources. Finally, we sought to demonstrate our compliance in this domain for accreditation.

Focus

As it is difficult to thoroughly evaluate an entire teaching curriculum at once, educators should focus their evaluation on key aspects of the curriculum. The focus, in turn, influences the data that will be collected.

We selected to focus on the adequacy of the residents’ consultation letters, their written and verbal communication with allied health professionals, their verbal communication with patients and families, and their presentation skills during visits.

There are many other areas where a communication skills curriculum can be important, such as teaching proper patient care handover,Citation28Citation30 discharge summary dictation,Citation31Citation33 and crisis resource management.Citation34Citation36 However, we chose to evaluate these other aspects of our curriculum in the future as we did not yet have the tools to assess the residents’ performance in these areas.

Data and evidence

After reflecting on the reasons, uses, and focus of the evaluation, educators must select the tools that they will use to collect the data and evidence necessary for the evaluation. This usually involves measuring how well the learners have achieved the learning objectivesCitation37 or measuring both the unintended and intended accomplishments of the learners.Citation38 Some even focus on the learners’ pursuit of accomplishment, rather than the actual accomplishment itself.Citation39 Typically, the data collection tools assess the achievements of the learners.Citation40 These assessments include observation of behavior in the clinical workplace,Citation41Citation44 observation of behavior during an objective structured clinical examination (OSCE) or simulation exercise,Citation45Citation51 and chart review or audit.Citation52Citation57 In addition, feedback can also be gathered from the learners and/or new graduates, through interviews or questionnaires, to determine if the teaching program has met their perceived needs.Citation58Citation60

Our data consisted of random audits of the residents’ consultation letters, multisource assessment of the residents’ communication skills by allied health professionals (specifically, the nurses and ward unit clerk), and faculty’s assessment of the residents’ communication skills during presentations. We chose these tools because we were already using them before implementing our new curriculum and had baseline results with which we could compare. We considered using feedback from patients and their families but had not yet implemented this form of assessment during our evaluation. Instead, we used the faculty’s direct observations of the residents’ communication with patients and families, realizing the potential limitations of this approach.Citation61 Additionally, we conducted exit interviews with the residents to look for any hidden curriculum or unintended consequences from our formal curriculum.

Audience

Educators should decide who the audience of the evaluation will be. In other words, who will see the results of the evaluation? The answer, in turn, depends on the level of decision-making informed by the evaluation.Citation62 Medical educators overseeing the teaching program can use the evaluation to determine the need for improvement. They may wish to share this information with their peers who can then enhance their own teaching programs. Accreditation organizations may use the information as proof of compliance for this facet of the whole academic program. And the evaluation results, if positive, can bolster the teaching faculty’s morale and validate their teaching efforts.

We chose to share our evaluation results with our residency program committee tasked with implementing and revising the curriculum. In addition, we decided to keep the results as part of our records for accreditation and also share them with our teaching faculty.

Timing

The timing of the evaluation is another decision that educators must contemplate. For example, we conducted our evaluation 6 months before the start of the upcoming academic year. This enabled us to ensure that we had collected and analyzed all the data and that we would have sufficient time to contemplate and implement the necessary changes at the beginning of the academic year involving the next cohort of learners.

Agency

Here, educators must consider who will conduct the evaluation. Both internal evaluators (ie, people already familiar with the program or who may have helped develop it) and external evaluators have their own advantages and disadvantages. For example, an internal evaluator acquainted with the intricacies of the program and local quirks of the educational environment may probe deeper to look for weaknesses or may be more available to conduct the evaluation. On the other hand, an external evaluator might possess greater objectivity or garner more legitimacy.Citation63 Potential disadvantages of an external evaluator include a greater cost (eg, for transportation and lodging) and difficulty in coordinating the schedule for evaluation.

After careful consideration, the program director of our residency training program, one of the developers of the curriculum, conducted the evaluation.

Evaluation results

We analyzed the data during the 12 months before and after the implementation of our communication skills curriculum.

For the evaluation of the residents’ consultation letters, we used a 9-item, 5-point scoring scale (for a maximum score of 45) developed by Keely et alCitation54 to analyze features such as content, style, organization, and educational value. We randomly selected 3 consultation letters written by each resident. Prior to implementation of our curriculum, the mean score was 32.8 with a SD of 4.4, whereas after implementation, the mean score improved to 40 with a SD of 2.9.

For the multisource assessment of the residents’ communication skills by allied health care professionals, we asked the nurses and pulmonary ward unit clerk to rate 4 aspects of the residents’ verbal and written communication as either unsatisfactory or satisfactory. Virtually, all residents achieved a satisfactory rating before and after implementation of our curriculum. On reflection, our rating scale had only 2 options – that is, satisfactory and unsatisfactory. Thus, we were likely unable to detect nuanced differences in performance. There did not appear to be any significant, unintended deterioration in the residents’ behavior with the new curriculum (within the limits of our evaluation).

Using direct observation, faculty rated the residents’ communication with patients and their families using a 5-point scale from 1 (poor) to 5 (excellent). Among 24 encounters observed before implementing our curriculum, 6 encounters received a score of 3 (adequate), 13 received a score of 4 (good), and 5 received a score of 5, with no encounters receiving a score of 1 or 2 (borderline). Among 20 encounters observed after implementing our curriculum, there seemed to be overall improvement with 2 encounters receiving a score of 3, 8 receiving a score of 4, and 10 receiving a score of 5.

Faculty also rated the residents’ communication skills during presentations using a 3-point scale (1, poor; 2, adequate; and 3, excellent). Out of 12 presentations given before implementing our curriculum, 1 received a score of poor, 7 received a score of adequate, and 4 received a score of excellent. Overall, scores improved after implementation where, among 10 presentations, 4 received a score of adequate and 6 received a score of excellent, with none receiving a score of poor.

During individual exit interviews, residents uniformly felt that the curriculum improved their written and verbal communication skills and felt that the faculty modeled the curriculum. We could find no instances where the formal curriculum conflicted with the learning environment.

Overall, we felt that we were meeting the accreditation standards, our residents were acquiring the communication skills that were the focus of our evaluation, and that there were no unintended consequences of our curriculum.

Conclusion

The RUFDATA framework is a simple tool that we used to evaluate our teaching program. It can be widely applied to evaluate other areas in education. For example, it has been used to evaluate other focused medical topics (such as a training module to teach intrapartum careCitation64) and even entire undergraduate university courses.Citation65 In the future, we plan to evaluate other aspects of our communication curriculum such as crisis resource management, discharge summary dictations, and resident-to-resident patient care handover. We will also refine our tool that assesses the residents’ communication with allied health professionals and explores options to involve patients and their families in the assessment.Citation66,Citation67

Although there are other frameworks that can be used to evaluate curricula, such Stufflebeam’s CIPPCitation68 and Kirkpatrick’s Four-Level model,Citation69 the RUFDATA framework’s reflexive nature helps foster a culture of curricular evaluation by enhancing reflection on the reasons for the evaluation. Medical educators can easily apply this framework to evaluate their own curriculum.

Disclosure

The author reports no conflicts of interest in this work.

References

  • GoldieJAMEE education guide no. 29: evaluating educational programmesMed Teach200628321022416753718
  • DiFlorioIDuncanPMartinBMiddlemissMACurriculum evaluationNurse Educ Today1989964024072601681
  • ColesCRGrantJGCurriculum evaluation in medical and health-care educationMed Educ19851954054224058341
  • OuelletLLRushKLForces influencing curriculum evaluationNurse Educ Today1989942192262779498
  • HuwendiekSMenninSDernPExpertise, needs and challenges of medical educators: results of an international web surveyMed Teach2010321191291821039102
  • ParitakulPWongwandeeMTantitemitTPumipichetSDennickRLevel of confidence in the 12 roles of medical teacher. A descriptive study at Faculty of Medicine, Srinakharinwirot University, ThailandJ Med Assoc Thai201598suppl 10S38S44
  • SaundersMBeginning an evaluation with RUFDATA: theorizing a practical approach to evaluation planningEvaluation200061721
  • FrankJRSnellLSherbinoJCanMEDS 2015 Physician Competency Framework2015 Available from: http://canmeds.royalcollege.ca/uploads/en/framework/CanMEDS%202015%20Framework_EN_Reduced.pdfAccessed May 26, 2016
  • ACGMEAccreditation Council for Graduate Medical Education Common Program Requirements2016 Available from: http://www.acgme.org/Portals/0/PFAssets/ProgramRequirements/CPRs_07012016.pdfAccessed May 4, 2016
  • BarringtonKCampbellBMUDD mapping: an interactive teaching-learning strategyNurse Educ200833415916318600156
  • AzerSAGuerreroAPWalshAEnhancing learning approaches: practical tips for students and teachersMed Teach201335643344323496121
  • KurtzSMSilvermanJDraperJTeaching and Learning Communication Skills in Medicine2nd edOxfordRadcliffe Pub2005
  • KnoxJDEBouchierIADCommunication skills teaching, learning and assessmentMed Educ19851942852894021854
  • StewartTLMyersACCulleyMREnhanced learning and retention through “writing to learn” in the psychology classroomTeach Psychol20093714649
  • BagnascoAPagnucciNTolottiARosaFTorreGSassoLThe role of simulation in developing communication and gestural skills in medical studentsBMC Med Educ20141411724387322
  • LewisRStrachanASmithMMIs high fidelity simulation the most effective method for the development of non-technical skills in nursing? A review of the current evidenceOpen Nurs J20126828922893783
  • Barry IssenbergSMcGaghieWCPetrusaERLee GordonDScaleseRJFeatures and uses of high-fidelity medical simulations that lead to effective learning: a BEME systematic reviewMed Teach2005271102816147767
  • BergerWThe power of inquiryA More Beautiful Question: The Power of Inquiry to Spark Breakthrough Ideas1st edBergerWNew York, NYBloomsbury USA20142048
  • ChelimskyEIntroductionChelimskyEShadishWREvaluation for the 21st Century: A HandbookThousand Oaks, CASAGE Publications199710
  • WoodsAExploring unplanned curriculum driftJ Nurs Educ2015541164164426517076
  • PaulDEwenSCJonesRCultural competence in medical education: aligning the formal, informal and hidden curriculaAdv Health Sci Educ Theory Pract201419575175824515602
  • HaflerJPOwnbyARThompsonBMDecoding the learning environment of medical education: a hidden curriculum perspective for faculty developmentAcad Med201186444044421346498
  • NyquistJGThe hidden curriculum in health professional educationJ Chiropr Educ2016301484926730504
  • HaffertyFWBeyond curriculum reform: confronting medicine’s hidden curriculumAcad Med19987344034079580717
  • HaffertyFWGaufbergEHO’DonnellJFThe role of the hidden curriculum in “on doctoring” coursesAMA J Ethics201517213013925676226
  • ScrivenMThe methodology of evaluationTylerRWGagneRMScrivensMPerspectives of Curriculum Evaluation1st edChicago, ILRand McNally19673983
  • FiolCMLylesMAOrganizational learningAcad Manage Rev1985104803813
  • PascoeHGillSDHughesAMcCall-WhiteMClinical handover: an audit from AustraliaAustralas Med J20147936337125324901
  • WayneJDTyagiRReinhardtGSimple standardized patient handoff system that increases accuracy and completenessJ Surg Educ200865647648519059181
  • GakharBSpencerALUsing direct observation, formal evaluation, and an interactive curriculum to improve the sign-out practices of internal medicine internsAcad Med20108571182118820375830
  • MyersJSJaipaulCKKoganJRKrekunSBelliniLMSheaJAAre discharge summaries teachable? The effects of a discharge summary curriculum on the quality of discharge summaries in an internal medicine residency programAcad Med20068110 supplS5S8
  • CarnahanJLFletcherKEDischarge education for residents: a study of trainee preparedness for hospital dischargeWMJ2015114518518926726338
  • OttoMSterlingMSieglerEEissBAssessing origins of quality gaps in discharge summaries: a survey of resident physician attitudesJ Biomed Educ2015201534175926523277
  • VolkMSWardJIriasNNavedoAPollartJWeinstockPHUsing medical simulation to teach crisis resource management and decision-making skills to otolaryngology housestaffOtolaryngol Head Neck Surg20111451354221493304
  • HolzmanRSCooperJBGabaDMPhilipJHSmallSDFeinsteinDAnesthesia crisis resource management: real-life simulation training in operating room crisesJ Clin Anesth1995786756878747567
  • RudySJPolomanoRMurrayWBHenryJMarineRTeam management training using crisis resource management results in perceived benefits by healthcare workersJ Contin Educ Nurs200738521922617907666
  • TylerRWHow can the effectiveness of learning experiences be evaluated?TylerRWHlebowitshPSBasic Principles of Curriculum and InstructionLondonThe University of Chicago Press Ltd2013117138
  • ScrivenMThe nature of evaluationNew Dir Program Eval1993199358548
  • CaulleyDNGrotelueschenADThe illusions of learner accomplishmentEduc Leadership1978354280
  • KellyAVAssessment, evaluation, appraisal and accountabilityThe Curriculum: Theory and Practice6th edKellyAVThousand Oaks, CASAGE Publications, Inc2009147186
  • Assis-HassidSHeartTReychavIPliskinJSReisSExisting instruments for assessing physician communication skills: are they valid in a computerized setting?Patient Educ Couns201393336336623623463
  • MakoulGThe SEGUE framework for teaching and assessing communication skillsPatient Educ Couns2001451233411602365
  • GustinJLWayDPWells-Di GregorioSMcCallisterJWValidation of the family meeting behavioral skills checklist: an instrument to assess fellows’ communication skillsAnn Am Thorac Soc20161381388139327249513
  • SkillingsJLPorcerelliJHMarkovaTContextualizing SEGUE: evaluating residents’ communication skills within the framework of a structured medical interviewJ Grad Med Educ20102110210721975894
  • PetersonEBCalhounAWRiderEAThe reliability of a modified Kalamazoo Consensus Statement Checklist for assessing the communication skills of multidisciplinary clinicians in the simulated environmentPatient Educ Couns201496341141825103180
  • YuasaMNagoshiMOshiro-WongCTinMWenAMasakiKStandardized patient and standardized interdisciplinary team meeting: validation of a new performance-based assessment toolJ Am Geriatr Soc201462117117424383978
  • SchmitzCCChipmanJGLuxenbergMGBeilmanGJProfessionalism and communication in the intensive care unit: reliability and validity of a simulated family conferenceSimul Healthc20083422423819088667
  • Ponton-CarssAHutchisonCViolatoCAssessment of communication, professionalism, and surgical skills in an objective structured performance-related examination (OSPRE): a psychometric studyAm J Surg2011202443344021861980
  • CohenDSColliverJARobbsRSSwartzMHA large-scale study of the reliabilities of checklist scores and ratings of interpersonal and communication skills evaluated on a standardized-patient examinationAdv Health Sci Educ Theory Pract19961320921324179020
  • CohenDSColliverJAMarcyMSFriedEDSwartzMHPsychometric properties of a standardized-patient checklist and rating-scale form used to assess interpersonal and communication skillsAcad Med1996711 supplS87S898546794
  • ComertMZillJMChristalleEDirmaierJHarterMSchollIAssessing communication skills of medical students in objective structured clinical examinations (OSCE) – a systematic review of rating scalesPLoS One2016113e015271727031506
  • TomlinsonPMcAndrewRAn audit of the standard of response letters sent by hospital specialists to general dental practitioners following a referred patient’s first appointmentAnn R Coll Surg Engl200587425125416053683
  • FoxATPalmerRDCrossleyJGSekaranDTrewavasESDaviesHAImproving the quality of outpatient clinic letters using the Sheffield Assessment Instrument for Letters (SAIL)Med Educ200438885285815271046
  • KeelyEMyersKDojeijiSCampbellCPeer assessment of outpatient consultation letters – feasibility and satisfactionBMC Med Educ200771317519024
  • CrossleyGMHoweANewbleDJollyBDaviesHASheffield Assessment Instrument for Letters (SAIL): performance assessment using outpatient lettersMed Educ200135121115112411895235
  • McCainGAMolineuxJEPedersonLStuartRKConsultation letters as a method for assessing in-training performance in a department of medicineEval Health Prof1988111214210286763
  • JackCHazelEBernatskySSomething’s missing here: a look at the quality of rheumatology referral lettersRheumatol Int20123241083108521340567
  • DarlowBDonovanSColemanKWhat makes an interprofessional education programme meaningful to students? Findings from focus group interviews with students based in New ZealandJ Interprof Care201630335536127152540
  • HenziDDavisEJasineviciusRHendricsonWIn the students’ own words: what are the strengths and weaknesses of the dental school curriculum?J Dent Educ200771563264517493972
  • CaveJGoldacreMLambertTWoolfKJonesADacreJNewly qualified doctors’ views about whether their medical school had trained them well: questionnaire surveysBMC Med Educ200773817945007
  • LeungKKWangWDChenYYMulti-source evaluation of interpersonal and communication skills of family medicine residentsAdv Health Sci Educ Theory Pract201217571772622240920
  • GephartWJWho will engage in curriculum evaluation?Educ Leadership1978354255
  • Conley-TylerMA fundamental choice: internal or external evaluation?Eval J Australas20054311
  • MukhopadhyaySSmithSCurriculum evaluation from the trainees’ perspective: application to the ALWP ATSMJ Obstet Gynaecol201030879579921126115
  • ShermanPDUsing RUFDATA to guide a logic model for a quality assurance process in an undergraduate university programEval Program Plann20165511211926788815
  • MyerholtzLAssessing family medicine residents’ communication skills from the patient’s perspective: evaluating the communication assessment toolJ Grad Med Educ20146349550026279775
  • StausmireJMCashenCPMyerholtzLBudererNMeasuring general surgery residents’ communication skills from the patient’s perspective using the communication assessment tool (CAT)J Surg Educ201572110811625139607
  • StufflebeamDLDaniel Stufflebeam’s CIPP model for evaluation: an improvement- and accountability-oriented approachStufflebeamDLCorynCLSEvaluation Theory, Models, and Applications2nd edSan Francisco, CAJossey-Bass2014204235
  • KirkpatrickDLKirkpatrickJDEvaluating Training Programs: The Four Levels3rd edOakland, CABerrett-Koehler Publishers2009