1,321
Views
4
CrossRef citations to date
0
Altmetric
Research Articles

Assessing medical student empathy in a family medicine clinical test: validity of the CARE measure

, , , &
Article: 27346 | Received 20 Jan 2015, Accepted 15 Jun 2015, Published online: 07 Jul 2015

Abstract

Introduction

The Consultation and Relational Empathy (CARE) measure developed and validated in primary care settings and used for general practitioner appraisal is a 10-item instrument used by patients to assess doctors’ empathy. The aim of this study is to investigate the validity of the CARE measure in assessing medical students’ empathy during a formative family medicine clinical test.

Method

All 158 final-year medical students were assessed by trained simulated patients (SPs) – who completed the CARE measure, the Jefferson Scale of Patient Perceptions of Physician Empathy (JSPPPE), and a global rating score to assess students’ empathy and history-taking ability.

Results

Exploratory and confirmatory factor analysis identified a unidimensional structure. The CARE measure strongly correlated with both convergent measures: global rating (ρ=0.79 and <0.001) and JSPPPE (ρ=0.77 and <0.001) and weakly correlated with the divergent measure: history-taking score (ρ=0.28 and <0.001). Internal consistency was excellent (Cronbach’s α=0.94).

Conclusion

The CARE measure had strong construct and internal reliability in a formative, undergraduate family medicine examination. Its role in higher stakes examinations and other educational settings should be explored.

At the heart of a meaningful doctor–patient relationship is empathy (Citation1). More than an expression of sympathy or a character trait, empathy in a clinical setting is a multifaceted concept. It includes emotive, moral, cognitive, and behavioral components (Citation2) that can be articulated as a professional skill or competency – which in turn, can be learned, demonstrated, and assessed. It has a direct, positive impact on the quality of patient care (Citation3) in terms of patient and doctor satisfaction, patient enablement, and possibly health outcomes (Citation4).

Given its recognized importance in patient care, nurturing empathy from the earliest stages of medical training has been widely advocated, and the Association of American Medical Colleges has recommended that empathy be an essential objective in undergraduate education (Citation5). One of the key aims of the undergraduate medical curriculum at the University of Hong Kong (HKU) is to develop students who will be able to ‘engage in productive, empathic relationships with patients, and display effective communication skills’ (Citation6). Indeed, researchers have found that medical student empathy predicts future doctor–patient empathy, underlining the importance of cultivating the development of empathy in medical students during their training (Citation7).

In terms of expected competencies within family medicine, medical students in their final year of undergraduate medical education at HKU are expected to be able to properly conduct a primary care consultation. This includes acquiring relevant information, generating diagnoses, and negotiating a management plan – all using a humanistic, patient-centered approach. Empathy is a central element in the patient-centered approach and key to the development of a therapeutic doctor–patient relationship. Since it is ultimately the patient’s perception which determines the success and effectiveness of the clinical relationship, patients’ perception of empathy is highly relevant.

On this premise, a patient-centered measure of empathy tailored to a primary care setting was developed in the UK. Known as the Consultation and Relational Empathy (CARE) measure, this 10-item questionnaire was designed to capture the set of physician competencies perceived by patients as important in holistic and empathic care (Citation8). It has been subsequently validated in primary care settings in the both the UK (Citation9) and Hong Kong (Citation10) and is capable of distinguishing between doctors’ interpersonal competencies (Citation11). In the UK, it also plays a role in quality assurance and training, where it is used for workplace appraisal and training of general practitioners (Citation12).

Other measures of empathy, mostly general self-report instruments (e.g., the Interpersonal Reactivity Index (Citation13), the Empathy Scale (Citation14), the Emotional Empathy Scale (Citation15)), have been used in a research context. The Jefferson Scale of Patient Perceptions of Physician Empathy (JSPPPE) (Citation16) is a generic scale, which has been used in medical education – but is not specifically designed for primary care. As our focus is on clinical consultations conducted within a family medicine framework, the CARE measure is a more fit-for-purpose instrument which, if valid, may be a useful assessment tool in identifying deficiencies in medical students’ relational empathy, as perceived by their future patients.

The aim of this study, then, is to establish the validity of the CARE measure in assessing the empathy of final-year medical students during a formative family medicine clinical competency assessment.

Method

Subjects and setting

All final year medical students taking the formative family medicine clinical competency test (CCT) in 2013 comprised the target population. Administered at the end of each of six annual family medicine rotations, the CCT requires students to conduct a 15-min consultation with a simulated patient (SP) in the presence of an examiner. Every SP is trained to assess students on their interpersonal skills and empathy, and to assess students’ acquisition of key history-taking information using a case-based checklist.

SP training sessions were conducted prior to each CCT, and the content of the CARE measure was reviewed to ensure SPs understood each element they were required to assess. The SPs were encouraged to respond according to how the student actually made them feel during the consultation.

A total of nine SPs (three males and six females) assessed 8–10 students across 1–4 clinical rotations; all SPs were 20–30 years of age and of Chinese descent. Different SPs were used depending on the gender requirement for the case and/or SP availability.

All cases were structured similarly and based on a common complaint encountered in family practice (e.g., cough, headache, and palpitations) – requiring students to identify and address (in a management plan) a biopsychosocial problem list. Although some scenarios were more conducive to showing empathy, elements of the CARE measure (e.g., Does the student make you feel at ease? Does the student really listen to you? Does the student explain things clearly?) pertained to general interpersonal skills required of any consultation.

Written informed consent was obtained from students prior to the CCT, permitting the use of their assessment scores in the study.

Study instrument

The CARE measure is a 10-item consultation process measure shown to produce valid scores of patients’ perceptions of relational empathy in primary care contexts (Citation9). A 5-point Likert scale ranging from 1 (poor) to 5 (excellent) is used to rate each item, which are summed into a total score ranging from 10 to 50. Missing values were handled as recommended in the guidance notes on the scoring (Citation12). Two or fewer missing values and ‘not applicable’ responses were replaced with the average score for the remaining items in that individual’s questionnaire. Questionnaires with more than two missing responses were excluded from the analysis.

Comparison instruments

The global rating of empathy is a single question, which asked patients to give their overall impression of the student’s empathy, interpersonal connection, and attitude on a 5-point Likert scale. This item is based on a global rating scale for empathy, which has been used to assess physician empathy in the domains of patient connectedness – allowing patients’ sharing of feelings and perspective and showing of empathic expression (Citation17). A similar summated global rating of senior medical student performance in the domains of empathy, coherence, and verbal/non-verbal expression has been shown to have good psychometric properties in an objective structured clinical examination (OSCE) setting (Citation18).

The JSPPPE is a 5-item scale rated on a 7-point Likert scale describing empathetic engagement of the physician as perceived by patients. Its use in medical education has been supported by psychometric evidence in studies involving post-graduate medical trainees (Citation16). It significantly correlates with patients’ satisfaction, interpersonal trust, and adherence to physicians’ recommendations (Citation19) – and has also been used in a US medical school to assess empathy during a third-year OSCE (Citation20).

A 10-item history-taking checklist documented student’s elicitation of key clinically relevant information from the SP. These items reflect solely factual information and are unrelated to interpersonal skills or empathy. Checklists completed by SPs or other observers have been useful in assessing history-taking and other domains in the realm of general medical practice (Citation21).

Ethics approval

Ethical approval of this study was granted by the Institutional Review Board of the University of Hong Kong/Hospital Authority Hong Kong West Cluster (Reference No.: UW 12-102).

Data analysis

To identify potential floor or ceiling effects in the CARE measure, the proportions of students receiving the minimum and maximum possible scores were calculated to see if either exceeded 15% (Citation22).

Using Spearman rank order correlation coefficients, construct validity of the CARE measure was established via its relationship to: 1) the JSPPPE and global empathy rating (convergent validity) and 2) the history-taking checklist (divergent validity). Convergent validity was supported if the CARE measure, the global empathy rating, and the JSPPPE scores were moderately to highly correlated (r≥0.3). Divergent validity was supported if the CARE measure was only correlated weakly (r<0.3) with the history-taking checklist score.

Exploratory factory analysis (EFA) utilizing a principal components method with Varimax rotation was used to establish the underlying factor structure of the CARE measure, and to compute the factor Eigenvalues and individual factor loadings. Factor loadings ≥0.5 reflected items’ correlation with a factor, while items which loaded <0.5 or loaded on multiple factors (i.e., cross-loaded) were removed from further investigation. Eigenvalues describe the amount of variance attributable to each factor; factors with eigenvalues of >1 were retained in the structure (Citation23).

Confirmatory factor analysis (CFA) was performed to further examine the construct validity of the factor structure proposed by the EFA and the one-factor solution of the original (UK) version of CARE measure (Citation9). Polychoric correlations measured the ordinal association between item scores, and maximum likelihood estimation explored the factor loadings and variance explained by one-factor solution. A chi-square test (Citation24), goodness-of-fit index (GFI) (Citation25), adjusted goodness-of-fit index (AGFI) (Citation25), root mean square error of approximation (RMSEA) (Citation26), and comparative fit index (CFI) were used to assess the model goodness-of-fit, which was considered adequate if: 1) chi-square test (p≥0.05); 2) RMSEA≤0.08; 3) GFI≥0.90; 4) AGFI≥0.80; and 5) CFI≥0.95 (Citation27).

For factor analysis, the sample was split into two subsamples comprising only of cases with complete data (no missing responses). Data from rotations 1–3 and 4–6 were used EFA and CFA (respectively) to identify subscales. Cronbach’s α coefficient was used to determine each subscales’ internal consistency relative to the expected standard of ≥0.7 (Citation28). The effect of imputed data substitutions (missing values) on internal consistency was undertaken in a sensitivity analysis.

Both the EFA and CFA were performed using LISREL 8.80 (Scientific Software International, Inc., Lincolnwood, IL, USA), while other statistical analyses were performed using IBM SPSS Window 21.0 program (SPSS, Inc., Chicago, IL, USA).

Results

Of the 158 study subjects, 97 (61.4%) were male and ranged in age from 22 to 37 (median=24). Based on the six rotations of 2013 CCT examinations, the mean CARE measure score was 35.8 out of a possible 50. No floor or ceiling effects were observed. Descriptive, univariate statistics of key variables are shown in .

Table 1 Descriptive statistics of key variables

Exploratory and confirmatory factor analysis

The validity of our data was first confirmed using EFA where the Kaiser–Meyer–Olkin measure of 0.94 and Bartlett’s test of sphericity [χ 2(45)=887.8, p<0.001] confirmed the sampling adequacy and variability. Using a principal components analysis, a one-factor solution was shown to explain 77.6% of the total variance. All 10 items loaded significantly on this single factor.

Based on conventional guidelines, a CFA found that this one-factor model met the criteria demonstrating excellent goodness of fit (RMSEA=0.06; GFI=0.89; AGFI=0.83; CFI=0.99). The null hypothesis of chi-square test was rejected (χ 2=46.72; p=0.09), suggesting an adequate fit of the data with the one-factor model. EFA and CFA loading are shown in .

Table 2 Exploratory and confirmatory factor loadings of CARE measure items

Convergent and divergent validity

Patients’ total CARE measure scores were strongly positively correlated with both their global empathy rating (ρ=0.79 and <0.001) and the JSPPPE scores (ρ=0.77 and <0.001), but only weakly associated with the history-taking score (ρ=0.28 and <0.001). This is shown in .

Table 3 Correlation of total CARE measure score with convergent and divergent constructs

Internal consistency

Internal consistency of the 10-item CARE measure was excellent, as evidenced by the Cronbach’s α of 0.94. A sensitivity analysis of mean substitutions of missing data yielded only a miniscule increase in internal consistency (α=0.95).

Discussion

The CARE measure is a widely used means of assessing primary care doctors’ relational empathy during a consultation, from the patient’s perspective. In this study, we extrapolated its validity to include medical students’ consultations in an undergraduate family medicine setting – showing that the CARE measure retained its original unidimensional structure (Citation9), excellent internal consistency, and had good convergent and divergent validity. These findings bring the patient perspective squarely into medical educational assessment and should encourage more objective and standardized assessment of a complex attribute, empathy, in a formative (low-stake), family medicine context.

As validated in this context, the CARE measure may have some educational benefits over shorter measures like the 1-item global empathy rating or the 5-item JSPPPE. Firstly, with 10 items, the CARE measure expands a complex concept into a set of concrete, practical elements that are clearly understood by students. Smaller components enable students to focus on particular aspects of the whole, analogous to learning a complex skill through microskill acquisition. Secondly, items better articulate the interpersonal skills needed by primary care doctors, so its applicability in primary care would be an advantage in teaching consultations in family medicine and other primarycare-oriented settings. Similar to some instruments used to measure healthcare outcomes – shorter, generic measures may not have the sensitivity to capture small differences or may be less responsive to capturing changes over time in a specified population (Citation29). Used formatively, where the focus is to help students learn and improve, the CARE measure can serve as a guiding rubric that represents the essential elements desired in a primary care consultation. This may be used for benchmarking and for generating student feedback to help identify specific clinical strengths and weaknesses.

Furthermore, the absence of a floor or ceiling effect in this context may make this instrument sensitive enough to differentiate among students’ performance. In contrast, when used in doctor–patient or therapist–patient settings, CARE measure scores tended toward the higher end of the distribution – with more than a quarter of targets receiving the maximum score (Citation9, Citation30). Real patients are likely to voluntarily seek out and establish relationships with doctors they find ‘acceptable’ and whom they may already know well. For students, this is a required interaction that represents a one-off visit. As well, SPs in an undergraduate exam setting may recognize the ‘developmental’ limitations of students, and hence refrain from awarding them the maximum score.

The excellent internal consistency of the CARE measure found in this study provides some preliminary evidence for its reliability. However, assessments of the same student by multiple examiners or over time would offer additional support of its reliability. In the primary care setting, it has been suggested that 50 completed assessments by patients using the CARE measure are required to reliably assess doctors’ empathy (Citation8), which would be impossible or impractical in most educational settings.

The value of assessment for learning (as opposed to assessment of learning) has been advocated in the learning of clinical competencies in medical education (Citation31), and students’ relational empathy may be best developed and improved if assessed in the same way. The CARE measure provides a valid way in which students can be assessed and learn to improve their relational empathy. This, combined with qualitative feedback from peers/supervisors and self- reflection, can provide a more solid indication of students’ acquisition of a core clinical consultation skill.

Strengths and limitations

An adequate and appropriate sample, as well as the use of external measures to establish convergent and divergent validity, is among the strengths of this study.

The main limitation relates to the unknown generalizability of the findings to other educational settings or activities. Even though our study included a low-stakes, formative emphasis, both students and patients may have behaved differently than those within a more realistic clinical setting. In addition, our study was conducted in a specific setting, in one curriculum, and at one institution, which necessitates further study to examine validity issues in other educational settings. Finally, although the internal consistency of the CARE measure was established, further psychometric examination in terms of test–retest and inter-rater reliability would greatly strengthen our findings.

Conclusion

The CARE measure was shown to have strong construct validity and excellent internal consistency in a formative, undergraduate family medicine examination. It also has some discriminatory potential in this context due to the absence of floor or ceiling effects and the ability of SPs to complete the measure under exam conditions. This study demonstrated that the CARE measure can be a useful tool to assess and generate feedback to students on specific interpersonal elements of the consultation – bringing patients’ perspective into the realm of primary care consultation. Further work is needed to explore its role in higher stakes clinical examinations and other educational settings.

Conflict of interest and funding

The authors have not received any funding or benefits from industry or elsewhere to conduct this study.

References

  • Shapiro H , Curnan M , Peschel E , St James D . Empathy and the practice of medicine: beyond pills and the scalpel. 1993; New Haven: Yale University Press.
  • Morse JM , Anderson G , Bottorff JL , Yonge O , O’Brien B , Solberg SM , etal. Exploring empathy: a conceptual fit for nursing practice?. Image J Nurs Sch. 1992; 24: 273–80.
  • Mercer SW , Reynolds WJ . Empathy and quality of care. Br J Gen Pract. 2002; 52: S9–S13. [PubMed Abstract] [PubMed CentralFull Text].
  • Price S , Mercer SW , MacPherson H . Practitioner empathy, patient enablement and health outcomes: a prospective study of acupuncture patients. Patient Educ Couns. 2006; 63: 239–45.
  • Medical School Objectives Project. Association of American Medical Colleges. Available from: https://www.aamc.org/initiatives/msop/ [cited 11 April 2014]..
  • The medical curriculum handbook. Li Ka Shing Faculty of Medicine. 2010; The University of Hong Kong.
  • Hojat M , Mangione S , Nasca TJ , Gonella JS , Magee M . Empathy scores in medical school and ratings of empathic behavior in residency training 3 years later. J Soc Psychol. 2005; 145: 663–72.
  • Mercer SW , McConnachie A , Maxwell M , Heaney D , Watt GC . Relevance and practical use of the Consultation and Relational Empathy (CARE) measure in general practice. Fam Pract. 2005; 22: 328–34.
  • Mercer SW , Maxwell M , Heaney D , Watt GC . The consultation and relational empathy (CARE) measure: development and preliminary validation and reliability of an empathy-based consultation process measure. Fam Pract. 2004; 21: 699–705.
  • Fung CSC , Hua A , Tam L , Mercer SW . Reliability and validity of the Chinese version of the CARE measure in a primary care setting in Hong Kong. Fam Pract. 2009; 26: 398–406.
  • Mercer S , Fung C , Chan F , Wong F , Wong S , Murphy D . The Chinese-version of the CARE measure reliably differentiates between doctors in primary care: a cross-sectional study in Hong Kong. BMC Fam Pract. 2011; 12: 1–9.
  • Consultation and Relational Empathy (CARE) measure – summary of research and current use. Section of General Practice and Primary Care, University of Glasgow. 2010. Available from: http://www.gla.ac.uk/departments/generalpracticeprimarycare/research/caremeasure/ [cited 11 April 2014]..
  • Davis MH . Measuring individual-differences in empathy – evidence for a multidimensional approach. J Pers Soc Psychol. 1983; 44: 113–26.
  • Persons JB , Burns DD . Mechanisms of action of cognitive therapy: the relative contributions of technical and interpersonal interventions. Cognit Ther Res. 1985; 9: 539–51.
  • Mehrabian A , Epstein N . A measure of emotional empathy. J Pers. 1972; 40: 525–43.
  • Kane G , Gotto J , Mangione S , West S , Hojat M . Jefferson Scale of Physician Empathy: preliminary psychometric data. Croat Med J. 2007; 48: 81–6.
  • Bonvicini KA , Perlin MJ , Bylund CL , Carroll G , Rouse RA , Goldstein MG . Impact of communication training on physician expression of empathy in patient encounters. Patient Educ Couns. 2009; 75: 3–10.
  • Hodges B , McIlroy JH . Analytic global OSCE ratings are sensitive to level of training. Med Educ. 2003; 37: 1012–16.
  • Hojat M , Louis DZ , Maxwell K , Markham F , Wender R , Gonnella JS . Patient perceptions of physician empathy, satisfaction with physician, interpersonal trust, and compliance. Int J Med Educ. 2010; 1: 83–7.
  • Berg K , Majdan JF , Berg D , Veloski J , Hojat M . A comparison of medical students’ self-reported empathy with simulated patients’ assessments of the students’ empathy. Med Teach. 2011; 33: 388–91.
  • Luck J , Peabody JW . Using standardised patients to measure physicians’ practice: validation study using audio recordings. BMJ. 2002; 325: 679.
  • Terwee CB , Bot SD , de Boer MR , van der Windt DA , Knol DL , Dekker J , etal. Quality criteria were proposed for measurement properties of health status questionnaires. J Clin Epidemiol. 2007; 60: 34–42.
  • Hair JF , Anderson RE , Tatham RL . Multivariate data analysis. 2006; Upper Saddle River, NJ: Prentice-Hall. 6th ed.
  • Marsh HW , Balla JR , McDonald RP . Goodness-of-fit indexes in confirmatory factor analysis: the effect of sample size. Psychol Bull. 1988; 103: 391–410.
  • Bentler PM . Quantitative methods in psychology: comparative fit indexes in structural models. Psychol Bull. 1990; 107: 238–46.
  • Cole DA . Methodological contributions to clinical research: utility of confirmatory factor analysis in test validation research. J Consult Clin Psychol. 1987; 55: 584–94.
  • Hu LT , Bentler PM . Cutoff criteria for fit indexes in covariance structure analysis: conventional criteria versus new alternatives. Struct Equ Model. 1999; 6: 1–55.
  • Nunnally JC . Psychometric theory. 1994; New York: McGraw Hill. 3rd ed.
  • Patrick DL , Deyo RA . Generic and disease-specific measures in assessing health status and quality of life. Med Care. 1989; 27: S217–S32.
  • Kersten P , White PJ , Tennant A . The consultation and relational empathy measure: an investigation of its scaling structure. Disabil Rehabil. 2012; 34: 503–9.
  • Schuwirth LWT , Van der Vleuten CPM . Programmatic assessment: from assessment of learning to assessment for learning. Med Teach. 2011; 33: 478–85.