224
Views
10
CrossRef citations to date
0
Altmetric
Original Research

Validating the modified System for Evaluation of Teaching Qualities: a teaching quality assessment instrument

, &
Pages 881-886 | Published online: 30 Nov 2018

Abstract

Construct

We assessed the validity of the modified System for Evaluation of Teaching Qualities (mSETQ) in evaluating clinical teachers in Bahrain.

Background

Clinical teacher assessment tools are essential for improving teaching quality. The mSETQ is a teaching quality measurement tool, and demonstrating the validity of this tool could provide a stronger evidence base for the utilization of this questionnaire for assessing medical teachers in Bahrain.

Approach

This study assessed the construct validity of this questionnaire in medical schools across Bahrain using 400 medical students and 149 clinical teachers. Data were analyzed using confirmatory factor analysis (CFA). The goodness-of-fit index (GFI), comparative fit index (CFI), root mean square residual, and standardized root mean square error of approximation (RMSEA) indices were used to evaluate the model fit. The internal consistency reliability was assessed using Cronbach’s alpha.

Results

The results of the CFA revealed an acceptable fit. All criteria for a good model fit were met except for the RMSEA fit index and the standardized root mean square residual (SRMR) value, which was very close to an acceptable value. Good overall reliability was found in the study (α=0.94).

Conclusion

The overall findings of this study provided some evidence supporting the reliability and validity of the mSETQ instrument.

Introduction

The quality of patient care is indirectly dependent on the quality of the training that the physicians receive. This student–teacher rapport necessitates the sharing of knowledge, attitudes, skills, experiences, influences, and interactions in the relationship in an appropriate manner.Citation1 In the recent years of learner-centered education, teaching hospitals around the world have placed greater emphasis on evaluating their clinical teachers based on their clinical competency, teaching skills, personal qualities, involvement of teachers with the students, involvement of students in the provision of patient care, and the provision of guidance and feedback.Citation2Citation4 Past literature reviews have identified many different instruments that have been used to assess clinical teachers.Citation5Citation7 The results from these reviews indicate that approximately 30–35 instruments were available in the form of questionnaires that included 1–58 items. The assessments were mainly based on responses from student learners or residents. These assessments were used to provide formative feedback on the students’ teaching efficiency, resource allocation, promotions, and performance review.Citation5Citation7 Considering the implications of these instruments, they should display high validity and reliability.Citation8 There are five sources of validity evidence which have been identified by the American Psychological and Educational Research Associations as follows: content, response process, internal structure, relation to other variables, and consequences.Citation9 These assessments are essential for continuous development and improvement in clinical teaching skills and need to be applied to a wide variety of samples at different points in the learning process.Citation8 The System for Evaluation of Teaching Qualities (SETQ) was developed and has been used extensively in the Netherlands to assess clinical teachers.Citation10Citation13 The SETQ consists of two sets of questionnaires containing the same items: one is for the supervisor’s self-evaluation and the other collects the learners’ assessments of their clinical teachers.Citation11 Studies of the SETQ using confirmatory factor analysis (CFA) in teaching hospitals in the Netherlands have concluded that this instrument is a highly reliable and valid tool for assessing physicians’ teaching performances.Citation14Citation16 This original instrument has been modified and tested for reliability and validity by Al Ansari et alCitation15 in a teaching school in Bahrain. However, to increase the strength of the evidence regarding the psychometric properties of the instrument, it must be applied in several settings.Citation8 The modified SETQ (mSETQ) created by Al Ansari et alCitation15 consists of 25 items in six domains that include the teaching and learning environment, professional attitude toward and support of residents, communication of goals, evaluation of students, feedback, and promotion of self-directed learning.Citation15 A previous study that evaluated the validity of the mSETQ in Bahrain had a sample size that was too small. Hence, the present study aimed to use a larger sample size to strengthen the validity of this questionnaire for assessing clinical teachers in Bah-rain to provide a stronger evidence base for the use of this questionnaire in this specific context.

Materials and methods

The mSETQ instrument

The 25-item mSETQ instrument was previously developed as a modification of the SETQ.Citation15 The mSETQ is composed of the following six domains: teaching and learning environment (six items), professional attitudes toward students (four items), communication of goals (three items), evaluation (five items), feedback (four items), and promotion of self-directed learning (three items). Each of the items is a statement about which the participants have to identify the extent to which they agree by providing their rating on a 5-point Likert scale ranging from “strongly disagree” to “strongly agree”. In this study, this questionnaire was presented several times in printed handouts, and the participants were shown a list of teachers that they could choose to evaluate. The students wrote the name of the tutor they would like to evaluate on top of each mSETQ questionnaire. They were informed they could evaluate as many (or as few) teachers as they liked using separate questionnaires.

Study population, setting, and data extraction

A total of 400 medical students from three different clinical years and 149 clinical teachers working at four different teaching hospitals in the Kingdom of Bahrain were invited to participate in the study from September 2016 until June 2017. The sample included students from the Arabian Gulf University (AGU) who were from different nationalities including Bahrain, Kuwait, Saudi Arabia, and Oman. The written informed consent to participate was obtained from the clinical teachers, and the verbal consent was obtained from students which was acceptable and approved by research ethical committee at the AGU of Bahrain. The study was conducted in accordance with the Declaration of Helsinki.Citation16 The study was approved by the research ethics committee at AGU, Bahrain. The students and teachers were aware that their data would be stored anonymously (using numbers rather than names) and that their data might be published. The clinical teachers were working in different departments that included internal medicine, obstetrics, gynecology, pediatrics, surgery, ophthalmology, psychiatry, and family medicine. The printed mSETQ questionnaire was distributed to the students by their clinical coordinators during their rotations in different hospitals and during their in-campus lectures. The students were requested to evaluate as many teachers as they liked (by completing several mSETQ questionnaires) and to return the completed questionnaires to their clinical coordinators. A total of 1,615 completed surveys were received. Incomplete questionnaires were excluded from the study. In total, the number of teachers included in the final analysis was 125, and 1,551 complete questionnaires were used in the analysis.

Statistical analyses

Missing data are common in much educational research, and we used pairwise deletion method to handle the missing data in the analysis. Pairwise deletion only removes the specific missing values from the analysis (not the entire case). Thus, pairwise deletion maximizes all data available by an analysis-by-analysis basis. The strength of this technique is that it increases the power of analyses. CFA was performed to evaluate the construct validity of the questionnaire. The analysis was conducted using EQS Structural Equation Modeling software (Multivariate software Inc, Broadway, CA, USA). The descriptive and reliability analyses were conducted by IBM SPSS version 23.0 (IBM Corporation, Armonk, NY, USA). In this study, the data were negatively skewed and violated the normality assumption. Therefore, we performed CFA with the robust maximum likelihood estimation (MLE) method.Citation17 The robust maximum likelihood (ML) estimator provides ML parameter estimates with standard errors and a chi-squared test statistic that are robust to non-normality. The fit indices used to determine the fit of the model were the comparative fit index (CFI), goodness-of-fit index (GFI), adjusted GFI (AGFI), root mean square error of approximation (RMSEA), and standardized root mean square residual (SRMR).Citation18 GFI, CFI, and AGFI values of >0.90 suggested a good model fit.Citation19,Citation20 An RMSEA value between 0.08 and 0.10 indicated an average model fit, and values below 0.06 demonstrated a good fit. SRMR values below 0.08 demonstrate a good fit.Citation21

The internal consistency reliability was assessed using the Cronbach’s alpha for each scale. A Cronbach’s alpha coefficient value of ≥0.70 was considered as an acceptable reliability.

Results

Descriptive data

The number of student raters varied from 3 to 41 for each clinical tutor. The response rates were 83.8% (N=125/149) for the clinical teachers and 100% (N=400/400) for the clinical raters.

In this study, 3.57% of the total values were missing (9.73% of the cases), and pairwise deletion method was used to handle the missing data. describes the students’ evaluation scores for each item and the overall score of each scale. The overall scores ranged from 3.76 to 3.90. The scale “promoting self-directed learning” was achieved the highest rating of an average score of 3.90 followed by “professional attitude toward students” scored 3.89 and “communication of goals” scored an average of 3.86.

Table 1 Mean (SD) of students’ responses and Cronbach’s alpha of different scales of the mSETQ instrument

Construct validity

The CFA results of the six-factor mSETQ are shown in . The path diagram shows the standardized regression weights (factor loadings) which explains the pattern of item–factor relationships, and all the factor-loading values are >0.40. The results of the CFA suggested an acceptable model fit. All values met the criteria for a good model fit (ie, GFI, CFI, and AGFI >0.90). However, a cutoff value close to 0.08 for SRMR and a cutoff value close to 0.06 for RMSEA are needed for a good model fit. This study reported an RMSEA (0.047) and SRMR (0.079). The chi-squared value was significant: c2 (385)=670.76, P<0.0001. summarizes the values of the fit indices.

Figure 1 Path analysis of different scales of the teaching quality, teaching and learning environment, professional attitude toward children, communication of goals, evaluation of students, feedback, and promoting self-directed learning.

Abbreviation: Q, question.
Figure 1 Path analysis of different scales of the teaching quality, teaching and learning environment, professional attitude toward children, communication of goals, evaluation of students, feedback, and promoting self-directed learning.

Table 2 CFA results: the model fit indices

Reliability

The overall reliability was measured using the internal consistency reliability (Cronbach’s alpha). The overall reliability of the instrument was 0.94 showing excellent reliability. The subscale reliabilities are described in .

Discussion

The present multicentre study is an attempt to assess the construct validity and internal structure of the mSETQ questionnaire that was used to evaluate the clinical teaching quality of clinical teachers in four different teaching hospitals in Bahrain. Four hundred medical students rated 125 clinical teachers using the questionnaire. All teachers included were involved with clinical teaching of the student raters. The results of this study indicate an overall teaching score between 3.76 and 3.90, which is considered satisfactory. Regarding the individual components, similar scores were observed across all components with teaching and learning environment, communication, and promoting self-directed learning exhibiting the highest values. According to our CFA, the mSETQ with 26 items in six factors was identified as a reliable and valid tool for assessing clinical teaching which showed an acceptable model fit.

The response rate is a major criterion that determines the overall outcome of questionnaire-based research. Previous studies have revealed that the average response rate for data collected from individuals is 40–80%.Citation22Citation24 The acceptable response rate is reported to be between 60% and 80%. Considering this rate, the questionnaire tool used should provide high response rates with limited numbers of evaluations, which is an indication of the feasibility of use. This is the greatest advantage of the mSETQ; the response rate of clinical teachers was 83.8% and that of the student raters was 100%.

Limitations

In this study, the learners were allowed to choose their objects of evaluation; this is a limitation as it could lead to selection bias. Future studies should avoid selection bias and ensure appropriate randomization by not allowing the learners to select their objects of evaluation. It is possible that allowing students to choose their objects of evaluation influenced the ceiling effect found in this study. The finding that most teachers scored highly in their evaluations in this study suggests that some alterations must be made to the methodology of the study. Moreover, another limitation of the current study is that the self-evaluation component of SETQ was not incorporated and utilized in the mSETQ. Therefore, future studies should replicate this study and include the self-evaluations as well as aim to reduce selection bias and the ceiling effect.

Conclusion

We provided some evidence supporting the reliability and the validity of the mSETQ tool for evaluating the clinical teaching quality of teachers. This system not only evaluates the current situation but also provides the teachers a scope for improvement in the future based on the feedback they receive. This feedback leads to the identification of learning needs and the development of innovative strategies for teaching and learning. Validity is concerned with meaningful associations, inferences, and their interpretations, and it is of utmost importance to validate any form of an assessment before interpreting it inferences. This notion is important considering the influence provided by the inferences. The validation of the mSETQ before its use in the current study will provide a reliable and meaningful interpretation of clinical teaching quality when the questionnaire is used for assessments in the future. This paper recommends the use of this questionnaire for assessing heterogeneous groups of clinical teachers from different departments in clinical education studies.

Acknowledgments

We would like to thank Professor Claudio Violato for reviewing and modifying the statistical analysis.

Disclosure

The authors report no conflicts of interest in this work.

References

  • SchmidtHGMamedeSHow to improve the teaching of clinical reasoning: a narrative review and a proposalMed Educ2015491096197326383068
  • WoolleySEmanuelCKoshyBA pilot study of the use and perceived utility of a scale to assess clinical dental teaching within a UK dental school restorative departmentEur J Dent Educ2009132737919368549
  • MookherjeeSMonashBWentworthKLSharpeBAFaculty development for hospitalists: structured peer observation of teachingJ Hosp Med20149424425024446215
  • BieryNBondWSmithABLeclairMFosterEUsing Telemedicine Technology to Assess Physician Outpatient TeachingFam Med2015471080781026545060
  • FluitCRBolhuisSGrolRLaanRWensingMAssessing the quality of clinical teachers: a systematic review of content and quality of questionnaires for assessing clinical teachersJ Gen Intern Med201025121337134520703952
  • LombartsMJHeinemanMJArahOAAssessing the quality of clinical teachersJ Gen Intern Med20112611420967506
  • BeckmanTJGhoshAKCookDAErwinPJMandrekarJNHow reliable are assessments of clinical teaching? A review of the published instrumentsJ Gen Intern Med200419997197715333063
  • SpencerJLearning and teaching in the clinical environment. ABC of learning and teaching in medicineBr Med J200332659159412637408
  • DowningSMValidity: on meaningful interpretation of assessment dataMed Educ200337983083714506816
  • BoerebachBCArahOABuschORLombartsKMReliable and valid tools for measuring surgeons’ teaching performance: residents’ vs self evaluationJ Surg Educ201269451152022677591
  • LombartsKMBucxMJArahOADevelopment of a system for the evaluation of the teaching qualities of anesthesiology facultyAnesthesiology2009111470971619707115
  • van der LeeuwRLombartsKHeinemanMJArahOSystematic evaluation of the teaching qualities of Obstetrics and Gynecology faculty: reliability and validity of the SETQ toolsPLoS One201165e1914221559275
  • van der LeeuwRLombartsKHeinemanMJArahOSystematic evaluation of the teaching qualities of Obstetrics and Gynecology faculty: reliability and validity of the SETQ toolsPLoS One201165e1914221559275
  • BoerebachBCLombartsKMArahOAConfirmatory Factor Analysis of the System for Evaluation of Teaching Qualities (SETQ) in Graduate Medical TrainingEval Health Prof2016391213225280728
  • Al AnsariAStrachanKHashimSOtoomSAnalysis of psychometric properties of the modified SETQ tool in undergraduate medical educationBMC Med Educ20171715628302151
  • World Medical Association Declaration of HelsinkiBull World Health Organ2001794
  • BentlerPEQS Structural Equations Program ManualEncino (Calif)Multivariate Software1995
  • HarringtonDConfirmatory Factor AnalysisOxfordOxford University Press2009
  • KlineRBPrinciples and Practice of Structural Equation Modeling3rd edNew YorkGuilford Publications2015
  • ByrneBStructural Equation Modeling with AMOSMahwahLawrence Erlbaum Associates2001
  • HuLBentlerPMCutoff criteria for fit indexes in covariance structure analysis: Conventional criteria versus new alternativesStruct Equ Modeling A Multidisciplinary J199961155
  • FeketeCSegererWGemperliABrinkhofMWSwiSCI Study GroupParticipation rates, response bias and response behaviours in the community survey of the Swiss Spinal Cord Injury Cohort Study (SwiSCI)BMC Med Res Methodol20151518026450702
  • EskanderAFreemanJRotsteinLResponse rates for mailout survey-driven studies in patients waiting for thyroid surgeryJ Otolaryngol Head Neck Surg201140646246722420433
  • ChristensenAIEkholmOKristensenPLThe effect of multiple reminders on response patterns in a Danish health surveyEur J Public Health201525115616124855288