90
Views
4
CrossRef citations to date
0
Altmetric
Brief Report

The Adapted Fresno test for speech pathologists, social workers, and dieticians/nutritionists: validation and reliability testing

, &
Pages 129-135 | Published online: 27 Feb 2014

Abstract

Purpose

The current versions of the Adapted Fresno test (AFT) are limited to physiotherapists and occupational therapists, and new scenarios and scoring rubrics are required for other allied health disciplines. The aim of this study was to examine the validity, reliability, and internal consistency of the AFT developed for speech pathologists (SPs), social workers (SWs), and dieticians/nutritionists (DNs).

Materials and methods

An expert panel from each discipline was formed to content-validate the AFT. A draft instrument, including clinical scenarios, questionnaire, and scoring rubric, was developed. The new versions were completed by ten SPs, 16 SWs, and 12 DNs, and scored by four raters. Interrater reliability was calculated using intraclass correlation coefficients (2,1) for the individual AFT items and the total score. The internal consistency of the AFT was examined using Cronbach’s α.

Results

Two new clinical scenarios and a revised scoring rubric were developed for each discipline. The reliability among raters was excellent for questions 1, 3, and 6 across all disciplines. Question 7 showed excellent reliability for SPs, but not for SWs and DNs. All other reliability coefficients increased to moderate or excellent levels following training. Cronbach’s α was 0.71 for SPs, 0.68 for SWs, and 0.74 for DNs, indicating that internal consistency was acceptable for all disciplines.

Conclusion

There is preliminary evidence to show that AFT is a valid and reliable tool for the assessment of evidence-based practice knowledge and skills of SPs, SWs, and DNs. Further research is required to establish its sensitivity to detect change in knowledge and skills following an educational program.

Introduction

The importance of evidence-based practice (EBP) in allied health is well documented in the literature.Citation1,Citation2 Clinical decisions that are based on patients’ unique circumstances, sound clinical expertise, and the best available research evidence are known to deliver the best outcomes for patients and their families.Citation3Citation5 Allied health practitioners hold positive attitudes toward EBP and believe in the value of research evidence in informing their clinical decisions. However, applying research findings to clinical decisions is not a simple process and is often difficult to achieve. One of the most commonly reported barriers to evidence uptake in allied health is the lack of knowledge of the EBP process and lack of skill in critically appraising research.Citation6Citation8 Teaching EBP is therefore an important step in promoting evidence-based clinical decision making. Allied health practitioners need to understand the principles of EBP before they can apply it.

Early EBP educational programs include the development of clinical questions, literature searches, and critical appraisal.Citation9 To evaluate the impact of such educational programs and document competence of individual practitioners, educators need objective and psychometrically sound instruments or assessment tools. Based on a review of the literature, the Fresno test is the only available instrument that comprehensively assesses EBP competence across all relevant domains.Citation10 The Fresno test consists of two clinical scenarios and 12 short-answer questions that require respondents to formulate a focused question, identify the most appropriate research design that will address the question, show knowledge of electronic database searching, identify issues important for determining the relevance and validity of a research paper, and discuss the magnitude and importance of research findings.Citation11 The test is scored by using a standardized grading rubric that describes explicit grading criteria. The Fresno test has content validity, good-to-excellent interrater reliability for all questions, and excellent internal consistency.Citation11 However, this tool focuses on assessing competence in medical students only, and therefore it cannot be used across different health disciplines.

In 2009, McCluskey and Bishop modified the Fresno test to measure the change in EBP skills and knowledge of occupational therapists following exposure to an EBP workshop.Citation12 New clinical scenarios (ie, versions 1 and 2) were developed to suit rehabilitation professionals, such as physiotherapists and occupational therapists. The 12 questions in the original Fresno test were reduced to seven (ie, questions 1–7), removing questions about diagnosis and complex statistics (ie, questions 8–12). The scoring rubric was also revised. Similar to the original Fresno test, the seven-item Adapted Fresno test (AFT) measures the following: the ability to develop a focused clinical question using the PICO (population, intervention, comparison, and outcome) format, the ability to develop a search strategy, the ability to interpret and critically appraise a research paper, and knowledge associated with understanding of the hierarchy of evidence and methodological biases in study designs, databases, and other sources of evidence and study designs. The AFT has been reported to have acceptable psychometric properties: interrater reliability ranged from good to excellent for individual items (version 1, intraclass correlation coefficient [ICC] 0.80–0.96; version 2, 0.68–0.94) and excellent for the total score (version 1, 0.96; version 2, 0.91); acceptable internal consistency (Cronbach’s α 0.74); and responsive to change in novice learners.Citation12

The current versions of the AFT are limited to physiotherapists and occupational therapists, and new scenarios and scoring rubrics are required for other allied health disciplines. Therefore, the aim of this study was to examine the validity, interrater reliability, and internal consistency of AFT versions developed for speech pathologists, social workers, and dieticians/nutritionists.

Materials and methods

This study was approved by the Human Research Ethics Committee of the University of South Australia and the Ethics Review Board of the University of Tasmania.

Development and content validation of AFT for speech pathology, social work, and dietetics/nutrition

An expert panel consisting of four practitioners from each discipline was formed to content-validate the AFT. Content validity refers to “… how well the combined elements used to construct the instrument truly describe the conceptual domain of interest”.Citation13 The panel represented practitioners with more than 10 years of clinical experience and with previous exposure to EBP training or research. The majority had graduate degrees in their respective disciplines or other clinical areas.

The panel members were presented with the original Fresno test and AFT, and were asked to examine the questionnaire and comment on which questions should be included in the new versions for speech pathologists, social workers, and dieticians/nutritionists. All members agreed that only questions adapted by AFT should be included for these disciplines. Following discussion, new clinical scenarios were developed for each discipline. The scoring rubric of the AFT was considered applicable to the new versions except for questions 1 (“Write a focused clinical question for one scenario to help you organize a search of the literature”), 2 (“Where might you find answers to these and other similar clinical questions? Name as many possible sources of information as you can, not just the ones you think are good sources”), and 4 (“If you were to search for Medline for original research to answer your question, describe the search strategy you might use”). Discipline-specific information was required to revise the scoring key for these questions.

Following consultation with the expert panel, a draft instrument including the clinical scenarios, questionnaire, and scoring rubric was prepared by the primary author. The draft instrument was emailed to the experts for feedback on the clarity of the entire instrument and completeness of the scoring rubric. The instrument and scoring rubric for each discipline were revised based on comments from the expert panel and returned to them for a final round of feedback. No further changes were required in the instrument.

Participants

The new AFT versions were completed by ten speech pathologists, 16 social workers, and 12 dieticians/nutritionists who agreed to participate in a larger study aimed at examining the impact of a journal club on the EBP knowledge and skills of allied health professionals.Citation14 They were asked to individually complete either a paper-and-pencil version or electronic version of the questionnaire at a time convenient for them. There were equal numbers of participants who held bachelor’s degrees and postgraduate degrees. Less than half had previous training in research or EBP, and the majority had been in clinical practice for less than 10 years.

Interrater reliability of the AFT

Interrater reliability is the “… degree to which measurements of the same phenomenon by different raters will yield the same results, or the consistency of results between raters”.Citation15 Interrater reliability was calculated for individual items and the total AFT score using ICCs (2,1) and 95% confidence intervals. For interpretation of results, ICC values of ≥0.80 indicate excellent reliability, values between 0.60 and 0.79 denote moderate reliability, and values <0.60 mean questionable reliability.Citation16

Four individuals experienced in research and teaching EBP for allied health students served as raters for the study. Before the study began, the raters reviewed and discussed the AFT test, and collaboratively scored a sample test for each discipline. They were then given a practice period, where they scored another set of sample tests, then compared and discussed their differences in scoring. Following discussion, the raters were instructed to score each test independently without conferring or comparing ratings. Raters were given 2 weeks to mark all questionnaires.

Initial examination of the interrater reliability showed poor reliability between raters for questions 2, 4, and 5 of all versions (ie, AFT for speech pathology, social work, and dietetics/nutrition) and question 7 for social work and dietetics/nutrition. This prompted the first author, who has experience in using the previous AFT versions, to provide further training and discussion of the scoring procedure to the raters. The training involved an explanation of the rating system, discussion of common rater errors, advice on process for decision making, and practice on interpreting the rubric. Those questions with poor reliability were rescored 2 weeks later.

Internal consistency of the AFT

Internal consistency reflects the coherence of the components of a scale or instrument.Citation17 The internal consistency of the AFT was examined using Cronbach’s α.

Results

Content validity of the AFT for speech pathologists, social workers, and dieticians/nutritionists

The content validity of the AFT instrument was established through formal feedback from the expert panel. The comments received were consistent across disciplines, and involved issues associated with the wording of the clinical scenarios. No comments were made on the questionnaire itself; however, additional possible answers were suggested for the scoring rubric. For example, in question 1, where respondents are asked to write a focused clinical question, the expert panel provided additional PICO terms or synonyms. Some members of the panel suggested further sources of research information for question 2, such as discipline-specific electronic databases, websites, and professional organizations.

Two new clinical scenarios and a revised scoring rubric were developed for each discipline. shows the final versions of the clinical scenarios. lists the questions included in the new AFT versions. A copy of the scoring rubric may be obtained from the primary author upon request.

Table 1 Discipline-specific clinical scenarios

Table 2 Questions in the Adapted Fresno test

Interrater reliability of the AFT

The reliability among raters was excellent for questions 1, 3, and 6 across all disciplines, as shown in . Question 7 showed excellent reliability for speech pathology, but not for social work or dietetics/nutrition. All other reliability coefficients increased to moderate or excellent levels following further training and discussion.

Table 3 Interrater reliability of the Adapted Fresno test (individual items and total score)

Internal consistency of the AFT

Cronbach’s α was 0.71 for speech pathology, 0.68 for social work, and 0.74 for dietetics/nutrition, indicating internal consistency was acceptable for all disciplines. Deletion of any of the items did not improve the internal consistency of the AFT for any discipline.

Discussion

The results provide preliminary evidence of the psychometric integrity of the AFT, and support its use in the assessment of EBP knowledge and skills of speech pathologists, social workers, and dieticians/nutritionists. Similar to the original AFT, the new versions assess knowledge and skills of the key processes involved in EBP, including the development of clinical questions, searching for literature, critical appraisal, and interpretation of research findings. The new AFT has content validity, moderate-to-excellent reliability and acceptable internal consistency. These results are consistent with the previously reported validity and reliability of the original Fresno testCitation11 and AFT versions for rehabilitation professionals (ie, occupational therapist and physiotherapist).Citation12

The importance of EBP training in facilitating an evidence-based approach to clinical practice has been highlighted by a number of systematic reviews.Citation18Citation21 Many of the training programs reported in these reviews relied on self-report data, which potentially reflect inaccuracies in actual knowledge.Citation22 Measuring the effectiveness of such training programs therefore requires objective and robust instruments to document changes in the competence of the individuals being trained. To the authors’ knowledge, the AFT is the only objective measure of EBP knowledge and skills that has been tested and applied in allied health. McCluskey and Bishop, who first reported about the validity and reliability of the AFT, urged researchers to develop new clinical scenarios and modify the instrument to suit other health disciplines.Citation12 The current study addressed this gap and provided researchers and educators an instrument to measure EBP skills and knowledge in speech pathologists, social workers, and dieticians/nutritionists. The new versions of the AFT were content-validated, and although the internal consistency of the different versions was slightly lower than the original AFT, the Cronbach’s α-values were still acceptable.

The reliability estimates for some of the items (questions 2, 4, 5, and 7) were questionable; however, after further training, the ICCs increased considerably, indicating moderate-to-excellent reliability of scores for these items. This finding highlights the importance of providing training to raters as a strategy to improve interrater reliability. Rater training has been shown to increase consistency of scoring between raters.Citation23 It emphasizes developing a common understanding among raters so they will apply the rating system as consistently as possible.Citation24 This common understanding, also called “frame of reference”, addresses the common sources of rater disagreements, which include lack of overlap among what is observed, discrepant interpretations of descriptor meanings, and personal beliefs or biases.Citation24 However, research also suggests that even comprehensive training will not ensure rater agreement.Citation25 Studies have suggested that a rater’s expertise may improve accuracy,Citation23,Citation26 which implies that rater characteristics are also an important consideration in ensuring consistency between raters. Reliability in examination scoring can be expected if the raters are highly knowledgeable in the domain in which ratings are made. Studies have found a relationship between rater expertise and rating accuracy, as well as the ability to differentiate between different domains in a rating scale.Citation24,Citation26 The raters involved in this study are experienced EBP educators and researchers, and these attributes could have contributed to the consistency in scoring. Because of their exposure to teaching, the raters may have already gained a wealth of experience in examination assessment, and could be expected to respond well to training. It is therefore not surprising to find that following training in AFT rating, the reliability estimates improved significantly for the previously questionable items. Based on the results of the current study, it appears that there are three important variables that can contribute to rater reliability: an explicit scoring criteria (ie, scoring rubric), raters’ training, and raters’ professional experience.

As with any study, this research has limitations that need to be considered when interpreting the results. First, the sample size may have been too small to produce sufficiently reliable results. Second, the expert panel was limited to four practitioners, which may not represent the collective set of views in the different professions. Third, the ability of the test to detect change following educational programs has not been tested.

Despite these limitations, the results of this study provide a valuable resource for EBP educators and researchers who require an objective instrument to measure knowledge and skills among social workers, speech pathologists, and dieticians/nutritionists.

Conclusion

The authors propose the use of AFT in evaluating the EBP knowledge and skills of social workers, speech pathologists, and dieticians/nutritionists. EBP educators and researchers should identify raters with experience in EBP teaching or those with previous EBP training, who should then receive training for AFT scoring. The reliability of raters should be evaluated before they participate in the actual assessment.

While the content validity, internal consistency, and reliability of the AFT have been shown in this study, further research is required to establish its sensitivity to detect change in knowledge and skills following an educational intervention for dieticians, speech pathologists, and social workers.

Disclosure

The authors report no conflicts of interest in this work.

References

  • MenonAKorner-BitenskyNKastnerMMcKibbonKAStrausSStrategies for rehabilitation professionals to move evidence-based knowledge into practice: a systematic reviewJ Rehabil Med200941131024103219893996
  • HeiweSKajermoKNTyni-LennéREvidence-based practice: attitudes, knowledge and behaviour among allied health care professionalsInt J Qual Health Care201123219820921242158
  • BahtsevaniCUdénGWillmanAOutcomes of evidence-based clinical practice guidelines: a systematic reviewInt J Technol Assess Health Car2004204427433
  • AlbertsMEastonDStroke best practices: a team approach to evidence-based careJ Natl Med Assoc200496Suppl 45S20S15104312
  • LeuferTCleary-HoldforthJEvidence-based practice: improving patient outcomesNurs Stand20092332353919441625
  • MetcalfeCLewinRWisherSPerrySBanniganKMoffettJKBarriers to implementing the evidence base in four NHS therapies: dietitians, occupational therapists, physiotherapists, speech and language therapistsPhysiotherapy2011878433441
  • JetteDUBaconKBattyCEvidence-based practice: beliefs, attitudes, knowledge, and behaviors of physical therapistsPhys Ther200383978680512940766
  • IlesRDavidsonMEvidence based practice: a survey of physiotherapists’ current practicePhysiother Res Int20061129310316808090
  • HockenberryMBrownTWaldenMBarreraPTeaching evidence-based practice skills in a hospitalJ Contin Educ Nurs2009401283219226996
  • IlicDAssessing competency in evidence based practice: strengths and limitations of current tools in practiceBMC Med Educ200995319656412
  • RamosKDSchaferSTraczSMValidation of Fresno test of competence in evidence based medicineBMJ2003326738431932112574047
  • McCluskeyABishopBThe Adapted Fresno Test of competence in evidence-based practiceJ Contin Educ Health Prof200929211912619530195
  • MastagliaBToyeCKristjansonLEnsuring content validity in instrument development: challenges and innovative approachesContemp Nurse200314328129112868667
  • LizarondoLMGrimmer-SomersKKumarSCrockettADoes journal club membership improve evidence uptake in different allied health disciplines: a pre-post studyBMC Res Notes2012558823106851
  • AmelangAInter-rater Reliability of the Clinical Practice Assessment System used to Evaluate Pre-service Teachers at Brigham Young University [master’s thesis]Provo (UT)Brigham Young University2009
  • RichmanJMakridesLPrinceBResearch methodology and applied statisticsPhysiother Can1980324253257
  • McCraeRRKurtzJEYamagataSTerraccianoAInternal consistency, retest reliability, and their implications for personality scale validityPers Soc Psychol Rev2011151285020435807
  • GreenMLGraduate medical education training in clinical epidemiology, critical appraisal, and evidence-based medicine: a critical review of curriculaAcad Med199974668669410386099
  • ParkesJHydeCDeeksJMilneRTeaching critical appraisal skills in health care settingsCochrane Database Syst Rev20013CD00127011686986
  • TaylorRSReevesBCEwingsPETaylorRJCritical appraisal skills training for health care professionals: a randomized controlled trial [ISRCTN46272378]BMC Med Educ2004413015585061
  • CoomarasamyAKhanKWhat is the evidence that postgraduate teaching in evidence based medicine changes anything? A systematic reviewBMJ20043297473101715514348
  • ShaneyfeltTBaumKBellDInstruments for evaluating education in evidence-based practice: a systematic reviewJAMA200629691116112716954491
  • BarrettSThe impact of training on rater variabilityInt Educ J2001214958
  • GrahamMMilanowskiAMillerJMeasuring and Promoting Inter-rater Agreement of Teacher and Principal Performance RatingsMadison (WI)Center for Educator Compensation Reform2012 Available from: http://cecr.ed.gov/pdfs/Inter_Rater.pdfAccessed June 13, 2013
  • HoytWTKernsMMagnitude and moderators of bias in observer ratings: a meta-analysisPsychol Methods199944403424
  • LizarondoLGrimmerKKumarSInter-rater reliability of Adapted Fresno Test across multiple ratersPhysiother Can201265213514024403674