8,337
Views
54
CrossRef citations to date
0
Altmetric
ARTICLES

Health Literacy INDEX: Development, Reliability, and Validity of a New Tool for Evaluating the Health Literacy Demands of Health Information Materials

, , , , , , , , , , , & show all
Pages 203-221 | Published online: 03 Oct 2012

Abstract

There is no consensus on how best to assess the health literacy demands of health information materials. Comprehensive, reliable, and valid assessment tools are needed. The authors report on the development, refinement, and testing of Health Literacy INDEX, a new tool reflecting empirical evidence and best practices. INDEX is comprised of 63 indicators organized into 10 criteria: plain language, clear purpose, supporting graphics, user involvement, skill-based learning, audience appropriateness, user instruction, development details, evaluation methods, and strength of evidence. In a sample of 100 materials, intercoder agreement was high: 90% or better for 52% of indicators, and above 80% for nearly all others. Overall scores generated by INDEX were highly correlated with average ratings from 12 health literacy experts (r = 0.89, p < .0001). Additional research is warranted to examine the association between evaluation ratings generated by INDEX and individual understanding, behaviors, and improved health. Health Literacy INDEX is a comprehensive tool with evidence for reliability and validity that can be used to evaluate the health literacy demands of health information materials. Although improvement in health information materials is just one aspect of mitigating the effects of limited health literacy on health outcomes, it is an essential step toward a more health literate public.

[Supplementary materials are available for this article. Go to the publisher's online edition of Journal of Health Communication for the following free supplemental resource: Health Literacy INDEX]

Health literacy, or the degree to which individuals have the capacity to obtain, communicate, process, and understand health information and services needed to make appropriate health decisions (U.S. Department of Health and Human Services, 2011), is an important predictor of health outcomes and health care utilization. The European Health Literacy Project has defined four dimensions of health literacy: accessing, understanding, appraising, and applying information to make everyday health decisions over the life span (Doyle, Cafferkey, & Fullam, Citation2012). The effects of limited health literacy on health and health care costs (Baker et al., Citation1996; Berkman, Pignone, DeWalt, & Sheridan, Citation2004; Gazmararian, Williams, Peel, & Baker, Citation2003; Nielsen-Bohlman, Panzer, & Kindig, 2004; Parker, Baker, Williams, & Nurss, Citation1995; Weiss, Hart, McGee, & D'Estelle, Citation1992; Weiss et al., Citation2005; Weiss & Palmer, Citation2004) have been conceptualized as being based on interactions between individuals' skill levels and the demands of health care and social systems (Nielsen-Bohlman et al., Citation2004). It is therefore critical to develop instruments that organizations can use to assess the health literacy demands of health materials in order to better understand the effects of health literacy on these important outcomes.

In past inquiries, investigating the health literacy demands of health information materials has often been narrowly defined as evaluating the reading level required to understand print materials. More than 300 articles have been published in public health and medical journals reporting on assessments of various types of health information materials, and these findings have been critical in drawing attention to the gap between materials' demands and the skills of the intended target audience (Nielsen-Bohlman et al., 2004; Rudd, Colton, & Schacht, Citation2000; Rudd, Comings, & Hyde, Citation2003). Because no consensus has been reached in the field on the best instrument with which to evaluate health information materials, various approaches have been used across prior studies. The most common approach has been to use readability formulas, which provide grade-level estimates of the reading difficulty of a piece of prose (i.e., words organized into sentence and paragraph format; Nielsen-Bohlman et al., 2004; Rudd et al., Citation2000; Rudd et al., Citation2003). Many different readability formulas exist, measuring somewhat different variables, but they are all generally based on aspects of writing style such as sentence and word length (Flesch, Citation1948; Fry, Citation1977; McLaughlin, Citation1969; Nielsen-Bohlman et al., Citation2004). However, these formulas capture only some of the cognitive demands that difficult health information materials place on users (Doak, Doak, & Root, Citation1996; Nielsen-Bohlman et al., 2004).

Therefore, other instruments have been developed in order to conduct more comprehensive evaluations of the health literacy demands of health information materials. The checklist developed by the National Cancer Institute, which describes key principles of effective low-literacy print materials (National Cancer Institute, 2003), and the Suitability Assessment of Materials instrument (Doak et al., Citation1996) capture multiple factors in addition to writing style that affect the demands of health information materials. The National Cancer Institute checklist assesses content, layout, and graphics; the Suitability Assessment of Materials instrument also includes the domains of learning motivation and cultural appropriateness. The Suitability Assessment of Materials instrument has provided useful assessments of health information materials, as described in the published literature (e.g., Kaphingst, Rudd, DeJong, & Daltroy, Citation2004; Kaphingst, Zanfini, & Emmons, Citation2006). However, these instruments do not have proven intercoder reliability, which can introduce subjective judgments into assessments of materials depending upon the characteristics of coders, and the validity of these existing instruments has not been systematically evaluated. In addition, neither the National Cancer Institute checklist nor the Suitability Assessment of Materials instrument include the domains of accessing and appraising health information materials. Although other instruments have been developed to evaluate health information materials, such as the EQIP tool (Moult, Franck, & Brady, Citation2004), and the Tool to Evaluate Materials Used in Patient Education (Clayton, Citation2009), these have not been widely adopted and do not comprehensively address all four domains of health literacy discussed earlier. Other instruments described in the literature are specific to certain types of health-related content, such as the DISCERN index (Charnock, Shepperd, Needham, & Gann, Citation1999; Khazaal et al., Citation2009) and the eHealth Engagement Scale (Lefebvre, Tada, Hilfiker, & Baur, Citation2010) or to single dimensions of health literacy, such as the Cultural Sensitivity Assessment Tool (Guidry & Walker, Citation1999) or the PMOSE/IKIRSCH instrument (Mosenthal & Kirsch, Citation1998), limiting their usefulness for other content or domains of health literacy.

Therefore, there is a need for a comprehensive, reliable, and valid instrument that can be used to evaluate the health literacy demands of health information materials that relate to accessing, understanding, appraising, and applying health information. Such an instrument could not only be critical in health literacy research but could also assist organizations and practitioners in choosing appropriate materials according to their objectives and audience. The present article describes the development of the Health Literacy INDEX, a comprehensive tool that can be used to evaluate the health literacy demands of health information materials. Our goals in creating the tool were twofold: (a) to develop a tool to be used in research examining the effects of health literacy demands of health information materials on health outcomes and (b) to help potential users of health information materials (i.e., organizations and practitioners) make informed decisions about which materials to use. By administering a standardized assessment of objective and observable criteria derived from health literacy science and practice, direct comparisons can be made between different materials. This article describes the development and testing of the tool, possible uses of the tool, and next steps to advance this area of the health literacy literature.

Method

Tool Development

Health Literacy INDEX was developed through an iterative process of literature review, operational definition, expert feedback, and reliability testing.

Literature Review

We conducted a literature review of the construct of health literacy to identify key concepts related to health literacy demands that might be included in the tool. In this literature review, we examined how health literacy demands of health information materials related to accessing (i.e., audience appropriateness, material development details), understanding (i.e., plain language, clearly stated purpose, use of graphics), appraising (i.e., evaluation findings), and applying (i.e., skill-based learning, user involvement, provision of instructions) health information. We searched the Medline, ERIC, PsycINFO, Social Work Abstracts, and Communication and Mass Media Complete databases using combinations of the following key words: health literacy, knowledge, comprehension, listening, speaking, writing, numeracy, reading, verbal communication skills, health information, risk information, access to health information, media literacy, health information seeking, discrimination in healthcare, cultural competency, language, and client–provider communication. We also reviewed existing assessment tools and guidelines developed for print and web-based materials (Bechtel & Davidhizar, Citation2000; Brach & Fraser, Citation2000; Center for Plain Language; Centers for Disease Control and Prevention, 1999, 2000; Davis, Williams, Marin, Parker, & Glass, Citation2002; Doak et al., Citation1996; DuBay, Citation2004; Gray & Leary, Citation1972; Grier & Bryant, Citation2005; Harvard School of Public Health; Klare, Citation1984; Kreuter, Lukwago, Bucholtz, Clark, & Sanders Thompson, 2003; Meade & Byrd, Citation1989; Meade, Diekmann, & Thornhill, Citation1992; Meyer, Citation1982; Michielutte, Bahnson, & Beal, Citation1990; Michielutte, Bahnson, Dignan, & Schroeder, Citation1992; Moriarty, Citation1995; National Cancer Institute, 2003; Office of Cancer Communications, 1989; Parks & Kreuter, Citation2007; Petty & Cacioppo, Citation1981; Pfizer Clear Communication Initiative, 2004; Plain Language Action and Information Network; Resnicow, Baranowski, Ahluwalia, & Braithwaite, Citation1999; Rudd, Kaphingst, Colton, Gregoire, & Hyde, 2004; Sabogal, Otero-Sabogal, Pasick, Jenkins, & Perez-Stable, Citation1996; Southern Institute on Children and Families, 2005; Stableford & Mettger, Citation2007; U.S. Securities and Exchange Commission, 1998; Wilson, Citation1995). We then developed a matrix of guidelines and practices. Although we did not quantify the strength of evidence for each guideline and/or practice because of wide variability in the evidence base, we considered whether research evidence existed for each guideline or practice and the number of studies or reviews that described and evaluated the guideline or practice. Best practices were defined as those practices consistently described in multiple resources or studies, or in reviews of the literature, as improving the health literacy demands of health information materials. We considered these best practices as possible indicators of good health literacy practice. This process led to an initial set of 76 indicators.

Operational Definitions

After the initial set of indicators was identified, we developed operational definitions for each that would allow coders to observe the indicator in health information materials. These operational definitions were refined during the reliability process, which is subsequently described. The first versions of these definitions were complex, using an ordinal response scale with different definitions for each level, and for materials delivered via different media. Because reliability was difficult to achieve using this approach, we adopted a simpler system in the third and fourth rounds of reliability testing. Each indicator had a single operational definition, stated in the positive (i.e., indicating good health literacy practice), and was coded as present or absent.

Reliability Testing

We conducted four rounds of reliability testing. For each round, we identified a convenience sample of 50 (rounds 1 and 2) or 100 (rounds 3 and 4) print, website, and audiovisual health information materials. The materials ranged widely on health topic (e.g., cancer, heart disease, sexually transmitted disease, mental health); format (e.g., brochures, toolkits); source/developer (e.g., government, non-profit organizations); target audience (e.g., public, patients); and outcome of interest (e.g., behavior change, awareness, education). All materials were independently coded by at least two master's-level coders who had been trained in using the INDEX tool. We calculated percent agreement and kappa value for each indicator to assess intercoder agreement (Cohen, Citation1960; Everitt & Hay, Citation1992).

The number of indicators used in each round of reliability testing and the characteristics of the samples are summarized in Table . After each round of reliability testing, the research team reviewed the results. Indicators with intercoder reliability falling below a priori thresholds of percent agreement (70%) and kappa (0.6) were reviewed and revised. Specifically, we identified items discordant between coders and reviewed the indicators and operational definitions as a team to identify and address possible sources of confusion. After conducting the first round of reliability testing, we split broad concepts into multiple, more specific indicators. After each subsequent round of reliability testing, indicator wording was then revised in order to make indicators and operational definitions less subjective and redundant indicators were deleted. The proportion of indicators that met the reliability threshold increased with each iteration of the tool and round of testing.

Table 1. Summary of reliability testing for Health Literacy INDEX tool

Expert Feedback

At five times during the development process, we sought feedback on the latest iteration of the tool from community and academic professionals directly involved in health literacy to refine the tool. A group (n = 12) of health literacy grantees and affiliated program staff from the Missouri Foundation for Health provided feedback twice, after the first and third rounds of reliability testing. A second group (n = 25) of community advisory network members from community-based agencies and organizations actively involved in health literacy provided feedback three times, before reliability testing and after the first and third rounds of reliability testing. These professionals critiqued the tool's criteria and indicators, suggested additions and deletions, and described how they might use the tool.

Examining Variability

After conducting the third round of intercoder reliability testing, we examined the level of variability across the sample on each indicator (i.e., what proportion of health information materials were meeting the standard set by the indicator). A major goal for the tool was to help users (e.g., organizations, practitioners) distinguish between health information materials with better and less favorable health literacy profiles, and indicators that were met by all resources would not help in making distinctions between materials. After calculating percent agreement and kappa for this round of reliability testing as described above, the two coders completed a consensus process to resolve all discrepancies in coding. Based on these consensus codes for 100 materials, we determined the proportion that met each indicator. We dropped indicators (n = 22) that were met by > 95% of materials from the tool to focus INDEX on criteria that varied across materials, thereby providing information that would inform the selection of health information materials.

Validity Testing

When we reached acceptable intercoder reliability and the final version of INDEX was established, we examined its validity. INDEX generates an overall score and criterion-specific scores for each material reviewed, as subsequently described. Using the overall score, we rank ordered 94 materials from highest (i.e., most favorable) to lowest (i.e., least favorable) score. We then selected eight materials from the top quartile and eight from the bottom quartile and created four distinct groups of materials. Each group contained two top quartile materials and two bottom quartile materials. We then recruited 12 academic researchers and practitioners currently working in health literacy to rate the materials, blind to the scores generated by INDEX. Each participant in the validity study rated one group of four materials on the basis of their own judgment, without using the INDEX tool. They were asked to sort the four materials into sets: two that they felt best reflected good health literacy practices and two that they felt did not reflect these practices. Each group of materials was evaluated by three different participants. We compared the expert ratings to ratings generated from the INDEX tool. The validity test activities were approved by the institutional review board at Washington University in St. Louis.

Analysis

We conducted all analyses using SAS 9.2 for Windows (Cary, NC). We calculated percent agreement and kappa values during reliability testing (Cohen, Citation1960; Everitt & Hay, Citation1992). In validity testing, we examined the association between INDEX scores and expert ratings using a Spearman correlation coefficient. After conducting reliability and validity testing, we examined descriptive statistics for a sample of materials that were evaluated using the final version of the INDEX tool.

Results

INDEX Tool

The final version of Health Literacy INDEX includes 63 indicators organized into 10 criteria: plain language, clear purpose, supporting graphics, user involvement, skill-based learning, audience appropriateness, user instruction, development details, evaluation methods, and strength of evidence (see Table ). These criteria address features of text (e.g., plain language, clear purpose), features of graphics (e.g., supporting graphics), and content characteristics (e.g., user involvement, skill-based learning, audience appropriateness, user instruction, developmental details, evaluation methods, and strength of evidence). The definitions for each criterion and indicator were created by the team on the basis of the literature review. When a health information material is evaluated using INDEX, it receives a “yes” or “no” score for each indicator. Some indicators are contingent upon whether a previous indicator was met (e.g., evaluation indicators are only rated for materials for which an evaluation was conducted). An overall INDEX score is computed by taking the proportion of “yes” answers out of the number of applicable indicators. This score can therefore range from 0% to 100%. In the final reliability test sample, overall INDEX scores ranged from 18% to 69%. The same methods are used to compute criterion-specific scores. The final criteria and indicators are subsequently described.

Table 2. Criteria addressed by the INDEX material evaluation tool

Plain Language

Eight indicators assessed reading grade level, writing style (i.e., use of active voice, second person, sentence length), use of jargon, and use and understandability of data. Intercoder agreement was high (range = 73–99%) and kappa was generally acceptable (above the 0.6 threshold for five indicators; see Table ). Among these indicators, materials were most likely to meet the indicators for presenting data (92–100%, when data were present) and least likely to meet the indicators for reading grade level (21%) and sentence length (61%).

Table 3. Reliability of INDEX material evaluation tool

Clear Purpose

Eight indicators assessed clarity of purpose, use of visual elements to communicate purpose, review of section topics, and number of key points. Intercoder agreement was high (range = 67–99%), although kappa values varied greatly. Most materials met the clear purpose indicators. However, less than half of materials met the indicator of reviewing major topics at the end of the material (45%), and only half (52%) limited the number of key points to five or fewer per section.

Supporting Graphics

Six indicators assessed whether graphics illustrated content, demonstrated behaviors, or helped readers understand abstract concepts, used data related to content, had explanatory labels, and had a simple design. Intercoder agreement was high for these indicators (range = 77–100%) as were kappa values (range = 0.6–1.0). Among these indicators, materials with graphics generally met the indicator of using graphics to illustrate content (89%). However, few materials with graphics included data as part of the graphics that were related to content (6%).

User Involvement

Seven indicators assessed the use of common user involvement strategies (i.e., advice column, frequently asked questions, quiz or game, story or narrative format or teasers), whether the audience was asked to do something in the material, and whether space was provided for audience response. Intercoder agreement was high (range = 86–100%) and kappa was generally acceptable (range = 0.51–0.78). All materials used a frequently asked questions format at least once. Materials were much less likely to meet the indicators of asking the audience to do an activity (21%) or providing space for audience response (22%).

Skill-Based Learning

Three indicators assessed whether materials explicitly addressed health behaviors, provided specific action steps, and illustrated these with examples. Intercoder agreement ranged from 79% to 82% and kappa values from 0.60 to 0.64. Although only about half (48%) met the indicator of explicitly addressing a health behavior, of those, almost all provided specific action steps and illustrated steps with examples.

Audience Appropriateness

Six indicators assessed whether the target audience was explicitly identified and whether the material used peripheral (visual cues), evidential (data), linguistic (appropriate language), and/or sociocultural (values, beliefs, experiences) strategies for cultural appropriateness. Intercoder agreement ranged from 65% to 95%, although kappa values varied from 0.14 to 0.70. Among materials that included graphics, most (77%) met the indicator of use of peripheral cues related to the target audience, and among materials that included data, most (80%) used evidential strategies. However, materials were less likely to meet indicators for stating the audience in the title (23%) or using sociocultural strategies on the basis of values, beliefs, and experiences of the audience (23%).

User Instructions

Six indicators assessed whether technical and user instructions were provided, easy to find, and comprehensive. Intercoder agreement was high (range = 76–86%) and kappa values were acceptable (0.62–0.80). Among those with instructions, materials were most likely to meet the indicator for providing comprehensive user instructions (53%), but were least likely to meet the indicators for providing comprehensive technical instructions (30%).

Development Details

Three indicators assessed whether a development date was provided, the recency of last review or development, and provision of contact information to reach the developer. Intercoder agreement ranged from 84 to 95%, with kappa values from 0.38 to 0.92. Materials were most likely to meet the indicator for providing contact information (84%) and least likely to meet the indicator for having reviewed or developed the materials within the last two years, of those that provided dates (66%).

Evaluation Methods

Thirteen indicators assessed whether a formative, process, or outcome evaluation was conducted and, if so, the evaluation methods used. Twenty materials had any type of evaluation; nine had a formative evaluation, four a process evaluation, and three an outcome evaluation. Of the few materials to which these criteria applied, intercoder reliability was high (range = 86–99%) and kappa values ranged from 0.47 to 1.0. Among materials with any type of evaluation information available, few met the indicators for conducting a process (27%) or outcome (13%) evaluation. Among materials subjected to an outcome evaluation, all met the four indicators for methods used.

Strength of Evidence

Three indicators assessed whether the materials incorporated findings from formative, process or outcome evaluations. Intercoder reliability ranged from 87% to 98% and kappa values ranged from 0.19 to 0.85. Of the few materials with outcome evaluation findings, all met the indicator of reporting a significant effect of the material. Materials with formative evaluation findings were less likely to meet the indicator that the material explicitly reflected these findings (46%).

Indicators Not Included in Final Version of INDEX

In the penultimate round of reliability testing, we eliminated 22 indicators of good health literacy practice because of limited variability (see Table ). Many of these indicators are widely recognized best practices in health literacy, and their exclusion from the final version of the tool in no way reflects doubts about their importance, but rather an effort to create a more parsimonious tool.

Table 4. Indicators removed during refinement of INDEX

Validity

We found that overall scores generated by INDEX were highly correlated with average ratings from 12 health literacy experts (r = 0.89, p < .0001).

Discussion

This article describes the development and testing of INDEX, a materials evaluation tool comprising 63 indicators organized into 10 criteria. INDEX is a comprehensive material evaluation tool for assessing the health literacy demands of health information materials. The tool assesses a number of major domains, including characteristics of text, characteristics of graphics, and content characteristics. The tool builds upon accrued knowledge from several decades of health literacy research and practice, but brings strengths in intercoder reliability, validity testing, and comprehensiveness to address gaps in the literature. Because INDEX can be reliably administered by trained coders and has demonstrated validity against expert ratings, it should have value to researchers and practitioners in health literacy.

Although it was not the primary goal of this project, our findings also highlight places for improvement in health information materials. The INDEX results showed that most materials required a reading level of greater than eighth grade, a finding that is consistent with many other assessments of health information materials (Nielsen-Bohlman et al., 2004; Rudd et al., Citation2000). These findings therefore suggest that the language used in materials could not be understood by many U.S. adults (Doak et al., Citation1996), and that there is still a need for improvement in the use of plain language and techniques such as definitions of technical jargon. Some aspects of graphics could be improved, particularly making use of graphics to illustrate health behaviors or show abstract concepts (e.g., Doak et al., Citation1996). Strategies to actively engage users were rare in these materials, and users could benefit from the development of materials that ask them to actively participate and respond to information.

The findings suggest that many materials may not be appropriate for their intended audience. The target audience was identified in less than half of the materials, which may make it more difficult for organizations to determine how well the material would meet their needs. Although peripheral strategies for audience appropriateness (e.g., showing pictures of the target audience) were sometimes used, deeper sociocultural strategies that reflect the audience's values, beliefs and experiences were rarely used. This is consistent with findings from previous reviews (Parks & Kreuter, Citation2007). Application of sociocultural strategies based on clearly defined target audiences could improve the effectiveness of health information materials (Kreuter et al., Citation2003; Kreuter & McClure, Citation2004; Resnicow et al., Citation1999). Greater use of easy-to-find and comprehensive instructions could also assist material users.

Last, few materials reported having conducted any type of evaluation activities, and reporting on outcome evaluations was particularly rare. There is a strong need to evaluate health information materials in order to ensure that they meet the needs of the target audience, are implemented well and reach the target audience, and have the desired outcomes (Office of Cancer Communications, 1989; Valente, Citation2002). We acknowledge the possibility that materials might have been evaluated without the information being made public. However, this would not allow practitioners selecting materials to draw upon such evidence; we encourage materials developers to make this information public. In nearly all areas of public health and medicine, the use of evidence-based approaches and programs is increasingly important (Brownson, Baker, Leet, Gilliespie, & True, Citation2010; Sackett, Rosenberg, Gray, Haynes, & Richardson, 1996; Victora, Habicht, & Bryce, Citation2004), and this emphasis will likely be of increasing importance in health literacy as well. The scores generated by INDEX do not constitute evidence of effectiveness in affecting health-related outcomes but are rather indicators of adherence to practices shown or believed to affect health literacy demands. In the absence of evidence about specific materials, these kinds of indicators may be useful in guiding material development and selection in practice.

During the process of developing INDEX, we found that the health information materials evaluated were consistently rated as having good organization of content (e.g., chunking, organization, consistency); effective use of some layout and typography features (e.g., consistent layout, visual cuing devices, adequate white space, color, and high contrast); and consistent use of clear headers. These results suggest that the extensive public and private efforts to develop and disseminate plain language and health literacy trainings and materials may be having an impact in improving health information materials. We have included the removed criteria in a supplement to INDEX so that they can be incorporated into document development processes. We consider these criteria to be “minimal qualifying criteria” for evaluation by INDEX. In other words, materials that do not meet these basic criteria are not suitable for evaluation by the tool and would not be considered to be useful in practice. This then focuses the INDEX criteria on a more parsimonious set of indicators found to differentiate between health information materials in terms of health literacy demands.

The INDEX tool builds on the existing literature in health literacy and provides a way to evaluate a fairly comprehensive set of indicators to evaluate the health literacy demands of health information materials, but limitations of the study should be considered in interpreting our results. Because the indicators are scored as present or absent, variations across a single material are not captured. Given the convenience nature of the samples of materials reviewed, findings about the extent to which different indicators were met in these samples should not be generalized to all health information materials. Almost all of the health information materials were developed in the United States, and the reliability and validity of the tool in other social contexts needs to be evaluated. In addition, some of the discarded indicators may still be applicable in other countries with less health literacy experience. Creating a supplement to INDEX that includes these discarded items and testing the usefulness of this information to materials developers is an important next step. Most of the materials reviewed were print materials, and the reliability and validity of INDEX for social and interactive games needs to be determined. In addition, while many of the indicators could be reliably measured after a number of iterations of testing and development, some indicators still had lower kappa values. Kappa values for some indicators were low primarily due to the distribution of the underlying phenomenon or small sample sizes for which the indicator was applicable; this may explain why kappa values were sometimes near zero despite high percentage agreement between coders. In the validity testing, we asked experts to sort materials from the top and bottom quartiles of health information materials into two groups. Further validity testing could explore whether INDEX scores are related to ranked or more finely graded expert ratings in order to explore the validity of the instrument for materials scoring in the middle of the scale. Future research could explore whether a subset of criteria or indicators are related to usefulness for organizational users who are focused on improving understanding rather than access to information.

Additional research is needed to examine the association between INDEX scores and individuals' experience with health information materials and health-related outcomes. While many of the indicators in the tool are based on best practices and existing literature, the research base needs to be strengthened. Prior studies have shown that while improving readability may increase comprehension (Nielsen-Bohlman et al., 2004), this is not always be the case (Davis et al., Citation1998) and it is not clear what factors affect the differences in this association. One existing study showed that a large-scale plain language effort did not show significant improvements in prescription drug utilization (Shrank et al., Citation2009), but the research base for this type of large-scale initiative is limited. More studies are needed to address how and when changing the health literacy demands of health information materials affects downstream outcomes. Examining the associations between overall and criterion-specific INDEX scores and understanding, health behaviors, and health outcomes as part of intervention research studies can help us to understand the role of health literacy demands of health information materials in affecting individual outcomes. Important practice questions about implementation must also be answered, for example, understanding how providers and organizations would use the tool and how to make INDEX-based information publically available and easily accessible.

Conclusion

Health Literacy INDEX is a comprehensive tool that can be used to evaluate the health literacy demands of health information materials, and has demonstrated intercoder reliability and validity against ratings of 12 experts. In addition to possible use as a tool for organizations selecting health information materials, INDEX can also be used as a checklist for developers of materials. The criteria-level scores generated by the tool can guide developers in refining drafts of materials. As with any coding instrument, we found that training and practice in applying the INDEX tool was critical for coding reliability. Developing a training and certification process for coders is an important next step to making INDEX practical for widespread use while assuring the integrity of its scores. This type of training could be web-based with some didactic review of INDEX, followed by practice coding of sample materials and comparison of coders' ratings to standard ratings. We expect that this training would require 3–4 hours total.

We welcome feedback and dialogue on the indicators and criteria in the INDEX tool, and look forward to further refining the tool and filling in gaps. Multiple efforts to develop material evaluation tools like INDEX are already underway and being led by experienced leaders in health literacy. The field will benefit from these efforts and from continued collaboration to generate an essential set of criteria on the basis of best practices, evidence, and consensus. While improvement in health information materials is just one aspect of mitigating the effects of limited health literacy on health outcomes, it is an essential step toward patient activation, a better equipped professional field, and a more health literate public.

Supplemental material

uhcm_a_712612_sup_28758589.docx

Download MS Word (93.9 KB)

Acknowledgments

This research was supported by a contract from the Missouri Foundation for Health and by funding from the Barnes-Jewish Hospital Foundation. The authors thank the health literacy grantees and affiliated program staff from the Missouri Foundation for Health and the Community Advisory Network members for providing feedback during the tool development process. The authors also thank those who agreed to participate in our validity testing process for their invaluable contribution.

Notes

a Number of health information materials to which criterion applied.

References

  • Baker , D. W. , Parker , R. M. , Williams , M. V. , Pitkin , K. , Parikh , N. S. , Coates , W. , & Imara , M. ( 1996 ). The health care experiences of patients with low literacy . Archives of Family Medicine , 5 , 329 – 334 .
  • Bechtel , G. , & Davidhizar , R. ( 2000 ). Integrating cultural diversity in patient education . Seminars in Nurse Management , 7 , 193 – 197 .
  • Berkman , N. , Pignone , M. P. , DeWalt , D. , & Sheridan , S. ( 2004 ). Health Literacy: Impact on Health Outcomes . Rockville , MD : Agency for Healthcare Research and Quality .
  • Brach , C. , & Fraser , I. ( 2000 ). Can cultural competency reduce racial and ethnic health disparities? A review and conceptual model . Medical Care Research and Review , 57 , 181 – 217 .
  • Brownson , R. C. , Baker , E. A. , Leet , T. L. , Gilliespie , K. N. , & True , W. R. ( 2010 ). Evidence-based public health () , 2nd ed. . New York , NY : Oxford University Press .
  • Center for Plain Language. (n. d.). Falls Church, VA: Author. Retrieved from http://www.centerforplainlanguage.org/index.html
  • Centers for Disease Control, & Prevention. ( 1999 ). Scientific and technical information: Simply put . Atlanta , GA : Author .
  • Centers for Disease Control, & Prevention. ( 2000 ). Beyond the Brochure: Alternative Approaches to Health Education . Atlanta , GA : Author .
  • Charnock , D. , Shepperd , S. , Needham , G. , & Gann , R. ( 1999 ). DISCERN: An instrument for judging the quality of written consumer health information on treatment choices . Journal of Epidemiology and Community Health , 53 , 105 – 111 .
  • Clayton , L. ( 2009 ). TEMPtEd: Development and psychometric properties of a tool to evaluate material used in patient education . Journal of Advanced Nursing , 65 , 2229 – 2238 .
  • Cohen , J. ( 1960 ). A coefficient of agreement for nominal scales . Educational and Psychological Measurement , 20 , 37 – 46 .
  • Davis , T. , Fredrickson , D. D. , Arnold , C. , Murphy , P. W. , Herbst , M. , & Bocchini , J. A. , Jr. ( 1998 ). A polio immunization pamphlet with increased appeal and simplified language does not improve comprehension to an acceptable level . Patient Education and Counseling , 33 , 25 – 37 .
  • Davis , T. , Williams , M. V. , Marin , E. , Parker , R. M. , & Glass , J. ( 2002 ). Health literacy and cancer communication . CA: A Cancer Journal for Clinicians , 52 , 134 – 149 .
  • Doak , C. C. , Doak , L. G. , & Root , J. H. ( 1996 ). Teaching patients with low literacy skills () , 2nd ed. . Philadelphia , PA : Lippincott .
  • Doyle , G. , Cafferkey , K. , & Fullam , J. ( 2012 ). The European Health Literacy Survey: Results from Ireland . Dublin , Ireland : University College Dublin .
  • DuBay , W. ( 2004 ). The principles of readability . Costa Mesa , CA : Impact Information .
  • Everitt , B. S. , & Hay , D. F. ( 1992 ). Talking about statistics: A psychologist's guide to data analysis . New York , NY : Halsted Press .
  • Flesch , R. ( 1948 ). A new readability yardstick . Journal of Applied Psychology , 32 , 221 – 223 .
  • Fry , E. ( 1977 ). Fry's readability graph: Clarifications, validity, and extensions to level 17 . Journal of Reading , 21 , 242 – 252 .
  • Gazmararian , J. A. , Williams , M. , Peel , J. , & Baker , D. ( 2003 ). Health literacy and knowledge of chronic disease . Patient Education and Counseling , 51 , 267 – 275 .
  • Gray , W. , & Leary , B. ( 1972 ). What makes a book readable . Chicago , IL : Chicago University Press .
  • Grier , S. , & Bryant , C. ( 2005 ). Social marketing in public health . Annual Review of Public Health , 26 , 319 – 339 .
  • Guidry , J. J. , & Walker , V. D. ( 1999 ). Assessing cultural sensitivity in printed cancer materials . Cancer Practice , 7 , 291 – 296 .
  • Harvard School of Public Health. (n. d.). Health Literacy Studies. Boston, MA: Harvard University. Retrieved from http://www.hsph.harvard.edu/healthliteracy
  • Kaphingst , K. A. , Rudd , R. E. , DeJong , W. , & Daltroy , L. H. ( 2004 ). Literacy demands of product information intended to supplement television direct-to-consumer prescription drug advertisements . Patient Education and Counseling , 55 , 293 – 300 .
  • Kaphingst , K. A. , Zanfini , C. , & Emmons , K. M. (2006). Accessibility of web sites containing colorectal cancer information to adults with limited literacy. Cancer Causes and Control , 17, 147–151.
  • Khazaal , Y. , Chatton , A. , Cochand , S. , Coquard , O. , Fernandez , S. , Khan , R. , … Zullino , D. ( 2009 ). Brief DISCERN, six questions for the evaluation of evidence-based content of health-related websites . Patient Education and Counseling , 77 , 33 – 37 .
  • Klare , G. R. ( 1984 ). Readability . In P. D. Pearson , R. Barr , M. L. Kamil & P. Mosenthal (Eds.), Handbook of Reading Research (pp. 681 – 744 ). New York , NY : Longman .
  • Kreuter , M. , Lukwago , S. , Bucholtz , R. , Clark , E. M. , & Sanders Thompson , V. ( 2003 ). Achieving cultural appropriateness in health promotion programs: Targeted and tailored approaches . Health Education & Behavior , 30 , 133 – 146 .
  • Kreuter , M. , & McClure , S. M. ( 2004 ). The role of culture in health communication . Annual Review of Public Health , 25 , 439 – 455 .
  • Lefebvre , R. C. , Tada , Y. , Hilfiker , S. W. , & Baur , C. ( 2010 ). The assessment of user engagement with eHealth content: The eHealth Engagement Scale . Journal of Computer-Mediated Communication , 15 , 666 – 681 .
  • McLaughlin , G. ( 1969 ). SMOG grading: A new readability formula . Journal of Reading , 12 , 639 – 646 .
  • Meade , C. D. , & Byrd , J. C. ( 1989 ). Patient literacy and the readability of smoking education literature . American Journal of Public Health , 79 , 204 – 206 .
  • Meade , C. D. , Diekmann , J. , & Thornhill , D. G. ( 1992 ). Readability of American Cancer Society patient education literature . Oncology Nursing Forum , 19 , 51 – 55 .
  • Meyer , B. ( 1982 ). Reading research and the composition teacher: The importance of plans . College Composition and Communication , 33 , 37 – 49 .
  • Michielutte , R. , Bahnson , J. , & Beal , P. ( 1990 ). Readability of the public education literature on cancer prevention and detection . Journal of Cancer Education , 5 , 55 – 61 .
  • Michielutte , R. , Bahnson , J. , Dignan , M. B. , & Schroeder , E. M. ( 1992 ). The use of illustrations and narrative text style to improve readability of a health education brochure . Journal of Cancer Education , 7 , 251 – 260 .
  • Moriarty , S. ( 1995 ). Visual communication as primary system . Journal of Visual Literacy , 14 , 11 – 21 .
  • Mosenthal , P. B. , & Kirsch , I. S. ( 1998 ). A new measure for assessing document complexity: The PMOSE/IKIRSCH document readability formula . Journal of Adolescent and Adult Literacy , 41 , 638 – 657 .
  • Moult , B. , Franck , L. S. , & Brady , H. ( 2004 ). Ensuring quality information for patients: Development and preliminary validation of a new instrument to improve the quality of written health care information . Health Expectations , 7 , 165 – 175 .
  • National Cancer Institute. ( 2003 ). Clear & simple: Developing effective print materials for low-literate readers. Retrieved from http://www.cancer.gov/aboutnci/oc/clear-and-simple
  • Nielsen-Bohlman , L. , Panzer , A. M. , Kindig , D. A. ( Eds.). (2004 ). Health Literacy: A Prescription to End Confusion . Washington , DC : National Academies Press .
  • Office of Cancer Communications. ( 1989 ). Making health communication programs work: A planner's guide . Bethesda , MD : Office of Cancer Communications, National Cancer Institute .
  • Parker , R. M. , Baker , D. W. , Williams , M. V. , & Nurss , J. R. ( 1995 ). The Test of Functional Health Literacy in Adults: A new instrument for measuring patients' literacy skills . Journal of General Internal Medicine , 10 , 537 – 541 .
  • Parks , S. , & Kreuter , M. ( 2007 ). Cultural appropriateness in interventions for racial and ethnic minorities . In L. Doll , S. Bonzo , D. Steet , J. Mercy & E. N. Haas (Eds.), Handbook of injury and violence prevention (pp. 449 – 462 ). New York , NY : Springer .
  • Petty , R. E. , & Cacioppo , J. T. ( 1981 ). Attitudes and Persuasion: Classic and Contemporary Approaches . Dubuque , IA : Brown .
  • Pfizer Clear Health Communication Initiative. ( 2004 ). Pfizer principles for clear health communication. New York, NY: Author. Retrieved from http://www.pfizerhealthliteracy.com/asset/pdf/PfizerPrinciples.pdf
  • Plain Language Action, & Information Network. (n. d.). Retrieved from http://www.plainlanguage.gov
  • Resnicow , K. , Baranowski , T. , Ahluwalia , J. , & Braithwaite , R. ( 1999 ). Cultural sensitivity in public health: Defined and demystified . Ethnicity and Disease , 9 , 10 – 21 .
  • Rudd , R. E. , Colton , T. , & Schacht , R. ( 2000 ). An overview of medical and public health literature addressing literacy issues: An annotated bibliography . Boston , MA : National Center for the Study of Adult Learning and Literacy .
  • Rudd , R. E. , Comings , J. P. , & Hyde , J. N. ( 2003 ). Leave no one behind: Improving health and risk communication through attention to literacy . Journal of Health Communication , 8 ( Suppl 1 ), 104 – 115 .
  • Rudd , R. E. , Kaphingst , K. , Colton , T. , Gregoire , J. , & Hyde , J. ( 2004 ). Rewriting public health information in plain language . Journal of Health Communication , 9 , 195 – 206 .
  • Sabogal , F. , Otero-Sabogal , E. , Pasick , R. , Jenkins , C. , & Perez-Stable , E. ( 1996 ). Printed health education materials for diverse communities: Suggestions learned from the field . Health Education Quarterly , 23 , S123 – S141 .
  • Sackett , D. L. , Rosenberg , W. M. , Gray , J. M. , Haynes , R. B. , & Richardson , W. S. ( 1996 ). Evidence based medicine: What it is and what it isn't . BMJ , 312 , 71 .
  • Shrank , W. H. , Gleason , P. P. , Canning , C. , Walters , C. , Heaton , A. H. , & Jan , S. , … Choudhry , N. K. ( 2009 ). Can improved prescription medication labeling influence adherence to chronic medications? An evaluation of the Target pharmacy label . Journal of General Internal Medicine , 24 , 570 – 578 .
  • Southern Institute on Children, & Families. ( 2005 ). Health literacy style manual. Reston, VA: Author. Retrieved from http://www.coveringkidsandfamilies.org/resources/docs/stylemanual.pdf
  • Stableford , S. , & Mettger , W. ( 2007 ). Plain language: A strategic response to the health literacy challenge . Journal of Public Health Policy , 28 , 71 – 93 .
  • U.S. Department of Health, & Human Services. ( 2011 ). Healthy people 2020. Retrieved from http://www.healthypeople.gov/2020/about/default.aspx
  • U.S. Securities, & Exchange Commission. ( 1998 ). A plain language handbook: How to create clear SEC disclosure documents . Washington , DC : Author .
  • Valente , T. W. ( 2002 ). Evaluating health promotion programs . New York , NY : Oxford University Press .
  • Victora , C. G. , Habicht , J.-P. , & Bryce , J. ( 2004 ). Evidence-based public health: Moving beyond randomized trials . American Journal of Public Health , 94 , 400 – 405 .
  • Weiss , B. D. , Hart , G. , McGee , D. L. , & D'Estelle , S. ( 1992 ). Health status of illiterate adults: Relation between literacy and health status among persons with low literacy skills . Journal of the American Board of Family Practice , 5 , 257 – 264 .
  • Weiss , B. D. , Mays , M. Z. , Martz , W. , Castro , K. M. , DeWalt , D. A. , Pignone , M. P. , … Hale , F. A. ( 2005 ). Quick assessment of literacy in primary care: The newest vital sign . Annals of Family Medicine , 3 , 514 – 522 .
  • Weiss , B. D. , & Palmer , R. ( 2004 ). Relationship between health care costs and very low literacy skills in a medically needy and indigent Medicaid population . Journal of the American Board of Family Medicine , 17 , 44 – 47 .
  • Wilson , F. L. ( 1995 ). Measuring patients' ability to read and comprehend: A first step in patient education . Nursing Connections , 8 , 17 – 25 .