Abstract
Comprehensive patient education is necessary for shared decision-making. While patient–provider conversations primarily drive patient education, patients also use published materials to enhance their understanding. In this investigation, we evaluated the readability of 2585 patient education materials published in high-impact medical journals from 1998 to 2018 and compared our findings to readability recommendations from national groups. For all materials, mean readability grade levels ranged from 11.2 to 13.8 by various metrics. Fifty-four (2.1%) materials met the American Medical Association recommendation of sixth grade reading level, and 215 (8.2%) met the National Institutes of Health recommendation of eighth grade level. When stratified by journal and material type, general medical education materials from Annals of Internal Medicine were the most readable (P < .001), with 79.8% meeting the eighth grade level. Readability did not differ significantly over time. Efforts to standardize publication practice with the incorporation of readability evaluation during the review process may improve patients’ understanding of their disease processes and treatment options.
Introduction
Health literacy has been shown to strongly correlate with patient outcomes (1,2). Low health literacy is associated with more hospitalizations, greater use of emergency care, and lower receipt of preventative care measures such as mammograms and influenza vaccines (3). Although the mechanisms underpinning these complex relationships remain incompletely understood, it has been suggested that a combination of patient-specific and systemic factors contribute to well-documented disparities among patients with varying levels of health literacy (4). One proposed systemic factor is the presence of communication barriers between health care consumers and medical professionals, which can disproportionately affect patients with low health literacy (5,6). Patients who are unable to fully comprehend medical information may be at a disadvantage as active participants in their own health care decisions, potentially compromising their access to the highest quality of care (7). Unfortunately, overcoming obstacles of communication can be increasingly difficult as the breadth and complexity of medical care continues to grow (8). It is therefore paramount that action be taken to improve patient–provider communication practices on a large scale.
While direct interfacing with medical professionals is often the most important source of health care education for patients, high-quality written educational material can serve as a valuable adjunct (9). As such, the creation of accurate and understandable written patient education materials represents an opportunity to promote informed medical decision-making. Although the design of effective written educational materials relies upon myriad factors such as content and organization, the readability, defined as the “ease of understanding or comprehension due to the style of writing” also contributes to patient comprehension (10). The American Medical Association (AMA) and National Institutes of Health (NIH) recommend that patient materials be written at the sixth and eighth grade reading level, respectively (11,12).
The aim of this investigation is to evaluate whether patient education materials from widely circulated, high-impact medical journals are written in a way that is understandable for the general public. We hypothesized that such materials would be written at reading levels above national recommendations and therefore may be suboptimal as tools to educate diverse patient populations.
Materials and Methods
Journal Query and Material Identification
The publication history and article submission guidelines of high-impact medical journals were searched in order to identify peer-reviewed patient education materials with potential for broad readership. High-impact journals were defined as those having an impact factor of 10.0 or higher according to the 2017 Journal Citation Report (JCR; Clarivate Analytics). Results were filtered to include only journals falling under 60 predetermined categories in the JCR database (Supplemental Table 1) with potential medical relevance determined upon manual review of all listed categories. For each journal, the entire publication history was searched to identify any patient education materials. Only documents published on or before December 31, 2018 which clearly designate patients or laypersons as the target audience were included. Because data collection was initiated in May 2019, this end date was selected in order to ensure that all time series data reflected full calendar year patterns. For each material, if an electronic version was available, the main text was copied directly into Word (Microsoft) for subsequent analysis. In cases where an electronic version was not readily available, materials were downloaded in PDF format and text was extracted to Word using Acrobat Reader (Adobe) and formatted for analysis.
Identified materials were further categorized as “general education” materials or “research lay summaries” according to their intended scope. For example, some journals publish materials designed to explain general medical conditions, therapies, or concepts to patients (eg, hypertension) and others publish materials designed to explain important research findings to a layperson audience (eg, describing results from a clinical trial). All materials were labeled accordingly and analyzed separately.
Readability Analysis
Readability analysis was conducted using Readability Studio 2012 (Oleandar Software). Indices included in this study were Degrees of Reading Power—Grade Equivalent (DRP-GE) (13); Flesch-Kincaid (FK) readability test (14); Ford, Caylor, Sticht (FORCAST) index (15); Fry score (16); Gunning Fog index (17); Raygor estimate (18); and Simple Measure of Gobbledygook grade (19). These metrics were chosen because they are well-validated and report in grade-level equivalents, therefore allowing for meaningful comparisons across scores. Readability indices are calculated based on textual parameters such as average sentence length, use of polysyllabic words, and frequency of difficult or uncommon words. Individual calculation formulae for each of the included metrics are provided in Supplemental Table 2.
Because every metric is calculated differently, each may be prone to specific biases. For example, the FK score, which is among the most commonly used due to ease of calculation, may underestimate difficulty of materials that contain unfamiliar but short words, as the FK score is dependent only on average sentence and word length. It is therefore often recommended, as was performed in the present study, to use multiple readability metrics in order to minimize the bias of any individual metric on overall results and data interpretation (20).
Statistical Analysis
One-sample t tests were used to compare readability scores of the sample population to national recommendations and 1-way analysis of variance was used to compare readability scores across categorical variables. Chi-square testing was used to compare proportions of materials meeting recommended readability grade levels from various national groups. For this analysis, the most permissive (eg, lowest) score for each material was used and scores were rounded to the nearest whole number grade level. For example, a reported grade level of 8.4 was considered to meet the eighth grade recommendation, while a reported grade level of 8.5 was not. Statistical analyses were performed with R 3.5.1 (R Foundation).
Results
Two thousand five hundred and eighty-five education materials for patients from 10 high-impact medical journals were identified. Six of the included journals were subspecialty journals of the Journal of the American Medical Association (JAMA) Network (JAMA, JAMA Cardiology, JAMA Dermatology, JAMA Internal Medicine, JAMA Oncology and JAMA Pediatrics). The others were the American Journal of Respiratory and Critical Care Medicine (AJRCC), Circulation, Annals of the Rheumatic Diseases (ARD), and Annals of Internal Medicine (AIM). The majority of journals published general education materials for patients; only ARD and AIM published research lay summaries. AIM was the only journal to publish both types of materials.
Mean readability grade levels for materials are summarized in Table 1, with results described individually by the included indices. Readability grade levels were above national recommendations for the majority of journals, regardless of metric. In aggregate, materials were written at a mean readability grade level of 11.2 to 13.8 according to the 7 included metrics. AIM general education materials had the lowest mean readability level by all indices, with means ranging from the 6.8 to 10.5 grade level by various scores. However, AIM research lay summaries were written at mean reading grade levels ranging from 11.2 to 13.7 by the included metrics.
Mean Readability Grade Level of Patient Education Educational Materials From High-Impact Medical Journals According to 7 Readability Indices.
Abbreviations: AIM (Gen Ed) = Annals of Internal Medicine, General Patient Education; AIM (Research), Annals of Internal Medicine, Research Article Lay Summaries; AJRCC, American Journal of Respiratory and Critical Care Medicine; ARD = Annals of the Rheumatic Diseases; DRP-GE, Degrees of Reading Power—Grade Equivalent; FORCAST, Ford, Caylor, Sticht Index; JAMA, Journal of the American Medical Association; JAMA IM, Journal of the American Medical Association, Internal Medicine; SMOG, Simple Measure of Gobbledygook.
The proportions of materials meeting national recommendations are summarized in Figure 1. Further journal-specific information is provided in Supplemental Table 3. The majority of materials failed to meet either the recommended sixth or eighth grade level. In total, 54 (2.1%) of 2585 met the sixth grade level and 215 (8.2%) met the eighth grade level. When analyzed separately by journal and material type, AIM general education materials had the greatest proportion of materials meeting either recommendation (P < .001), with 42.6% meeting the AMA-recommended sixth grade level and 79.8% meeting the NIH-recommended eighth grade level.

Proportion of materials meeting national readability recommendations. AIM (Gen Ed) indicates Annals of Internal Medicine, General Patient Education; AIM (Research), Annals of Internal Medicine, Research Article Lay Summaries; AJRCC, American Journal of Respiratory and Critical Care Medicine; ARD, Annals of the Rheumatic Diseases; JAMA, Journal of the American Medical Association; JAMA IM, Journal of the American Medical Association, Internal Medicine. *Small sample size (n = 2).
Figure 2 describes the readability grade levels of published materials over time from 1998 through 2018, again stratified by journal. Individual points represent the readability grade level for materials published in each year, calculated as the mean score of all included readability metrics. Publication dates of the AIM general education materials were not readily available so these materials were excluded from Figure 2. Regardless of journal or material type, readability scores were consistently well above recommendations throughout the analyzed time period, without significant changes over time.

Mean readability grade levels of materials over time, stratified by journal. The AMA recommends that patient materials be written at the sixth grade level (red dotted line); the NIH recommends the eighth grade level (blue dotted line). AIM (Gen Ed) indicates Annals of Internal Medicine, general patient education; AIM (Research), Annals of Internal Medicine, Research Article Lay Summaries; AJRCC, American Journal of Respiratory and Critical Care Medicine; ARD, Annals of the Rheumatic Diseases; JAMA, Journal of the American Medical Association; JAMA IM, Journal of the American Medical Association, Internal Medicine; NIH, National Institutes of Health.
Raygor distributions describing the readability of materials are presented in Figure 3, with all materials from the JAMA network combined in panel F. The Raygor estimate is calculated by plotting the number of long words (defined as 6 or more characters) per 100 words on the x-axis and the number of sentences per 100 words on the y-axis. The location of the resulting point on the Raygor grid provides an estimate of the readability of that particular text. Use of short, simple sentences and avoidance of long words therefore translates into a lower grade level score. The Raygor estimate distributions in Figure 3 redemonstrate that the majority of materials are written at inappropriately high reading levels and fail to meet readability recommendations. Furthermore, the majority of points fall below the central demarcation in each Raygor grid, suggesting that use of long words is more likely to be driving increased scores by this metric. Raygor estimates of readability, along with all other metrics used the current study, have been widely applied in the setting of patient education and other clinical scenarios such as evaluation of consent forms or templates for recording patient-reported outcomes (21 –24).

Readability of materials as measured by the Raygor estimate. A, American Journal of Respiratory and Critical Care Medicine; (B) Annals of Internal Medicine (General Education); (C) Annals of Internal Medicine (Research Lay Summaries); (D) Annals of the Rheumatic Diseases; (E) Circulation; (F) Journal of the American Medical Association Network, including all subspecialty journals. Raygor estimates with high percentages of long words or short sentences are considered invalid.
Discussion
In this investigation evaluating the readability of peer-reviewed patient education materials published in high-impact medical journals, we found that the vast majority are written at reading levels far above national recommendations of sixth and eighth grade. Other than AIM general education materials, which encouragingly had a large proportion (79.8%) meeting the NIH target of an eighth grade level, this trend was apparent in all journals and spanned the entire study period from 1998 through 2018. These results reveal an important opportunity to enhance patient education and communication practices on a large scale. Modification of publication procedures by medical journals, for example, through the introduction of readability requirements for patient materials, might lead to meaningful improvements in how patients, particularly those with low health literacy, understand and make medical decisions within an increasingly complex health care system.
Comprehensive and accurate education of patients is a fundamental prerequisite for medical decision-making. Traditional paternalistic models of decision-making held physicians as the sole authoritative party responsible for all aspects of every medical interaction, from diagnosis to treatment choice (25). Under this model, there is minimal need for patient education, as physicians made most decisions (26). However, the shared decision-making model has more recently emerged as the preferred approach for achieving high-quality patient-centered care (27,28). The principles of shared decision-making propose that patients and physicians participate equally in medical decisions, weighing available evidence to choose optimal treatment paths that incorporate patients’ personal values and beliefs (29). Therefore, it is necessary that medical knowledge be communicated in an easily understandable manner so that patients can become active participants who are empowered as fully informed directors of their own care.
Patient education can be delivered through a variety of mechanisms, perhaps most commonly in the form of direct interface with providers. However, written educational material has been shown to enhance patient understanding of medical conditions and possible treatments (30). Written materials can be used as a framework to guide shared decision-making conversations to ensure that all aspects of a particular treatment decision are discussed (31). Furthermore, written materials provide patients and their families a reliable reference after leaving a health care visit. Prior research suggests that use of written materials in conjunction with other forms of patient education can lead to significant improvements in long-term retention of knowledge (32).
This investigation provides evidence that a minority of currently available patient education materials from widely circulated medical journals are written in a way that may be accessible to many health care consumers. The average United States adult reads at the eighth grade level and the average Medicaid enrollee reads at the fifth grade level (33,34). However, the mean reading level of the 2585 identified materials was 11.2 to 13.8 grade by the various indices. Even according to the most forgiving readability metric, only 8.2% of the identified materials met the NIH-recommended eighth grade level and 2.1% met the more aggressive recommendation of the sixth grade level.
When analyzed separately by journal and material type, a high reading level was noted for all materials other than the general education materials from AIM, of which nearly 80% were at or below an eighth grade level. In contrast, research lay summaries published in AIM were written at a mean grade level of 11.2 to 13.7, with only 4.7% of materials meeting the eighth grade recommendation. These differences might exist for several reasons. First, it may be possible that even within a single journal or journal network, different publication requirements may exist for different types of patient education materials. For instance, if readability evaluation is required for all general education materials prior to publication but is not a routine requirement for research lay summaries, then significant differences are likely to arise. Second, it may be possible that the nature of medical concepts themselves may bias the readability of education materials. As compared to research topics, general education materials might more often explain concepts, for example, obesity or stroke, that are easily described with lay terms. Lastly, differences may exist in the scope of research and general education materials, which might affect the flexibility of language that can be used for each type of text. As an example, a document describing a scientific study might require the use of complex terms in order to communicate the message of that study without compromising accuracy. By contrast, a material describing a general medical topic might be more flexible in the use of language, which would allow authors more opportunity to write materials at reading levels consistent with national recommendations. Because readability is universally important for effective written communication, these observed differences, regardless of the precise causal mechanism, should encourage targeted efforts to evaluate and improve patient materials which may be at higher risk of worse readability. For example, as illustrated by our analysis, careful consideration of readability should be given for those materials describing complex topics such as research studies or other scientific subjects.
Increasing the accessibility of widely circulated patient education materials represents a promising opportunity to improve patient outcomes (2). As described previously, low health literacy is disproportionately prevalent in patients with lower levels of educational attainment and is associated with lower rates of adoption of basic health-promotion behaviors such as taking medications as directed, which translate into worse outcomes overall (35,36). One central mechanism to explain these associations is the inability of patients with limited health literacy to fully comprehend medical information that is communicated either verbally or through written text. Efforts to strengthen communication through the development of easily readable health education resources would therefore increase the ability of patients to gain and retain medical knowledge. Improvements in comprehension would promote patient autonomy and encourage active patient participation in medical decisions according to the shared decision-making model.
Although designing easily readable materials without compromising accuracy or breadth of content is challenging, this study provides strategies to guide the development of effective educational texts. Simply promoting awareness of readability as an important contributor to patient understanding of materials is essential. Our results suggest that many journals are not routinely evaluating the readability of patient materials prior to publication. Modification of peer-review or editorial practices to require readability assessment may lead to improvements in the proportions of materials meeting AMA and NIH readability recommendations. However, awareness alone is not sufficient; for authors unfamiliar with principles of readability, it can be difficult to describe complex medical concepts in an easily understandable manner.
The majority of readability metrics are dependent on parameters such as average sentence length, word length, and use of difficult or uncommon terms (Supplemental Table 2). The use of short words and sentences will generally translate into improved readability (20). Our Raygor score analysis (Figure 3) shows that the majority of materials fall below the central demarcation line along the Raygor grid, suggesting that frequent use of complex or long words is driving worse readability scores more often than long sentences. This likely reflects the inherent bias of medical education materials, which often require the use of terminology that may be uncommon or foreign to the lay audience. If possible, intentional efforts to minimize unnecessary use of complex terms, possibly through incorporation of medical abbreviations, could help mitigate this bias. However, abbreviations themselves should be employed with caution, as they may actually increase difficulty of comprehension for readers unfamiliar with medical terminology despite an apparent improvement in readability scores. In some instances, it may simply not be possible to write a comprehensive and accurate description of a medical concept without using certain requisite terms. In such cases, it may be best practice to target the less aggressive eighth grade reading level suggested by the NIH. For authors interested in evaluating readability of text, free online calculators are available (https://readable.com, https://readabilityformulas.com), and some word-processing programs have built-in readability testing capability. However, for the most comprehensive readability evaluations or for analysis of large cohorts of materials, we suggest more rigorous software packages.
Limitations
Although this study evaluates readability of patient education materials from a variety of high-impact medical journals and includes a broad study period, it nonetheless is limited by factors related to the experimental design. First, the goal of the investigation was to identify and analyze peer-reviewed materials with potentially wide readership. In order to objectively identify such materials, we only searched medical journals with an impact factor greater than 10. While impact factor may be a useful and reproducible metric, it is possible that many patient education materials with potentially large readership were not included. Second, determination of the target audience (eg, whether a published material is intended to educate laypersons or medical professionals) may not be explicitly stated, and therefore, decisions regarding material inclusion may vary across reviewers. Third, although each journal’s publication history was reviewed in its entirety, it is possible that some materials in a given journal which would have met inclusion criteria were not included. However, because this study draws from a large sample (n = 2585) of materials, it is unlikely that any missed documents would have led to meaningful differences in the study conclusions.
Additionally, appropriate readability levels alone do not guarantee comprehension of materials. Many other facets of material design, such as layout, font size, and use of graphics, also contribute to the effectiveness of written educational text (24). Evaluation of these components was not attempted in this study. Further, although improvement in readability of materials would theoretically improve education for diverse populations, it is possible that certain individuals may have limited access to these published materials in the first place. Inequal access to education resources could therefore introduce or exacerbate disparities across populations, and efforts to improve readability might not affect comprehension for all patients equally. Lastly, though the included readability metrics have been externally validated, they are not necessarily dependent on direct input from end users (eg, patients). Ideally, the effectiveness of materials would be evaluated by individuals from the target population; however, such evaluation may often not be feasible and therefore surrogate markers such as readability metrics may be required. Nonetheless, this is an important consideration for interpretation and contextualization of this study.
Conclusion
This investigation analyzes the readability of peer-reviewed patient education materials from high-impact medical journals and demonstrates that only a small minority of these materials are written at a grade level appropriate for the general population. Promisingly, general education materials published in AIM had a large proportion of materials (79.8%) meeting this goal. These results suggest that significant differences in review and publication processes exist across medical journals, with some appearing to emphasize readability of patient education materials more than others. This investigation highlights an important opportunity to enhance the large-scale education of patients through improvements in the readability and patient-centered design of written educational material. Efforts to address this need may allow patients from diverse backgrounds and varying levels of educational attainment to better participate as autonomous decision makers in an increasingly complex medical system.
Supplemental Material
Supplemental Material, sj-pdf-1-jpx-10.1177_2374373521998847 - Readability of Patient Education Materials From High-Impact Medical Journals: A 20-Year Analysis
Supplemental Material, sj-pdf-1-jpx-10.1177_2374373521998847 for Readability of Patient Education Materials From High-Impact Medical Journals: A 20-Year Analysis by Michael K Rooney, Gaia Santiago, Subha Perni, David P Horowitz, Anne R McCall, Andrew J Einstein, Reshma Jagsi and Daniel W Golden in Journal of Patient Experience
Footnotes
Authors’ Note
Ethics approval is not applicable for this article. It does not contain any studies with human or animal subjects. As there are no human subjects, informed consent is not applicable.
Declaration of Conflicting Interests
The author(s) declared the following potential conflicts of interest with respect to the research, authorship, and/or publication of this article. All authors have completed and submitted the ICMJE Form for Disclosure of Potential Conflicts of Interest. Dr. Horowitz reported receiving consulting fees and travel reimbursement from Carl Zeiss and consulting fees from Champions Oncology. Dr. Einstein reported receiving grant funding for unrelated research from the National Heart Lung, and Blood Institute, the National Cancer Institute, the International Atomic Energy Agency, Canon Medical Systems, Roche Medical Systems, and W. L. Gore & Associates; he has received consulting fees from GE Healthcare and W. L. Gore & Associates. Dr. Jagsi has stock options as compensation for her advisory board role in Equity Quotient, a company that evaluates culture in health care companies; she has received personal fees from Amgen and Vizient and grants for unrelated work from the National Institutes of Health, the Doris Duke Foundation, the Greenwall Foundation, the Komen Foundation, and Blue Cross Blue Shield of Michigan for the Michigan Radiation Oncology Quality Consortium. She has a contract to conduct an investigator initiated study with Genentech. She has served as an expert witness for Sherinian and Hasso and Dressman Benzinger LaVelle. She is an uncompensated founding member of TIME’S UP Healthcare and a member of the Board of Directors of ASCO. Dr. Golden reported receiving grant funding from the National Institutes of Health, Radiation Oncology Institute, and Bucksbaum Institute for Clinical Excellence. He is manager of RadOncQuestions LLC and HemOncReview LLC. No other disclosures were reported.
Funding
The author(s) received no financial support for the research, authorship, and/or publication of this article.
Supplemental Material
Supplemental material for this article is available online.
References
Supplementary Material
Please find the following supplemental material available below.
For Open Access articles published under a Creative Commons License, all supplemental material carries the same license as the article it is associated with.
For non-Open Access articles published, all supplemental material carries a non-exclusive license, and permission requests for re-use of supplemental material or any part of supplemental material shall be sent directly to the copyright owner as specified in the copyright notice associated with the article.
