Abstract

In Canada, our provincial and territorial health systems are increasingly challenged to meet the needs of patients, including mental health needs. Research is needed to find ways to provide more effective and efficient care. Waste in research, however, presents a formidable barrier to achieving this. 1 –6 One important source of waste in research is the incomplete and inaccurate reporting of published research results. Poor and inaccurate reporting pose substantial barriers to effectively translating research into improved patient care, both in psychiatry and other areas of medicine. 6 Assessing the utility and applicability of published findings requires an understanding of what actually occurred in a study and knowledge of the study’s complete set of results. 7 This information, however, is often either not provided in published study reports or is not provided with enough clarity to be useful. 6,8 As a result, billions of dollars in research funding are wasted every year, and the risk of inadequate and misinformed care is needlessly heightened. 1 –6
Biomedical research reporting guidelines have been developed with the goal of increasing the value and quality of research. These guidelines typically describe a minimum set of information that should be clearly reported, provide examples of guideline-consistent reporting, and include a checklist to facilitate compliance. 9 The first Consolidated Standards of Reporting Trials (CONSORT) statement on the reporting of parallel group randomized controlled trials was published in 1996. 10 The EQUATOR (Enhancing the QUAlity and Transparency Of health Research) initiative, which grew out of the work of CONSORT and other groups, was founded in 2008 with the mission of improving health research by promoting transparent, accurate, and complete reporting. 11 Other major reporting guidelines that have been developed by EQUATOR include the Preferred Reporting Items for Systematic Reviews and Meta-analyses (PRISMA) statement, 12,13 the Strengthening the Reporting of Observational Studies in Epidemiology (STROBE) statement, 14,15 and the Standards for Reporting of Diagnostic Accuracy (STARD) statement. 16,17 In recent years, EQUATOR reporting guidelines have expanded to include an extensive list of guidelines intended for specific trial designs (e.g., cluster randomized trials), 18 for reporting specific types of outcomes (e.g., patient-reported outcomes), 19 or for reporting specific types of systematic reviews and meta-analyses, including the Preferred Reporting Items for a Systematic Review and Meta-Analysis of Diagnostic Test Accuracy Studies (PRISMA-DTA) statement, 20 among other examples.
The publication of reporting guidelines has increased the completeness and quality of reporting. Pre-post studies of several reporting guidelines have shown that studies published after the publication of CONSORT, STARD, and PRISMA include more elements from the relevant guidelines than studies published before the guidelines. 21 –23 Given the positive impact of guideline publication, it is not surprising that many medical journals encourage researchers to refer to guidelines when preparing their manuscripts for peer review, and some journals require authors to adhere to them and submit a completed checklist along with their manuscript. Journal endorsement of guidelines has improved the completeness of reporting as well. Across journals, studies published in journals endorsing the CONSORT statement report more CONSORT items than studies published in journals that do not endorse the CONSORT statement. 24,25 Within journals, similar results have been found, with more CONSORT items reported in studies published post- versus pre-CONSORT endorsement by the journals. 25 However, enforcement appears to be more effective than endorsement alone. One study found that journals that actively enforced the CONSORT for Abstracts guideline had a significant increase in number of items reported over time, whereas no such trends were found among journals that endorsed, but did not enforce, the guideline. 26
Despite evidence that reporting guidelines are effective in improving the completeness of study reports across different study designs, many journals do not require that authors use them. Indeed, we reviewed the author instructions of the top 10 psychiatry journals based on impact factor per Thomson Reuters InCites Journal Citation Reports (March 23, 2018) and found that only 3 (
In psychiatry, the use of screening and assessment tools is increasingly encouraged for the identification of patients with unrecognized conditions or as case-finding tools for patients suspected of having a disorder, including major depressive disorder. 27,28 Concerns have been raised, however, about the quality of primary studies in this area 29 –31 and about the methodological quality and reporting of meta-analyses. 23,32 The recently published PRISMA-DTA, if followed, will help this situation.
The original PRISMA statement 12,13 is a reporting guideline for systematic reviews and meta-analyses that contains an evidence-based minimum set of items (27 items) that should be reported, along with a flow chart for illustrating the review process. The main focus of the original PRISMA guideline, however, was on systematic reviews and meta-analysis of randomized trials. Many elements of the conduct and reporting of interventional studies and DTA studies are common to both study designs. There are, however, important differences.
Thus, the PRISMA-DTA statement 20 was developed as an extension of the original PRISMA statement and aimed to reflect specific concepts, methods, language, and requirements for reporting of systematic reviews and meta-analyses of diagnostic test accuracy studies. It was designed to help ensure the applicability of reviews in practice by multiple potential users and to improve transparency and completeness of reporting of DTA systematic reviews and meta-analyses.
The PRISMA-DTA checklist consists of 27 items, 8 of which are identical to the original PRISMA items (3, 5, 7, 9, 10, 16, 17, and 27), 17 that were modified based on PRISMA items (1, 2, 4, 6, 8, 11-14, 18-21, and 23-26) due to ambiguity in wording in the original PRISMA items or to address reporting concerns specific to DTA systematic reviews and meta-analyses, and 2 new items. The first new item (D1) regards the statement of the scientific and clinical background for using the test, including its intended use and clinical applicability and, if relevant, a rationale for minimally accepted accuracy. The same test can be used for different purposes, and in psychiatry, a depression screening tool, for instance, could be tested as a general population screening tool or as a case-finding tool among people suspected of having depression. The same tool could also be used for tracking progress of patients undergoing treatment. How the test is used can be an important source of variability in diagnostic accuracy. Furthermore, drawing conclusions about how a test may be used in practice depends on understanding minimal levels of acceptable test accuracy and reasons for setting such thresholds. The second new item (D2) requires the reporting of the statistical methods used for meta-analyses, if performed. Rigorously conducted meta-analyses of DTA studies require more sophisticated multivariate models (e.g., bivariate and hierarchical summary receiver-operating characteristic) than is often the case for meta-analyses of other types of studies in order to model the tradeoff between sensitivity and specificity and to account for correlation between sensitivity and specificity. Thus, greater attention to reporting of modeling procedures was considered essential. 20
The PRISMA-DTA statement also removed 2 items, items 15 and 22, from the original PRISMA statement. These items pertain to methods and results related to the risk of bias that may affect the cumulative evidence, including publication bias or selective reporting of results within studies. The rationale for this decision was that there is only limited evidence suggesting that publication bias and reporting bias are problematic for DTA studies.
However, in psychiatry research, selective cutoff reporting, whereby authors report only DTA results for cutoffs with reasonable sensitivity and specificity, was recently shown to result in biased estimates of pooled sensitivity and specificity when primary studies are meta-analyzed. A conventional meta-analysis of the Patient Health Questionnaire–9 (PHQ-9) depression screening tool found that pooled estimates of sensitivity actually improved as cutoff thresholds increased from 9 (less severe symptoms) to 11 (more severe symptoms), which would be mathematically impossible if complete data were analyzed. 33 This occurred due to selective reporting of results for cutoff thresholds only where the PHQ-9 performed well in individual primary studies but nonreporting for thresholds where it performed poorly. Members of our team compared results from conventional meta-analysis of published PHQ-9 diagnostic accuracy using aggregate study–level data to results using individual patient data meta-analysis (IPDMA). 34 Using IPDMA allowed us to incorporate accuracy results from both published and unpublished cutoff thresholds from the same set of studies. We found that due to selective cutoff reporting in primary studies, estimates of pooled sensitivity were distorted; when only published accuracy results were analyzed, sensitivity was underestimated for cutoff thresholds below the generally accepted “standard” cutoff of 10 and overestimated for cutoff thresholds greater than10 compared with the complete IPDMA results. 34
Concerns that selective cutoff reporting in primary studies may be biasing results in meta-analyses were also recently raised by the authors of a conventional meta-analysis of the PHQ-9. 35 They pointed out that no firm conclusions could be drawn about cutoff thresholds other than the standard threshold of 10 because authors of different studies often reported inconsistent cutoffs, and this appeared to be driven by selective reporting. 35 Another recent meta-analysis of brief versions of the Geriatric Depression Scale warned readers that results should be interpreted cautiously due to the likely influence of selective cutoff reporting on synthesized accuracy results. 36 More work is needed to understand the scope of this problem, and IPDMA is one approach to overcome it, although statistical approaches may also work. In the meantime, we recommend that authors of systematic reviews and meta-analyses of the accuracy of mental health–screening tools evaluate the possibility of selective cutoff reporting and describe their methods and results per the PRISMA-DTA items on conducting additional analyses (items 16 and 23).
In sum, reporting guidelines improve the transparency and completeness of published study reports. Authors who submit manuscripts to the
Footnotes
Declaration of Conflicting Interests
The authors declared the following potential conflicts of interest with respect to the research, authorship, and/or publication of this article: All authors have completed the Unified Competing Interest form at
(available on request from the corresponding author). Dr. Thombs declared that he was an author of the PRISMA-DTA reporting guideline and that he has no other competing interests. All other authors declared that they have no competing interests.
Funding
The authors disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: Drs. Thombs and Benedetti were supported by Fonds de recherche du Québec–Santé researcher salary awards. Ms. Rice was supported by a Canadian Institutes of Health Research (CIHR) Vanier Graduate Scholarship. Ms. Levis was supported by a CIHR Frederick Banting and Charles Best Canada Graduate Scholarship Doctoral Award.
