Abstract
We investigated whether statistical-reporting inconsistencies could be avoided if journals implement the tool statcheck in the peer-review process. In a preregistered pretest–posttest quasi-experiment covering more than 7,000 articles and more than 147,000 extracted statistics, we compared the prevalence of reported p values that were inconsistent with their degrees of freedom and test statistics in two journals that implemented statcheck in their peer-review process (Psychological Science and Journal of Experimental and Social Psychology) and two matched control journals (Journal of Experimental Psychology: General and Journal of Personality and Social Psychology) before and after statcheck was implemented. Preregistered multilevel logistic regression analyses showed that the decrease in both inconsistencies and decision inconsistencies around p = .05 is considerably steeper in statcheck journals than in control journals, offering preliminary support for the notion that statcheck can be a useful tool for journals to avoid statistical-reporting inconsistencies in published articles. We discuss limitations and implications of these findings.
Many conclusions in psychological research are based on the results of null hypothesis significance tests (NHSTs; Cumming et al., 2007; Nuijten et al., 2016). It is important that results of NHSTs are reported correctly: If one cannot rely on the reported numbers, one cannot rely on the overall robustness of the findings (Nosek et al., 2022; Nuijten, 2022; Nuijten et al., 2018). Unfortunately, in previous research, we found a high prevalence of statistical-reporting inconsistencies in published psychology articles (Nuijten et al., 2016). Specifically, ≈50% of the articles with statistical results contained at least one p value that did not match its accompanying test statistic and degrees of freedom. In ≈12.5% of the articles with statistics, we found at least one “decision inconsistency” (or “gross inconsistency”) in which the reported p value was statistically significant, whereas the recomputed p value was not or vice versa (Nuijten et al., 2016).
In response to these high inconsistency rates, several journals started recommending or even requiring authors to scan their submitted manuscripts with statcheck (Nuijten & Epskamp, 2023). Statcheck is a free software tool that can automatically scan academic articles and detect inconsistencies in reported NHST results; effectively, it acts as a “spellchecker” for statistics. Of course, the hope of the journals that implemented statcheck was that inconsistency rates would decrease, but this remains an empirical question. In this preregistered study, we empirically estimated the effectiveness of using statcheck in the peer-review process in two major psychology journals that started using statcheck in 2016 (Psychological Science [PS]) and 2017 (Journal of Experimental Social Psychology [JESP]). Specifically, we compared the trends in inconsistencies of these two journals against trends in two matched journals that did not have a specific policy to use statcheck to test the following hypothesis:
Hypothesis 1: Articles published in journals that include statcheck in their peer-review process show a steeper decline in statistical-reporting inconsistencies and decision inconsistencies compared with matched journals that do not use statcheck.
Method
Disclosures
The hypothesis and study protocol (rationale, methods, and analysis plan, including a power analysis) were preregistered before data collection on recent articles 1 on the OSF study registry on November 30, 2022 (https://osf.io/umwea). All departures from that protocol are explicitly acknowledged. We report how we determined our sample size, all data exclusions, all manipulations, and all measures in the study. The data and analysis scripts are available at https://osf.io/q84jn/. We cannot share the specific articles we scraped because of copyright restrictions, but our description of our sample should allow anyone with the relevant journal subscriptions to obtain the same sample. This study was approved by the Ethics Review Board of the Tilburg School of Social and Behavioral Sciences on November 18, 2022 (Approval TSB_RP773).
Design
In this study, we compared statistical-reporting inconsistencies in two sets of journals over time: journals that use statcheck in their peer-review process and journals that do not. This makes our design a pretest–posttest quasi-experiment.
Sample
Two journals that implemented statcheck in their peer-review process are PS and JESP. PS implemented statcheck in July 2016 (Nuijten, Borghuis, et al., 2017, p. 9). Their policy states: StatCheck is an R program that is designed to detect inconsistencies between different components of inferential statistics (e.g., t value, df, and p). StatCheck is not designed to detect fraud, but rather to catch typographical errors (see https://mbnuijten.com/statcheck/ for more about StatCheck). Authors of accepted manuscripts must also provide a StatCheck report run on the accepted version of the manuscript that indicates a clean (i.e., error-free) result [italics added]. A web app version of StatCheck can be accessed at http://statcheck.io/. If StatCheck does detect errors in the accepted version of the manuscript, authors should contact the action editor directly to determine the best course of action. (Association for Psychological Science, 2022)
We included all articles published in PS from 2003—the first year that articles in html format were available, which is more suitable for text mining than pdf files—to October 2022 (the latest complete issue at the time of data collection), excluding retractions, errata, and editorials. PS articles up until 2013 had already been downloaded in previous research (Nuijten et al., 2016).
The second journal that implemented statcheck in their peer-review process, JESP, announced the new policy in August 2017 (JESP Piloting the Use of Statcheck, 2017). Their policy states: For all manuscripts that are deemed to fit within the Aims and Scope of the journal, the editorial team will be using statcheck as part of their initial triage of manuscripts [italics added]. For any manuscripts found to have important discrepancies in reporting, we will ask authors to resolve these in the manuscript before they can be sent on for further review. The pilot is intended to help editors and authors to work together to decrease the number of errors in published articles in the journal. Before submitting, authors are invited to run a HTML or PDF version of their APA-formatted manuscript through statcheck prior to submitting their manuscripts [italics added], via this link: http://statcheck.io/. This will be the same portal that the JESP Editorial Team will be using. (JESP, 2022)
We included all articles published in JESP between 2003 and November 2022 (the latest complete issue at the time of data collection), excluding retractions, errata, and editorials.
Any decrease in statistical-reporting inconsistencies in these journals after implementation of statcheck could in principle also be due to other factors (e.g., an increased awareness of reporting errors and/or statcheck in general). To control for this (at least in part), we also include two matched comparison journals that did not implement statcheck.
We looked for journals to match with PS and JESP that were similar in content and impact and were likely to contain results reported in American Psychological Association (APA) style to facilitate the automated detection of statistical results. As a matched control for PS, we chose the Journal of Experimental Psychology: General (JEPG). Both journals focus on general psychology and often publish experimental designs. As a matched control for JESP, we chose the Journal of Personality and Social Psychology (JPSP), but only articles from the subsection “Attitudes and Social Cognition” to ensure a closer match with the articles in JESP.
From both JEPG and JPSP, we included articles from the same years as PS and JESP: 2003 to October 2022 (in both cases, the latest complete issue at the time of data collection). Here, too, we excluded retractions, corrections, and editorials. Thirty-nine articles from different volumes and publication years of JEPG did not seem to be available in html format, so these were excluded. 2 Given the large overall sample size, we deem it unlikely that these missing cases would influence our conclusions. JEPG and JPSP articles up until 2013 had already been downloaded in previous research (Nuijten et al., 2016).
We used regular expressions to extract the submission date of each article to classify them as submitted before (T1) or after (T2) statcheck was implemented in the journal or its matched control journal (July 1, 2016, for PS and JEPG and August 1, 2017, for JESP and JPSP).
Detecting statistical-reporting inconsistencies with statcheck
We used the R package statcheck (Version 1.4.1-beta.2; Nuijten & Epskamp, 2023) to automatically detect statistical-reporting inconsistencies in the included articles. 3 Statcheck’s algorithm works in roughly four steps:
Statcheck first converts an article in pdf or html format to plain text.
Next, statcheck uses “regular expressions” to search for specific patterns of letters, numbers, and symbols that signal an NHST result. Specifically, statcheck can detect results of t tests, F tests, Z tests, χ2 tests, correlation tests, and Q tests as long as they are reported completely (i.e., test statistic, degrees of freedom if applicable, and p value), in text (statcheck usually misses results reported in tables) and in American Psychological Association style (APA, 2019).
Third, statcheck uses the reported test statistic and degrees of freedom to recalculate the (two-tailed) p value.
In the final step, statcheck compares the reported and recalculated p values. If these two values do not match, statcheck labels the result as an “inconsistency.” In cases where the reported p value is statistically significant (assuming a = .05) and the recalculated one is not, or vice versa, statcheck labels the result as a “decision inconsistency.”
Statcheck takes into account one-tailed p values (that are half the size of what statcheck would expect by default) that are explicitly identified as such. Specifically, if the word “one-tailed,” “one-sided,” or “directional” appears anywhere in the article and the result would be internally consistent if the p value was one-sided, statcheck treats it as such and counts it as correct.
Statcheck also takes correct rounding of the test statistic into account. Consider, for instance, the result, t(48) = 1.43, p = .158. Recalculation would give a p value of .159, not .158, seemingly an inconsistency. However, the true t value could lie in the interval [1.425, 1.435], with p values ranging from .158 to .161. To take this into account, statcheck will consider any p value within this range as consistent.
Finally, statcheck counts p = .000 and p < .000 as inconsistent. A p value of exactly zero is mathematically impossible, so the APA manual (APA, 2019) advises to report very small p values as p < .001.
We estimate that statcheck detects roughly 60% of all reported NHST results (Nuijten et al., 2016). The results it does not detect are usually reported in tables instead of full text or the results are not reported in APA style.
We systematically assessed the validity of earlier versions of statcheck in classifying inconsistencies in previous research and concluded it is high (Nuijten et al., 2016, 2017). The interrater reliability between manual coding and statcheck was .76 for inconsistencies and .89 for decision inconsistencies. Furthermore, statcheck’s sensitivity (true positive rate) and specificity (true negative rate) are high: between 85.3% and 100% and between 96.0% and 100%, respectively, depending on the assumptions and settings (e.g., the one-tailed test detection, as described above, increases specificity). The overall accuracy of statcheck ranged from 96.2% to 99.9% (Nuijten et al., 2017). Solving several issues in the previous statcheck versions 4 has likely further improved its accuracy.
Since the implementation of the statcheck policy in PS (July 1, 2016) and JESP (August 1, 2017), the statcheck web app has depended on different versions of the statcheck R package: Version 1.0.1 (released February 4, 2015), Version 1.2.2 (released August 18, 2016), and Version 1.3.0 (released May 4, 2018). For the current study, we used the latest version of statcheck: Version 1.4.1-beta.2. The subsequent versions became increasingly accurate in detecting statistical results and classifying them as consistent or not. Some of the most notable updates that improved the detection rate of statistical results and/or the classification of inconsistencies are improvements in detecting χ2 tests and minus signs, improvements in the detection and recalculation of one-tailed tests, the addition of the Q test for heterogeneity in meta-analyses, and improvements in determining correct rounding of test statistics. 5
Results
Descriptives
Table 1 describes the number of available articles per journal, categorized by journal type (statcheck or matched control). For our analyses, we had to determine whether an article was submitted before or after statcheck was implemented in the (matched) journal’s peer-review process. We therefore report the number of articles we could download and the number of articles from which we could extract the date submitted/received. 6 We were able to extract submission dates from almost all articles in our sample (97.4%–99.6%, depending on the journal).
Number of Downloaded Articles and Number of Articles for Which We Could Automatically Extract Information About the Date the Article Was Submitted/Received
Note: PS = Psychological Science; JESP = Journal of Experimental Social Psychology; JEPG = Journal of Experimental Psychology: General; JPSP = Journal of Personality and Social Psychology.
Table 2 shows the most important descriptive statistics, split up by journal type (statcheck or control) and period: before (T1) and after (T2) statcheck was implemented. In total, we extracted 147,784 NHST results from 7,314 articles. Of these results, 10,160 were inconsistent (6.9%), and 1,226 were a decision inconsistency (0.8%). To get a general sense of the inconsistency rates in the different journal types and periods, we can look at the mean percentage of (decision) inconsistencies in articles with NHST results. This statistic is calculated as follows: Say that statcheck detects 10 NHST results in a given article, of which two are inconsistent. The percentage of inconsistencies in this article is then 20%. We calculated this percentage for each article with NHST results and averaged these percentages per journal type and period. For more detailed descriptive results, split up per journal, see Table 3.
Descriptive Statistics Split Up for Journals Using Statcheck and Their Matched Control Journals Before (T1) and After (T2) Statcheck Was Implemented
Note: NHST = null hypothesis significance test.
Total number of downloaded articles from which we could extract the date submitted/received. For details, see Table 1.
Note that in a previous preprint of this article (https://doi.org/10.31234/osf.io/bxau9; Version 1), we reported fewer articles with NHST results in journals using statcheck (4,012 in T1 and 876 in T2). This was mainly caused by the use of a specific style of spaces in statistical results in 2019 Journal of Experimental Social Psychology articles that statcheck Version 1.4.0-beta.7, which we used at that time, did not recognize. In the current statcheck Version 1.4.1-beta.2, this issue was fixed, hence the higher detection rate of NHST results in statcheck journals.
Descriptive Statistics Split Up for Journals Using Statcheck and Their Matched Control Journals Before (T1) and After (T2) Statcheck Was Implemented
Note: NHST = null hypothesis significance test; PS = Psychological Science; JESP = Journal of Experimental Social Psychology; JEPG = Journal of Experimental Psychology: General; JPSP = Journal of Personality and Social Psychology.
Total number of downloaded articles from which we could extract the date submitted/received. For details, see Table 1.
We found that the mean inconsistency percentage decreased more steeply in the statcheck journals (8.8 – 4.3 = 4.5 percentage points) than in the control journals (7.4 – 6.4 = 1.0 percentage point). We see a similar pattern in the decision inconsistencies: 1.2 – 0.3 = 0.9 percentage points in the statcheck journals and 1.0 – 0.8 = 0.2 percentage points in the control journals. We note that the distributions of these percentages are highly skewed: In most articles, only a small percentage of reported results is inconsistent, but there are a few articles in which up to 100% of reported results are inconsistent. This skewness can be seen in Figures 1a and 1b, where the y-axis in Figure 1b is truncated to zoom in on the bulk of the observed percentages. To more clearly illustrate the difference in inconsistency percentages between journals, Figure 2 shows the change in the mean inconsistency percentages per type of journal and period and split up for the journal pairs separately.

Distribution of the percentages of inconsistencies and decision inconsistencies per article with NHST results, per type of journal and period. (a) The full distributions. (b) The y-axis is truncated at 25% to improve readability. The number of articles with NHST results from the control journals before and after statcheck implementation were N = 1,445 and N = 763, respectively, and for the statcheck journals, they were N = 4,053 and N = 1,053, respectively. NHST = null hypothesis significance test.

Illustrative representation of the mean percentage of inconsistencies (top row) and decision inconsistencies (bottom row) in articles with null hypothesis significance test results before and after statcheck implementation. Solid lines represent statcheck journals, and dashed lines represent matched control journals. The first column depicts mean percentages for the journals combined, and the second and third columns show the results for the journal pairs separately. Preregistered multilevel logistic regressions showed that the interaction between journal type and period was statistically significant when predicting inconsistencies and decision inconsistencies. Exploratory follow-up analyses of the journal pairs separately found a significant interaction only when predicting inconsistencies in Psychological Science/Journal of Experimental Psychology: General.
Note that these descriptives are mainly reported for illustrative purposes: because the number of NHST results differs greatly per article (which in part also explains the skewness of the distributions), it is hard to directly interpret these percentages. The same holds for Figures 1 and 2: these average percentages are mainly useful to get a general image of the observed patterns. We appropriately test differences in the next section.
Confirmatory analyses
Following our preregistration, we tested our main hypotheses using two multilevel logistic models:
where subscript i indicates article, statcheckJournal indicates if the journal implemented statcheck in peer review [no = 0 [JEPG and JPSP], yes = 1 [PS and JESP]), Period is the period in which an article is submitted (before [0] or after [1] statcheck was implemented in peer review), and
We hypothesized that in both models, the coefficient
When predicting the inconsistencies, we found a significant interaction effect of Period × statcheckJournal in the predicted direction, b3 = −0.71, SE = 0.11, 95% confidence interval [CI] = [–0.92, −0.51], Z = −6.73, p < .001. This corresponds to an odds of exp(–0.71) = 0.49, 95% CI = [0.40, 0.60] and a probability of 0.49 / (1 + 0.49) = 0.33, 95% CI = [0.28, 0.38]. In addition, when predicting decision inconsistencies, we found a significant interaction effect of Period × statcheckJournal in the predicted direction, b3 = −0.82, SE = 0.35, 95% CI = [−1.51, −0.14], Z = −2.37, p = .018. This corresponds to an odds of 0.44, 95% CI = [0.22, 0.87] and a probability of 0.31, 95% CI = [0.18, 0.47]. See Table 4 for detailed results.
Results of Estimating the Multilevel Logistic Regression Models That Predict Whether a Statistical Result Is an Inconsistency (Model 1) or Decision Inconsistency (Model 2) Based on Period and Type of Journal That the Article Was Published in, With Random Intercepts for Articles
Note: CI = confidence interval.
As a simplified illustration of the coefficients in Table 4, consider an imaginary article with 100 NHST results. This article would show a decrease from four to three inconsistencies in a control journal and a decrease from five to two inconsistencies in a statcheck journal. In this same imaginary article, the number of decision inconsistencies would decrease from 0.03 to 0.02 in a control journal and from 0.04 to 0.01 in a statcheck journal. Note, however, that these inconsistency rates are lower than the ones reported in our descriptive statistics. This discrepancy is due to the estimation of the b coefficients, which takes into account the random intercept, resulting in a lower probability of an inconsistency than observed directly in the data.
These findings indicate that the prevalence of inconsistencies and decision inconsistencies decreased more steeply in statcheck journals than in control journals. This finding is in line with the notion that implementing statcheck in the peer-review process could decrease the prevalence of reporting inconsistencies but note that causality cannot be established and alternative explanations are possible because of the nature of the design. We list potential alternative explanations for these results in the Discussion.
Exploratory analyses
Bayesian analysis
Next to conducting a frequentist hypothesis test, we also computed Bayesian hypothesis tests. We computed approximated adjusted fractional Bayes factors (Gu et al., 2018) using the default implementation in the R package BFpack (Version 1.0.0; Mulder et al., 2021). The approximated adjusted fractional Bayes factor uses a minimal fraction of the available data to train a noninformative normally distributed prior and approximate the marginal likelihood of the tested hypotheses. We compared two models with each other: Period × statcheckJournal < 0 (which would indicate that inconsistencies decreased stronger in statcheck journals than in non-statcheck journals) versus its unconstrained complement.
Given our data, we found that the model predicting inconsistencies where Period × statcheckJournal < 0 was 1.16e11 times more likely than a model where Period × statcheckJournal was not < 0 (BF10 = 1.16e11, posterior probability = 1). When predicting decision inconsistencies, we found that the model where Period × statcheckJournal < 0 was 110.2 times more likely than its complement (BF10 = 110.2, posterior probability = .991). These Bayesian hypothesis tests indicated very strong and strong evidence, respectively (Kass & Raftery, 1995), in favor of our hypothesis that both inconsistencies and decision inconsistencies decreased more steeply in statcheck journals after statcheck implementation than in non-statcheck journals.
Inconsistency rates in journal pairs separately
Exploratively, we also looked at the inconsistency rates in the journal pairs separately. Table 3 and Figure 2 show the mean percentages of inconsistencies and decision inconsistencies in an article per journal and time. We fitted the multilevel logistic models explained above for the two journal pairs separately. When predicting the inconsistencies, we did find a statistically significant interaction term for the pair PS/JEPG (b3 = −0.92, 95% CI = [–1.20, –0.64], Z = −6.42, p < .001) but not for JESP/JPSP (b3 = −0.12, 95% CI = [–0.51, 0.28], Z = −0.58, p = .561). When predicting decision inconsistencies, we did not find statistically significant effects for either journal pair: PS/JEPG (b3 = −0.80, 95% CI = [–1.84, 0.24], Z = −1.51, p = .130) and JESP/JPSP (b3 = −0.37, 95% CI = [–1.38, 0.63], Z = −0.73, p = .466). These exploratory results could indicate that the overall significant interaction term of Period × statcheckJournal in the confirmatory results may be driven by a strong effect of implementing statcheck in PS, specifically. Note, however, that we did not formally test this three-way interaction to assess if there is a significant difference in interaction effects between the pairs PS/JEPG and JESP/JPSP. Such an analysis would likely be underpowered, and because of its post hoc nature, any resulting p values would be hard to interpret.
Trends in the prevalence of reporting inconsistencies over time
Following Nuijten et al. (2016), we considered trends over time in the prevalence of (decision) inconsistencies. Figure 3 shows the percentage of articles with NHST results that contained at least one inconsistency (turquoise line) or decision inconsistency (purple line) over time, split up per journal. We did not do a regression analysis to estimate the overall trends because in the statcheck journals, we would not expect a linear decrease but, rather, a difference in trends before and after statcheck implementation. Visual inspection of the trends shows a clear drop in the prevalence of articles with inconsistencies in PS after statcheck implementation: Before, roughly 40% of articles contained at least one inconsistency, but after statcheck implementation, this halved to about 20% of articles. Other patterns in this graph do not jump out as much, but overall, the results seem to be in line with Nuijten et al., who found an overall decrease in statistical-reporting inconsistencies and decision inconsistencies over time. After implementation of statcheck during review at PS, the prevalence of reporting inconsistencies visibly diminished further.

Percentage of articles with null hypothesis significance test results that contained at least one inconsistency (turquoise line) or decision inconsistency (purple line) over time, split up per journal. The vertical dotted line indicates the moment that statcheck was implemented in a journal or its counterpart.
Looking at the remaining inconsistencies in detail
Notwithstanding the introduction of statcheck during peer review, journals continued to publish articles with statistical-reporting inconsistencies (see Table 3). On the face of it, this seems surprising: If both journals require a “clean” statcheck report for publication, how could statistical inconsistencies remain? Four potential explanations are that (1) the remaining inconsistencies would not have been picked up by statcheck Version 1.3.0 (the latest published version at the time of writing) or older versions compared with Version 1.4.1-beta.2 that we used in this study; (2) the remaining inconsistencies are intentional and can be explained by, for example, the use of statistical corrections, use of one-tailed tests that are not explicitly identified in the way statcheck expects, or p values that were deliberately reported as p = .000; (3) statcheck incorrectly flagged a correct result as an inconsistency; or (4) some remaining errors slipped through the review process, for example, if an editorial team did not routinely (or correctly) follow through on their statcheck policy.
To better understand the remaining inconsistencies, we analyzed the full text of a subsample of articles in more detail. Specifically, we randomly selected 10 articles that had at least one inconsistency from each of the two statcheck journals published after statcheck was implemented and categorized likely reasons for remaining errors. We summarize the results in Table 5.
Classification of Likely Reasons for Remaining Inconsistencies in a Random Subsample of 20 Articles From the Two Statcheck Journals After Statcheck Was Implemented
Note: PS = Psychological Science; JESP = Journal of Experimental Social Psychology.
In total, these 20 articles contained 737 NHST results, of which 113 were an inconsistency (15.3%) and eight were a decision inconsistency (1.1%). The majority of remaining inconsistencies seemed to be due to statistical corrections to either the p value or the degrees of freedom (64 cases; 57%). In previous work, we have argued that such corrections can and should be reported in a way that does not render the full result inconsistent (Nuijten et al., 2017), but this is not a convention yet. Next, we classified 35 inconsistencies as actual statistical errors that seemed to have slipped through the review process. These included cases in which a p value appeared to be rounded incorrectly; a “<” sign was reported, but “=” would have been correct; or digits were switched. Finally, in 14 cases, all from the same article, statcheck made an error in extracting the results from the article: It added two additional digits from the next row of a table to the reported p value, rendering it inconsistent. This illustrates that despite statcheck’s high performance in classifying inconsistencies (Nuijten et al., 2017), it is not perfect, and its results should not be acted on blindly. We noted that none of the remaining inconsistencies could be attributed to statcheck version updates.
Discussion
In this pretest-posttest quasi-experiment, we compared the prevalence of statistical-reporting inconsistencies in two journals that implemented statcheck in their peer-review process and two matched control journals before and after the statcheck implementation. We found a steeper decline of both inconsistencies (b3 = −0.71, 95% CI = [−0.92, −0.51]) and decision inconsistencies (b3 = −0.82, 95% CI = [−1.51, −0.14]) in the statcheck journals than in the control journals, which is in line with the notion that implementing statcheck in the peer-review process can be effective in avoiding reporting errors in published articles. Exploratory Bayesian hypothesis tests provided strong evidence in line with this notion.
In additional exploratory analyses, we observed a decrease in the prevalence of articles with (decision) inconsistencies across all included years in all but one journal (JEPG). This may indicate that without any interventions, the prevalence of statistical-reporting inconsistencies may decrease somewhat over time, but our confirmatory results do give an initial, potential indication that certain interventions may reduce reporting errors a lot faster.
An important limitation is that our study was observational. We did not randomly assign journals or manuscripts to be checked by statcheck or not, which means there can be selection effects and other potential confounding factors that could explain the observed effect. For example, it is possible that we see a steeper decline in reporting errors in the statcheck journals because the implementation of statcheck signaled a commitment to improved reporting practices, which inspired the more conscientious authors to submit to those journals. Or conversely, authors with a tendency toward less diligent reporting may have been deterred from submitting to statcheck journals because of the increased likelihood of errors being detected by statcheck, which might be a good outcome for the journal at hand but might also mean that reporting errors remain appearing elsewhere.
The implementation of statcheck in PS and JESP has also not occurred in a vacuum. In both cases, statcheck was implemented relatively shortly after a new editor-in-chief was installed. New editors often install new policies, which could affect the type of articles submitted to these journals and the way that submissions are handled. Indeed, around the same time that statcheck was introduced, both PS and JESP published editorials emphasizing the importance of open practices and replication (Giner-Sorolla, 2016; Lindsay, 2015). We cannot rule out that this emphasis on responsible research practices has affected the prevalence of reporting inconsistencies. However, the control journal JPSP also published a similar editorial promoting open practices around the same time (Cooper, 2016), and there, the observed decline in reporting inconsistencies is less stable than in PS and JESP (see Fig. 3).
A related limitation is that we picked two comparison journals based on relatively subjective criteria based on similarity in subfields and impact. Alternative comparison journals may also have been suitable and could potentially have shown other results. Furthermore, several other journals besides PS and JESP have recently started to use statcheck. 7 In time, when a sufficient number of articles has been published in these journals after statcheck’s implementation, it would be interesting to assess declines in reporting inconsistencies across a wider range of journals.
Another limitation is that we looked only at reporting inconsistencies that could be picked up by statcheck. That means that inconsistencies in non-APA reported results, errors in data entry, or other types of statistical problems have not been detected. Because our initial goal was to have an indication whether implementing statcheck could be effective in decreasing “statcheck-type inconsistencies,” this limitation has no direct bearing on our conclusion. However, for a deeper understanding of the different types of problems in statistical reporting, we would need richer data and much more checking of the many ways in which analyses and reporting of results can go awry.
A final, related limitation is that statcheck itself is not flawless. Based on exploratory, in-depth analyses of a subsample of articles in which statcheck flagged inconsistencies, we found that statcheck wrongly extracted statistics from a table in one article, resulting in falsely flagged inconsistencies. In addition, in this subsample, statcheck flagged some cases that turned out to be results that were corrected for multiple testing or violations of assumptions. Whether this is problematic is debatable, but we previously argued that such corrections can and should be reported in a way that does not render the full result inconsistent (Nuijten et al., 2017) to allow a reproducible report of applied corrections.
Even when taking the limitations of this study and of statcheck into account, we would still tentatively recommend journals that adhere to APA reporting guidelines to consider using statcheck in their peer-review process. Human peer reviewers seem to often overlook statistical-reporting inconsistencies (judging from the high prevalence of inconsistencies in the literature and recent experimental work; Augusteijn et al., 2023), so including automated tools such as statcheck in the peer-review process might be a viable solution.
Even though we did not assess the direct effect of the use of statcheck during peer review, we argue that any mechanism that lowers the currently high prevalence of reporting errors in the literature is preferred. Given its ease of use and provided that statcheck is used carefully and in a collaborative manner by reviewers, authors, and editors, we see few potential downsides in implementing statcheck more widely to improve the quality of reporting of statistical results. We do caution editors and peer reviewers not to rely solely on the results of statcheck (or any automated tool, for that matter) when deciding to accept or reject an article: Software can be useful in reducing workload or human errors but is not free of its own pitfalls. That said, we think statcheck can be a quick and easy way to help journals avoid statistical-reporting inconsistencies in their articles and increase their overall quality and robustness.
Footnotes
Acknowledgements
We thank Afra Kiliç and Tsz Keung Wong for their assistance in downloading the articles. A shortened version of this article (excluding exploratory analyses and based on an older version of statcheck) was posted as a preprint on January 31, 2023, at https://psyarxiv.com/bxau9. A previous version of the current, full article was posted as a preprint on July 19, 2023, at
(Version 1). The article in its current form was posted as a preprint on April 26, 2024, also at https://psyarxiv.com/qejhk (Version 2).
Transparency
Action Editor: Katie Corker
Editor: David A. Sbarra
Author Contributions
