Nuijen and colleagues studied reporting errors in a sample of over 250,000 p-values reported in eight major psychology journals. The authors developed the new R package Statcheck. This set of programs was used to retrieve null-hypothesis significance testing (NHST) results. The paper reports that one in eight papers contained a grossly inconsistent p-value that may have affected the statistical conclusion. The prevalence of gross inconsistencies was higher in p-values reported as significant. The authors conclude: "This could indicate a systematic bias in favor of significant results. Possible solutions for the high prevalence of reporting inconsistencies could be to encourage sharing data, to let co-authors check results in a so-called co-pilot model, and to use statcheck to flag possible inconsistencies in one's own manuscript or during the review process."