scispace - formally typeset
Open Access

Preprint of Too good to be false: Nonsignificant results revisited

Reads0
Chats0
TLDR
This article examined evidence for false negatives in nonsignificant results in three different ways and concluded that false negatives deserve more attention in the current debate on statistical practices in psychology, and they also proposed the adapted Fisher test to detect the presence of at least one false negative in a set of statistically nonsignificantly results.
Abstract
Due to its probabilistic nature, Null Hypothesis Significance Testing (NHST) is subject to decision errors. The concern for false positives has overshadowed the concern for false negatives in the recent debates in psychology. This might be unwarranted, since reported statistically nonsignificant findings may just be ‘too good to be false’. We examined evidence for false negatives in nonsignificant results in three different ways. We adapted the Fisher test to detect the presence of at least one false negative in a set of statistically nonsignificant results. Simulations show that the adapted Fisher method generally is a powerful method to detect false negatives. We examined evidence for false negatives in the psychology literature in three applications of the adapted Fisher method. These applications indicate that (i) the observed effect size distribution of nonsignificant effects exceeds the expected distribution assuming a null-effect, and approximately two out of three (66.7%) psychology articles reporting nonsignificant results contain evidence for at least one false negative, (ii) nonsignificant results on gender effects contain evidence of true nonzero effects, and (iii) the statistically nonsignificant replications from the Reproducibility Project Psychology (RPP) do not warrant strong conclusions about the absence or presence of true zero effects underlying these nonsignificant results. We conclude that false negatives deserve more attention in the current debate on statistical practices in psychology. Potentially neglecting effects due to a lack of statistical power can lead to a waste of research resources and stifle the scientific discovery process.

read more

Citations
More filters
Journal ArticleDOI

Mechanisms underlying training-induced cognitive change

TL;DR: The authors provide an overview of cognitive training to induce transfer to untrained tasks, with a particular focus on the theoretical mechanisms that have been proposed to underlie training and transfer effects, and conclude that there is relatively little evidence that training enhances cognitive capacity, that is, the overall cognitive resources available to an individual.
Posted ContentDOI

Contributions towards understanding and building sustainable science

TL;DR: This dissertation argues that "after-the-fact" research papers do not help alleviate issues of access, selective publication, and reproducibility, but actually cause some of these threats because the chronology of the research cycle is lost in a research paper.

Statistical Analyses for Studying Replication: Meta-Analytic Perspectives.

TL;DR: The authors provide alternative conceptual frameworks that lead to different statistical analyses to test hypotheses about replication, including whether the burden of proof is placed on replication or nonreplication, whether replication is exact or allows for a small amount of "negligible heterogeneity," and whether the studies observed are assumed to be fixed (constituting the entire body of relevant evidence) or are a sample from a universe of possibly relevant studies.
Journal ArticleDOI

Published correlational effect sizes in social and developmental psychology

TL;DR: This article evaluated 12 412 manually collected correlation effect sizes (Sample 1) and 31 157 computer-extracted correlation effect size (Sample 2) published in journals focused on social or developmental psychology and found that large sample sizes were associated with smaller reported correlations.
Journal ArticleDOI

Estudios de casos y controles: Propuesta de robustez de análisis para ciencias de la conducta

TL;DR: In el campo de las ciencias de la conducta resulta indispensable realizar interpretaciones adecuadas de los resultados obtenidos que permitan el rechazo de la hipótesis nula as mentioned in this paper .
References
More filters
Book

Statistical Power Analysis for the Behavioral Sciences

TL;DR: The concepts of power analysis are discussed in this paper, where Chi-square Tests for Goodness of Fit and Contingency Tables, t-Test for Means, and Sign Test are used.
Journal ArticleDOI

The file drawer problem and tolerance for null results

TL;DR: Quantitative procedures for computing the tolerance for filed and future null results are reported and illustrated, and the implications are discussed.
Journal ArticleDOI

Estimating the reproducibility of psychological science

Alexander A. Aarts, +290 more
- 28 Aug 2015 - 
TL;DR: A large-scale assessment suggests that experimental reproducibility in psychology leaves a lot to be desired, and correlational tests suggest that replication success was better predicted by the strength of original evidence than by characteristics of the original and replication teams.
Journal ArticleDOI

The Kolmogorov-Smirnov Test for Goodness of Fit

TL;DR: In this paper, the maximum difference between an empirical and a hypothetical cumulative distribution is calculated, and confidence limits for a cumulative distribution are described, showing that the test is superior to the chi-square test.