scispace - formally typeset
Open Access

Robust misinterpretation of confidence intervals

Reads0
Chats0
TLDR
Although all six statements were false, both researchers and students endorsed, on average, more than three statements, indicating a gross misunderstanding of CIs, which suggests that many researchers do not know the correct interpretation of a CI.
Abstract
Null hypothesis significance testing (NHST) is undoubtedly the most common inferential technique used to justify claims in the social sciences. However, even staunch defenders of NHST agree that its outcomes are often misinterpreted. Confidence intervals (CIs) have frequently been proposed as a more useful alternative to NHST, and their use is strongly encouraged in the APA Manual. Nevertheless, little is known about how researchers interpret CIs. In this study, 120 researchers and 442 students—all in the field of psychology—were asked to assess the truth value of six particular statements involving different interpretations of a CI. Although all six statements were false, both researchers and students endorsed, on average, more than three statements, indicating a gross misunderstanding of CIs. Self-declared experience with statistics was not related to researchers’ performance, and, even more surprisingly, researchers hardly outperformed the students, even though the students had not received any education on statistical inference whatsoever. Our findings suggest that many researchers do not know the correct interpretation of a CI. The misunderstandings surrounding p-values and CIs are particularly unfortunate because they constitute the main tools by which psychologists draw conclusions from data.

read more

Citations
More filters
Journal ArticleDOI

How much evidence is p > .05? Stimulus pre-testing and null primary outcomes in violent video games research.

TL;DR: In this paper, the authors apply a more appropriate Bayesian analysis to measure evidence for or against the null hypothesis relative to reasonable alternative hypotheses, and conclude that current methodological standards cannot rule out substantial confounds between violent and nonviolent video games.
Journal Article

After p Values: The New Statistics for Undergraduate Neuroscience Education.

TL;DR: A promising alternative approach to statistics curricula for neuroscience majors is what Cumming has dubbed the "New Statistics", an approach that emphasizes effect sizes, confidence intervals, meta-analysis, and open science.
Journal ArticleDOI

Does antecedent complexity affect ellipsis processing? An empirical investigation

TL;DR: The results are compatible with pointer-based approaches while casting doubt on the notion that changes antecedent complexity lead to measurable differences in ellipsis processing speed.
Proceedings ArticleDOI

Null hypothesis significance testing in simulation

Marko Hofmann
TL;DR: A critical reflection of the arguments contra NHST shows that although NHST is indeed ill-suited for many simulation applications and objectives it is by no means superfluous, neither in general, nor in particular for simulation.
Dissertation

Reliability, replicability and reproducibility in PET imaging

TL;DR: This thesis explores themes of reliability, replicability and reproducibility for PET research, which allows researchers to more effectively gauge the feasibility of new between-individual studies before collection of any data, and to focus their efforts on research questions which can be expected to yield more interpretable outcomes.
References
More filters
Journal ArticleDOI

The earth is round (p < .05)

TL;DR: The authors reviewed the problems with null hypothesis significance testing, including near universal misinterpretation of p as the probability that H is false, the misinterpretation that its complement is the probability of successful replication, and the mistaken assumption that if one rejects H₀ one thereby affirms the theory that led to the test.
Journal ArticleDOI

Statistical Methods in Psychology Journals: Guidelines and Explanations

TL;DR: The Task Force on Statistical Inference (TFSI) of the American Psychological Association (APA) as discussed by the authors was formed to discuss the application of significance testing in psychology journals and its alternatives, including alternative underlying models and data transformation.
Journal ArticleDOI

The Abuse of Power: The Pervasive Fallacy of Power Calculations for Data Analysis

TL;DR: The problem of post-experiment power calculation is discussed in this paper. But, the problem is extensive and present arguments to demonstrate the flaw in the logic, which is fundamentally flawed.
Journal ArticleDOI

Publication Manual of the American Psychological AssociationPublication Manual of the American Psychological Association.

TL;DR: The book provides stronger standards for maintaining the participant confidentiality and for reducing bias in language describing participants and suggests that researchers avoid the use of derogatory language such as using “minority” for “non-white” populations.
Book

Introduction to Probability and Statistics

TL;DR: The twelve edition of the Introduction to Probability and Statistics (INTRODUCTION TO PROBABILITY and STATISTICS) as discussed by the authors has been used by hundreds of thousands of students since its first edition.
Related Papers (5)