scispace - formally typeset
Open Access

Robust misinterpretation of confidence intervals

Reads0
Chats0
TLDR
Although all six statements were false, both researchers and students endorsed, on average, more than three statements, indicating a gross misunderstanding of CIs, which suggests that many researchers do not know the correct interpretation of a CI.
Abstract
Null hypothesis significance testing (NHST) is undoubtedly the most common inferential technique used to justify claims in the social sciences. However, even staunch defenders of NHST agree that its outcomes are often misinterpreted. Confidence intervals (CIs) have frequently been proposed as a more useful alternative to NHST, and their use is strongly encouraged in the APA Manual. Nevertheless, little is known about how researchers interpret CIs. In this study, 120 researchers and 442 students—all in the field of psychology—were asked to assess the truth value of six particular statements involving different interpretations of a CI. Although all six statements were false, both researchers and students endorsed, on average, more than three statements, indicating a gross misunderstanding of CIs. Self-declared experience with statistics was not related to researchers’ performance, and, even more surprisingly, researchers hardly outperformed the students, even though the students had not received any education on statistical inference whatsoever. Our findings suggest that many researchers do not know the correct interpretation of a CI. The misunderstandings surrounding p-values and CIs are particularly unfortunate because they constitute the main tools by which psychologists draw conclusions from data.

read more

Citations
More filters
Journal ArticleDOI

The fallacy of placing confidence in confidence intervals

TL;DR: It is shown in a number of examples that CIs do not necessarily have any of the properties of confidence intervals, and can lead to unjustified or arbitrary inferences, and is suggested that other theories of interval estimation should be used instead.
Journal ArticleDOI

Ordinal Regression Models in Psychology: A Tutorial

TL;DR: In psychology, ordinal variables, although extremely common in psychology, are almost exclusively analyzed with statistical models that falsely assume them to be metric as discussed by the authors, which can lead to distorted effect.
Journal ArticleDOI

Painfree and accurate Bayesian estimation of psychometric functions for (potentially) overdispersed data

TL;DR: It is shown that the use of the beta-binomial model makes it possible to determine accurate credible intervals even in data which exhibit substantial overdispersion, and Bayesian inference methods are used for estimating the posterior distribution of the parameters of the psychometric function.
Journal ArticleDOI

The philosophy of Bayes’ factors and the quantification of statistical evidence

TL;DR: In this article, the authors explore the concept of statistical evidence and how it can be quantified using the Bayes factor, and discuss the philosophical issues inherent in the use of the BFA.
References
More filters
Journal ArticleDOI

Calibration of p Values for Testing Precise Null Hypotheses

Abstract: P values are the most commonly used tool to measure evidence against a hypothesis or hypothesized model. Unfortunately, they are often incorrectly viewed as an error probability for rejection of the hypothesis or, even worse, as the posterior probability that the hypothesis is true. The fact that these interpretations can be completely misleading when testing precise hypotheses is first reviewed, through consideration of two revealing simulations. Then two calibrations of a ρ value are developed, the first being interpretable as odds and the second as either a (conditional) frequentist error probability or as the posterior probability of the hypothesis.
Journal ArticleDOI

Multiple comparisons : philosophies and illustrations

TL;DR: In this review, the statistical issue embedded in multiple comparisons is demonstrated, the philosophies of handling this issue are summarized, and the false discovery rate procedure may be the best practical solution to the problems of multiple comparisons that exist within physiology and other scientific disciplines.
BookDOI

What if there were no significance tests

TL;DR: Significance testing has been a controversial topic in the analysis of scientific data as discussed by the authors, with many opponents arguing that it should be replaced by confidence intervals instead of statistical significance tests.
Journal ArticleDOI

The case for objective Bayesian analysis

James O. Berger
- 01 Sep 2006 - 
TL;DR: It is suggested that the statistical community should accept formal objective Bayesian techniques with confidence, but should be more cautious about casual objectiveBayesian techniques.
Journal ArticleDOI

A primer on the understanding, use, and calculation of confidence intervals that are based on central and noncentral distributions

TL;DR: In this article, the authors discuss four reasons for promoting use of confidence intervals: they are readily interpretable, are linked to familiar statistical significance tests, can encourage meta-analytic thinking, and give information about precision.
Related Papers (5)