scispace - formally typeset
Open AccessJournal ArticleDOI

Confirmatory Methods, or Huge Samples, Are Required to Obtain Power for the Evaluation of Theories

TLDR
It is shown that many of the common choices in hypothesis testing led to a severely underpowered form of theory evaluation and that confirmatory methods are required in the context of theory Evaluation and that the scientific literature would benefit from a clearer distinction between confirmatory and exploratory findings.
Abstract
Experimental studies are usually designed with specific expectations about the results in mind. However, most researchers apply some form of omnibus test to test for any differences, with follow up tests like pairwise comparisons or simple effects analyses for further investigation of the effects. The power to find full support for the theory with such an exploratory approach which is usually based on multiple testing is, however, rather disappointing. With the simulations in this paper we showed that many of the common choices in hypothesis testing led to a severely underpowered form of theory evaluation. Furthermore, some less commonly used approaches were presented and a comparison of results in terms of power to find support for the theory was made. We concluded that confirmatory methods are required in the context of theory evaluation and that the scientific literature would benefit from a clearer distinction between confirmatory and exploratory findings. Also, we emphasis the importance of reporting all tests, significant or not, including the appropriate sample statistics like means and standard deviations. Another recommendation is related to the fact that researchers, when they discuss the conclusions of their own study, seem to underestimate the role of sampling variability. The execution of more replication studies in combination with proper reporting of all results provides insight in between study variability and the amount of chance findings.

read more

Content maybe subject to copyright    Report

Citations
More filters
Journal ArticleDOI

Hidden multiplicity in exploratory multiway ANOVA: Prevalence and remedies.

TL;DR: This work explains the multiple-comparison problem and demonstrates that researchers almost never correct for it, and describes four remedies: the omnibus F test, control of the familywise error rate, controls of the false discovery rate, and preregistration of the hypotheses.
Journal ArticleDOI

A tutorial on testing hypotheses using the Bayes factor.

TL;DR: After reading this tutorial and executing the associated code, researchers will be able to use their own data for the evaluation of hypotheses by means of the Bayes factor, not only in thecontext of ANOVA models, but also in the context of other statistical models.
Posted Content

Hidden Multiplicity in Multiway ANOVA: Prevalence, Consequences, and Remedies

TL;DR: In this article, the authors explain the multiple comparison problem in multiway analysis of variance and demonstrate that researchers almost never correct for it, and propose several correction procedures (i.e., sequential Bonferroni) and show that their application alters at least one of the substantive conclusions in 45 of 60 articles considered.
Posted Content

Hidden Multiplicity in Multiway ANOVA: Prevalence and Remedies

TL;DR: In this paper, the omnibus F test, the control of familywise error rate, the controlling of false discovery rate, and the preregistration of hypotheses are proposed to mitigate the multiple comparison problem.
Journal ArticleDOI

Do we fail to exert self-control because we lack resources or motivation? Competing theories to explain a debated phenomenon.

TL;DR: This paper found evidence that self-control performance was impaired after a high- versus a low-demand task in the no-incentive conditions, whereas participants in the incentive conditions showed higher intrinsic, autonomous motivation, they did not exert greater effort.
References
More filters
Journal ArticleDOI

G*Power 3: A flexible statistical power analysis program for the social, behavioral, and biomedical sciences

TL;DR: G*Power 3 provides improved effect size calculators and graphic options, supports both distribution-based and design-based input modes, and offers all types of power analyses in which users might be interested.
Journal ArticleDOI

The earth is round (p < .05)

TL;DR: The authors reviewed the problems with null hypothesis significance testing, including near universal misinterpretation of p as the probability that H is false, the misinterpretation that its complement is the probability of successful replication, and the mistaken assumption that if one rejects H₀ one thereby affirms the theory that led to the test.
Journal ArticleDOI

Stereotype Threat and Women's Math Performance

TL;DR: This article found that when the test was described as producing gender differences and stereotype threat was high, women performed substantially worse than equally qualified men did on difficult (but not easy) math tests among a highly selected sample of men and women.
Journal ArticleDOI

Null hypothesis significance testing : A review of an old and continuing controversy

TL;DR: The concluding opinion is that NHST is easily misunderstood and misused but that when applied with good judgment it can be an effective aid to the interpretation of experimental data.
Book

Statistical Evidence: A Likelihood Paradigm

TL;DR: The first principle of the Law of Likelihood as discussed by the authors is that the strength of evidence is limited by the expectation of the researcher's expectation, and the importance of the evidence is determined by the test of significance.
Related Papers (5)