scispace - formally typeset
Search or ask a question

Showing papers by "Jelte M. Wicherts published in 2020"


Journal ArticleDOI
Eötvös Loránd University1, University of Amsterdam2, University of Southern California3, Cardiff University4, Columbia University5, University of Wisconsin-Madison6, Stanford University7, Maastricht University8, Springer Science+Business Media9, Emory University10, University of Melbourne11, University of Victoria12, University of Bristol13, University of New South Wales14, University of California, San Diego15, University College London16, University of Illinois at Chicago17, Tilburg University18, University of Toronto19, University of Lausanne20, Ohio State University21, University of Münster22, University of North Carolina at Charlotte23, University of Texas at Austin24, Old Dominion University25, University of California, Davis26, Georgia Southern University27, University of Modena and Reggio Emilia28, University of Massachusetts Boston29, Michigan State University30, University of Massachusetts Amherst31, Royal College of Surgeons in Ireland32, University of Nebraska Omaha33, University of Florence34, University of Missouri35, The College of New Jersey36, Leiden University37, Georgia Institute of Technology38, University of Western Australia39, Carnegie Mellon University40, Radboud University Nijmegen41, University of Zurich42, University of York43, University of Kent44, Case Western Reserve University45, University of Rhode Island46, University of California, Berkeley47, Northeastern University48, Boston College49, University of Miami50, Vanderbilt University51, University of California, San Francisco52, University of Michigan53, North Carolina State University54, University of Cincinnati55
TL;DR: A consensus-based checklist to improve and document the transparency of research reports in social and behavioural research and to submit with their manuscript or post to a public repository is presented.
Abstract: We present a consensus-based checklist to improve and document the transparency of research reports in social and behavioural research. An accompanying online application allows users to complete the form and generate a report that they can submit with their manuscript or post to a public repository.

72 citations


Journal ArticleDOI
TL;DR: It is concluded that effective preregistration is challenging, and registration formats that provide effective guidance may improve the quality of research.
Abstract: Researchers face many, often seemingly arbitrary, choices in formulating hypotheses, designing protocols, collecting data, analyzing data, and reporting results. Opportunistic use of “researcher degrees of freedom” aimed at obtaining statistical significance increases the likelihood of obtaining and publishing false-positive results and overestimated effect sizes. Preregistration is a mechanism for reducing such degrees of freedom by specifying designs and analysis plans before observing the research outcomes. The effectiveness of preregistration may depend, in part, on whether the process facilitates sufficiently specific articulation of such plans. In this preregistered study, we compared 2 formats of preregistration available on the OSF: Standard Pre-Data Collection Registration and Prereg Challenge Registration (now called “OSF Preregistration,” http://osf.io/prereg/). The Prereg Challenge format was a “structured” workflow with detailed instructions and an independent review to confirm completeness; the “Standard” format was “unstructured” with minimal direct guidance to give researchers flexibility for what to prespecify. Results of comparing random samples of 53 preregistrations from each format indicate that the “structured” format restricted the opportunistic use of researcher degrees of freedom better (Cliff’s Delta = 0.49) than the “unstructured” format, but neither eliminated all researcher degrees of freedom. We also observed very low concordance among coders about the number of hypotheses (14%), indicating that they are often not clearly stated. We conclude that effective preregistration is challenging, and registration formats that provide effective guidance may improve the quality of research.

44 citations


Journal ArticleDOI
27 May 2020-PLOS ONE
TL;DR: Recommendations are provided to improve transparency in the reporting of the entire meta-analytic process, including the use of preregistration, data and workflow sharing, and explicit coding practices.
Abstract: To determine the reproducibility of psychological meta-analyses, we investigated whether we could reproduce 500 primary study effect sizes drawn from 33 published meta-analyses based on the information given in the meta-analyses, and whether recomputations of primary study effect sizes altered the overall results of the meta-analysis. Results showed that almost half (k = 224) of all sampled primary effect sizes could not be reproduced based on the reported information in the meta-analysis, mostly because of incomplete or missing information on how effect sizes from primary studies were selected and computed. Overall, this led to small discrepancies in the computation of mean effect sizes, confidence intervals and heterogeneity estimates in 13 out of 33 meta-analyses. We provide recommendations to improve transparency in the reporting of the entire meta-analytic process, including the use of preregistration, data and workflow sharing, and explicit coding practices.

31 citations


Journal ArticleDOI
TL;DR: A meta-analytic review of the literature on sex differences in the trust game (174 effect sizes) and the related gift-exchange game (35 effect sizes), based on parental investment theory and social role theory, is presented in this article.

29 citations


Journal ArticleDOI
TL;DR: The findings show little evidence of widespread heterogeneity in direct replication studies in social and cognitive psychology, suggesting that minor changes in sample population and settings are unlikely to affect research outcomes in these fields of psychology.
Abstract: We examined the evidence for heterogeneity (of effect sizes) when only minor changes to sample population and settings were made between studies and explored the association between heterogeneity and average effect size in a sample of 68 meta-analyses from 13 preregistered multilab direct replication projects in social and cognitive psychology. Among the many examined effects, examples include the Stroop effect, the "verbal overshadowing" effect, and various priming effects such as "anchoring" effects. We found limited heterogeneity; 48/68 (71%) meta-analyses had nonsignificant heterogeneity, and most (49/68; 72%) were most likely to have zero to small heterogeneity. Power to detect small heterogeneity (as defined by Higgins, Thompson, Deeks, & Altman, 2003) was low for all projects (mean 43%), but good to excellent for medium and large heterogeneity. Our findings thus show little evidence of widespread heterogeneity in direct replication studies in social and cognitive psychology, suggesting that minor changes in sample population and settings are unlikely to affect research outcomes in these fields of psychology. We also found strong correlations between observed average effect sizes (standardized mean differences and log odds ratios) and heterogeneity in our sample. Our results suggest that heterogeneity and moderation of effects is unlikely for a 0 average true effect size, but increasingly likely for larger average true effect size. (PsycInfo Database Record (c) 2020 APA, all rights reserved).

25 citations


Journal ArticleDOI
TL;DR: It is concluded that intelligence research does show signs of low power and publication bias, but that these problems seem less severe than in many other scientific fields.
Abstract: In this meta-study, we analyzed 2442 effect sizes from 131 meta-analyses in intelligence research, published from 1984 to 2014, to estimate the average effect size, median power, and evidence for bias. We found that the average effect size in intelligence research was a Pearson’s correlation of 0.26, and the median sample size was 60. Furthermore, across primary studies, we found a median power of 11.9% to detect a small effect, 54.5% to detect a medium effect, and 93.9% to detect a large effect. We documented differences in average effect size and median estimated power between different types of intelligence studies (correlational studies, studies of group differences, experiments, toxicology, and behavior genetics). On average, across all meta-analyses (but not in every meta-analysis), we found evidence for small-study effects, potentially indicating publication bias and overestimated effects. We found no differences in small-study effects between different study types. We also found no convincing evidence for the decline effect, US effect, or citation bias across meta-analyses. We concluded that intelligence research does show signs of low power and publication bias, but that these problems seem less severe than in many other scientific fields.

24 citations


Journal ArticleDOI
31 Jul 2020-PLOS ONE
TL;DR: It was found that PCRs and IRBs more often included sample size decisions based on power analyses than the SPRs, which did not result in larger planned sample sizes, and there is ample room for improvements in the quality of the registrations.
Abstract: In this preregistered study, we investigated whether the statistical power of a study is higher when researchers are asked to make a formal power analysis before collecting data. We compared the sample size descriptions from two sources: (i) a sample of pre-registrations created according to the guidelines for the Center for Open Science Preregistration Challenge (PCRs) and a sample of institutional review board (IRB) proposals from Tilburg School of Behavior and Social Sciences, which both include a recommendation to do a formal power analysis, and (ii) a sample of pre-registrations created according to the guidelines for Open Science Framework Standard Pre-Data Collection Registrations (SPRs) in which no guidance on sample size planning is given. We found that PCRs and IRBs (72%) more often included sample size decisions based on power analyses than the SPRs (45%). However, this did not result in larger planned sample sizes. The determined sample size of the PCRs and IRB proposals (Md = 90.50) was not higher than the determined sample size of the SPRs (Md = 126.00; W = 3389.5, p = 0.936). Typically, power analyses in the registrations were conducted with G*power, assuming a medium effect size, α = .05 and a power of .80. Only 20% of the power analyses contained enough information to fully reproduce the results and only 62% of these power analyses pertained to the main hypothesis test in the pre-registration. Therefore, we see ample room for improvements in the quality of the registrations and we offer several recommendations to do so.

10 citations