Please use this identifier to cite or link to this item: http://hdl.handle.net/10071/20766
Author(s): Landy, J. F.
Jia; M.
Ding, I. L.
Viganola, D.
Tierney, W
Dreber, A.
Johannesson, M.
Pfeiffer, T.
Ebersole, C.
Gronau, Q. F.
Ly, A.
van den Bergh, D.
Marsman, M.
Derks, K.
Wagenmakers, E.-J.
Proctor, A.
Bartels, D. M.
Bauman, C. W.
Brady, W. J.
Cheung, F.
Cimpian, A.
Dohle, S.
Donnellan, M. B.
Hahn, A.
Hall, M. P.
Jiménez-Leal, W.
Johnson, D. J.
Lucas, R. E.
Monin, B.
Montealegre, A.
Mullen, E.
Pang, J.
Ray, J.
Reinero, D. A.
Reynolds, J.
Sowden, W.
Storage, D.
Su, R.
Tworek, C. M.
Walco, D.
Wills, J.
Van Bavel, J. J.
Xu, X.
Yam, K. C.
Yang, X.
Cunningham, W. A.
Schweinsberg, M.
Urwitz, M.
Uhlmann, Eric L.
Horchak, O.V.
Crowdsourcing Hypothesis Tests Col
Date: 2020
Title: Crowdsourcing hypothesis tests: making transparent how design choices shape research results
Volume: 146
Number: 5
Pages: 451 - 479
ISSN: 0033-2909
DOI (Digital Object Identifier): 10.1037/bul0000220
Keywords: Conceptual replications
Crowdsourcing
Forecasting
Research robustness
Scientific transparency
Abstract: To what extent are research results influenced by subjective decisions that scientists make as they design studies? Fifteen research teams independently designed studies to answer five original research questions related to moral judgments, negotiations, and implicit cognition. Participants from 2 separate large samples (total N > 15,000) were then randomly assigned to complete 1 version of each study. Effect sizes varied dramatically across different sets of materials designed to test the same hypothesis: Materials from different teams rendered statistically significant effects in opposite directions for 4 of 5 hypotheses, with the narrowest range in estimates being d = -0.37 to + 0.26. Meta-analysis and a Bayesian perspective on the results revealed overall support for 2 hypotheses and a lack of support for 3 hypotheses. Overall, practically none of the variability in effect sizes was attributable to the skill of the research team in designing materials, whereas considerable variability was attributable to the hypothesis being tested. In a forecasting survey, predictions of other scientists were significantly correlated with study results, both across and within hypotheses. Crowdsourced testing of research hypotheses helps reveal the true consistency of empirical support for a scientific claim.
Peerreviewed: yes
Access type: Open Access
Appears in Collections:CIS-RI - Artigos em revistas científicas internacionais com arbitragem científica

Files in This Item:
File Description SizeFormat 
Landy-et-al.-2020-PsychologicalBulletin.pdfVersão Submetida2,31 MBAdobe PDFView/Open


FacebookTwitterDeliciousLinkedInDiggGoogle BookmarksMySpaceOrkut
Formato BibTex mendeley Endnote Logotipo do DeGóis Logotipo do Orcid 

Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.