1
|
Protzko J, Krosnick J, Nelson L, Nosek BA, Axt J, Berent M, Buttrick N, DeBell M, Ebersole CR, Lundmark S, MacInnis B, O'Donnell M, Perfecto H, Pustejovsky JE, Roeder SS, Walleczek J, Schooler JW. High replicability of newly discovered social-behavioural findings is achievable. Nat Hum Behav 2024; 8:311-319. [PMID: 37945809 PMCID: PMC10896719 DOI: 10.1038/s41562-023-01749-9] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/25/2023] [Accepted: 10/05/2023] [Indexed: 11/12/2023]
Abstract
Failures to replicate evidence of new discoveries have forced scientists to ask whether this unreliability is due to suboptimal implementation of methods or whether presumptively optimal methods are not, in fact, optimal. This paper reports an investigation by four coordinated laboratories of the prospective replicability of 16 novel experimental findings using rigour-enhancing practices: confirmatory tests, large sample sizes, preregistration and methodological transparency. In contrast to past systematic replication efforts that reported replication rates averaging 50%, replication attempts here produced the expected effects with significance testing (P < 0.05) in 86% of attempts, slightly exceeding the maximum expected replicability based on observed effect sizes and sample sizes. When one lab attempted to replicate an effect discovered by another lab, the effect size in the replications was 97% that in the original study. This high replication rate justifies confidence in rigour-enhancing methods to increase the replicability of new discoveries.
Collapse
Affiliation(s)
- John Protzko
- Department of Psychological & Brain Sciences, University of California, Santa Barbara, Santa Barbara, CA, USA.
- Department of Psychological Science, Central Connecticut State University, New Britain, CT, USA.
| | - Jon Krosnick
- Institute for Research in the Social Sciences, Stanford University, Stanford, CA, USA
| | - Leif Nelson
- Haas School of Business, University of California, Berkeley, Berkeley, CA, USA
| | - Brian A Nosek
- Center for Open Science, Charlottesville, VA, USA
- Department of Psychology, University of Virginia, Charlottesville, VA, USA
| | - Jordan Axt
- Department of Psychology, McGill University, Montreal, Quebec, Canada
| | | | - Nicholas Buttrick
- Department of Psychology, University of Wisconsin-Madison, Madison, WI, USA
| | - Matthew DeBell
- Institute for Research in the Social Sciences, Stanford University, Stanford, CA, USA
| | - Charles R Ebersole
- Department of Psychology, University of Virginia, Charlottesville, VA, USA
| | | | - Bo MacInnis
- Institute for Research in the Social Sciences, Stanford University, Stanford, CA, USA
| | - Michael O'Donnell
- McDonough School of Business, Georgetown University, Washington, DC, USA
| | - Hannah Perfecto
- Olin School of Business, Washington University in St. Louis, St. Louis, MO, USA
| | - James E Pustejovsky
- Educational Psychology Department, University of Wisconsin-Madison, Madison, WI, USA
| | - Scott S Roeder
- Darla Moore School of Business, University of South Carolina, Columbia, SC, USA
| | | | - Jonathan W Schooler
- Department of Psychological & Brain Sciences, University of California, Santa Barbara, Santa Barbara, CA, USA
| |
Collapse
|
2
|
Botvinik-Nezer R, Wager TD. Reproducibility in Neuroimaging Analysis: Challenges and Solutions. BIOLOGICAL PSYCHIATRY. COGNITIVE NEUROSCIENCE AND NEUROIMAGING 2023; 8:780-788. [PMID: 36906444 DOI: 10.1016/j.bpsc.2022.12.006] [Citation(s) in RCA: 7] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/15/2022] [Revised: 11/27/2022] [Accepted: 12/11/2022] [Indexed: 12/23/2022]
Abstract
Recent years have marked a renaissance in efforts to increase research reproducibility in psychology, neuroscience, and related fields. Reproducibility is the cornerstone of a solid foundation of fundamental research-one that will support new theories built on valid findings and technological innovation that works. The increased focus on reproducibility has made the barriers to it increasingly apparent, along with the development of new tools and practices to overcome these barriers. Here, we review challenges, solutions, and emerging best practices with a particular emphasis on neuroimaging studies. We distinguish 3 main types of reproducibility, discussing each in turn. Analytical reproducibility is the ability to reproduce findings using the same data and methods. Replicability is the ability to find an effect in new datasets, using the same or similar methods. Finally, robustness to analytical variability refers to the ability to identify a finding consistently across variation in methods. The incorporation of these tools and practices will result in more reproducible, replicable, and robust psychological and brain research and a stronger scientific foundation across fields of inquiry.
Collapse
Affiliation(s)
- Rotem Botvinik-Nezer
- Department of Psychological and Brain Sciences, Dartmouth College, Hanover, New Hampshire.
| | - Tor D Wager
- Department of Psychological and Brain Sciences, Dartmouth College, Hanover, New Hampshire
| |
Collapse
|
3
|
Baumeister RF, Tice DM, Bushman BJ. A Review of Multisite Replication Projects in Social Psychology: Is It Viable to Sustain Any Confidence in Social Psychology's Knowledge Base? PERSPECTIVES ON PSYCHOLOGICAL SCIENCE 2023; 18:912-935. [PMID: 36442681 DOI: 10.1177/17456916221121815] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/20/2023]
Abstract
Multisite (multilab/many-lab) replications have emerged as a popular way of verifying prior research findings, but their record in social psychology has prompted distrust of the field and a sense of crisis. We review all 36 multisite social-psychology replications (plus three articles reporting multiple ministudies). We start by assuming that both the original and the multisite replications were conducted in honest and diligent fashion, despite often yielding different conclusions. Four of the 36 (11%) were clearly successful in terms of providing significant support for the original hypothesis, and five others (14%) had mixed results. The remaining 27 (75%) were failures. Multiple explanations for the generally poor record of replications are considered, including the possibility that the original hypothesis was wrong; operational failure; low engagement of participants; and bias toward failure. The relevant evidence is assessed as well. There was evidence for each of the possibilities listed above, with low engagement emerging as a widespread problem (reflected in high rates of discarded data and weak manipulation checks). The few procedures with actual interpersonal interaction fared much better than others. We discuss implications in relation to manipulation checks, effect sizes, and impact on the field and offer recommendations for improving future multisite projects.
Collapse
|
4
|
Schauer JM. On the Accuracy of Replication Failure Rates. MULTIVARIATE BEHAVIORAL RESEARCH 2023; 58:598-615. [PMID: 37339430 DOI: 10.1080/00273171.2022.2066500] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/22/2023]
Abstract
A prominent approach to studying the replication crisis has been to conduct replications of several different scientific findings as part of the same research effort. The reported proportion of findings that these programs determined failed to replicate have become important statistics in the replication crisis. However, these "failure rates" are based on decisions about whether individual studies replicated, which are themselves subject to statistical uncertainty. In this article, we examine how that uncertainty impacts the accuracy of reported failure rates and find that the reported failure rates can be substantially biased and highly variable. Indeed, very high or very low failure rates could arise from chance alone.
Collapse
|
5
|
Protzko J, Schooler JW. Moral contamination: Perceptions of good (but not bad) deeds depend on the ethical history of the actor. Front Psychol 2023; 13:1025214. [PMID: 36743620 PMCID: PMC9892465 DOI: 10.3389/fpsyg.2022.1025214] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/22/2022] [Accepted: 12/05/2022] [Indexed: 01/20/2023] Open
Abstract
In the majority of moral decision-making research, we are asked to consider the action of someone we know little about-an anonymous actor. This is inconsistent with our everyday judgments of the actions of others. Here we test the novel prediction of whether actions are considered as comparably virtuous or malignant when performed by a good person, an immoral person, or the standard anonymous actor. Across four sets of experiments (nine studies in total), we show that the moral status of the actor contaminates peoples' evaluations of the virtue of their actions. Even without ulterior motives, people do not judge good acts consistently across actors. We also discover a dose-response relationship where the more immoral the actor has been in the past-the less credit they are given for a good action in the present. This process does not occur for good people performing bad acts, however. Bad acts are bad regardless of who commits them. These results give new insights into the way people evaluate the behaviors of others.
Collapse
Affiliation(s)
- John Protzko
- Department of Psychological Science, Central Connecticut State University, New Britain, CT, United States,*Correspondence: John Protzko,
| | - Jonathan W. Schooler
- Department of Brain and Biological Science, University of California, Santa Barbara, Santa Barbara, CA, United States
| |
Collapse
|
6
|
Sternisko A, Delouvée S, Van Bavel JJ. Clarifying the relationship between randomness dismissal and conspiracist ideation: A preregistered replication and meta-analysis. JOURNAL OF EXPERIMENTAL SOCIAL PSYCHOLOGY 2022. [DOI: 10.1016/j.jesp.2022.104357] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
|
7
|
Dougherty MR, Horne Z. Citation counts and journal impact factors do not capture some indicators of research quality in the behavioural and brain sciences. ROYAL SOCIETY OPEN SCIENCE 2022; 9:220334. [PMID: 35991336 PMCID: PMC9382220 DOI: 10.1098/rsos.220334] [Citation(s) in RCA: 14] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/21/2022] [Accepted: 07/22/2022] [Indexed: 05/29/2023]
Abstract
Citation data and journal impact factors are important components of faculty dossiers and figure prominently in both promotion decisions and assessments of a researcher's broader societal impact. Although these metrics play a large role in high-stakes decisions, the evidence is mixed about whether they are strongly correlated with indicators of research quality. We use data from a large-scale dataset comprising 45 144 journal articles with 667 208 statistical tests and data from 190 replication attempts to assess whether citation counts and impact factors predict three indicators of research quality: (i) the accuracy of statistical reporting, (ii) the evidential value of the reported data and (iii) the replicability of a given experimental result. Both citation counts and impact factors were weak and inconsistent predictors of research quality, so defined, and sometimes negatively related to quality. Our findings raise the possibility that citation data and impact factors may be of limited utility in evaluating scientists and their research. We discuss the implications of these findings in light of current incentive structures and discuss alternative approaches to evaluating research.
Collapse
Affiliation(s)
| | - Zachary Horne
- Department of Psychology, University of Edinburgh, Edinburgh, UK
| |
Collapse
|
8
|
Abstract
The extent to which results from complex datasets generalize across contexts is critically important to numerous scientific fields as well as to practitioners who rely on such analyses to guide important strategic decisions. Our initiative systematically investigated whether findings from the field of strategic management would emerge in new time periods and new geographies. Original findings that were statistically reliable in the first place were typically obtained again in novel tests, suggesting surprisingly little sensitivity to context. For some social scientific areas of inquiry, results from a specific time and place can be a meaningful guide as to what will be observed more generally. This initiative examined systematically the extent to which a large set of archival research findings generalizes across contexts. We repeated the key analyses for 29 original strategic management effects in the same context (direct reproduction) as well as in 52 novel time periods and geographies; 45% of the reproductions returned results matching the original reports together with 55% of tests in different spans of years and 40% of tests in novel geographies. Some original findings were associated with multiple new tests. Reproducibility was the best predictor of generalizability—for the findings that proved directly reproducible, 84% emerged in other available time periods and 57% emerged in other geographies. Overall, only limited empirical evidence emerged for context sensitivity. In a forecasting survey, independent scientists were able to anticipate which effects would find support in tests in new samples.
Collapse
|
9
|
Nosek BA, Hardwicke TE, Moshontz H, Allard A, Corker KS, Dreber A, Fidler F, Hilgard J, Struhl MK, Nuijten MB, Rohrer JM, Romero F, Scheel AM, Scherer LD, Schönbrodt FD, Vazire S. Replicability, Robustness, and Reproducibility in Psychological Science. Annu Rev Psychol 2021; 73:719-748. [PMID: 34665669 DOI: 10.1146/annurev-psych-020821-114157] [Citation(s) in RCA: 107] [Impact Index Per Article: 35.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Replication-an important, uncommon, and misunderstood practice-is gaining appreciation in psychology. Achieving replicability is important for making research progress. If findings are not replicable, then prediction and theory development are stifled. If findings are replicable, then interrogation of their meaning and validity can advance knowledge. Assessing replicability can be productive for generating and testing hypotheses by actively confronting current understandings to identify weaknesses and spur innovation. For psychology, the 2010s might be characterized as a decade of active confrontation. Systematic and multi-site replication projects assessed current understandings and observed surprising failures to replicate many published findings. Replication efforts highlighted sociocultural challenges such as disincentives to conduct replications and a tendency to frame replication as a personal attack rather than a healthy scientific practice, and they raised awareness that replication contributes to self-correction. Nevertheless, innovation in doing and understanding replication and its cousins, reproducibility and robustness, has positioned psychology to improve research practices and accelerate progress. Expected final online publication date for the Annual Review of Psychology, Volume 73 is January 2022. Please see http://www.annualreviews.org/page/journal/pubdates for revised estimates.
Collapse
Affiliation(s)
- Brian A Nosek
- Department of Psychology, University of Virginia, Charlottesville, Virginia 22904, USA; .,Center for Open Science, Charlottesville, Virginia 22903, USA
| | - Tom E Hardwicke
- Department of Psychology, University of Amsterdam, 1012 ZA Amsterdam, The Netherlands
| | - Hannah Moshontz
- Addiction Research Center, University of Wisconsin-Madison, Madison, Wisconsin 53706, USA
| | - Aurélien Allard
- Department of Psychology, University of California, Davis, California 95616, USA
| | - Katherine S Corker
- Psychology Department, Grand Valley State University, Allendale, Michigan 49401, USA
| | - Anna Dreber
- Department of Economics, Stockholm School of Economics, 113 83 Stockholm, Sweden
| | - Fiona Fidler
- School of Biosciences, University of Melbourne, Parkville VIC 3010, Australia
| | - Joe Hilgard
- Department of Psychology, Illinois State University, Normal, Illinois 61790, USA
| | | | - Michèle B Nuijten
- Meta-Research Center, Tilburg University, 5037 AB Tilburg, The Netherlands
| | - Julia M Rohrer
- Department of Psychology, Leipzig University, 04109 Leipzig, Germany
| | - Felipe Romero
- Department of Theoretical Philosophy, University of Groningen, 9712 CP, The Netherlands
| | - Anne M Scheel
- Department of Industrial Engineering and Innovation Sciences, Eindhoven University of Technology, 5612 AZ Eindhoven, The Netherlands
| | - Laura D Scherer
- University of Colorado Anschutz Medical Campus, Aurora, Colorado 80045, USA
| | - Felix D Schönbrodt
- Department of Psychology, Ludwig Maximilian University of Munich, 80539 Munich, Germany
| | - Simine Vazire
- School of Psychological Sciences, University of Melbourne, Parkville VIC 3052, Australia
| |
Collapse
|
10
|
Chinman M, Acosta J, Ebener P, Shearer A. "What We Have Here, Is a Failure to [Replicate]": Ways to Solve a Replication Crisis in Implementation Science. PREVENTION SCIENCE : THE OFFICIAL JOURNAL OF THE SOCIETY FOR PREVENTION RESEARCH 2021; 23:739-750. [PMID: 34312769 DOI: 10.1007/s11121-021-01286-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 07/14/2021] [Indexed: 10/20/2022]
Abstract
Adapting the classic line from the 1967 film Cool Hand Luke, the title is meant to convey that implementation science (IS), like other fields, has not been embracing replication studies, which is a key component to the open science movement. The purpose of this article is to review what is known about replication of implementation trials and identify the gaps and next steps to continue increasing the transparency, openness, and replicability of implementation research. After presenting an overview of study replication and how it is a key component of open science, the article will examine how replication of implementation studies has (or more accurately has not) been approached in IS. As will be discussed, replication in IS shares some challenges with studies that attempt to replicate interventions, but also presents unique challenges. This article discusses different types of replications (e.g., direct vs. conceptual) and how they can benefit the field of IS. The article then presents a specific example of an implementation strategy called Getting To Outcomes© to describe how to design a replication study and interpret the results. The article ends with multiple options implementation scientists could consider to improve the likelihood and quality of replication studies. The discussion also envisions how implementation science can enable researchers and practitioners to work together in real-world contexts to encourage wide replication of implementation studies and advance the goal of improving public health.
Collapse
Affiliation(s)
- Matthew Chinman
- RAND Corporation, 4570 Fifth Ave #600, Pittsburgh, PA, 15213, USA.
| | - Joie Acosta
- RAND Corporation, 4570 Fifth Ave #600, Pittsburgh, PA, 15213, USA
| | - Patricia Ebener
- RAND Corporation, 4570 Fifth Ave #600, Pittsburgh, PA, 15213, USA
| | - Amy Shearer
- RAND Corporation, 4570 Fifth Ave #600, Pittsburgh, PA, 15213, USA
| |
Collapse
|
11
|
Klein J. Improving the reproducibility of findings by updating research methodology. ACTA ACUST UNITED AC 2021; 56:1597-1609. [PMID: 34257468 PMCID: PMC8265723 DOI: 10.1007/s11135-021-01196-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 06/26/2021] [Indexed: 11/30/2022]
Abstract
The literature discusses causes of low reproducibility of scientific publications. Our article adds another main cause—uncritical adherence to accepted research procedures. This is evident in: (1) anachronistically requiring researchers to base themselves on theoretical background even if the studies cited were not tested for reproducibility; (2) conducting studies suffering from a novelty effect bias; (3) forcing researchers who use data mining methods and field-based theory, with no preliminary theoretical rationale, to present a theoretical background that allegedly guided their work—as a precondition for publication of their findings. It is possible to increase research validity in relation to the above problems by the following means: (1) Conducting a longitudinal study on the same participants and only on them; (2) Trying to shorten the time period between laboratory experiments and those on humans, based on cost–benefit considerations, anchored in ethical norms; (3) Reporting the theoretical background in a causal modular format; (4) Giving incentives to those who meet the above criteria while moderating the pressure for fast output.
Collapse
Affiliation(s)
- Joseph Klein
- School of Education, Bar-Ilan University, 52900 Ramat-Gan, Israel
| |
Collapse
|
12
|
Gordon M, Viganola D, Dreber A, Johannesson M, Pfeiffer T. Predicting replicability-Analysis of survey and prediction market data from large-scale forecasting projects. PLoS One 2021; 16:e0248780. [PMID: 33852589 PMCID: PMC8046229 DOI: 10.1371/journal.pone.0248780] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/18/2020] [Accepted: 03/05/2021] [Indexed: 12/21/2022] Open
Abstract
The reproducibility of published research has become an important topic in science policy. A number of large-scale replication projects have been conducted to gauge the overall reproducibility in specific academic fields. Here, we present an analysis of data from four studies which sought to forecast the outcomes of replication projects in the social and behavioural sciences, using human experts who participated in prediction markets and answered surveys. Because the number of findings replicated and predicted in each individual study was small, pooling the data offers an opportunity to evaluate hypotheses regarding the performance of prediction markets and surveys at a higher power. In total, peer beliefs were elicited for the replication outcomes of 103 published findings. We find there is information within the scientific community about the replicability of scientific findings, and that both surveys and prediction markets can be used to elicit and aggregate this information. Our results show prediction markets can determine the outcomes of direct replications with 73% accuracy (n = 103). Both the prediction market prices, and the average survey responses are correlated with outcomes (0.581 and 0.564 respectively, both p < .001). We also found a significant relationship between p-values of the original findings and replication outcomes. The dataset is made available through the R package "pooledmaRket" and can be used to further study community beliefs towards replications outcomes as elicited in the surveys and prediction markets.
Collapse
Affiliation(s)
| | | | - Anna Dreber
- Stockholm School of Economics, Stockholm, Sweden
| | | | | |
Collapse
|
13
|
Erlandsson A, Wingren M, Andersson PA. Type and amount of help as predictors for impression of helpers. PLoS One 2020; 15:e0243808. [PMID: 33306708 PMCID: PMC7732071 DOI: 10.1371/journal.pone.0243808] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/26/2020] [Accepted: 11/26/2020] [Indexed: 11/20/2022] Open
Abstract
Impression of helpers can vary as a function of the magnitude of helping (amount of help) and of situational and motivational aspects (type of help). Over three studies conducted in Sweden and the US, we manipulated both the amount and the type of help in ten diverse vignettes and measured participants' impressions of the described helpers. Impressions were almost unaffected when increasing the amount of help by 500%, but clearly affected by several type of help-manipulations. Particularly, helpers were less positively evaluated if they had mixed motives for helping, did not experience intense emotions or empathy, or if helping involved no personal sacrifice. In line with the person-centered theory of moral judgment, people seem to form impressions of helpers primarily based on the presumed underlying processes and motives of prosociality rather than its consequences.
Collapse
Affiliation(s)
- Arvid Erlandsson
- Department of Behavioral Sciences and Learning, Linköping University, Linköping, Sweden
| | - Mattias Wingren
- Department of Behavioral Sciences and Learning, Linköping University, Linköping, Sweden
| | - Per A. Andersson
- Department of Behavioral Sciences and Learning, Linköping University, Linköping, Sweden
| |
Collapse
|
14
|
Baker ZG, Gentzis EA, Watlington EM, Castejon S, Petit WE, Britton M, Haddad S, DiBello AM, Rodriguez LM, Derrick JL, Knee CR. Reflections on a registered report replicating a body of dyadic cross-sectional research. PERSONAL RELATIONSHIPS 2020; 27:907-938. [PMID: 36419735 PMCID: PMC9681012 DOI: 10.1111/pere.12343] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/16/2023]
Abstract
This article reflects on a new kind of registered report (RR) that replicated the work of an early career researcher. The research items targeted in this RR were peer-reviewed, cross-sectional, dyadic studies to which the first author of this RR had contributed. The findings being replicated are not noteworthy for their prestige or representativeness of the wider field. Instead, this method of replication may have several benefits and less desirable qualities for the researcher and research team whose work is being replicated, for science more broadly, and for relationship science specifically, reviewed herein. The authors hope that this reflection inspires researchers to improve upon their methodology by incorporating replication of their work early and often into their own research process.
Collapse
Affiliation(s)
| | | | | | | | | | - Maggie Britton
- Department of Psychology, University of Houston, Houston, Texas
| | - Sana Haddad
- Department of Psychology, University of Houston, Houston, Texas
| | - Angelo M. DiBello
- Department of Psychology, City University of New York, Brooklyn College, Brooklyn, New York
- Center for Alcohol and Addiction Studies, Department of Behavioral and Social Sciences, Brown University, Providence, Rhode Island
| | - Lindsey M. Rodriguez
- Department of Psychology, University of South Florida St Petersburg, St. Petersburg, Florida
| | - Jaye L. Derrick
- Department of Psychology, University of Houston, Houston, Texas
| | - C. Raymond Knee
- Department of Psychology, University of Houston, Houston, Texas
| |
Collapse
|
15
|
Tierney W, Hardy JH, Ebersole CR, Leavitt K, Viganola D, Clemente EG, Gordon M, Dreber A, Johannesson M, Pfeiffer T, Uhlmann EL. Creative destruction in science. ORGANIZATIONAL BEHAVIOR AND HUMAN DECISION PROCESSES 2020. [DOI: 10.1016/j.obhdp.2020.07.002] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/23/2022]
|
16
|
Reinero DA, Wills JA, Brady WJ, Mende-Siedlecki P, Crawford JT, Van Bavel JJ. Is the Political Slant of Psychology Research Related to Scientific Replicability? PERSPECTIVES ON PSYCHOLOGICAL SCIENCE 2020; 15:1310-1328. [DOI: 10.1177/1745691620924463] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Social science researchers are predominantly liberal, and critics have argued this representation may reduce the robustness of research by embedding liberal values into the research process. In an adversarial collaboration, we examined whether the political slant of research findings in psychology is associated with lower rates of scientific replicability. We analyzed 194 original psychology articles reporting studies that had been subject to a later replication attempt ( N = 1,331,413 participants across replications) by having psychology doctoral students (Study 1) and an online sample of U.S. residents (Study 2) from across the political spectrum code the political slant (liberal vs. conservative) of the original research abstracts. The methods and analyses were preregistered. In both studies, the liberal or conservative slant of the original research was not associated with whether the results were successfully replicated. The results remained consistent regardless of the ideology of the coder. Political slant was unrelated to both subsequent citation patterns and the original study’s effect size and not consistently related to the original study’s sample size. However, we found modest evidence that research with greater political slant—whether liberal or conservative—was less replicable, whereas statistical robustness consistently predicted replication success. We discuss the implications for social science, politics, and replicability.
Collapse
Affiliation(s)
| | | | - William J. Brady
- Department of Psychology, New York University
- Department of Psychology, Yale University
| | | | | | - Jay J. Van Bavel
- Department of Psychology, New York University
- Center for Neural Science, New York University
| |
Collapse
|
17
|
Ghelfi E, Christopherson CD, Urry HL, Lenne RL, Legate N, Ann Fischer M, Wagemans FMA, Wiggins B, Barrett T, Bornstein M, de Haan B, Guberman J, Issa N, Kim J, Na E, O’Brien J, Paulk A, Peck T, Sashihara M, Sheelar K, Song J, Steinberg H, Sullivan D. Reexamining the Effect of Gustatory Disgust on Moral Judgment: A Multilab Direct Replication of Eskine, Kacinik, and Prinz (2011). ADVANCES IN METHODS AND PRACTICES IN PSYCHOLOGICAL SCIENCE 2020. [DOI: 10.1177/2515245919881152] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/22/2022]
Abstract
Eskine, Kacinik, and Prinz’s (2011) influential experiment demonstrated that gustatory disgust triggers a heightened sense of moral wrongness. We report a large-scale multisite direct replication of this study conducted by labs in the Collaborative Replications and Education Project. Subjects in each sample were randomly assigned to one of three beverage conditions: bitter (disgusting), control (neutral), or sweet. Then, subjects made a series of judgments about the moral wrongness of the behavior depicted in six vignettes. In the original study ( N = 57), drinking the bitter beverage led to higher ratings of moral wrongness than did drinking the control or sweet beverage; a contrast between the bitter condition and the other two conditions was significant among conservative ( n = 19) but not liberal ( n = 25) subjects. In the current project, random-effects meta-analyses across all subjects ( N = 1,137, k = 11 studies), conservative subjects ( n = 142, k = 5), and liberal subjects ( n = 635, k = 9) revealed standardized overall effect sizes across replications that were smaller than reported in the original study. Some were in the opposite of the predicted direction; all had 95% confidence intervals containing zero, and all were smaller than the effect size the original authors could have meaningfully detected. Results of linear mixed-effects regressions revealed that drinking the bitter beverage led to higher ratings of moral wrongness than did drinking the control beverage but not the sweet beverage. Bayes factor tests revealed greater relative support for the null than for the replication hypothesis. The overall pattern provides little to no support for the theory that physical disgust via taste perception harshens judgments of moral wrongness.
Collapse
Affiliation(s)
- Eric Ghelfi
- Department of Psychology, Brigham Young University
| | | | | | - Richie L. Lenne
- Department of Psychology, University of Minnesota, Minneapolis
| | - Nicole Legate
- Department of Psychology, Illinois Institute of Technology
| | | | - Fieke M. A. Wagemans
- Department of Psychology, Tilburg University
- Institute for Socio-Economics, University of Duisburg-Essen
| | - Brady Wiggins
- Department of Psychology, Brigham Young University–Idaho
| | | | | | | | | | - Nada Issa
- Department of Psychology, Indiana University Northwest
| | - Joan Kim
- Department of Medicine, Massachusetts General Hospital, Boston, Massachusetts
| | - Elim Na
- Department of Medicine, Boston University School of Medicine
| | | | - Aidan Paulk
- Department of Acupuncture, Oregon College of Oriental Medicine
| | - Tayler Peck
- Department of Psychology, Southern Oregon University
| | | | - Karen Sheelar
- Department of Psychology, Southern Oregon University
| | | | - Hannah Steinberg
- Department of Psychology, Pacific Graduate School of Psychology/Stanford University Doctor of Psychology Consortium
| | | |
Collapse
|
18
|
Grahe JE, Cuccolo K, Leighton DC, Cramblet Alvarez LD. Open Science Promotes Diverse, Just, and Sustainable Research and Educational Outcomes. PSYCHOLOGY LEARNING AND TEACHING-PLAT 2019. [DOI: 10.1177/1475725719869164] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Open science initiatives, which are often collaborative efforts focused on making research more transparent, have experienced increasing popularity in the past decade. Open science principles of openness and transparency provide opportunities to advance diversity, justice, and sustainability by promoting diverse, just, and sustainable outcomes among both undergraduate and senior researchers. We review models that demonstrate the importance of greater diversity, justice, and sustainability in psychological science before describing how open science initiatives promote these values. Open science initiatives also promote diversity, justice, and sustainability through increased levels of inclusion and access, equitable distribution of opportunities and dissemination of knowledge, and increased sustainability stemming from increased generalizability. In order to provide an application of the concepts discussed, we offer a set of diversity, justice, and sustainability lens questions for individuals to use while assessing research projects and other organizational systems and consider concrete classroom applications for these initiatives.
Collapse
Affiliation(s)
- Jon E Grahe
- Psychology Department, Pacific Lutheran University, USA
| | - Kelly Cuccolo
- Psychology Department, University of North Dakota, USA
| | - Dana C Leighton
- College of Arts, Science, and Education, Texas A&M University – Texarkana, USA
| | | |
Collapse
|
19
|
Uhlmann EL, Ebersole CR, Chartier CR, Errington TM, Kidwell MC, Lai CK, McCarthy RJ, Riegelman A, Silberzahn R, Nosek BA. Scientific Utopia III: Crowdsourcing Science. PERSPECTIVES ON PSYCHOLOGICAL SCIENCE 2019; 14:711-733. [PMID: 31260639 DOI: 10.1177/1745691619850561] [Citation(s) in RCA: 39] [Impact Index Per Article: 7.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Most scientific research is conducted by small teams of investigators who together formulate hypotheses, collect data, conduct analyses, and report novel findings. These teams operate independently as vertically integrated silos. Here we argue that scientific research that is horizontally distributed can provide substantial complementary value, aiming to maximize available resources, promote inclusiveness and transparency, and increase rigor and reliability. This alternative approach enables researchers to tackle ambitious projects that would not be possible under the standard model. Crowdsourced scientific initiatives vary in the degree of communication between project members from largely independent work curated by a coordination team to crowd collaboration on shared activities. The potential benefits and challenges of large-scale collaboration span the entire research process: ideation, study design, data collection, data analysis, reporting, and peer review. Complementing traditional small science with crowdsourced approaches can accelerate the progress of science and improve the quality of scientific research.
Collapse
Affiliation(s)
| | | | | | | | | | - Calvin K Lai
- 6 Department of Psychological and Brain Sciences, Washington University in St. Louis
| | - Randy J McCarthy
- 7 Center for the Study of Family Violence and Sexual Assault, Northern Illinois University
| | | | | | - Brian A Nosek
- 2 Department of Psychology, University of Virginia.,4 Center for Open Science, Charlottesville, Virginia
| |
Collapse
|
20
|
Button KS, Chambers CD, Lawrence N, Munafò MR. Grassroots Training for Reproducible Science: A Consortium-Based Approach to the Empirical Dissertation. PSYCHOLOGY LEARNING AND TEACHING-PLAT 2019. [DOI: 10.1177/1475725719857659] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
There is a widely acknowledged need to improve the reliability and efficiency of scientific research to increase the credibility of the published scientific literature and accelerate discovery. Widespread improvement requires a cultural shift in both thinking and practice, and better education will be instrumental to achieve this. Here we argue that education in reproducible science should start at the grassroots. We present our model of consortium-based student projects to train undergraduates in reproducible team science. We discuss how with careful design we have aligned collaboration with the current conventions for individual student assessment. We reflect on our experiences of several years running the GW4 Undergraduate Psychology Consortium offering insights we hope will be of practical use to others wishing to adopt a similar approach. We consider the pedagogical benefits of our approach in equipping students with 21st-century skills. Finally, we reflect on the need to shift incentives to reward to team science in global research and how this applies to the reward structures of student assessment.
Collapse
Affiliation(s)
| | - Christopher D. Chambers
- Cardiff University Brain Research Imaging Centre, School of Psychology, Cardiff University, UK
| | | | - Marcus R. Munafò
- MRC Integrative Epidemiology Unit at the University of Bristol, UK
- UK Centre for Tobacco and Alcohol Studies, School of Experimental Psychology, University of Bristol, UK
| |
Collapse
|
21
|
|
22
|
Laraway S, Snycerski S, Pradhan S, Huitema BE. An Overview of Scientific Reproducibility: Consideration of Relevant Issues for Behavior Science/Analysis. Perspect Behav Sci 2019; 42:33-57. [PMID: 31976420 PMCID: PMC6701706 DOI: 10.1007/s40614-019-00193-3] [Citation(s) in RCA: 37] [Impact Index Per Article: 7.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022] Open
Abstract
For over a decade, the failure to reproduce findings in several disciplines, including the biomedical, behavioral, and social sciences, have led some authors to claim that there is a so-called "replication (or reproducibility) crisis" in those disciplines. The current article examines: (a) various aspects of the reproducibility of scientific studies, including definitions of reproducibility; (b) published concerns about reproducibility in the scientific literature and public press; (c) variables involved in assessing the success of attempts to reproduce a study; (d) suggested factors responsible for reproducibility failures; (e) types of validity of experimental studies and threats to validity as they relate to reproducibility; and (f) evidence for threats to reproducibility in the behavior science/analysis literature. Suggestions for improving the reproducibility of studies in behavior science and analysis are described throughout.
Collapse
Affiliation(s)
- Sean Laraway
- Department of Psychology, San José State University, San José, CA 95192-0120 USA
| | - Susan Snycerski
- Department of Psychology, San José State University, San José, CA 95192-0120 USA
| | | | | |
Collapse
|
23
|
Hartshorne JK, Skorb L, Dietz SL, Garcia CR, Iozzo GL, Lamirato KE, Ledoux JR, Mu J, Murdock KN, Ravid J, Savery AA, Spizzirro JE, Trimm KA, van Horne KD, Vidal J. The Meta-Science of Adult Statistical Word Segmentation: Part 1. COLLABRA: PSYCHOLOGY 2019. [DOI: 10.1525/collabra.181] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022] Open
Abstract
We report the first set of results in a multi-year project to assess the robustness – and the factors promoting robustness – of the adult statistical word segmentation literature. This includes eight total experiments replicating six different experiments. The purpose of these replications is to assess the reproducibility of reported experiments, examine the replicability of their results, and provide more accurate effect size estimates. Reproducibility was mixed, with several papers either lacking crucial details or containing errors in the description of method, making it difficult to ascertain what was done. Replicability was also mixed: although in every instance we confirmed above-chance statistical word segmentation, many theoretically important moderations of that learning failed to replicate. Moreover, learning success was generally much lower than in the original studies. In the General Discussion, we consider whether these differences are due to differences in subject populations, low power in the original studies, or some combination of these and other factors. We also consider whether these findings are likely to generalize to the broader statistical word segmentation literature.
Collapse
Affiliation(s)
| | | | | | | | | | | | | | - Jesse Mu
- Department of Psychology, Boston College, US
| | | | - Jon Ravid
- Department of Psychology, Boston College, US
| | | | | | | | | | | |
Collapse
|
24
|
Moshontz H, Campbell L, Ebersole CR, IJzerman H, Urry HL, Forscher PS, Grahe JE, McCarthy RJ, Musser ED, Antfolk J, Castille CM, Evans TR, Fiedler S, Flake JK, Forero DA, Janssen SMJ, Keene JR, Protzko J, Aczel B, Solas SÁ, Ansari D, Awlia D, Baskin E, Batres C, Borras-Guevara ML, Brick C, Chandel P, Chatard A, Chopik WJ, Clarance D, Coles NA, Corker KS, Dixson BJW, Dranseika V, Dunham Y, Fox NW, Gardiner G, Garrison SM, Gill T, Hahn AC, Jaeger B, Kačmár P, Kaminski G, Kanske P, Kekecs Z, Kline M, Koehn MA, Kujur P, Levitan CA, Miller JK, Okan C, Olsen J, Oviedo-Trespalacios O, Özdoğru AA, Pande B, Parganiha A, Parveen N, Pfuhl G, Pradhan S, Ropovik I, Rule NO, Saunders B, Schei V, Schmidt K, Singh MM, Sirota M, Steltenpohl CN, Stieger S, Storage D, Sullivan GB, Szabelska A, Tamnes CK, Vadillo MA, Valentova JV, Vanpaemel W, Varella MAC, Vergauwe E, Verschoor M, Vianello M, Voracek M, Williams GP, Wilson JP, Zickfeld JH, Arnal JD, Aydin B, Chen SC, DeBruine LM, Fernandez AM, Horstmann KT, Isager PM, Jones B, Kapucu A, Lin H, Mensink MC, Navarrete G, Silan MA, Chartier CR. The Psychological Science Accelerator: Advancing Psychology through a Distributed Collaborative Network. ADVANCES IN METHODS AND PRACTICES IN PSYCHOLOGICAL SCIENCE 2018; 1:501-515. [PMID: 31886452 PMCID: PMC6934079 DOI: 10.1177/2515245918797607] [Citation(s) in RCA: 126] [Impact Index Per Article: 21.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
Concerns have been growing about the veracity of psychological research. Many findings in psychological science are based on studies with insufficient statistical power and nonrepresentative samples, or may otherwise be limited to specific, ungeneralizable settings or populations. Crowdsourced research, a type of large-scale collaboration in which one or more research projects are conducted across multiple lab sites, offers a pragmatic solution to these and other current methodological challenges. The Psychological Science Accelerator (PSA) is a distributed network of laboratories designed to enable and support crowdsourced research projects. These projects can focus on novel research questions, or attempt to replicate prior research, in large, diverse samples. The PSA's mission is to accelerate the accumulation of reliable and generalizable evidence in psychological science. Here, we describe the background, structure, principles, procedures, benefits, and challenges of the PSA. In contrast to other crowdsourced research networks, the PSA is ongoing (as opposed to time-limited), efficient (in terms of re-using structures and principles for different projects), decentralized, diverse (in terms of participants and researchers), and inclusive (of proposals, contributions, and other relevant input from anyone inside or outside of the network). The PSA and other approaches to crowdsourced psychological science will advance our understanding of mental processes and behaviors by enabling rigorous research and systematically examining its generalizability.
Collapse
|
25
|
Chartier CR, Riegelman A, McCarthy RJ. StudySwap: A Platform for Interlab Replication, Collaboration, and Resource Exchange. ADVANCES IN METHODS AND PRACTICES IN PSYCHOLOGICAL SCIENCE 2018. [DOI: 10.1177/2515245918808767] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
The resources needed to conduct psychological research (i.e., time, access to participants, equipment, expertise, etc.) are sometimes distributed inefficiently: Resources going unused at one lab may be needed to complete projects at another lab. This inefficient distribution of resources can be an impediment to scientific progress and to individuals’ careers. StudySwap ( https://osf.io/view/StudySwap ) is an online platform where researchers can post brief de-scriptions of research resources that are available for use ( haves) or that they need and another researcher may have ( needs). This Tutorial provides instructions for posting haves and needs on StudySwap, responding to the posts of other researchers, and creating exchange agreements that define expectations of all collaborative parties prior to data collection or other resource exchanges. Ultimately, we hope that StudySwap can be used to increase the efficiency with which psychology’s collective research resources are being used.
Collapse
Affiliation(s)
| | | | - Randy J. McCarthy
- Center for the Study of Family Violence and Assault, Northern Illinois University
| |
Collapse
|
26
|
Schaerer M, du Plessis C, Yap AJ, Thau S. Low power individuals in social power research: A quantitative review, theoretical framework, and empirical test. ORGANIZATIONAL BEHAVIOR AND HUMAN DECISION PROCESSES 2018. [DOI: 10.1016/j.obhdp.2018.08.004] [Citation(s) in RCA: 46] [Impact Index Per Article: 7.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
27
|
Kerr NL, Ao X, Hogg MA, Zhang J. Addressing replicability concerns via adversarial collaboration: Discovering hidden moderators of the minimal intergroup discrimination effect. JOURNAL OF EXPERIMENTAL SOCIAL PSYCHOLOGY 2018. [DOI: 10.1016/j.jesp.2018.05.001] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|
28
|
Abstract
AbstractThe widespread replication of research findings in independent laboratories prior to publication is suggested as a complement to traditional replication approaches. The pre-publication independent replication approach further addresses three key concerns from replication skeptics by systematically taking context into account, reducing reputational costs for original authors and replicators, and increasing the theoretical value of failed replications.
Collapse
|
29
|
Ernst AF, Hoekstra R, Wagenmakers EJ, Gelman A, van Ravenzwaaij D. Do Researchers Anchor Their Beliefs on the Outcome of an Initial Study? Exp Psychol 2018; 65:158-169. [PMID: 29905114 DOI: 10.1027/1618-3169/a000402] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
Abstract
As a research field expands, scientists have to update their knowledge and integrate the outcomes of a sequence of studies. However, such integrative judgments are generally known to fall victim to a primacy bias where people anchor their judgments on the initial information. In this preregistered study we tested the hypothesis that people anchor on the outcome of a small initial study, reducing the impact of a larger subsequent study that contradicts the initial result. Contrary to our expectation, undergraduates and academics displayed a recency bias, anchoring their judgment on the research outcome presented last. This recency bias is due to the fact that unsuccessful replications decreased trust in an effect more than did unsuccessful initial experiments. We recommend the time-reversal heuristic to account for temporal order effects during integration of research results.
Collapse
Affiliation(s)
- Anja Franziska Ernst
- 1 Department of Psychometrics and Statistics, University of Groningen, The Netherlands
| | - Rink Hoekstra
- 2 Department Educational Science, University of Groningen, The Netherlands
| | | | - Andrew Gelman
- 4 Department of Statistics, Columbia University, New York, NY, USA
| | - Don van Ravenzwaaij
- 1 Department of Psychometrics and Statistics, University of Groningen, The Netherlands
| |
Collapse
|
30
|
When positive ends tarnish the means: The morality of nonprofit more than of for-profit organizations is tainted by the use of compliance techniques. JOURNAL OF EXPERIMENTAL SOCIAL PSYCHOLOGY 2018. [DOI: 10.1016/j.jesp.2017.12.007] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
31
|
Why Do Some Psychology Researchers Resist Adopting Proposed Reforms to Research Practices? A Description of Researchers’ Rationales. ADVANCES IN METHODS AND PRACTICES IN PSYCHOLOGICAL SCIENCE 2018. [DOI: 10.1177/2515245918757427] [Citation(s) in RCA: 31] [Impact Index Per Article: 5.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
In response to the replication crisis, many psychologists recommended that the field adopt several proposed reforms to research practices, such as preregistration, to make research more replicable. However, how researchers have received these proposals is not well known because, to our knowledge, no systematic investigation into use of these reforms has been conducted. We wanted to learn about the rationales researchers have for not adopting the proposed reforms. We analyzed survey data of 1,035 researchers in social and personality psychology who were asked to indicate whether they thought it was acceptable to not follow four specific proposed reforms and to explain their reasoning when they thought it was acceptable to not adopt these reforms. The four reforms were preregistering hypotheses and methods, making data publicly available online, conducting formal power analyses, and reporting effect sizes. Our results suggest that (a) researchers have adopted some of the proposed reforms (e.g., reporting effect sizes) more than others (e.g., preregistering studies) and (b) rationales for not adopting them reflect a need for more discussion and education about their utility and feasibility.
Collapse
|
32
|
Introduction to the special issue – A replication project in personality psychology. JOURNAL OF RESEARCH IN PERSONALITY 2018. [DOI: 10.1016/j.jrp.2017.11.004] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
|
33
|
Crone DL, Bode S, Murawski C, Laham SM. The Socio-Moral Image Database (SMID): A novel stimulus set for the study of social, moral and affective processes. PLoS One 2018; 13:e0190954. [PMID: 29364985 PMCID: PMC5783374 DOI: 10.1371/journal.pone.0190954] [Citation(s) in RCA: 14] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/30/2017] [Accepted: 12/22/2017] [Indexed: 12/24/2022] Open
Abstract
A major obstacle for the design of rigorous, reproducible studies in moral psychology is the lack of suitable stimulus sets. Here, we present the Socio-Moral Image Database (SMID), the largest standardized moral stimulus set assembled to date, containing 2,941 freely available photographic images, representing a wide range of morally (and affectively) positive, negative and neutral content. The SMID was validated with over 820,525 individual judgments from 2,716 participants, with normative ratings currently available for all images on affective valence and arousal, moral wrongness, and relevance to each of the five moral values posited by Moral Foundations Theory. We present a thorough analysis of the SMID regarding (1) inter-rater consensus, (2) rating precision, and (3) breadth and variability of moral content. Additionally, we provide recommendations for use aimed at efficient study design and reproducibility, and outline planned extensions to the database. We anticipate that the SMID will serve as a useful resource for psychological, neuroscientific and computational (e.g., natural language processing or computer vision) investigations of social, moral and affective processes. The SMID images, along with associated normative data and additional resources are available at https://osf.io/2rqad/.
Collapse
Affiliation(s)
- Damien L. Crone
- Melbourne School of Psychological Sciences, University of Melbourne, Melbourne, Australia
| | - Stefan Bode
- Melbourne School of Psychological Sciences, University of Melbourne, Melbourne, Australia
| | - Carsten Murawski
- Department of Finance, University of Melbourne, Melbourne, Australia
| | - Simon M. Laham
- Melbourne School of Psychological Sciences, University of Melbourne, Melbourne, Australia
| |
Collapse
|
34
|
Shrout PE, Rodgers JL. Psychology, Science, and Knowledge Construction: Broadening Perspectives from the Replication Crisis. Annu Rev Psychol 2018; 69:487-510. [DOI: 10.1146/annurev-psych-122216-011845] [Citation(s) in RCA: 279] [Impact Index Per Article: 46.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Affiliation(s)
- Patrick E. Shrout
- Department of Psychology, New York University, New York, New York 10003
| | - Joseph L. Rodgers
- Department of Psychology and Human Development, Peabody College, Vanderbilt University, Nashville, Tennessee 37205
| |
Collapse
|
35
|
Abstract
AbstractMany philosophers of science and methodologists have argued that the ability to repeat studies and obtain similar results is an essential component of science. A finding is elevated from single observation to scientific evidence when the procedures that were used to obtain it can be reproduced and the finding itself can be replicated. Recent replication attempts show that some high profile results – most notably in psychology, but in many other disciplines as well – cannot be replicated consistently. These replication attempts have generated a considerable amount of controversy, and the issue of whether direct replications have value has, in particular, proven to be contentious. However, much of this discussion has occurred in published commentaries and social media outlets, resulting in a fragmented discourse. To address the need for an integrative summary, we review various types of replication studies and then discuss the most commonly voiced concerns about direct replication. We provide detailed responses to these concerns and consider different statistical ways to evaluate replications. We conclude there are no theoretical or statistical obstacles to making direct replication a routine aspect of psychological science.
Collapse
|
36
|
DeSoto KA, Schweinsberg M. Replication data collection highlights value in diversity of replication attempts. Sci Data 2017; 4:170028. [PMID: 28291224 PMCID: PMC5349248 DOI: 10.1038/sdata.2017.28] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/06/2016] [Accepted: 02/10/2017] [Indexed: 11/08/2022] Open
Abstract
Researchers agree that replicability and reproducibility are key aspects of science. A collection of Data Descriptors published in Scientific Data presents data obtained in the process of attempting to replicate previously published research. These new replication data describe published and unpublished projects. The different papers in this collection highlight the many ways that scientific replications can be conducted, and they reveal the benefits and challenges of crucial replication research. The organizers of this collection encourage scientists to reuse the data contained in the collection for their own work, and also believe that these replication examples can serve as educational resources for students, early-career researchers, and experienced scientists alike who are interested in learning more about the process of replication.
Collapse
Affiliation(s)
- K. Andrew DeSoto
- Association for Psychological Science, 1800 Massachusetts Ave NW, Ste 402, Washington, District Of Columbia 20036-1218, USA
| | - Martin Schweinsberg
- Department of Organizational Behavior, ESMT Berlin, Schlossplatz 1, Berlin 10178, Germany
| |
Collapse
|
37
|
Tierney W, Schweinsberg M, Jordan J, Kennedy DM, Qureshi I, Sommer SA, Thornley N, Madan N, Vianello M, Awtrey E, Zhu LL, Diermeier D, Heinze JE, Srinivasan M, Tannenbaum D, Bivolaru E, Dana J, Davis-Stober CP, du Plessis C, Gronau QF, Hafenbrack AC, Liao EY, Ly A, Marsman M, Murase T, Schaerer M, Tworek CM, Wagenmakers EJ, Wong L, Anderson T, Bauman CW, Bedwell WL, Brescoll V, Canavan A, Chandler JJ, Cheries E, Cheryan S, Cheung F, Cimpian A, Clark MA, Cordon D, Cushman F, Ditto PH, Amell A, Frick SE, Gamez-Djokic M, Grady RH, Graham J, Gu J, Hahn A, Hanson BE, Hartwich NJ, Hein K, Inbar Y, Jiang L, Kellogg T, Legate N, Luoma TP, Maibeucher H, Meindl P, Miles J, Mislin A, Molden DC, Motyl M, Newman G, Ngo HH, Packham H, Ramsay PS, Ray JL, Sackett AM, Sellier AL, Sokolova T, Sowden W, Storage D, Sun X, Van Bavel JJ, Washburn AN, Wei C, Wetter E, Wilson CT, Darroux SC, Uhlmann EL. Data from a pre-publication independent replication initiative examining ten moral judgement effects. Sci Data 2016; 3:160082. [PMID: 27727246 PMCID: PMC5058337 DOI: 10.1038/sdata.2016.82] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/04/2016] [Accepted: 08/12/2016] [Indexed: 11/09/2022] Open
Abstract
We present the data from a crowdsourced project seeking to replicate findings in independent laboratories before (rather than after) they are published. In this Pre-Publication Independent Replication (PPIR) initiative, 25 research groups attempted to replicate 10 moral judgment effects from a single laboratory's research pipeline of unpublished findings. The 10 effects were investigated using online/lab surveys containing psychological manipulations (vignettes) followed by questionnaires. Results revealed a mix of reliable, unreliable, and culturally moderated findings. Unlike any previous replication project, this dataset includes the data from not only the replications but also from the original studies, creating a unique corpus that researchers can use to better understand reproducibility and irreproducibility in science.
Collapse
Affiliation(s)
- Warren Tierney
- INSEAD, Fontainebleau 77305, France and Singapore 138676, Singapore
| | | | | | | | - Israr Qureshi
- IE Business School, IE University, Madrid 28006, Spain
| | | | - Nico Thornley
- INSEAD, Fontainebleau 77305, France and Singapore 138676, Singapore
| | - Nikhil Madan
- INSEAD, Fontainebleau 77305, France and Singapore 138676, Singapore
| | | | - Eli Awtrey
- University of Washington, Seattle 98195, USA
| | - Luke Lei Zhu
- University of Manitoba, Winnipeg R3T 5V4, Canada
| | | | | | | | | | - Eliza Bivolaru
- INSEAD, Fontainebleau 77305, France and Singapore 138676, Singapore
| | | | | | | | | | - Andrew C Hafenbrack
- UCP-Católica Lisbon School of Business &Economics, Lisbon 1649-023, Portugal
| | - Eko Yi Liao
- Hang Seng Management College, Hong Kong, Hong Kong
| | - Alexander Ly
- University of Amsterdam, Amsterdam 1001 NK, The Netherlands
| | | | | | - Michael Schaerer
- INSEAD, Fontainebleau 77305, France and Singapore 138676, Singapore
| | | | | | - Lynn Wong
- INSEAD, Fontainebleau 77305, France and Singapore 138676, Singapore
| | | | | | | | | | | | - Jesse J Chandler
- Institute for Social Research, University of Michigan, Ann Arbor 48104, USA
| | - Erik Cheries
- University of Massachusetts Amherst, Amherst 1003, USA
| | | | - Felix Cheung
- Washington University in St Louis, St Louis 63130, USA.,University of Hong Kong, Hong Kong, Hong Kong
| | - Andrei Cimpian
- Department of Psychology, New York University, New York 10003, USA
| | | | - Diana Cordon
- Illinois Institute of Technology, Chicago 60616, USA
| | | | | | - Alice Amell
- Illinois Institute of Technology, Chicago 60616, USA
| | | | | | | | - Jesse Graham
- University of Southern California, Los Angeles 90089, USA
| | - Jun Gu
- Monash University, Melbourne 3145, Australia
| | - Adam Hahn
- Social Cognition Center Cologne, University of Cologne, Koeln 50931, Germany
| | | | - Nicole J Hartwich
- Social Cognition Center Cologne, University of Cologne, Koeln 50931, Germany
| | - Kristie Hein
- Illinois Institute of Technology, Chicago 60616, USA
| | - Yoel Inbar
- University of Toronto, Toronto ON M5S, Canada
| | - Lily Jiang
- University of Washington, Seattle 98195, USA
| | | | - Nicole Legate
- Illinois Institute of Technology, Chicago 60616, USA
| | - Timo P Luoma
- Social Cognition Center Cologne, University of Cologne, Koeln 50931, Germany
| | | | - Peter Meindl
- University of Pennsylvania, Philadelphia 19104, USA
| | | | | | | | - Matt Motyl
- University of Illinois at Chicago, Chicago 60607, USA
| | | | - Hoai Huong Ngo
- Université Paris Ouest Nanterre La Défense, Nanterre 92000, France
| | | | | | - Jennifer L Ray
- Department of Psychology, New York University, New York 10003, USA
| | | | | | | | - Walter Sowden
- Centre for Psychiatry and Neuroscience, Walter Reed Army Institute of Research (WRAIR), Silver Spring 20910, USA
| | - Daniel Storage
- University of Illinois at Urbana-Champaign, Champaign 61820, USA
| | - Xiaomin Sun
- Beijing Normal University, Beijing 100875, China
| | - Jay J Van Bavel
- Department of Psychology, New York University, New York 10003, USA
| | | | - Cong Wei
- Beijing Normal University, Beijing 100875, China
| | - Erik Wetter
- Stockholm School of Economics, Stockholm 11383, Sweden
| | | | | | | |
Collapse
|
38
|
Dovidio JF. Commentary: A big problem requires a foundational change. JOURNAL OF EXPERIMENTAL SOCIAL PSYCHOLOGY 2016. [DOI: 10.1016/j.jesp.2016.01.008] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
|
39
|
Schaller M. The empirical benefits of conceptual rigor: Systematic articulation of conceptual hypotheses can reduce the risk of non-replicable results (and facilitate novel discoveries too). JOURNAL OF EXPERIMENTAL SOCIAL PSYCHOLOGY 2016. [DOI: 10.1016/j.jesp.2015.09.006] [Citation(s) in RCA: 19] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|
40
|
Stroebe W. Are most published social psychological findings false? JOURNAL OF EXPERIMENTAL SOCIAL PSYCHOLOGY 2016. [DOI: 10.1016/j.jesp.2015.09.017] [Citation(s) in RCA: 37] [Impact Index Per Article: 4.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
41
|
Exploring Small, Confirming Big: An alternative system to The New Statistics for advancing cumulative and replicable psychological research. JOURNAL OF EXPERIMENTAL SOCIAL PSYCHOLOGY 2016. [DOI: 10.1016/j.jesp.2015.09.013] [Citation(s) in RCA: 44] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
|
42
|
Berinsky AJ, Margolis MF, Sances MW. Can we turn shirkers into workers? JOURNAL OF EXPERIMENTAL SOCIAL PSYCHOLOGY 2016. [DOI: 10.1016/j.jesp.2015.09.010] [Citation(s) in RCA: 30] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
|
43
|
Stangor C, Lemay EP. Introduction to the Special Issue on Methodological Rigor and Replicability. JOURNAL OF EXPERIMENTAL SOCIAL PSYCHOLOGY 2016. [DOI: 10.1016/j.jesp.2016.02.006] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/16/2023]
|
44
|
Baumeister RF. Charting the future of social psychology on stormy seas: Winners, losers, and recommendations. JOURNAL OF EXPERIMENTAL SOCIAL PSYCHOLOGY 2016. [DOI: 10.1016/j.jesp.2016.02.003] [Citation(s) in RCA: 25] [Impact Index Per Article: 3.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
|
45
|
Van Bavel JJ, Mende-Siedlecki P, Brady WJ, Reinero DA. Contextual sensitivity in scientific reproducibility. Proc Natl Acad Sci U S A 2016; 113:6454-9. [PMID: 27217556 PMCID: PMC4988618 DOI: 10.1073/pnas.1521897113] [Citation(s) in RCA: 168] [Impact Index Per Article: 21.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
In recent years, scientists have paid increasing attention to reproducibility. For example, the Reproducibility Project, a large-scale replication attempt of 100 studies published in top psychology journals found that only 39% could be unambiguously reproduced. There is a growing consensus among scientists that the lack of reproducibility in psychology and other fields stems from various methodological factors, including low statistical power, researcher's degrees of freedom, and an emphasis on publishing surprising positive results. However, there is a contentious debate about the extent to which failures to reproduce certain results might also reflect contextual differences (often termed "hidden moderators") between the original research and the replication attempt. Although psychologists have found extensive evidence that contextual factors alter behavior, some have argued that context is unlikely to influence the results of direct replications precisely because these studies use the same methods as those used in the original research. To help resolve this debate, we recoded the 100 original studies from the Reproducibility Project on the extent to which the research topic of each study was contextually sensitive. Results suggested that the contextual sensitivity of the research topic was associated with replication success, even after statistically adjusting for several methodological characteristics (e.g., statistical power, effect size). The association between contextual sensitivity and replication success did not differ across psychological subdisciplines. These results suggest that researchers, replicators, and consumers should be mindful of contextual factors that might influence a psychological process. We offer several guidelines for dealing with contextual sensitivity in reproducibility.
Collapse
Affiliation(s)
- Jay J Van Bavel
- Department of Psychology, New York University, New York, NY 10003
| | | | - William J Brady
- Department of Psychology, New York University, New York, NY 10003
| | - Diego A Reinero
- Department of Psychology, New York University, New York, NY 10003
| |
Collapse
|
46
|
Abstract
Modern psychology is apparently in crisis and the prevailing view is that this partly reflects an inability to replicate past findings. If a crisis does exists, then it is some kind of ‘chronic’ crisis, as psychologists have been censuring themselves over replicability for decades. While the debate in psychology is not new, the lack of progress across the decades is disappointing. Recently though, we have seen a veritable surfeit of debate alongside multiple orchestrated and well-publicised replication initiatives. The spotlight is being shone on certain areas and although not everyone agrees on how we should interpret the outcomes, the debate is happening and impassioned. The issue of reproducibility occupies a central place in our whig history of psychology.
Collapse
|
47
|
Fiske ST. How to Publish Rigorous Experiments in the 21 st Century. JOURNAL OF EXPERIMENTAL SOCIAL PSYCHOLOGY 2016; 66:145-147. [PMID: 30555180 DOI: 10.1016/j.jesp.2016.01.006] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
Abstract
Crises provide an opportunity for the field to take stock, as do the articles in this special issue. Constructive advice for 21st century publication standards includes appropriate theory, internal validity, and external validity. First, well-grounded theory can produce a priori plausibility, testable logic, and a focus on the ideas involved, all cumulatively informed by meta-analysis across studies. Second, internal validity benefits from both exploratory work and confirmatory analyses on well-powered samples that require systematic detection and principled decisions about data quality. Inferences benefit from manipulated mediation analysis and from careful interpretation without over-claiming. Finally, external validity profits from a variety of exact and conceptual replications, best evaluated by meta-analysis.
Collapse
|