1
|
Raymond M, Crochet PA. Carving Non-Proximal Explanations for Same-Sex Sexual Orientation. ARCHIVES OF SEXUAL BEHAVIOR 2023; 52:3007-3012. [PMID: 36469147 DOI: 10.1007/s10508-022-02497-z] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/14/2022] [Revised: 11/19/2022] [Accepted: 11/21/2022] [Indexed: 06/17/2023]
Affiliation(s)
- Michel Raymond
- CNRS, EPHE, IRD, Institut des Sciences de l'Evolution, Univ Montpellier, Place E. Bataillon, 34095, Montpellier Cedex 05, France.
| | | |
Collapse
|
2
|
Regenwetter M, Robinson MM, Wang C. Four Internal Inconsistencies in Tversky and Kahneman’s (1992) Cumulative Prospect Theory Article: A Case Study in Ambiguous Theoretical Scope and Ambiguous Parsimony. ADVANCES IN METHODS AND PRACTICES IN PSYCHOLOGICAL SCIENCE 2022. [DOI: 10.1177/25152459221074653] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Scholars heavily rely on theoretical scope as a tool to challenge existing theory. We advocate that scientific discovery could be accelerated if far more effort were invested into also overtly specifying and painstakingly delineating the intended purview of any proposed new theory at the time of its inception. As a case study, we consider Tversky and Kahneman (1992). They motivated their Nobel-Prize-winning cumulative prospect theory with evidence that in each of two studies, roughly half of the participants violated independence, a property required by expected utility theory (EUT). Yet even at the time of inception, new theories may reveal signs of their own limited scope. For example, we show that Tversky and Kahneman’s findings in their own test of loss aversion provide evidence that at least half of their participants violated their theory, in turn, in that study. We highlight a combination of conflicting findings in the original article that make it ambiguous to evaluate both cumulative prospect theory’s scope and its parsimony on the authors’ own evidence. The Tversky and Kahneman article is illustrative of a social and behavioral research culture in which theoretical scope plays an extremely asymmetric role: to call existing theory into question and motivate surrogate proposals.
Collapse
Affiliation(s)
- Michel Regenwetter
- Department of Psychology, University of Illinois at Urbana-Champaign, Champaign, Illinois
- Department of Political Science, University of Illinois at Urbana-Champaign, Urbana, Illinois
- Department of Electrical & Computer Engineering, University of Illinois at Urbana-Champaign, Urbana, Illinois
| | - Maria M. Robinson
- Department of Psychology, University of California San Diego, La Jolla, California
| | - Cihang Wang
- Department of Economics, University of Illinois at Urbana-Champaign, Urbana, Illinois
| |
Collapse
|
3
|
How Can Big Data Science Transform the Psychological Sciences? SPANISH JOURNAL OF PSYCHOLOGY 2020; 23:e44. [PMID: 33148362 DOI: 10.1017/sjp.2020.45] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/17/2022]
Abstract
Big data and related technologies are radically altering our society. In a similar way, these approaches can transform the psychological sciences. The goal of this commentary is to motivate psychologists to embrace big data science for the betterment of the field. Big data sources, algorithmic methods, and a culture that embraces prediction has the potential to advance our science, improve the robustness and replicability of our research, and allow us to focus more centrally on actual behaviors. We highlight these key transformations, acknowledge criticisms of big data approaches, and emphasize specific ways psychologists can contribute to the big data science revolution.
Collapse
|
4
|
Vollmer HC, Domma J. [Differences in the success rates of inpatient therapy for alcohol and illegal drug abusers. A Replication Study]. Encephale 2019; 46:102-109. [PMID: 31866104 DOI: 10.1016/j.encep.2019.09.008] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/22/2019] [Revised: 09/15/2019] [Accepted: 09/28/2019] [Indexed: 11/25/2022]
Abstract
INTRODUCTION Irrespective of the type of psychotherapy used, the abstinence-oriented treatment of drug abusers is less successful than that for alcohol abusers. If, on the other hand, the two groups are parallelized in such a way that the patients are identical with respect to the five characteristics of gender, age, schooling, work situation and partner situation, then there is no difference between the success rates of the drug and alcohol abusers. The aim of this study is to determine whether this result can be replicated in another therapeutic institution. METHOD Retrospective field study of 320 abusers of illegal drugs and 320 alcohol abusers who were treated with behaviour therapy. By combining the binary characteristics gender, work situation and age, the drug-dependent patients were divided into 23=8 groups, and the same number of alcohol abusers were randomly selected for each group. The scheduled period of inpatient treatment was 90 days for the alcohol abusers and 120 days for the drug abusers. Every week the patients had one session of individual psychotherapy and four to five group therapy sessions. According to the indications, the certified behaviour therapists implemented the following interventions including behaviour analysis, relapse prevention, cognitive therapy, self-management and behavioural family therapy. Comparison of the success rates was carried out using the Chi2 test, and changes in the psychological findings were tested with one-way variance analysis. RESULTS There was no difference between drug and alcohol abusers with respect to the rate of therapy termination according to plan (around 80%). A total of 48% of the drug abusers and 41 % of the alcohol abusers who could be followed up had been continuously abstinent at the one-year catamnesis without a single relapse. There were also no differences between the two groups when it was assumed that the patients who could not be followed up had relapsed. In the case of both the drug and alcohol abusers the abstinence rate was highest in over-29-year-old employed men (57.6%; 48.4%). The abstinence rate was lowest in employed female drug abusers (27.8%) and young, unemployed female drug abusers (0%, n=11). DISCUSSION What appears to influence the abstinence rate after inpatient treatment is not only the type of substance consumed but also sociodemographic characteristics. In addition to individually tailored therapy, our results confirm the importance of a highly differentiated presentation of the outcomes of therapy in the specialist literature. An average rate of abstinence (e.g. 30%) is insufficient to evaluate an intervention unless information is also provided about the patients for which the intervention is suitable and those for which it is not. In accordance with the Reproducibility Project, we consider replication studies essential in psychotherapy, even though in practice the considerable methodical requirements can only be partially fulfilled.
Collapse
Affiliation(s)
- H C Vollmer
- IFT Institut für Therapieforschung, Leopoldstr. 175, 80804 München, Allemagne.
| | - J Domma
- Salus Klinik, Rodderstr. 7a, 50354 Hürth, Allemagne
| |
Collapse
|
5
|
On the Replicability of Abductive Research in Management and Organizations: Internal Replication and Its Alternatives. ACADEMY OF MANAGEMENT DISCOVERIES 2019. [DOI: 10.5465/amd.2019.0121] [Citation(s) in RCA: 23] [Impact Index Per Article: 4.6] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
|
6
|
Uher J. Quantitative Data From Rating Scales: An Epistemological and Methodological Enquiry. Front Psychol 2018; 9:2599. [PMID: 30622493 PMCID: PMC6308206 DOI: 10.3389/fpsyg.2018.02599] [Citation(s) in RCA: 20] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/14/2018] [Accepted: 12/03/2018] [Indexed: 11/13/2022] Open
Abstract
Rating scales are popular methods for generating quantitative data directly by persons rather than automated technologies. But scholars increasingly challenge their foundations. This article contributes epistemological and methodological analyses of the processes involved in person-generated quantification. They are crucial for measurement because data analyses can reveal information about study phenomena only if relevant properties were encoded systematically in the data. The Transdisciplinary Philosophy-of-Science Paradigm for Research on Individuals (TPS-Paradigm) is applied to explore psychological and social-science concepts of measurement and quantification, including representational measurement theory, psychometric theories and their precursors in psychophysics. These are compared to theories from metrology specifying object-dependence of measurement processes and subject-independence of outcomes as key criteria, which allow tracing data to the instances measured and the ways they were quantified. Separate histories notwithstanding, the article's basic premise is that general principles of scientific measurement and quantification should apply to all sciences. It elaborates principles by which these metrological criteria can be implemented also in psychology and social sciences, while considering their research objects' peculiarities. Application of these principles is illustrated by quantifications of individual-specific behaviors ('personality'). The demands rating methods impose on data-generating persons are deconstructed and compared with the demands involved in other quantitative methods (e.g., ethological observations). These analyses highlight problematic requirements for raters. Rating methods sufficiently specify neither the empirical study phenomena nor the symbolic systems used as data nor rules of assignment between them. Instead, pronounced individual differences in raters' interpretation and use of items and scales indicate considerable subjectivity in data generation. Together with recoding scale categories into numbers, this introduces a twofold break in the traceability of rating data, compromising interpretability of findings. These insights question common reliability and validity concepts for ratings and provide novel explanations for replicability problems. Specifically, rating methods standardize only data formats but not the actual data generation. Measurement requires data generation processes to be adapted to the study phenomena's properties and the measurement-executing persons' abilities and interpretations, rather than to numerical outcome formats facilitating statistical analyses. Researchers must finally investigate how people actually generate ratings to specify the representational systems underlying rating data.
Collapse
Affiliation(s)
- Jana Uher
- London School of Economics and Political Science, London, United Kingdom
| |
Collapse
|
7
|
Vollmer HC, Domma J. Selbstwirksamkeitserwartung, psychische Belastung und Therapieerfahrung als Prädiktoren der Abstinenz nach stationärer Behandlung – eine Replikationsstudie. SUCHT 2018. [DOI: 10.1024/0939-5911/a000551] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
Abstract
Zusammenfassung. Fragestellung: Gestatten die Merkmale „Anzahl vorheriger Entgiftungen“, „Depressivität“ und „Selbstwirksamkeitserwartung, SWE“ eine Prognose der Abstinenz zur Ein-Jahres Katamnese? Methode: Prospektive Replikations-Feldstudie in einer anderen Klinik, in der mittels binärer logistischer Regression und Chi Quadrat Tests Unterschiede in Patienten-Merkmalen zwischen zur Ein-Jahres-Katamnese durchgehend abstinenten (N = 285) vs. rückfällig gewordenen Alkoholabhängigen (N = 274) analysiert wurden. Ergebnisse: Ebenso wie in unserer vorherigen Studie waren Alter, Geschlecht, Schulbildung, Arbeitslosigkeit, Familienstand, Partnersituation, suchtspezifische und psychische Komorbidität prognostisch nicht relevant, Ausnahme Persönlichkeitsstörungen. Wiederum hatten Patienten mit weniger als zwei Entgiftungen und einer hohen SWE die höchste Wahrscheinlichkeit ein Jahr durchgehend abstinent zu leben (82 %). Ebenso bestätigt wurde, dass Verbesserungen in der psychischen Belastung nicht mit Abstinenz korrelieren. Depressivität und vorherige Entwöhnungsbehandlungen wurden nicht als Prädiktoren repliziert. Schlussfolgerungen: Vorherige Entgiftungen, SWE und Persönlichkeitsstörungen könnten für den hier untersuchten Kliniktyp allgemeinverbindliche Prädiktoren sein. Inwiefern der Ausschluss der beiden Prädiktoren Depressivität und vorherige Entwöhnungsbehandlungen eine Folge neu eingeführter Interventionen für Depressive und Therapieerfahrene ist, wäre zu prüfen. Wie in unserer vorherigen Studie gestattet das Regressionsmodell trotz geringer Varianzaufklärung und mittlerer Effektstärken die Ableitung kausaler Hypothesen zur klinikspezifischen Verbesserung der Behandlung. Replikationsstudien sollten ebenso wie die empirische Orientierung ein fester Bestandteil verhaltenstherapeutischer Behandlung sein.
Collapse
|
8
|
Ferguson CJ. The problem of false positives and false negatives in violent video game experiments. INTERNATIONAL JOURNAL OF LAW AND PSYCHIATRY 2018; 56:35-43. [PMID: 29701597 DOI: 10.1016/j.ijlp.2017.11.001] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/04/2017] [Revised: 08/20/2017] [Accepted: 11/01/2017] [Indexed: 06/08/2023]
Abstract
The problem of false positives and negatives has received considerable attention in behavioral research in recent years. The current paper uses video game violence research as an example of how such issues may develop in a field. Despite decades of research, evidence on whether violent video games (VVGs) contribute to aggression in players has remained mixed. Concerns have been raised in recent years that experiments regarding VVGs may suffer from both "false positives" and "false negatives." The current paper examines this issue in three sets of video game experiments, two sets of video game experiments on aggression and prosocial behaviors identified in meta-analysis, and a third group of recent null studies. Results indicated that studies of VVGs and aggression appear to be particularly prone to false positive results. Studies of VVGs and prosocial behavior, by contrast are heterogeneous and did not demonstrate any indication of false positive results. However, their heterogeneous nature made it difficult to base solid conclusions on them. By contrast, evidence for false negatives in null studies was limited, and little evidence emerged that null studies lacked power in comparison those highlighted in past meta-analyses as evidence for effects. These results are considered in light of issues related to false positives and negatives in behavioral science more broadly.
Collapse
Affiliation(s)
- Christopher J Ferguson
- Department of Psychology, Stetson University, 421 N. Woodland Blvd., DeLand, FL, 32729, United States.
| |
Collapse
|
9
|
Abstract
In this article, we present a model for determining how total research payoff depends on researchers' choices of sample sizes, α levels, and other parameters of the research process. The model can be used to quantify various trade-offs inherent in the research process and thus to balance competing goals, such as (a) maximizing both the number of studies carried out and also the statistical power of each study, (b) minimizing the rates of both false positive and false negative findings, and (c) maximizing both replicability and research efficiency. Given certain necessary information about a research area, the model can be used to determine the optimal values of sample size, statistical power, rate of false positives, rate of false negatives, and replicability, such that overall research payoff is maximized. More specifically, the model shows how the optimal values of these quantities depend upon the size and frequency of true effects within the area, as well as the individual payoffs associated with particular study outcomes. The model is particularly relevant within current discussions of how to optimize the productivity of scientific research, because it shows which aspects of a research area must be considered and how these aspects combine to determine total research payoff.
Collapse
|
10
|
Abstract
Although many researchers agree that scientific data should be open to scrutiny to ferret out poor analyses and outright fraud, most raw data sets are not available on demand. There are many reasons researchers do not open their data, and one is technical. It is often time consuming to prepare and archive data. In response, my laboratory has automated the process such that our data are archived the night they are created without any human approval or action. All data are versioned, logged, time stamped, and uploaded including aborted runs and data from pilot subjects. The archive is GitHub, github.com, the world's largest collection of open-source materials. Data archived in this manner are called born open. In this paper, I discuss the benefits of born-open data and provide a brief technical overview of the process. I also address some of the common concerns about opening data before publication.
Collapse
|
11
|
Rouder JN, Morey RD, Verhagen J, Province JM, Wagenmakers E. Is There a Free Lunch in Inference? Top Cogn Sci 2016; 8:520-47. [DOI: 10.1111/tops.12214] [Citation(s) in RCA: 53] [Impact Index Per Article: 6.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/26/2015] [Revised: 03/07/2016] [Accepted: 05/18/2016] [Indexed: 11/30/2022]
|
12
|
Interpreting confidence intervals: A comment on Hoekstra, Morey, Rouder, and Wagenmakers (2014). Psychon Bull Rev 2016; 23:124-30. [PMID: 26558758 DOI: 10.3758/s13423-015-0859-7] [Citation(s) in RCA: 8] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
Hoekstra, Morey, Rouder, and Wagenmakers (Psychonomic Bulletin & Review 21(5), 1157-1164 2014) reported the results of a questionnaire designed to assess students' and researchers' understanding of confidence intervals (CIs). They interpreted their results as evidence that these groups "have no reliable knowledge about the correct interpretation of CIs" (Hoekstra et al. Psychonomic Bulletin & Review 21(5), 1157-1164 2014, p. 1161). We argue that their data do not substantiate this conclusion and that their report includes misleading suggestions about the correct interpretations of confidence intervals.
Collapse
|
13
|
Observations versus assessments of personality: A five-method multi-species study reveals numerous biases in ratings and methodological limitations of standardised assessments. JOURNAL OF RESEARCH IN PERSONALITY 2016. [DOI: 10.1016/j.jrp.2016.02.003] [Citation(s) in RCA: 27] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022]
|
14
|
Wiernik BM, Dilchert S, Ones DS. Age and Employee Green Behaviors: A Meta-Analysis. Front Psychol 2016; 7:194. [PMID: 26973550 PMCID: PMC4773609 DOI: 10.3389/fpsyg.2016.00194] [Citation(s) in RCA: 33] [Impact Index Per Article: 4.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/04/2015] [Accepted: 02/01/2016] [Indexed: 11/20/2022] Open
Abstract
Recent economic and societal developments have led to an increasing emphasis on organizational environmental performance. At the same time, demographic trends are resulting in increasingly aging labor forces in many industrialized nations. Commonly held stereotypes suggest that older workers are less likely to be environmentally responsible than younger workers. To evaluate the degree to which such age differences are present, we meta-analyzed 132 independent correlations and 336 d-values based on 4676 professional workers from 22 samples in 11 countries. Contrary to popular stereotypes, age showed small positive relationships with pro-environmental behaviors, suggesting that older adults engaged in these workplace behaviors slightly more frequently. Relationships with age appeared to be linear for overall, Conserving, Avoiding Harm, and Taking Initiative pro-environmental behaviors, but non-linear trends were observed for Transforming and Influencing Others behaviors.
Collapse
Affiliation(s)
- Brenton M Wiernik
- Department of Psychology, University of Minnesota Minneapolis, MN, USA
| | - Stephan Dilchert
- Department of Management, Baruch College, City University of New York New York, NY, USA
| | - Deniz S Ones
- Department of Psychology, University of Minnesota Minneapolis, MN, USA
| |
Collapse
|
15
|
Camerer CF, Dreber A, Forsell E, Ho TH, Huber J, Johannesson M, Kirchler M, Almenberg J, Altmejd A, Chan T, Heikensten E, Holzmeister F, Imai T, Isaksson S, Nave G, Pfeiffer T, Razen M, Wu H. Evaluating replicability of laboratory experiments in economics. Science 2016; 351:1433-6. [PMID: 26940865 DOI: 10.1126/science.aaf0918] [Citation(s) in RCA: 286] [Impact Index Per Article: 35.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/16/2015] [Accepted: 02/19/2016] [Indexed: 01/05/2023]
Abstract
The replicability of some scientific findings has recently been called into question. To contribute data about replicability in economics, we replicated 18 studies published in the American Economic Review and the Quarterly Journal of Economics between 2011 and 2014. All of these replications followed predefined analysis plans that were made publicly available beforehand, and they all have a statistical power of at least 90% to detect the original effect size at the 5% significance level. We found a significant effect in the same direction as in the original study for 11 replications (61%); on average, the replicated effect size is 66% of the original. The replicability rate varies between 67% and 78% for four additional replicability indicators, including a prediction market measure of peer beliefs.
Collapse
Affiliation(s)
- Colin F Camerer
- Division of Humanities and Social Sciences, California Institute of Technology, 1200 East California Boulevard, MC 228-77, Pasadena, CA 91125, USA
| | - Anna Dreber
- Department of Economics, Stockholm School of Economics, Box 6501, SE-113 83 Stockholm, Sweden
| | - Eskil Forsell
- Department of Economics, Stockholm School of Economics, Box 6501, SE-113 83 Stockholm, Sweden
| | - Teck-Hua Ho
- Haas School of Business, University of California-Berkeley, Berkeley, CA 94720-1900, USA. NUS Business School, National University of Singapore, Singapore 119245
| | - Jürgen Huber
- Department of Banking and Finance, University of Innsbruck, Universitätsstrasse 15, 6020 Innsbruck, Austria
| | - Magnus Johannesson
- Department of Economics, Stockholm School of Economics, Box 6501, SE-113 83 Stockholm, Sweden
| | - Michael Kirchler
- Department of Banking and Finance, University of Innsbruck, Universitätsstrasse 15, 6020 Innsbruck, Austria. Centre for Finance, Department of Economics, University of Göteborg, SE-40530 Göteborg, Sweden
| | | | - Adam Altmejd
- Department of Economics, Stockholm School of Economics, Box 6501, SE-113 83 Stockholm, Sweden
| | - Taizan Chan
- Office of the Deputy President (Research and Technology), National University of Singapore, Singapore 119077
| | - Emma Heikensten
- Department of Economics, Stockholm School of Economics, Box 6501, SE-113 83 Stockholm, Sweden
| | - Felix Holzmeister
- Department of Banking and Finance, University of Innsbruck, Universitätsstrasse 15, 6020 Innsbruck, Austria
| | - Taisuke Imai
- Division of Humanities and Social Sciences, California Institute of Technology, 1200 East California Boulevard, MC 228-77, Pasadena, CA 91125, USA
| | - Siri Isaksson
- Department of Economics, Stockholm School of Economics, Box 6501, SE-113 83 Stockholm, Sweden
| | - Gideon Nave
- Division of Humanities and Social Sciences, California Institute of Technology, 1200 East California Boulevard, MC 228-77, Pasadena, CA 91125, USA
| | - Thomas Pfeiffer
- New Zealand Institute for Advanced Study, Private Bag 102904, North Shore Mail Centre, Auckland 0745, New Zealand. Wissenschaftskolleg zu Berlin, Institute for Advanced Study, D-14193 Berlin, Germany
| | - Michael Razen
- Department of Banking and Finance, University of Innsbruck, Universitätsstrasse 15, 6020 Innsbruck, Austria
| | - Hang Wu
- NUS Business School, National University of Singapore, Singapore 119245
| |
Collapse
|
16
|
Sijtsma K. Playing with Data--Or How to Discourage Questionable Research Practices and Stimulate Researchers to Do Things Right. PSYCHOMETRIKA 2016; 81:1-15. [PMID: 25820980 DOI: 10.1007/s11336-015-9446-0] [Citation(s) in RCA: 27] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/04/2023]
Abstract
Recent fraud cases in psychological and medical research have emphasized the need to pay attention to Questionable Research Practices (QRPs). Deliberate or not, QRPs usually have a deteriorating effect on the quality and the credibility of research results. QRPs must be revealed but prevention of QRPs is more important than detection. I suggest two policy measures that I expect to be effective in improving the quality of psychological research. First, the research data and the research materials should be made publicly available so as to allow verification. Second, researchers should more readily consider consulting a methodologist or a statistician. These two measures are simple but run against common practice to keep data to oneself and overestimate one's methodological and statistical skills, thus allowing secrecy and errors to enter research practice.
Collapse
Affiliation(s)
- Klaas Sijtsma
- Department of Methodology and Statistics, Tilburg School of Social and Behavioral Sciences, Tilburg University, PO Box 90153, 5000 LE , Tilburg, The Netherlands.
| |
Collapse
|
17
|
Using prediction markets to estimate the reproducibility of scientific research. Proc Natl Acad Sci U S A 2015; 112:15343-7. [PMID: 26553988 DOI: 10.1073/pnas.1516179112] [Citation(s) in RCA: 88] [Impact Index Per Article: 9.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022] Open
Abstract
Concerns about a lack of reproducibility of statistically significant results have recently been raised in many fields, and it has been argued that this lack comes at substantial economic costs. We here report the results from prediction markets set up to quantify the reproducibility of 44 studies published in prominent psychology journals and replicated in the Reproducibility Project: Psychology. The prediction markets predict the outcomes of the replications well and outperform a survey of market participants' individual forecasts. This shows that prediction markets are a promising tool for assessing the reproducibility of published scientific results. The prediction markets also allow us to estimate probabilities for the hypotheses being true at different testing stages, which provides valuable information regarding the temporal dynamics of scientific discovery. We find that the hypotheses being tested in psychology typically have low prior probabilities of being true (median, 9%) and that a "statistically significant" finding needs to be confirmed in a well-powered replication to have a high probability of being true. We argue that prediction markets could be used to obtain speedy information about reproducibility at low cost and could potentially even be used to determine which studies to replicate to optimally allocate limited resources into replications.
Collapse
|
18
|
Nakagawa S, Parker TH. Replicating research in ecology and evolution: feasibility, incentives, and the cost-benefit conundrum. BMC Biol 2015; 13:88. [PMID: 26510635 PMCID: PMC4624660 DOI: 10.1186/s12915-015-0196-3] [Citation(s) in RCA: 62] [Impact Index Per Article: 6.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022] Open
Abstract
We believe that replicating studies in ecology and evolution is extremely valuable, but replication within species and systems is troublingly rare, and even ‘quasi-replications’ in different systems are often insufficient. We make a case for supporting multiple types of replications and point out that the current incentive structure needs to change if ecologists and evolutionary biologist are to value scientific replication sufficiently.
Collapse
Affiliation(s)
- Shinichi Nakagawa
- School of Biological, Earth and Environmental Sciences, University of New South Wales, Sydney, NSW, 2052, Australia.
| | - Timothy H Parker
- Biology Department, Whitman College, Walla Walla, WA, 99362, USA
| |
Collapse
|
19
|
Saint-Mont U. Randomization Does Not Help Much, Comparability Does. PLoS One 2015; 10:e0132102. [PMID: 26193621 PMCID: PMC4507867 DOI: 10.1371/journal.pone.0132102] [Citation(s) in RCA: 32] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/21/2013] [Accepted: 04/27/2015] [Indexed: 11/25/2022] Open
Abstract
According to R.A. Fisher, randomization "relieves the experimenter from the anxiety of considering innumerable causes by which the data may be disturbed." Since, in particular, it is said to control for known and unknown nuisance factors that may considerably challenge the validity of a result, it has become very popular. This contribution challenges the received view. First, looking for quantitative support, we study a number of straightforward, mathematically simple models. They all demonstrate that the optimism surrounding randomization is questionable: In small to medium-sized samples, random allocation of units to treatments typically yields a considerable imbalance between the groups, i.e., confounding due to randomization is the rule rather than the exception. In the second part of this contribution, the reasoning is extended to a number of traditional arguments in favour of randomization. This discussion is rather non-technical, and sometimes touches on the rather fundamental Frequentist/Bayesian debate. However, the result of this analysis turns out to be quite similar: While the contribution of randomization remains doubtful, comparability contributes much to a compelling conclusion. Summing up, classical experimentation based on sound background theory and the systematic construction of exchangeable groups seems to be advisable.
Collapse
Affiliation(s)
- Uwe Saint-Mont
- Nordhausen University of Applied Sciences, Nordhausen, Germany
| |
Collapse
|
20
|
Pashler H, Wagenmakers EJ. Editors' Introduction to the Special Section on Replicability in Psychological Science: A Crisis of Confidence? PERSPECTIVES ON PSYCHOLOGICAL SCIENCE 2015; 7:528-30. [PMID: 26168108 DOI: 10.1177/1745691612465253] [Citation(s) in RCA: 578] [Impact Index Per Article: 64.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
|
21
|
Makel MC, Plucker JA, Hegarty B. Replications in Psychology Research: How Often Do They Really Occur? PERSPECTIVES ON PSYCHOLOGICAL SCIENCE 2015; 7:537-42. [PMID: 26168110 DOI: 10.1177/1745691612460688] [Citation(s) in RCA: 270] [Impact Index Per Article: 30.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Recent controversies in psychology have spurred conversations about the nature and quality of psychological research. One topic receiving substantial attention is the role of replication in psychological science. Using the complete publication history of the 100 psychology journals with the highest 5-year impact factors, the current article provides an overview of replications in psychological research since 1900. This investigation revealed that roughly 1.6% of all psychology publications used the term replication in text. A more thorough analysis of 500 randomly selected articles revealed that only 68% of articles using the term replication were actual replications, resulting in an overall replication rate of 1.07%. Contrary to previous findings in other fields, this study found that the majority of replications in psychology journals reported similar findings to their original studies (i.e., they were successful replications). However, replications were significantly less likely to be successful when there was no overlap in authorship between the original and replicating articles. Moreover, despite numerous systemic biases, the rate at which replications are being published has increased in recent decades.
Collapse
|
22
|
Kirkland T, Gruber J, Cunningham WA. Comparing Happiness and Hypomania Risk: A Study of Extraversion and Neuroticism Aspects. PLoS One 2015; 10:e0132438. [PMID: 26161562 PMCID: PMC4498734 DOI: 10.1371/journal.pone.0132438] [Citation(s) in RCA: 7] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/21/2014] [Accepted: 06/16/2015] [Indexed: 01/26/2023] Open
Abstract
Positive affect has long been considered a hallmark of subjective happiness. Yet, high levels of positive affect have also been linked with hypomania risk: a set of cognitive, affective, and behavioral characteristics that constitute a dispositional risk for future episodes of hypomania and mania. At a personality level, two powerful predictors of affective experience are extraversion and neuroticism: extraversion has been linked to positive affect, and neuroticism to negative affect. As such, a single personality trait--extraversion--has been linked to both beneficial and harmful outcomes associated with positivity. It is clear that positive affect, in different forms, has divergent consequences for well-being, but previous research has struggled to articulate the nature of these differences. We suggest that the relationship between affect and well-being needs to be situated within the psychological context of the individual--both in terms of more specific forms of extraversion and neuroticism, but also in terms of interactions among personality aspects. Consistent with this idea, we found that two aspects of extraversion (enthusiasm and assertiveness) differentially predicted subjective happiness from hypomania risk and two aspects of neuroticism (volatility and withdrawal) interacted to predict hypomania risk: the highest levels of hypomania risk were associated with the combination of high volatility and low withdrawal. These findings underscore the importance of examining personality at the right level of resolution to understand well-being and dysfunction.
Collapse
Affiliation(s)
- Tabitha Kirkland
- Department of Psychology, Bellevue College, Bellevue, Washington, United States of America
- Department of Psychology, The Ohio State University, Columbus, Ohio, United States of America
| | - June Gruber
- Department of Psychology and Neuroscience, University of Colorado Boulder, Boulder, Colorado, United States of America
| | - William A. Cunningham
- Department of Psychology, The Ohio State University, Columbus, Ohio, United States of America
- Department of Psychology, University of Toronto, Toronto, Ontario, Canada
- Department of Marketing, University of Toronto, Toronto, Ontario, Canada
| |
Collapse
|
23
|
Ivanova IV, Tasca GA, Hammond N, Balfour L, Ritchie K, Koszycki D, Bissada H. Negative affect mediates the relationship between interpersonal problems and binge-eating disorder symptoms and psychopathology in a clinical sample: a test of the interpersonal model. EUROPEAN EATING DISORDERS REVIEW 2015; 23:133-8. [PMID: 25582510 DOI: 10.1002/erv.2344] [Citation(s) in RCA: 46] [Impact Index Per Article: 5.1] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2014] [Revised: 12/08/2014] [Accepted: 12/10/2014] [Indexed: 11/11/2022]
Abstract
This study evaluated the validity of the interpersonal model of binge-eating disorder (BED) psychopathology in a clinical sample of women with BED. Data from a cross-sectional sample of 255 women with BED were examined for the direct effects of interpersonal problems on BED symptoms and psychopathology, and indirect effects mediated by negative affect. Structural equation modelling analyses demonstrated that higher levels of interpersonal problems were associated with greater negative affect, and greater negative affect was associated with higher frequency of BED symptoms and psychopathology. There was a significant indirect effect of interpersonal problems on BED symptoms and psychopathology mediated through negative affect. Interpersonal problems may lead to greater BED symptoms and psychopathology, and this relationship may be partially explained by elevated negative affect. The results of the study are the first to provide support for the interpersonal model of BED symptoms and psychopathology in a clinical sample of women.
Collapse
|
24
|
Parker TH, Nakagawa S. Mitigating the epidemic of type I error: ecology and evolution can learn from other disciplines. Front Ecol Evol 2014. [DOI: 10.3389/fevo.2014.00076] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/13/2022] Open
|
25
|
Aydinoglu AU, Suomela T, Malone J. Data management in astrobiology: challenges and opportunities for an interdisciplinary community. ASTROBIOLOGY 2014; 14:451-461. [PMID: 24840364 PMCID: PMC4060838 DOI: 10.1089/ast.2013.1127] [Citation(s) in RCA: 6] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/06/2013] [Accepted: 03/10/2014] [Indexed: 06/03/2023]
Abstract
Data management and sharing are growing concerns for scientists and funding organizations throughout the world. Funding organizations are implementing requirements for data management plans, while scientists are establishing new infrastructures for data sharing. One of the difficulties is sharing data among a diverse set of research disciplines. Astrobiology is a unique community of researchers, containing over 110 different disciplines. The current study reports the results of a survey of data management practices among scientists involved in the astrobiology community and the NASA Astrobiology Institute (NAI) in particular. The survey was administered over a 2-month period in the first half of 2013. Fifteen percent of the NAI community responded (n=114), and additional (n=80) responses were collected from members of an astrobiology Listserv. The results of the survey show that the astrobiology community shares many of the same concerns for data sharing as other groups. The benefits of data sharing are acknowledged by many respondents, but barriers to data sharing remain, including lack of acknowledgement, citation, time, and institutional rewards. Overcoming technical, institutional, and social barriers to data sharing will be a challenge into the future.
Collapse
Affiliation(s)
- Arsev Umur Aydinoglu
- NASA Astrobiology Institute, NASA Ames Research Center, Moffett Field, California
| | - Todd Suomela
- School of Information Sciences, University of Tennessee, Knoxville, Tennessee
| | - Jim Malone
- School of Information Sciences, University of Tennessee, Knoxville, Tennessee
| |
Collapse
|
26
|
Read JCA. The place of human psychophysics in modern neuroscience. Neuroscience 2014; 296:116-29. [PMID: 24880153 DOI: 10.1016/j.neuroscience.2014.05.036] [Citation(s) in RCA: 23] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/19/2014] [Revised: 05/14/2014] [Accepted: 05/14/2014] [Indexed: 10/25/2022]
Abstract
Human psychophysics is the quantitative measurement of our own perceptions. In essence, it is simply a more sophisticated version of what humans have done since time immemorial: noticed and reflected upon what we can see, hear, and feel. In the 21st century, when hugely powerful techniques are available that enable us to probe the innermost structure and function of nervous systems, is human psychophysics still relevant? I argue that it is, and that in combination with other techniques, it will continue to be a key part of neuroscience for the foreseeable future. I discuss these points in detail using the example of binocular stereopsis, where human psychophysics in combination with physiology and computational vision, has made a substantial contribution.
Collapse
Affiliation(s)
- J C A Read
- Institute of Neuroscience, Newcastle University, Henry Wellcome Building, Framlington Place, Newcastle upon Tyne NE2 4HH, UK.
| |
Collapse
|
27
|
Bouwer FL, Van Zuijen TL, Honing H. Beat processing is pre-attentive for metrically simple rhythms with clear accents: an ERP study. PLoS One 2014; 9:e97467. [PMID: 24870123 PMCID: PMC4037171 DOI: 10.1371/journal.pone.0097467] [Citation(s) in RCA: 43] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/28/2013] [Accepted: 04/20/2014] [Indexed: 12/03/2022] Open
Abstract
The perception of a regular beat is fundamental to music processing. Here we examine whether the detection of a regular beat is pre-attentive for metrically simple, acoustically varying stimuli using the mismatch negativity (MMN), an ERP response elicited by violations of acoustic regularity irrespective of whether subjects are attending to the stimuli. Both musicians and non-musicians were presented with a varying rhythm with a clear accent structure in which occasionally a sound was omitted. We compared the MMN response to the omission of identical sounds in different metrical positions. Most importantly, we found that omissions in strong metrical positions, on the beat, elicited higher amplitude MMN responses than omissions in weak metrical positions, not on the beat. This suggests that the detection of a beat is pre-attentive when highly beat inducing stimuli are used. No effects of musical expertise were found. Our results suggest that for metrically simple rhythms with clear accents beat processing does not require attention or musical expertise. In addition, we discuss how the use of acoustically varying stimuli may influence ERP results when studying beat processing.
Collapse
Affiliation(s)
- Fleur L. Bouwer
- Institute for Logic, Language and Computation, University of Amsterdam, Amsterdam, The Netherlands
- Amsterdam Brain and Cognition (ABC), University of Amsterdam, Amsterdam, The Netherlands
| | - Titia L. Van Zuijen
- Research Institute of Child Development and Education, University of Amsterdam, Amsterdam, The Netherlands
| | - Henkjan Honing
- Institute for Logic, Language and Computation, University of Amsterdam, Amsterdam, The Netherlands
- Amsterdam Brain and Cognition (ABC), University of Amsterdam, Amsterdam, The Netherlands
| |
Collapse
|
28
|
Braver SL, Thoemmes FJ, Rosenthal R. Continuously Cumulating Meta-Analysis and Replicability. PERSPECTIVES ON PSYCHOLOGICAL SCIENCE 2014; 9:333-42. [DOI: 10.1177/1745691614529796] [Citation(s) in RCA: 187] [Impact Index Per Article: 18.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/17/2022]
Abstract
The current crisis in scientific psychology about whether our findings are irreproducible was presaged years ago by Tversky and Kahneman (1971), who noted that even sophisticated researchers believe in the fallacious Law of Small Numbers—erroneous intuitions about how imprecisely sample data reflect population phenomena. Combined with the low power of most current work, this often leads to the use of misleading criteria about whether an effect has replicated. Rosenthal (1990) suggested more appropriate criteria, here labeled the continuously cumulating meta-analytic (CCMA) approach. For example, a CCMA analysis on a replication attempt that does not reach significance might nonetheless provide more, not less, evidence that the effect is real. Alternatively, measures of heterogeneity might show that two studies that differ in whether they are significant might have only trivially different effect sizes. We present a nontechnical introduction to the CCMA framework (referencing relevant software), and then explain how it can be used to address aspects of replicability or more generally to assess quantitative evidence from numerous studies. We then present some examples and simulation results using the CCMA approach that show how the combination of evidence can yield improved results over the consideration of single studies.
Collapse
|
29
|
|
30
|
Bresin K, Robinson MD. Losing control, literally: Relations between anger control, trait anger, and motor control. Cogn Emot 2013; 27:995-1012. [DOI: 10.1080/02699931.2012.755119] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/27/2022]
|
31
|
Chabris CF, Lee JJ, Benjamin DJ, Beauchamp JP, Glaeser EL, Borst G, Pinker S, Laibson DI. Why it is hard to find genes associated with social science traits: theoretical and empirical considerations. Am J Public Health 2013; 103 Suppl 1:S152-66. [PMID: 23927501 DOI: 10.2105/ajph.2013.301327] [Citation(s) in RCA: 42] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/23/2023]
Abstract
OBJECTIVES We explain why traits of interest to behavioral scientists may have a genetic architecture featuring hundreds or thousands of loci with tiny individual effects rather than a few with large effects and why such an architecture makes it difficult to find robust associations between traits and genes. METHODS We conducted a genome-wide association study at 2 sites, Harvard University and Union College, measuring more than 100 physical and behavioral traits with a sample size typical of candidate gene studies. We evaluated predictions that alleles with large effect sizes would be rare and most traits of interest to social science are likely characterized by a lack of strong directional selection. We also carried out a theoretical analysis of the genetic architecture of traits based on R.A. Fisher's geometric model of natural selection and empirical analyses of the effects of selection bias and phenotype measurement stability on the results of genetic association studies. RESULTS Although we replicated several known genetic associations with physical traits, we found only 2 associations with behavioral traits that met the nominal genome-wide significance threshold, indicating that physical and behavioral traits are mainly affected by numerous genes with small effects. CONCLUSIONS The challenge for social science genomics is the likelihood that genes are connected to behavioral variation by lengthy, nonlinear, interactive causal chains, and unraveling these chains requires allying with personal genomics to take advantage of the potential for large sample sizes as well as continuing with traditional epidemiological studies.
Collapse
Affiliation(s)
- Christopher F Chabris
- Christopher F. Chabris is with the Department of Psychology, Union College, Schenectady, NY. James J. Lee, Gregoire Borst, and Steven Pinker are with the Department of Psychology, Harvard University, Cambridge, MA. Daniel J. Benjamin is with the Department of Economics, Cornell University, Ithaca, NY. Jonathan P. Beauchamp, Edward L. Glaeser, and David I. Laibson are with the Department of Economics, Harvard University
| | | | | | | | | | | | | | | |
Collapse
|
32
|
LeBel EP, Borsboom D, Giner-Sorolla R, Hasselman F, Peters KR, Ratliff KA, Smith CT. PsychDisclosure.org. PERSPECTIVES ON PSYCHOLOGICAL SCIENCE 2013; 8:424-32. [DOI: 10.1177/1745691613491437] [Citation(s) in RCA: 63] [Impact Index Per Article: 5.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/12/2022]
Abstract
There is currently an unprecedented level of doubt regarding the reliability of research findings in psychology. Many recommendations have been made to improve the current situation. In this article, we report results from PsychDisclosure.org , a novel open-science initiative that provides a platform for authors of recently published articles to disclose four methodological design specification details that are not required to be disclosed under current reporting standards but that are critical for accurate interpretation and evaluation of reported findings. Grassroots sentiment—as manifested in the positive and appreciative response to our initiative—indicates that psychologists want to see changes made at the systemic level regarding disclosure of such methodological details. Almost 50% of contacted researchers disclosed the requested design specifications for the four methodological categories (excluded subjects, nonreported conditions and measures, and sample size determination). Disclosed information provided by participating authors also revealed several instances of questionable editorial practices, which need to be thoroughly examined and redressed. On the basis of these results, we argue that the time is now for mandatory methods disclosure statements for all psychology journals, which would be an important step forward in improving the reliability of findings in psychology.
Collapse
Affiliation(s)
| | | | | | - Fred Hasselman
- Behavioural Science Institute, Radboud University Nijmegen
| | | | | | | |
Collapse
|
33
|
Brembs B, Button K, Munafò M. Deep impact: unintended consequences of journal rank. Front Hum Neurosci 2013. [PMID: 23805088 DOI: 10.3389/fnhum.2013.00291.] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/13/2022] Open
Abstract
Most researchers acknowledge an intrinsic hierarchy in the scholarly journals ("journal rank") that they submit their work to, and adjust not only their submission but also their reading strategies accordingly. On the other hand, much has been written about the negative effects of institutionalizing journal rank as an impact measure. So far, contributions to the debate concerning the limitations of journal rank as a scientific impact assessment tool have either lacked data, or relied on only a few studies. In this review, we present the most recent and pertinent data on the consequences of our current scholarly communication system with respect to various measures of scientific quality (such as utility/citations, methodological soundness, expert ratings or retractions). These data corroborate previous hypotheses: using journal rank as an assessment tool is bad scientific practice. Moreover, the data lead us to argue that any journal rank (not only the currently-favored Impact Factor) would have this negative impact. Therefore, we suggest that abandoning journals altogether, in favor of a library-based scholarly communication system, will ultimately be necessary. This new system will use modern information technology to vastly improve the filter, sort and discovery functions of the current journal system.
Collapse
Affiliation(s)
- Björn Brembs
- Institute of Zoology-Neurogenetics, University of Regensburg Regensburg, Germany
| | | | | |
Collapse
|
34
|
Brembs B, Button K, Munafò M. Deep impact: unintended consequences of journal rank. Front Hum Neurosci 2013; 7:291. [PMID: 23805088 PMCID: PMC3690355 DOI: 10.3389/fnhum.2013.00291] [Citation(s) in RCA: 132] [Impact Index Per Article: 12.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/25/2013] [Accepted: 06/03/2013] [Indexed: 12/27/2022] Open
Abstract
Most researchers acknowledge an intrinsic hierarchy in the scholarly journals (“journal rank”) that they submit their work to, and adjust not only their submission but also their reading strategies accordingly. On the other hand, much has been written about the negative effects of institutionalizing journal rank as an impact measure. So far, contributions to the debate concerning the limitations of journal rank as a scientific impact assessment tool have either lacked data, or relied on only a few studies. In this review, we present the most recent and pertinent data on the consequences of our current scholarly communication system with respect to various measures of scientific quality (such as utility/citations, methodological soundness, expert ratings or retractions). These data corroborate previous hypotheses: using journal rank as an assessment tool is bad scientific practice. Moreover, the data lead us to argue that any journal rank (not only the currently-favored Impact Factor) would have this negative impact. Therefore, we suggest that abandoning journals altogether, in favor of a library-based scholarly communication system, will ultimately be necessary. This new system will use modern information technology to vastly improve the filter, sort and discovery functions of the current journal system.
Collapse
Affiliation(s)
- Björn Brembs
- Institute of Zoology-Neurogenetics, University of Regensburg Regensburg, Germany
| | | | | |
Collapse
|
35
|
Magical thinking in predictions of negative events: Evidence for tempting fate but not for a protection effect. JUDGMENT AND DECISION MAKING 2013. [DOI: 10.1017/s1930297500004496] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/04/2023]
Abstract
AbstractIn this paper we test two hypotheses regarding magical thinking about the perceived likelihood of future events. The first is that people believe that those who “tempt fate” by failing to take necessary precautions are more likely to suffer negative outcomes. The second is the “protection effect”, where reminding people of precautions they have taken leads them to see related risks as less likely. To this end, we describe the results from three attempted direct replications of a protection effect experiment reported in Tykocinski (2008) and two replications of a tempting fate experiment reported in Risen and Gilovich (2008) in which we add a test of the protection effect. We did not replicate the protection effect but did replicate the tempting fate effect.
Collapse
|
36
|
Kern SE. Why your new cancer biomarker may never work: recurrent patterns and remarkable diversity in biomarker failures. Cancer Res 2012; 72:6097-101. [PMID: 23172309 DOI: 10.1158/0008-5472.can-12-3232] [Citation(s) in RCA: 158] [Impact Index Per Article: 13.2] [Reference Citation Analysis] [Abstract] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/01/2023]
Abstract
Less than 1% of published cancer biomarkers actually enter clinical practice. Although best practices for biomarker development are published, optimistic investigators may not appreciate the statistical near-certainty and diverse modes by which the other 99% (likely including your favorite new marker) do indeed fail. Here, patterns of failure were abstracted for classification from publications and an online database detailing marker failures. Failure patterns formed a hierarchical logical structure, or outline, of an emerging, deeply complex, and arguably fascinating science of biomarker failure. A new cancer biomarker under development is likely to have already encountered one or more of the following fatal features encountered by prior markers: lack of clinical significance, hidden structure in the source data, a technically inadequate assay, inappropriate statistical methods, unmanageable domination of the data by normal variation, implausibility, deficiencies in the studied population or in the investigator system, and its disproof or abandonment for cause by others. A greater recognition of the science of biomarker failure and its near-complete ubiquity is constructive and celebrates a seemingly perpetual richness of biologic, technical, and philosophical complexity, the full appreciation of which could improve the management of scarce research resources.
Collapse
Affiliation(s)
- Scott E Kern
- The Sidney Kimmel Comprehensive Cancer Center at Johns Hopkins University, Department of Oncology, 1650 Orleans Avenue, Baltimore, MD 21287, USA.
| |
Collapse
|
37
|
Abstract
Although replications are vital to scientific progress, psychologists rarely engage in systematic replication efforts. In this article, we consider psychologists’ narrative approach to scientific publications as an underlying reason for this neglect and propose an incentive structure for replications within psychology. First, researchers need accessible outlets for publishing replications. To accomplish this, psychology journals could publish replication reports in files that are electronically linked to reports of the original research. Second, replications should get cited. This can be achieved by cociting replications along with original research reports. Third, replications should become a valued collaborative effort. This can be realized by incorporating replications in teaching programs and by stimulating adversarial collaborations. The proposed incentive structure for replications can be developed in a relatively simple and cost-effective manner. By promoting replications, this incentive structure may greatly enhance the dependability of psychology’s knowledge base.
Collapse
|
38
|
Wagenmakers EJ, Wetzels R, Borsboom D, van der Maas HLJ, Kievit RA. An Agenda for Purely Confirmatory Research. PERSPECTIVES ON PSYCHOLOGICAL SCIENCE 2012; 7:632-8. [PMID: 26168122 DOI: 10.1177/1745691612463078] [Citation(s) in RCA: 419] [Impact Index Per Article: 34.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
The veracity of substantive research claims hinges on the way experimental data are collected and analyzed. In this article, we discuss an uncomfortable fact that threatens the core of psychology’s academic enterprise: almost without exception, psychologists do not commit themselves to a method of data analysis before they see the actual data. It then becomes tempting to fine tune the analysis to the data in order to obtain a desired result—a procedure that invalidates the interpretation of the common statistical tests. The extent of the fine tuning varies widely across experiments and experimenters but is almost impossible for reviewers and readers to gauge. To remedy the situation, we propose that researchers preregister their studies and indicate in advance the analyses they intend to conduct. Only these analyses deserve the label “confirmatory,” and only for these analyses are the common statistical tests valid. Other analyses can be carried out but these should be labeled “exploratory.” We illustrate our proposal with a confirmatory replication attempt of a study on extrasensory perception.
Collapse
|
39
|
Pashler H, Harris CR. Is the Replicability Crisis Overblown? Three Arguments Examined. PERSPECTIVES ON PSYCHOLOGICAL SCIENCE 2012; 7:531-6. [PMID: 26168109 DOI: 10.1177/1745691612463401] [Citation(s) in RCA: 281] [Impact Index Per Article: 23.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
We discuss three arguments voiced by scientists who view the current outpouring of concern about replicability as overblown. The first idea is that the adoption of a low alpha level (e.g., 5%) puts reasonable bounds on the rate at which errors can enter the published literature, making false-positive effects rare enough to be considered a minor issue. This, we point out, rests on statistical misunderstanding: The alpha level imposes no limit on the rate at which errors may arise in the literature (Ioannidis, 2005b). Second, some argue that whereas direct replication attempts are uncommon, conceptual replication attempts are common—providing an even better test of the validity of a phenomenon. We contend that performing conceptual rather than direct replication attempts interacts insidiously with publication bias, opening the door to literatures that appear to confirm the reality of phenomena that in fact do not exist. Finally, we discuss the argument that errors will eventually be pruned out of the literature if the field would just show a bit of patience. We contend that there are no plausible concrete scenarios to back up such forecasts and that what is needed is not patience, but rather systematic reforms in scientific practice.
Collapse
|
40
|
Nosek BA, Spies JR, Motyl M. Scientific Utopia: II. Restructuring Incentives and Practices to Promote Truth Over Publishability. PERSPECTIVES ON PSYCHOLOGICAL SCIENCE 2012; 7:615-31. [PMID: 26168121 PMCID: PMC10540222 DOI: 10.1177/1745691612459058] [Citation(s) in RCA: 543] [Impact Index Per Article: 45.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/11/2023]
Abstract
An academic scientist's professional success depends on publishing. Publishing norms emphasize novel, positive results. As such, disciplinary incentives encourage design, analysis, and reporting decisions that elicit positive results and ignore negative results. Prior reports demonstrate how these incentives inflate the rate of false effects in published science. When incentives favor novelty over replication, false results persist in the literature unchallenged, reducing efficiency in knowledge accumulation. Previous suggestions to address this problem are unlikely to be effective. For example, a journal of negative results publishes otherwise unpublishable reports. This enshrines the low status of the journal and its content. The persistence of false findings can be meliorated with strategies that make the fundamental but abstract accuracy motive-getting it right-competitive with the more tangible and concrete incentive-getting it published. This article develops strategies for improving scientific practices and knowledge accumulation that account for ordinary human motivations and biases.
Collapse
|