1
|
Li W, Liu X, Zhang Q, Shi L, Zhang JX, Zhang X, Luan J, Li Y, Xu T, Zhang R, Han X, Lei J, Wang X, Wang Y, Lan H, Chen X, Wu Y, Wu Y, Xia L, Liao H, Shen C, Yu Y, Xu X, Deng C, Liu P, Feng Z, Huang CJ, Chen Z. Formalistic data and code availability policy in high-profile medical journals and pervasive policy-practice gaps in published articles: A meta-research study. Account Res 2025:1-25. [PMID: 40130560 DOI: 10.1080/08989621.2025.2481943] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/30/2024] [Accepted: 03/17/2025] [Indexed: 03/26/2025]
Abstract
BACKGROUND Poor data and code (DAC) sharing undermines open science principles. This study evaluates the stringency of DAC availability policies in high-profile medical journals and identifies policy-practice gaps (PPG) in published articles. METHODS DAC availability policies of 931 Q1 medical journals (Clarivate JCR 2021) were evaluated, with PPGs quantified across 3,191 articles from The BMJ, JAMA, NEJM, and The Lancet. RESULTS Only 9.1% (85/931) of journals mandated DAC sharing and availability statements, with 70.6% of these lacking mechanisms to verify authenticity, and 61.2% allowing publication despite invalid sharing. Secondary analysis revealed a disproportionate distribution of policies across subspecialties, with 18.6% (11/59) of subspecialties having >20% journals with mandated policies. Journal impact factors exhibited positive correlations with the stringency of availability statement policies (ρ = 0.20, p < 0.001) but not with sharing policies (ρ = 0.01, p = 0.737). Among the 3,191 articles, PPGs were observed in over 90% of cases. Specifically, 33.7% lacked DAC availability statements, 23.3% refused sharing (58.4% of which without justification in public statements), and 13.5% declared public sharing, with 39.0% being unreachable. Finally, only 0.5% achieved full computational reproducibility. CONCLUSIONS Formalistic policies and prevalent PPGs undermine DAC transparency, necessitating a supportive publication ecosystem that empowers authors to uphold scientific responsibility and integrity.
Collapse
Affiliation(s)
- Wei Li
- School of Psychology, Army Medical University, Chongqing, China
| | - Xuerong Liu
- School of Psychology, Army Medical University, Chongqing, China
| | - Qianyu Zhang
- School of Psychology, Army Medical University, Chongqing, China
| | - Liping Shi
- School of Psychology, Army Medical University, Chongqing, China
| | - Jing-Xuan Zhang
- School of Psychology, Army Medical University, Chongqing, China
| | - Xiaolin Zhang
- School of Psychology, Army Medical University, Chongqing, China
| | - Jia Luan
- Editorial Board, The Journal of Third Military Medical University China, China
| | - Yue Li
- Editorial Board, The Journal of Third Military Medical University China, China
| | - Ting Xu
- School of Psychology, Southwest University, Chongqing, China
| | - Rong Zhang
- School of Psychology, Southwest University, Chongqing, China
| | - Xiaodi Han
- School of Psychology, Army Medical University, Chongqing, China
| | - Jingyu Lei
- School of Psychology, Army Medical University, Chongqing, China
| | - Xueqian Wang
- School of Psychology, Army Medical University, Chongqing, China
| | - Yaozhi Wang
- School of Education, Sichuan Normal University, Chengdu, China
| | - Hai Lan
- School of Psychology, Sichuan Normal University, Chengdu, China
| | - Xiaohan Chen
- President Office, The Chengdu University of Traditional Chinese Medicine China, China
| | - Yi Wu
- School of Management, Army Medical University, Chongqing, China
| | - Yan Wu
- School of Architecture, Zhengzhou University, Zhengzhou, China
| | - Lei Xia
- School of Psychology, Army Medical University, Chongqing, China
| | - Haiping Liao
- School of Psychology, Army Medical University, Chongqing, China
| | - Chang Shen
- School of Psychology, Army Medical University, Chongqing, China
| | - Yang Yu
- School of Psychology, Army Medical University, Chongqing, China
| | - Xinyu Xu
- School of Psychology, Army Medical University, Chongqing, China
| | - Chao Deng
- School of Psychology, Army Medical University, Chongqing, China
| | - Pei Liu
- School of Psychology, Army Medical University, Chongqing, China
| | - Zhengzhi Feng
- School of Psychology, Army Medical University, Chongqing, China
| | - Chun-Ji Huang
- Presidential Office, Army Medical University, Chongqing, China
| | - Zhiyi Chen
- School of Psychology, Army Medical University, Chongqing, China
| |
Collapse
|
2
|
Bannach-Brown A, Rackoll T, Macleod MR, McCann SK. Building a synthesis-ready research ecosystem: fostering collaboration and open science to accelerate biomedical translation. BMC Med Res Methodol 2025; 25:66. [PMID: 40065205 PMCID: PMC11892198 DOI: 10.1186/s12874-025-02524-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2024] [Accepted: 02/27/2025] [Indexed: 03/14/2025] Open
Abstract
In this review article, we provide a comprehensive overview of current practices and challenges associated with research synthesis in preclinical biomedical research. We identify critical barriers and roadblocks that impede effective identification, utilisation, and integration of research findings to inform decision making in research translation. We examine practices at each stage of the research lifecycle, including study design, conduct, and publishing, that can be optimised to facilitate the conduct of timely, accurate, and comprehensive evidence synthesis. These practices are anchored in open science and engaging with the broader research community to ensure evidence is accessible and useful to all stakeholders. We underscore the need for collective action from researchers, synthesis specialists, institutions, publishers and journals, funders, infrastructure providers, and policymakers, who all play a key role in fostering an open, robust and synthesis-ready research environment, for an accelerated trajectory towards integrated biomedical research and translation.
Collapse
Affiliation(s)
- Alexandra Bannach-Brown
- QUEST Center for Responsible Research, Berlin Institute of Health at Charité - Universitätsmedizin Berlin, Charitéplatz 1, 10117, Berlin, Germany
| | - Torsten Rackoll
- QUEST Center for Responsible Research, Berlin Institute of Health at Charité - Universitätsmedizin Berlin, Charitéplatz 1, 10117, Berlin, Germany
| | - Malcolm R Macleod
- Centre for Clinical Brain Sciences, The University of Edinburgh Medical School, Edinburgh, UK
| | - Sarah K McCann
- QUEST Center for Responsible Research, Berlin Institute of Health at Charité - Universitätsmedizin Berlin, Charitéplatz 1, 10117, Berlin, Germany.
| |
Collapse
|
3
|
Ng JY. The case for data sharing in traditional, complementary, and integrative medicine research. Integr Med Res 2025; 14:101101. [PMID: 39834890 PMCID: PMC11742618 DOI: 10.1016/j.imr.2024.101101] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/15/2024] [Revised: 11/06/2024] [Accepted: 11/13/2024] [Indexed: 01/22/2025] Open
Abstract
Traditional, complementary, and integrative medicine (TCIM) research encompasses a diverse range of health practices rooted in various cultural, philosophical, and historical frameworks. As global interest in conducting research in this field grows, the need for rigorous research to support the integration of evidence-based TCIM therapies into mainstream healthcare has become essential. Data sharing is critical to advancing TCIM research by enhancing reproducibility, fostering interdisciplinary collaboration, promoting ethical practices, and addressing global health challenges. Despite its benefits, numerous challenges are associated with data sharing in TCIM, including a lack of standardized practices, cultural sensitivity, intellectual property concerns, and technical barriers in resource-limited settings. This editorial explores the unique nature of TCIM research, emphasizing the importance of data sharing while acknowledging the complexities it entails. Implementing the CARE Principles for Indigenous Data Governance, which prioritize collective benefit, authority to control, responsibility, and ethics, offers a framework for ensuring that data sharing respects indigenous knowledge and cultural sensitivities. Strategies for overcoming barriers to data sharing include developing standardized protocols, investing in research infrastructure, and fostering a culture of openness and collaboration within the TCIM community and beyond. By advancing data sharing practices, TCIM research can contribute to evidence-based healthcare solutions and address global health disparities, ultimately improving health outcomes worldwide.
Collapse
Affiliation(s)
- Jeremy Y. Ng
- Institute of General Practice and Interprofessional Care, University Hospital Tübingen, Tübingen, Germany
- Robert Bosch Center for Integrative Medicine and Health, Bosch Health Campus, Stuttgart, Germany
- Centre for Journalology, Ottawa Hospital Research Institute, Ottawa, Canada
| |
Collapse
|
4
|
Wells CA, Guhr A, Bairoch A, Chen Y, Hu M, Löser P, Ludwig TE, Mah N, Mueller SC, Seiler Wulczyn AEM, Seltmann S, Rossbach B, Kurtz A. Guidelines for managing and using the digital phenotypes of pluripotent stem cell lines. Stem Cell Reports 2024; 19:1369-1378. [PMID: 39332404 PMCID: PMC11561460 DOI: 10.1016/j.stemcr.2024.08.009] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/09/2024] [Revised: 08/26/2024] [Accepted: 08/27/2024] [Indexed: 09/29/2024] Open
Abstract
Each pluripotent stem cell line has a physical entity as well as a digital phenotype, but linking the two unambiguously is confounded by poor naming practices and assumed knowledge. Registration gives each line a unique and persistent identifier that links to phenotypic data generated over the lifetime of that line. Registration is a key recommendation of the 2023 ISSCR Standards for the use of human stem cells in research. Here we consider how community adoption of stem cell line registration could facilitate the establishment of integrated digital phenotypes of specific human pluripotent stem cell (hPSC) lines.
Collapse
Affiliation(s)
- Christine A Wells
- Stem Cell Systems, Department of Anatomy and Physiology, Medical, Dentistry and Health Sciences, University of Melbourne, Parkville, VIC 3010, Australia.
| | - Anke Guhr
- Robert Koch Institute, 13353 Berlin, Germany
| | - Amos Bairoch
- University of Geneva and SIB Swiss Institute of Bioinformatics, CMU, 1 Rue Michel Servet, 1211 Geneva, Switzerland
| | - Ying Chen
- Fraunhofer Institute for Biomedical Engineering (IBMT), Joseph-von-Fraunhofer Weg 1, 66280 Sulzbach, Germany
| | - Mengqi Hu
- Stem Cell Systems, Department of Anatomy and Physiology, Medical, Dentistry and Health Sciences, University of Melbourne, Parkville, VIC 3010, Australia
| | - Peter Löser
- Robert Koch Institute, 13353 Berlin, Germany
| | | | - Nancy Mah
- Fraunhofer Institute for Biomedical Engineering (IBMT), Joseph-von-Fraunhofer Weg 1, 66280 Sulzbach, Germany
| | - Sabine C Mueller
- Fraunhofer Institute for Biomedical Engineering (IBMT), Joseph-von-Fraunhofer Weg 1, 66280 Sulzbach, Germany
| | | | - Stefanie Seltmann
- Fraunhofer Institute for Biomedical Engineering (IBMT), Joseph-von-Fraunhofer Weg 1, 66280 Sulzbach, Germany
| | - Bella Rossbach
- Fraunhofer Institute for Biomedical Engineering (IBMT), Joseph-von-Fraunhofer Weg 1, 66280 Sulzbach, Germany
| | - Andreas Kurtz
- Fraunhofer Institute for Biomedical Engineering (IBMT), Joseph-von-Fraunhofer Weg 1, 66280 Sulzbach, Germany; Berlin Institute of Health Center for Regenerative Therapies at Charité, Berlin, Germany.
| |
Collapse
|
5
|
Iarkaeva A, Nachev V, Bobrov E. Workflow for detecting biomedical articles with underlying open and restricted-access datasets. PLoS One 2024; 19:e0302787. [PMID: 38718077 PMCID: PMC11078384 DOI: 10.1371/journal.pone.0302787] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/06/2023] [Accepted: 04/11/2024] [Indexed: 05/12/2024] Open
Abstract
To monitor the sharing of research data through repositories is increasingly of interest to institutions and funders, as well as from a meta-research perspective. Automated screening tools exist, but they are based on either narrow or vague definitions of open data. Where manual validation has been performed, it was based on a small article sample. At our biomedical research institution, we developed detailed criteria for such a screening, as well as a workflow which combines an automated and a manual step, and considers both fully open and restricted-access data. We use the results for an internal incentivization scheme, as well as for a monitoring in a dashboard. Here, we describe in detail our screening procedure and its validation, based on automated screening of 11035 biomedical research articles, of which 1381 articles with potential data sharing were subsequently screened manually. The screening results were highly reliable, as witnessed by inter-rater reliability values of ≥0.8 (Krippendorff's alpha) in two different validation samples. We also report the results of the screening, both for our institution and an independent sample from a meta-research study. In the largest of the three samples, the 2021 institutional sample, underlying data had been openly shared for 7.8% of research articles. For an additional 1.0% of articles, restricted-access data had been shared, resulting in 8.3% of articles overall having open and/or restricted-access data. The extraction workflow is then discussed with regard to its applicability in different contexts, limitations, possible variations, and future developments. In summary, we present a comprehensive, validated, semi-automated workflow for the detection of shared research data underlying biomedical article publications.
Collapse
Affiliation(s)
- Anastasiia Iarkaeva
- QUEST Center for Responsible Research, Berlin Institute of Health (BIH) at Charité –Universitätsmedizin Berlin, Berlin, Germany
| | - Vladislav Nachev
- QUEST Center for Responsible Research, Berlin Institute of Health (BIH) at Charité –Universitätsmedizin Berlin, Berlin, Germany
| | - Evgeny Bobrov
- QUEST Center for Responsible Research, Berlin Institute of Health (BIH) at Charité –Universitätsmedizin Berlin, Berlin, Germany
| |
Collapse
|
6
|
Ohmann C, Panagiotopoulou M, Canham S, Felder G, Verde PE. An assessment of the informative value of data sharing statements in clinical trial registries. BMC Med Res Methodol 2024; 24:61. [PMID: 38461273 PMCID: PMC10924983 DOI: 10.1186/s12874-024-02168-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/18/2023] [Accepted: 02/02/2024] [Indexed: 03/11/2024] Open
Abstract
BACKGROUND The provision of data sharing statements (DSS) for clinical trials has been made mandatory by different stakeholders. DSS are a device to clarify whether there is intention to share individual participant data (IPD). What is missing is a detailed assessment of whether DSS are providing clear and understandable information about the conditions for data sharing of IPD for secondary use. METHODS A random sample of 200 COVID-19 clinical trials with explicit DSS was drawn from the ECRIN clinical research metadata repository. The DSS were assessed and classified, by two experienced experts and one assessor with less experience in data sharing (DS), into different categories (unclear, no sharing, no plans, yes but vague, yes on request, yes with specified storage location, yes but with complex conditions). RESULTS Between the two experts the agreement was moderate to substantial (kappa=0.62, 95% CI [0.55, 0.70]). Agreement considerably decreased when these experts were compared with a third person who was less experienced and trained in data sharing ("assessor") (kappa=0.33, 95% CI [0.25, 0.41]; 0.35, 95% CI [0.27, 0.43]). Between the two experts and under supervision of an independent moderator, a consensus was achieved for those cases, where both experts had disagreed, and the result was used as "gold standard" for further analysis. At least some degree of willingness of DS (data sharing) was expressed in 63.5% (127/200) cases. Of these cases, around one quarter (31/127) were vague statements of support for data sharing but without useful detail. In around half of the cases (60/127) it was stated that IPD could be obtained by request. Only in in slightly more than 10% of the cases (15/127) it was stated that the IPD would be transferred to a specific data repository. In the remaining cases (21/127), a more complex regime was described or referenced, which could not be allocated to one of the three previous groups. As a result of the consensus meetings, the classification system was updated. CONCLUSION The study showed that the current DSS that imply possible data sharing are often not easy to interpret, even by relatively experienced staff. Machine based interpretation, which would be necessary for any practical application, is currently not possible. Machine learning and / or natural language processing techniques might improve machine actionability, but would represent a very substantial investment of research effort. The cheaper and easier option would be for data providers, data requestors, funders and platforms to adopt a clearer, more structured and more standardised approach to specifying, providing and collecting DSS. TRIAL REGISTRATION The protocol for the study was pre-registered on ZENODO ( https://zenodo.org/record/7064624#.Y4DIAHbMJD8 ).
Collapse
Affiliation(s)
- Christian Ohmann
- European Clinical Research Infrastructures Network (ECRIN), Kaiserswerther Strasse 70, 40477, Düsseldorf, Germany.
| | | | - Steve Canham
- European Clinical Research Infrastructure Network (ECRIN), 75014, Paris, France
| | - Gerd Felder
- European Clinical Research Infrastructure Network (ECRIN), 40764, Langenfeld, Germany
| | - Pablo Emilio Verde
- Coordination Centre for Clinical Trials, Heinrich Heine University Düsseldorf, 40225, Düsseldorf, Nordrhein-Westfalen, Germany
| |
Collapse
|
7
|
Hamilton DG, Hong K, Fraser H, Rowhani-Farid A, Fidler F, Page MJ. Prevalence and predictors of data and code sharing in the medical and health sciences: systematic review with meta-analysis of individual participant data. BMJ 2023; 382:e075767. [PMID: 37433624 PMCID: PMC10334349 DOI: 10.1136/bmj-2023-075767] [Citation(s) in RCA: 12] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Accepted: 06/07/2023] [Indexed: 07/13/2023]
Abstract
OBJECTIVES To synthesise research investigating data and code sharing in medicine and health to establish an accurate representation of the prevalence of sharing, how this frequency has changed over time, and what factors influence availability. DESIGN Systematic review with meta-analysis of individual participant data. DATA SOURCES Ovid Medline, Ovid Embase, and the preprint servers medRxiv, bioRxiv, and MetaArXiv were searched from inception to 1 July 2021. Forward citation searches were also performed on 30 August 2022. REVIEW METHODS Meta-research studies that investigated data or code sharing across a sample of scientific articles presenting original medical and health research were identified. Two authors screened records, assessed the risk of bias, and extracted summary data from study reports when individual participant data could not be retrieved. Key outcomes of interest were the prevalence of statements that declared that data or code were publicly or privately available (declared availability) and the success rates of retrieving these products (actual availability). The associations between data and code availability and several factors (eg, journal policy, type of data, trial design, and human participants) were also examined. A two stage approach to meta-analysis of individual participant data was performed, with proportions and risk ratios pooled with the Hartung-Knapp-Sidik-Jonkman method for random effects meta-analysis. RESULTS The review included 105 meta-research studies examining 2 121 580 articles across 31 specialties. Eligible studies examined a median of 195 primary articles (interquartile range 113-475), with a median publication year of 2015 (interquartile range 2012-2018). Only eight studies (8%) were classified as having a low risk of bias. Meta-analyses showed a prevalence of declared and actual public data availability of 8% (95% confidence interval 5% to 11%) and 2% (1% to 3%), respectively, between 2016 and 2021. For public code sharing, both the prevalence of declared and actual availability were estimated to be <0.5% since 2016. Meta-regressions indicated that only declared public data sharing prevalence estimates have increased over time. Compliance with mandatory data sharing policies ranged from 0% to 100% across journals and varied by type of data. In contrast, success in privately obtaining data and code from authors historically ranged between 0% and 37% and 0% and 23%, respectively. CONCLUSIONS The review found that public code sharing was persistently low across medical research. Declarations of data sharing were also low, increasing over time, but did not always correspond to actual sharing of data. The effectiveness of mandatory data sharing policies varied substantially by journal and type of data, a finding that might be informative for policy makers when designing policies and allocating resources to audit compliance. SYSTEMATIC REVIEW REGISTRATION Open Science Framework doi:10.17605/OSF.IO/7SX8U.
Collapse
Affiliation(s)
- Daniel G Hamilton
- MetaMelb Research Group, School of BioSciences, University of Melbourne, Melbourne, VIC, Australia
- Melbourne Medical School, Faculty of Medicine, Dentistry, and Health Sciences, University of Melbourne, Melbourne, VIC, Australia
| | - Kyungwan Hong
- Department of Practice, Sciences, and Health Outcomes Research, University of Maryland School of Pharmacy, Baltimore, MD, USA
| | - Hannah Fraser
- MetaMelb Research Group, School of BioSciences, University of Melbourne, Melbourne, VIC, Australia
| | - Anisa Rowhani-Farid
- Department of Practice, Sciences, and Health Outcomes Research, University of Maryland School of Pharmacy, Baltimore, MD, USA
| | - Fiona Fidler
- MetaMelb Research Group, School of BioSciences, University of Melbourne, Melbourne, VIC, Australia
- School of Historical and Philosophical Studies, University of Melbourne, Melbourne, VIC, Australia
| | - Matthew J Page
- Methods in Evidence Synthesis Unit, School of Public Health and Preventive Medicine, Monash University, Melbourne, VIC, Australia
| |
Collapse
|
8
|
Hair K, Wilson E, Wong C, Tsang A, Macleod M, Bannach-Brown A. Systematic online living evidence summaries: emerging tools to accelerate evidence synthesis. Clin Sci (Lond) 2023; 137:773-784. [PMID: 37219941 PMCID: PMC10220429 DOI: 10.1042/cs20220494] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/19/2022] [Revised: 01/30/2023] [Accepted: 03/06/2023] [Indexed: 05/24/2023]
Abstract
Systematic reviews and meta-analysis are the cornerstones of evidence-based decision making and priority setting. However, traditional systematic reviews are time and labour intensive, limiting their feasibility to comprehensively evaluate the latest evidence in research-intensive areas. Recent developments in automation, machine learning and systematic review technologies have enabled efficiency gains. Building upon these advances, we developed Systematic Online Living Evidence Summaries (SOLES) to accelerate evidence synthesis. In this approach, we integrate automated processes to continuously gather, synthesise and summarise all existing evidence from a research domain, and report the resulting current curated content as interrogatable databases via interactive web applications. SOLES can benefit various stakeholders by (i) providing a systematic overview of current evidence to identify knowledge gaps, (ii) providing an accelerated starting point for a more detailed systematic review, and (iii) facilitating collaboration and coordination in evidence synthesis.
Collapse
Affiliation(s)
- Kaitlyn Hair
- Centre for Clinical Brain Sciences, The University of Edinburgh, Edinburgh, U.K
| | - Emma Wilson
- Centre for Clinical Brain Sciences, The University of Edinburgh, Edinburgh, U.K
| | - Charis Wong
- Anne Rowling Regenerative Neurology Clinic, University of Edinburgh, Edinburgh, U.K
- Euan Macdonald Centre for Motor Neuron Disease Research, University of Edinburgh, Edinburgh, U.K
| | - Anthony Tsang
- King’s Technology Evaluation Centre, King’s College London, U.K
| | - Malcolm Macleod
- Centre for Clinical Brain Sciences, The University of Edinburgh, Edinburgh, U.K
| | - Alexandra Bannach-Brown
- Charité Universitaetsmedizin Berlin, Berlin Institute of Health – QUEST Center, Berlin, Germany
| |
Collapse
|
9
|
Haven TL, Abunijela S, Hildebrand N. Biomedical supervisors' role modeling of open science practices. eLife 2023; 12:83484. [PMID: 37211820 DOI: 10.7554/elife.83484] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2022] [Accepted: 05/04/2023] [Indexed: 05/23/2023] Open
Abstract
Supervision is one important way to socialize Ph.D. candidates into open and responsible research. We hypothesized that one should be more likely to identify open science practices (here publishing open access and sharing data) in empirical publications that were part of a Ph.D. thesis when the Ph.D. candidates' supervisors engaged in these practices compared to those whose supervisors did not or less often did. Departing from thesis repositories at four Dutch University Medical centers, we included 211 pairs of supervisors and Ph.D. candidates, resulting in a sample of 2062 publications. We determined open access status using UnpaywallR and Open Data using Oddpub, where we also manually screened publications with potential open data statements. Eighty-three percent of our sample was published openly, and 9% had open data statements. Having a supervisor who published open access more often than the national average was associated with an odds of 1.99 to publish open access. However, this effect became nonsignificant when correcting for institutions. Having a supervisor who shared data was associated with 2.22 (CI:1.19-4.12) times the odds to share data compared to having a supervisor that did not. This odds ratio increased to 4.6 (CI:1.86-11.35) after removing false positives. The prevalence of open data in our sample was comparable to international studies; open access rates were higher. Whilst Ph.D. candidates spearhead initiatives to promote open science, this study adds value by investigating the role of supervisors in promoting open science.
Collapse
Affiliation(s)
- Tamarinde L Haven
- Danish Centre for Studies in Research and Research Policy, Department of Political Science, Aarhus University, Aarhus, Denmark
| | - Susan Abunijela
- QUEST Center for Responsible Research, Berlin Institute of Health at Charité - Universitätsmedizin Berlin, Berlin, Germany
| | - Nicole Hildebrand
- QUEST Center for Responsible Research, Berlin Institute of Health at Charité - Universitätsmedizin Berlin, Berlin, Germany
| |
Collapse
|
10
|
Franzen DL, Carlisle BG, Salholz-Hillel M, Riedel N, Strech D. Institutional dashboards on clinical trial transparency for University Medical Centers: A case study. PLoS Med 2023; 20:e1004175. [PMID: 36943836 PMCID: PMC10030018 DOI: 10.1371/journal.pmed.1004175] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/28/2022] [Accepted: 01/18/2023] [Indexed: 03/23/2023] Open
Abstract
BACKGROUND University Medical Centers (UMCs) must do their part for clinical trial transparency by fostering practices such as prospective registration, timely results reporting, and open access. However, research institutions are often unaware of their performance on these practices. Baseline assessments of these practices would highlight where there is room for change and empower UMCs to support improvement. We performed a status quo analysis of established clinical trial registration and reporting practices at German UMCs and developed a dashboard to communicate these baseline assessments with UMC leadership and the wider research community. METHODS AND FINDINGS We developed and applied a semiautomated approach to assess adherence to established transparency practices in a cohort of interventional trials and associated results publications. Trials were registered in ClinicalTrials.gov or the German Clinical Trials Register (DRKS), led by a German UMC, and reported as complete between 2009 and 2017. To assess adherence to transparency practices, we identified results publications associated to trials and applied automated methods at the level of registry data (e.g., prospective registration) and publications (e.g., open access). We also obtained summary results reporting rates of due trials registered in the EU Clinical Trials Register (EUCTR) and conducted at German UMCs from the EU Trials Tracker. We developed an interactive dashboard to display these results across all UMCs and at the level of single UMCs. Our study included and assessed 2,895 interventional trials led by 35 German UMCs. Across all UMCs, prospective registration increased from 33% (n = 58/178) to 75% (n = 144/193) for trials registered in ClinicalTrials.gov and from 0% (n = 0/44) to 79% (n = 19/24) for trials registered in DRKS over the period considered. Of trials with a results publication, 38% (n = 714/1,895) reported the trial registration number in the publication abstract. In turn, 58% (n = 861/1,493) of trials registered in ClinicalTrials.gov and 23% (n = 111/474) of trials registered in DRKS linked the publication in the registration. In contrast to recent increases in summary results reporting of drug trials in the EUCTR, 8% (n = 191/2,253) and 3% (n = 20/642) of due trials registered in ClinicalTrials.gov and DRKS, respectively, had summary results in the registry. Across trial completion years, timely results reporting (within 2 years of trial completion) as a manuscript publication or as summary results was 41% (n = 1,198/2,892). The proportion of openly accessible trial publications steadily increased from 42% (n = 16/38) to 74% (n = 72/97) over the period considered. A limitation of this study is that some of the methods used to assess the transparency practices in this dashboard rely on registry data being accurate and up-to-date. CONCLUSIONS In this study, we observed that it is feasible to assess and inform individual UMCs on their performance on clinical trial transparency in a reproducible and publicly accessible way. Beyond helping institutions assess how they perform in relation to mandates or their institutional policy, the dashboard may inform interventions to increase the uptake of clinical transparency practices and serve to evaluate the impact of these interventions.
Collapse
Affiliation(s)
- Delwen L. Franzen
- Berlin Institute of Health at Charité - Universitätsmedizin Berlin, QUEST Center for Responsible Research, Berlin, Germany
| | - Benjamin Gregory Carlisle
- Berlin Institute of Health at Charité - Universitätsmedizin Berlin, QUEST Center for Responsible Research, Berlin, Germany
| | - Maia Salholz-Hillel
- Berlin Institute of Health at Charité - Universitätsmedizin Berlin, QUEST Center for Responsible Research, Berlin, Germany
| | - Nico Riedel
- Berlin Institute of Health at Charité - Universitätsmedizin Berlin, QUEST Center for Responsible Research, Berlin, Germany
| | - Daniel Strech
- Berlin Institute of Health at Charité - Universitätsmedizin Berlin, QUEST Center for Responsible Research, Berlin, Germany
| |
Collapse
|
11
|
Collins A, Alexander R. Reproducibility of COVID-19 pre-prints. Scientometrics 2022; 127:4655-4673. [PMID: 35813409 PMCID: PMC9252536 DOI: 10.1007/s11192-022-04418-2] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/22/2021] [Accepted: 05/20/2022] [Indexed: 01/24/2023]
Abstract
To examine the reproducibility of COVID-19 research, we create a dataset of pre-prints posted to arXiv, bioRxiv, and medRxiv between 28 January 2020 and 30 June 2021 that are related to COVID-19. We extract the text from these pre-prints and parse them looking for keyword markers signaling the availability of the data and code underpinning the pre-print. For the pre-prints that are in our sample, we are unable to find markers of either open data or open code for 75% of those on arXiv, 67% of those on bioRxiv, and 79% of those on medRxiv.
Collapse
|
12
|
Menke J, Eckmann P, Ozyurt IB, Roelandse M, Anderson N, Grethe J, Gamst A, Bandrowski A. Establishing Institutional Scores With the Rigor and Transparency Index: Large-scale Analysis of Scientific Reporting Quality. J Med Internet Res 2022; 24:e37324. [PMID: 35759334 PMCID: PMC9274430 DOI: 10.2196/37324] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2022] [Revised: 05/10/2022] [Accepted: 05/23/2022] [Indexed: 12/11/2022] Open
Abstract
BACKGROUND Improving rigor and transparency measures should lead to improvements in reproducibility across the scientific literature; however, the assessment of measures of transparency tends to be very difficult if performed manually. OBJECTIVE This study addresses the enhancement of the Rigor and Transparency Index (RTI, version 2.0), which attempts to automatically assess the rigor and transparency of journals, institutions, and countries using manuscripts scored on criteria found in reproducibility guidelines (eg, Materials Design, Analysis, and Reporting checklist criteria). METHODS The RTI tracks 27 entity types using natural language processing techniques such as Bidirectional Long Short-term Memory Conditional Random Field-based models and regular expressions; this allowed us to assess over 2 million papers accessed through PubMed Central. RESULTS Between 1997 and 2020 (where data were readily available in our data set), rigor and transparency measures showed general improvement (RTI 2.29 to 4.13), suggesting that authors are taking the need for improved reporting seriously. The top-scoring journals in 2020 were the Journal of Neurochemistry (6.23), British Journal of Pharmacology (6.07), and Nature Neuroscience (5.93). We extracted the institution and country of origin from the author affiliations to expand our analysis beyond journals. Among institutions publishing >1000 papers in 2020 (in the PubMed Central open access set), Capital Medical University (4.75), Yonsei University (4.58), and University of Copenhagen (4.53) were the top performers in terms of RTI. In country-level performance, we found that Ethiopia and Norway consistently topped the RTI charts of countries with 100 or more papers per year. In addition, we tested our assumption that the RTI may serve as a reliable proxy for scientific replicability (ie, a high RTI represents papers containing sufficient information for replication efforts). Using work by the Reproducibility Project: Cancer Biology, we determined that replication papers (RTI 7.61, SD 0.78) scored significantly higher (P<.001) than the original papers (RTI 3.39, SD 1.12), which according to the project required additional information from authors to begin replication efforts. CONCLUSIONS These results align with our view that RTI may serve as a reliable proxy for scientific replicability. Unfortunately, RTI measures for journals, institutions, and countries fall short of the replicated paper average. If we consider the RTI of these replication studies as a target for future manuscripts, more work will be needed to ensure that the average manuscript contains sufficient information for replication attempts.
Collapse
Affiliation(s)
- Joe Menke
- Center for Research in Biological Systems, University of California, San Diego, La Jolla, CA, United States
- SciCrunch Inc., San Diego, CA, United States
| | - Peter Eckmann
- SciCrunch Inc., San Diego, CA, United States
- Department of Neuroscience, University of California, San Diego, La Jolla, CA, United States
| | - Ibrahim Burak Ozyurt
- SciCrunch Inc., San Diego, CA, United States
- Department of Neuroscience, University of California, San Diego, La Jolla, CA, United States
| | | | | | - Jeffrey Grethe
- SciCrunch Inc., San Diego, CA, United States
- Department of Neuroscience, University of California, San Diego, La Jolla, CA, United States
| | - Anthony Gamst
- Department of Mathematics, University of California, San Diego, CA, United States
| | - Anita Bandrowski
- SciCrunch Inc., San Diego, CA, United States
- Department of Neuroscience, University of California, San Diego, La Jolla, CA, United States
| |
Collapse
|
13
|
Schulz R, Barnett A, Bernard R, Brown NJL, Byrne JA, Eckmann P, Gazda MA, Kilicoglu H, Prager EM, Salholz-Hillel M, Ter Riet G, Vines T, Vorland CJ, Zhuang H, Bandrowski A, Weissgerber TL. Is the future of peer review automated? BMC Res Notes 2022; 15:203. [PMID: 35690782 PMCID: PMC9188010 DOI: 10.1186/s13104-022-06080-6] [Citation(s) in RCA: 26] [Impact Index Per Article: 8.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/04/2022] [Accepted: 05/18/2022] [Indexed: 12/19/2022] Open
Abstract
The rising rate of preprints and publications, combined with persistent inadequate reporting practices and problems with study design and execution, have strained the traditional peer review system. Automated screening tools could potentially enhance peer review by helping authors, journal editors, and reviewers to identify beneficial practices and common problems in preprints or submitted manuscripts. Tools can screen many papers quickly, and may be particularly helpful in assessing compliance with journal policies and with straightforward items in reporting guidelines. However, existing tools cannot understand or interpret the paper in the context of the scientific literature. Tools cannot yet determine whether the methods used are suitable to answer the research question, or whether the data support the authors' conclusions. Editors and peer reviewers are essential for assessing journal fit and the overall quality of a paper, including the experimental design, the soundness of the study's conclusions, potential impact and innovation. Automated screening tools cannot replace peer review, but may aid authors, reviewers, and editors in improving scientific papers. Strategies for responsible use of automated tools in peer review may include setting performance criteria for tools, transparently reporting tool performance and use, and training users to interpret reports.
Collapse
Affiliation(s)
- Robert Schulz
- BIH QUEST Center for Responsible Research, Berlin Institute of Health at Charité Universitätsmedizin Berlin, Berlin, Germany
| | - Adrian Barnett
- Australian Centre for Health Services Innovation and Centre for Healthcare Transformation, School of Public Health & Social Work, Queensland University of Technology, Brisbane, QLD, Australia
| | - René Bernard
- NeuroCure Cluster of Excellence, Charité Universitätsmedizin Berlin, Berlin, Germany
| | | | - Jennifer A Byrne
- Faculty of Medicine and Health, New South Wales Health Pathology, The University of Sydney, New South Wales, Australia
| | - Peter Eckmann
- Department of Neuroscience, University of California, San Diego, La Jolla, CA, USA
| | - Małgorzata A Gazda
- UMR 3525, Institut Pasteur, Université de Paris, CNRS, INSERM UA12, Comparative Functional Genomics group, Paris, France
| | - Halil Kilicoglu
- School of Information Sciences, University of Illinois Urbana-Champaign, Champaign, IL, USA
| | - Eric M Prager
- Translational Research and Development, Cohen Veterans Bioscience, New York, NY, USA
| | - Maia Salholz-Hillel
- BIH QUEST Center for Responsible Research, Berlin Institute of Health at Charité Universitätsmedizin Berlin, Berlin, Germany
| | - Gerben Ter Riet
- Faculty of Health, Center of Expertise Urban Vitality, Amsterdam University of Applied Science, Amsterdam, The Netherlands
| | - Timothy Vines
- DataSeer Research Data Services Ltd, Vancouver, BC, Canada
| | - Colby J Vorland
- Indiana University School of Public Health-Bloomington, Bloomington, IN, USA
| | - Han Zhuang
- School of Information Studies, Syracuse University, Syracuse, NY, USA
| | - Anita Bandrowski
- Department of Neuroscience, University of California, San Diego, La Jolla, CA, USA
| | - Tracey L Weissgerber
- BIH QUEST Center for Responsible Research, Berlin Institute of Health at Charité Universitätsmedizin Berlin, Berlin, Germany.
| |
Collapse
|
14
|
Uribe SE, Sofi-Mahmudi A, Raittio E, Maldupa I, Vilne B. Dental Research Data Availability and Quality According to the FAIR Principles. J Dent Res 2022; 101:1307-1313. [PMID: 35656591 PMCID: PMC9516597 DOI: 10.1177/00220345221101321] [Citation(s) in RCA: 12] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/05/2023] Open
Abstract
According to the FAIR principles, data produced by scientific research should be findable, accessible, interoperable, and reusable-for instance, to be used in machine learning algorithms. However, to date, there is no estimate of the quantity or quality of dental research data evaluated via the FAIR principles. We aimed to determine the availability of open data in dental research and to assess compliance with the FAIR principles (or FAIRness) of shared dental research data. We downloaded all available articles published in PubMed-indexed dental journals from 2016 to 2021 as open access from Europe PubMed Central. In addition, we took a random sample of 500 dental articles that were not open access through Europe PubMed Central. We assessed data sharing in the articles and compliance of shared data to the FAIR principles programmatically. Results showed that of 7,509 investigated articles, 112 (1.5%) shared data. The average (SD) level of compliance with the FAIR metrics was 32.6% (31.9%). The average for each metric was as follows: findability, 3.4 (2.7) of 7; accessibility, 1.0 (1.0) of 3; interoperability, 1.1 (1.2) of 4; and reusability, 2.4 (2.6) of 10. No considerable changes in data sharing or quality of shared data occurred over the years. Our findings indicated that dental researchers rarely shared data, and when they did share, the FAIR quality was suboptimal. Machine learning algorithms could understand 1% of available dental research data. These undermine the reproducibility of dental research and hinder gaining the knowledge that can be gleaned from machine learning algorithms and applications.
Collapse
Affiliation(s)
- S E Uribe
- Bioinformatics Lab, Riga Stradins University, Riga, Latvia.,Department of Conservative Dentistry and Oral Health, Riga Stradins University, Riga, Latvia.,School of Dentistry, Universidad Austral de Chile, Valdivia, Chile.,Baltic Biomaterials Centre of Excellence, Riga Technical University, Riga, Latvia
| | - A Sofi-Mahmudi
- Seqiz Health Network, Kurdistan University of Medical Sciences, Seqiz, Kurdistan.,Cochrane Iran Associate Centre, National Institute for Medical Research Development, Tehran, Iran
| | - E Raittio
- Institute of Dentistry, University of Eastern Finland, Kuopio, Finland
| | - I Maldupa
- Department of Conservative Dentistry and Oral Health, Riga Stradins University, Riga, Latvia
| | - B Vilne
- Bioinformatics Lab, Riga Stradins University, Riga, Latvia
| |
Collapse
|
15
|
Weissgerber T, Riedel N, Kilicoglu H, Labbé C, Eckmann P, Ter Riet G, Byrne J, Cabanac G, Capes-Davis A, Favier B, Saladi S, Grabitz P, Bannach-Brown A, Schulz R, McCann S, Bernard R, Bandrowski A. Automated screening of COVID-19 preprints: can we help authors to improve transparency and reproducibility? Nat Med 2021; 27:6-7. [PMID: 33432174 PMCID: PMC8177099 DOI: 10.1038/s41591-020-01203-7] [Citation(s) in RCA: 30] [Impact Index Per Article: 7.5] [Reference Citation Analysis] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Affiliation(s)
- Tracey Weissgerber
- Quality | Ethics | Open Science | Translation (QUEST), Berlin Institute of Health, Berlin, Germany.
- Charité-Universitätsmedizin Berlin, Berlin, Germany.
| | - Nico Riedel
- Quality | Ethics | Open Science | Translation (QUEST), Berlin Institute of Health, Berlin, Germany
| | - Halil Kilicoglu
- School of Information Sciences, University of Illinois at Urbana-Champaign, Champaign, IL, USA
| | - Cyril Labbé
- University Grenoble Alpes, CNRS, Grenoble INP, LIG, Grenoble, France
| | - Peter Eckmann
- Department of Neuroscience, University of California, San Diego, La Jolla, CA, USA
- SciCrunch Inc., San Diego, CA, USA
| | - Gerben Ter Riet
- Department of Cardiology, Amsterdam UMC, University of Amsterdam, Amsterdam, the Netherlands
- Urban Vitality Center of Expertise, Amsterdam University of Applied Sciences, Amsterdam, the Netherlands
| | - Jennifer Byrne
- New South Wales Health Statewide Biobank, New South Wales Health Pathology, Sydney, New South Wales, Australia
- Faculty of Medicine and Health, University of Sydney, Sydney, New South Wales, Australia
| | | | - Amanda Capes-Davis
- CellBank Australia, Children's Medical Research Institute and The University of Sydney, Westmead, NSW, Australia
| | | | - Shyam Saladi
- California Institute of Technology, Pasadena, CA, USA
| | - Peter Grabitz
- Quality | Ethics | Open Science | Translation (QUEST), Berlin Institute of Health, Berlin, Germany
- Charité-Universitätsmedizin Berlin, Berlin, Germany
| | - Alexandra Bannach-Brown
- Quality | Ethics | Open Science | Translation (QUEST), Berlin Institute of Health, Berlin, Germany
| | - Robert Schulz
- Quality | Ethics | Open Science | Translation (QUEST), Berlin Institute of Health, Berlin, Germany
- Charité-Universitätsmedizin Berlin, Berlin, Germany
| | - Sarah McCann
- Quality | Ethics | Open Science | Translation (QUEST), Berlin Institute of Health, Berlin, Germany
- Charité-Universitätsmedizin Berlin, Berlin, Germany
| | - Rene Bernard
- NeuroCure Cluster of Excellence, Charité-Universitätsmedizin Berlin, corporate member of the Freie Universität Berlin, Humboldt-Universität zu Berlin and Berlin Institute of Health, Berlin, Germany
| | - Anita Bandrowski
- Department of Neuroscience, University of California, San Diego, La Jolla, CA, USA
- SciCrunch Inc., San Diego, CA, USA
| |
Collapse
|