51
|
Bullock GS, Hughes T, Arundale AH, Ward P, Collins GS, Kluzek S. Black Box Prediction Methods in Sports Medicine Deserve a Red Card for Reckless Practice: A Change of Tactics is Needed to Advance Athlete Care. Sports Med 2022; 52:1729-1735. [PMID: 35175575 DOI: 10.1007/s40279-022-01655-6] [Citation(s) in RCA: 14] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 01/26/2022] [Indexed: 01/22/2023]
Abstract
There is growing interest in the role of predictive analytics in sport, where such extensive data collection provides an exciting opportunity for the development and utilisation of prediction models for medical and performance purposes. Clinical prediction models have traditionally been developed using regression-based approaches, although newer machine learning methods are becoming increasingly popular. Machine learning models are considered 'black box'. In parallel with the increase in machine learning, there is also an emergence of proprietary prediction models that have been developed by researchers with the aim of becoming commercially available. Consequently, because of the profitable nature of proprietary systems, developers are often reluctant to transparently report (or make freely available) the development and validation of their prediction algorithms; the term 'black box' also applies to these systems. The lack of transparency and unavailability of algorithms to allow implementation by others of 'black box' approaches is concerning as it prevents independent evaluation of model performance, interpretability, utility, and generalisability prior to implementation within a sports medicine and performance environment. Therefore, in this Current Opinion article, we: (1) critically examine the use of black box prediction methodology and discuss its limited applicability in sport, and (2) argue that black box methods may pose a threat to delivery and development of effective athlete care and, instead, highlight why transparency and collaboration in prediction research and product development are essential to improve the integration of prediction models into sports medicine and performance.
Collapse
Affiliation(s)
- Garrett S Bullock
- Department of Orthopaedic Surgery & Rehabilitation, Wake Forest School of Medicine, Winston-Salem, NC, USA.
- Centre for Sport, Exercise and Osteoarthritis Research Versus Arthritis, University of Oxford, Oxford, UK.
- Nuffield Department of Orthopaedics, Rheumatology, and Musculoskeletal Sciences, University of Oxford, Oxford, UK.
| | - Tom Hughes
- Department of Health Professions, Manchester Metropolitan University, Manchester, UK
- Manchester United Football Club, Manchester, UK
| | - Amelia H Arundale
- Red Bull Athlete Performance Center, Thalgua, Austria
- Icahn School of Medicine, Mount Sinai Health System, New York, NY, USA
| | | | - Gary S Collins
- Nuffield Department of Orthopaedics, Rheumatology, and Musculoskeletal Sciences, University of Oxford, Oxford, UK
- Centre for Statistics in Medicine, University of Oxford, Oxford, UK
- Oxford University Hospitals NHS Foundation Trust, Oxford, UK
| | - Stefan Kluzek
- Centre for Sport, Exercise and Osteoarthritis Research Versus Arthritis, University of Oxford, Oxford, UK
- Nuffield Department of Orthopaedics, Rheumatology, and Musculoskeletal Sciences, University of Oxford, Oxford, UK
- University of Nottingham, Nottingham, UK
| |
Collapse
|
52
|
Belbasis L, Panagiotou OA. Reproducibility of prediction models in health services research. BMC Res Notes 2022; 15:204. [PMID: 35690767 PMCID: PMC9188254 DOI: 10.1186/s13104-022-06082-4] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/31/2022] [Accepted: 05/18/2022] [Indexed: 12/23/2022] Open
Abstract
The field of health services research studies the health care system by examining outcomes relevant to patients and clinicians but also health economists and policy makers. Such outcomes often include health care spending, and utilization of care services. Building accurate prediction models using reproducible research practices for health services research is important for evidence-based decision making. Several systematic reviews have summarized prediction models for outcomes relevant to health services research, but these systematic reviews do not present a thorough assessment of reproducibility and research quality of the prediction modelling studies. In the present commentary, we discuss how recent advances in prediction modelling in other medical fields can be applied to health services research. We also describe the current status of prediction modelling in health services research, and we summarize available methodological guidance for the development, update, external validation and systematic appraisal of prediction models.
Collapse
Affiliation(s)
- Lazaros Belbasis
- Meta-Research Innovation Center Berlin, QUEST Center, Berlin Institute of Health, Charité - Universitätsmedizin Berlin, Berlin, Germany.
| | - Orestis A Panagiotou
- Center for Evidence Synthesis in Health, School of Public Health, Brown University, Providence, RI, USA.,Department of Health Services, Policy and Practice, School of Public Health, Brown University, Providence, RI, USA.,Department of Epidemiology, School of Public Health, Brown University, Providence, RI, USA
| |
Collapse
|
53
|
Abdulaziz KE, Perry JJ, Yadav K, Dowlatshahi D, Stiell IG, Wells GA, Taljaard M. Quality and transparency of reporting derivation and validation prognostic studies of recurrent stroke in patients with TIA and minor stroke: a systematic review. Diagn Progn Res 2022; 6:9. [PMID: 35585563 PMCID: PMC9118704 DOI: 10.1186/s41512-022-00123-z] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/14/2021] [Accepted: 03/01/2022] [Indexed: 12/23/2022] Open
Abstract
BACKGROUND Clinical prediction models/scores help clinicians make optimal evidence-based decisions when caring for their patients. To critically appraise such prediction models for use in a clinical setting, essential information on the derivation and validation of the models needs to be transparently reported. In this systematic review, we assessed the quality of reporting of derivation and validation studies of prediction models for the prognosis of recurrent stroke in patients with transient ischemic attack or minor stroke. METHODS MEDLINE and EMBASE databases were searched up to February 04, 2020. Studies reporting development or validation of multivariable prognostic models predicting recurrent stroke within 90 days in patients with TIA or minor stroke were included. Included studies were appraised for reporting quality and conduct using a select list of items from the Transparent Reporting of a Multivariable Prediction Model for Individual Prognosis or Diagnosis (TRIPOD) Statement. RESULTS After screening 7026 articles, 60 eligible articles were retained, consisting of 100 derivation and validation studies of 27 unique prediction models. Four models were newly derived while 23 were developed by validating and updating existing models. Of the 60 articles, 15 (25%) reported an informative title. Among the 100 derivation and validation studies, few reported whether assessment of the outcome (24%) and predictors (12%) was blinded. Similarly, sample size justifications (49%), description of methods for handling missing data (16.1%), and model calibration (5%) were seldom reported. Among the 96 validation studies, 17 (17.7%) clearly reported on similarity (in terms of setting, eligibility criteria, predictors, and outcomes) between the validation and the derivation datasets. Items with the highest prevalence of adherence were the source of data (99%), eligibility criteria (93%), measures of discrimination (81%) and study setting (65%). CONCLUSIONS The majority of derivation and validation studies for the prognosis of recurrent stroke in TIA and minor stroke patients suffer from poor reporting quality. We recommend that all prediction model derivation and validation studies follow the TRIPOD statement to improve transparency and promote uptake of more reliable prediction models in practice. TRIAL REGISTRATION The protocol for this review was registered with PROSPERO (Registration number CRD42020201130 ).
Collapse
Affiliation(s)
- Kasim E. Abdulaziz
- grid.412687.e0000 0000 9606 5108Clinical Epidemiology Program, Ottawa Hospital Research Institute, Ottawa, Ontario Canada
- grid.28046.380000 0001 2182 2255School of Epidemiology and Public Health, Faculty of Medicine, University of Ottawa, Ottawa, Ontario Canada
| | - Jeffrey J. Perry
- grid.412687.e0000 0000 9606 5108Clinical Epidemiology Program, Ottawa Hospital Research Institute, Ottawa, Ontario Canada
- grid.28046.380000 0001 2182 2255School of Epidemiology and Public Health, Faculty of Medicine, University of Ottawa, Ottawa, Ontario Canada
- grid.28046.380000 0001 2182 2255Department of Emergency Medicine, University of Ottawa, Ottawa, Ontario Canada
| | - Krishan Yadav
- grid.412687.e0000 0000 9606 5108Clinical Epidemiology Program, Ottawa Hospital Research Institute, Ottawa, Ontario Canada
- grid.28046.380000 0001 2182 2255Department of Emergency Medicine, University of Ottawa, Ottawa, Ontario Canada
| | - Dar Dowlatshahi
- grid.28046.380000 0001 2182 2255School of Epidemiology and Public Health, Faculty of Medicine, University of Ottawa, Ottawa, Ontario Canada
- grid.28046.380000 0001 2182 2255Department of Emergency Medicine, University of Ottawa, Ottawa, Ontario Canada
- grid.412687.e0000 0000 9606 5108Department of Medicine (Neurology), University of Ottawa, Ottawa Hospital Research Institute, Ottawa, Ontario Canada
| | - Ian G. Stiell
- grid.412687.e0000 0000 9606 5108Clinical Epidemiology Program, Ottawa Hospital Research Institute, Ottawa, Ontario Canada
- grid.28046.380000 0001 2182 2255School of Epidemiology and Public Health, Faculty of Medicine, University of Ottawa, Ottawa, Ontario Canada
- grid.28046.380000 0001 2182 2255Department of Emergency Medicine, University of Ottawa, Ottawa, Ontario Canada
| | - George A. Wells
- grid.412687.e0000 0000 9606 5108Clinical Epidemiology Program, Ottawa Hospital Research Institute, Ottawa, Ontario Canada
- grid.28046.380000 0001 2182 2255School of Epidemiology and Public Health, Faculty of Medicine, University of Ottawa, Ottawa, Ontario Canada
- grid.28046.380000 0001 2182 2255Cardiovascular Research Methods Centre, University of Ottawa Heart Institute, Ottawa, Ontario Canada
| | - Monica Taljaard
- grid.412687.e0000 0000 9606 5108Clinical Epidemiology Program, Ottawa Hospital Research Institute, Ottawa, Ontario Canada
- grid.28046.380000 0001 2182 2255School of Epidemiology and Public Health, Faculty of Medicine, University of Ottawa, Ottawa, Ontario Canada
| |
Collapse
|
54
|
Huang AW, Haslberger M, Coulibaly N, Galárraga O, Oganisian A, Belbasis L, Panagiotou OA. Multivariable prediction models for health care spending using machine learning: a protocol of a systematic review. Diagn Progn Res 2022; 6:4. [PMID: 35321760 PMCID: PMC8943988 DOI: 10.1186/s41512-022-00119-9] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 11/09/2021] [Accepted: 01/18/2022] [Indexed: 12/11/2022] Open
Abstract
BACKGROUND With rising cost pressures on health care systems, machine-learning (ML)-based algorithms are increasingly used to predict health care costs. Despite their potential advantages, the successful implementation of these methods could be undermined by biases introduced in the design, conduct, or analysis of studies seeking to develop and/or validate ML models. The utility of such models may also be negatively affected by poor reporting of these studies. In this systematic review, we aim to evaluate the reporting quality, methodological characteristics, and risk of bias of ML-based prediction models for individual-level health care spending. METHODS We will systematically search PubMed and Embase to identify studies developing, updating, or validating ML-based models to predict an individual's health care spending for any medical condition, over any time period, and in any setting. We will exclude prediction models of aggregate-level health care spending, models used to infer causality, models using radiomics or speech parameters, models of non-clinically validated predictors (e.g., genomics), and cost-effectiveness analyses without predicting individual-level health care spending. We will extract data based on the Checklist for Critical Appraisal and Data Extraction for Systematic Reviews of Prediction Modeling Studies (CHARMS), previously published research, and relevant recommendations. We will assess the adherence of ML-based studies to the Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD) statement and examine the inclusion of transparency and reproducibility indicators (e.g. statements on data sharing). To assess the risk of bias, we will apply the Prediction model Risk Of Bias Assessment Tool (PROBAST). Findings will be stratified by study design, ML methods used, population characteristics, and medical field. DISCUSSION Our systematic review will appraise the quality, reporting, and risk of bias of ML-based models for individualized health care cost prediction. This review will provide an overview of the available models and give insights into the strengths and limitations of using ML methods for the prediction of health spending.
Collapse
Affiliation(s)
- Andrew W Huang
- Department of Health Services, Policy and Practice, Brown University School of Public Health, Rhode Island, Providence, USA.
| | - Martin Haslberger
- QUEST Center, Berlin Institute of Health, Charité-Universitätsmedizin Berlin, Berlin, Germany
| | - Neto Coulibaly
- Department of Health Services, Policy and Practice, Brown University School of Public Health, Rhode Island, Providence, USA
| | - Omar Galárraga
- Department of Health Services, Policy and Practice, Brown University School of Public Health, Rhode Island, Providence, USA
| | - Arman Oganisian
- Department of Biostatistics, Brown University School of Public Health, Providence, Rhode Island, USA
| | - Lazaros Belbasis
- Meta-Research Innovation Center Berlin, QUEST Center, Berlin Institute of Health, Charité-Universitätsmedizin Berlin, Berlin, Germany
| | - Orestis A Panagiotou
- Department of Health Services, Policy and Practice, Brown University School of Public Health, Rhode Island, Providence, USA
- Center for Evidence Synthesis in Health, Brown University School of Public Health, Providence, Rhode Island, USA
| |
Collapse
|
55
|
Lu SC, Xu C, Nguyen CH, Geng Y, Pfob A, Sidey-Gibbons C. Machine Learning-Based Short-Term Mortality Prediction Models for Patients With Cancer Using Electronic Health Record Data: Systematic Review and Critical Appraisal. JMIR Med Inform 2022; 10:e33182. [PMID: 35285816 PMCID: PMC8961346 DOI: 10.2196/33182] [Citation(s) in RCA: 20] [Impact Index Per Article: 6.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/26/2021] [Revised: 01/23/2022] [Accepted: 01/31/2022] [Indexed: 01/17/2023] Open
Abstract
BACKGROUND In the United States, national guidelines suggest that aggressive cancer care should be avoided in the final months of life. However, guideline compliance currently requires clinicians to make judgments based on their experience as to when a patient is nearing the end of their life. Machine learning (ML) algorithms may facilitate improved end-of-life care provision for patients with cancer by identifying patients at risk of short-term mortality. OBJECTIVE This study aims to summarize the evidence for applying ML in ≤1-year cancer mortality prediction to assist with the transition to end-of-life care for patients with cancer. METHODS We searched MEDLINE, Embase, Scopus, Web of Science, and IEEE to identify relevant articles. We included studies describing ML algorithms predicting ≤1-year mortality in patients of oncology. We used the prediction model risk of bias assessment tool to assess the quality of the included studies. RESULTS We included 15 articles involving 110,058 patients in the final synthesis. Of the 15 studies, 12 (80%) had a high or unclear risk of bias. The model performance was good: the area under the receiver operating characteristic curve ranged from 0.72 to 0.92. We identified common issues leading to biased models, including using a single performance metric, incomplete reporting of or inappropriate modeling practice, and small sample size. CONCLUSIONS We found encouraging signs of ML performance in predicting short-term cancer mortality. Nevertheless, no included ML algorithms are suitable for clinical practice at the current stage because of the high risk of bias and uncertainty regarding real-world performance. Further research is needed to develop ML models using the modern standards of algorithm development and reporting.
Collapse
Affiliation(s)
- Sheng-Chieh Lu
- Department of Symptom Research, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | - Cai Xu
- Department of Symptom Research, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | - Chandler H Nguyen
- McGovern Medical School, University of Texas Health Science Center, Houston, TX, United States
| | - Yimin Geng
- Research Medical Library, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| | - André Pfob
- Department of Obstetrics and Gynecology, Heidelberg University Hospital, Heidelberg, Germany
| | - Chris Sidey-Gibbons
- Department of Symptom Research, The University of Texas MD Anderson Cancer Center, Houston, TX, United States
| |
Collapse
|
56
|
Machine Learning Applications for Differentiation of Glioma from Brain Metastasis-A Systematic Review. Cancers (Basel) 2022; 14:cancers14061369. [PMID: 35326526 PMCID: PMC8946855 DOI: 10.3390/cancers14061369] [Citation(s) in RCA: 14] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/19/2022] [Revised: 02/22/2022] [Accepted: 03/01/2022] [Indexed: 12/19/2022] Open
Abstract
Simple Summary We present a systematic review of published reports on machine learning (ML) applications for the differentiation of gliomas from brain metastases by summarizing study characteristics, strengths, and pitfalls. Based on these findings, we present recommendations for future research in this field. Abstract Glioma and brain metastasis can be difficult to distinguish on conventional magnetic resonance imaging (MRI) due to the similarity of imaging features in specific clinical circumstances. Multiple studies have investigated the use of machine learning (ML) models for non-invasive differentiation of glioma from brain metastasis. Many of the studies report promising classification results, however, to date, none have been implemented into clinical practice. After a screening of 12,470 studies, we included 29 eligible studies in our systematic review. From each study, we aggregated data on model design, development, and best classifiers, as well as quality of reporting according to the TRIPOD statement. In a subset of eligible studies, we conducted a meta-analysis of the reported AUC. It was found that data predominantly originated from single-center institutions (n = 25/29) and only two studies performed external validation. The median TRIPOD adherence was 0.48, indicating insufficient quality of reporting among surveyed studies. Our findings illustrate that despite promising classification results, reliable model assessment is limited by poor reporting of study design and lack of algorithm validation and generalizability. Therefore, adherence to quality guidelines and validation on outside datasets is critical for the clinical translation of ML for the differentiation of glioma and brain metastasis.
Collapse
|
57
|
Weaver CGW, Basmadjian RB, Williamson T, McBrien K, Sajobi T, Boyne D, Yusuf M, Ronksley PE. Reporting of Model Performance and Statistical Methods in Studies That Use Machine Learning to Develop Clinical Prediction Models: Protocol for a Systematic Review. JMIR Res Protoc 2022; 11:e30956. [PMID: 35238322 PMCID: PMC8931652 DOI: 10.2196/30956] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/13/2021] [Revised: 12/09/2021] [Accepted: 12/31/2021] [Indexed: 11/13/2022] Open
Abstract
BACKGROUND With the growing excitement of the potential benefits of using machine learning and artificial intelligence in medicine, the number of published clinical prediction models that use these approaches has increased. However, there is evidence (albeit limited) that suggests that the reporting of machine learning-specific aspects in these studies is poor. Further, there are no reviews assessing the reporting quality or broadly accepted reporting guidelines for these aspects. OBJECTIVE This paper presents the protocol for a systematic review that will assess the reporting quality of machine learning-specific aspects in studies that use machine learning to develop clinical prediction models. METHODS We will include studies that use a supervised machine learning algorithm to develop a prediction model for use in clinical practice (ie, for diagnosis or prognosis of a condition or identification of candidates for health care interventions). We will search MEDLINE for studies published in 2019, pseudorandomly sort the records, and screen until we obtain 100 studies that meet our inclusion criteria. We will assess reporting quality with a novel checklist developed in parallel with this review, which includes content derived from existing reporting guidelines, textbooks, and consultations with experts. The checklist will cover 4 key areas where the reporting of machine learning studies is unique: modelling steps (order and data used for each step), model performance (eg, reporting the performance of each model compared), statistical methods (eg, describing the tuning approach), and presentation of models (eg, specifying the predictors that contributed to the final model). RESULTS We completed data analysis in August 2021 and are writing the manuscript. We expect to submit the results to a peer-reviewed journal in early 2022. CONCLUSIONS This review will contribute to more standardized and complete reporting in the field by identifying areas where reporting is poor and can be improved. TRIAL REGISTRATION PROSPERO International Prospective Register of Systematic Reviews CRD42020206167; https://www.crd.york.ac.uk/PROSPERO/display_record.php?RecordID=206167. INTERNATIONAL REGISTERED REPORT IDENTIFIER (IRRID) RR1-10.2196/30956.
Collapse
Affiliation(s)
- Colin George Wyllie Weaver
- Department of Community Health Sciences, Cumming School of Medicine, University of Calgary, Calgary, AB, Canada
| | - Robert B Basmadjian
- Department of Community Health Sciences, Cumming School of Medicine, University of Calgary, Calgary, AB, Canada
| | - Tyler Williamson
- Department of Community Health Sciences, Cumming School of Medicine, University of Calgary, Calgary, AB, Canada
| | - Kerry McBrien
- Department of Community Health Sciences, Cumming School of Medicine, University of Calgary, Calgary, AB, Canada.,Department of Family Medicine, Cumming School of Medicine, University of Calgary, Calgary, AB, Canada
| | - Tolu Sajobi
- Department of Community Health Sciences, Cumming School of Medicine, University of Calgary, Calgary, AB, Canada
| | - Devon Boyne
- Department of Oncology, Cumming School of Medicine, University of Calgary, Calgary, AB, Canada
| | - Mohamed Yusuf
- Faculty of Science & Engineering, Manchester Metropolitan University, Manchester, United Kingdom
| | - Paul Everett Ronksley
- Department of Community Health Sciences, Cumming School of Medicine, University of Calgary, Calgary, AB, Canada
| |
Collapse
|
58
|
Appelt AL, Elhaminia B, Gooya A, Gilbert A, Nix M. Deep Learning for Radiotherapy Outcome Prediction Using Dose Data - A Review. Clin Oncol (R Coll Radiol) 2022; 34:e87-e96. [PMID: 34924256 DOI: 10.1016/j.clon.2021.12.002] [Citation(s) in RCA: 17] [Impact Index Per Article: 5.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/06/2021] [Revised: 11/26/2021] [Accepted: 12/01/2021] [Indexed: 12/23/2022]
Abstract
Artificial intelligence, and in particular deep learning using convolutional neural networks, has been used extensively for image classification and segmentation, including on medical images for diagnosis and prognosis prediction. Use in radiotherapy prognostic modelling is still limited, however, especially as applied to toxicity and tumour response prediction from radiation dose distributions. We review and summarise studies that applied deep learning to radiotherapy dose data, in particular studies that utilised full three-dimensional dose distributions. Ten papers have reported on deep learning models for outcome prediction utilising spatial dose information, whereas four studies used reduced dimensionality (dose volume histogram) information for prediction. Many of these studies suffer from the same issues that plagued early normal tissue complication probability modelling, including small, single-institutional patient cohorts, lack of external validation, poor data and model reporting, use of late toxicity data without taking time-to-event into account, and nearly exclusive focus on clinician-reported complications. They demonstrate, however, how radiation dose, imaging and clinical data may be technically integrated in convolutional neural networks-based models; and some studies explore how deep learning may help better understand spatial variation in radiosensitivity. In general, there are a number of issues specific to the intersection of radiotherapy outcome modelling and deep learning, for example translation of model developments into treatment plan optimisation, which will require further combined effort from the radiation oncology and artificial intelligence communities.
Collapse
Affiliation(s)
- A L Appelt
- Leeds Institute of Medical Research at St James's, University of Leeds, Leeds, UK.
| | - B Elhaminia
- Centre for Computational Imaging and Simulation Technologies in Biomedicine (CISTIB), Department of Electronic and Electrical Engineering, University of Leeds, Leeds, UK
| | - A Gooya
- Centre for Computational Imaging and Simulation Technologies in Biomedicine (CISTIB), Department of Electronic and Electrical Engineering, University of Leeds, Leeds, UK
| | - A Gilbert
- Leeds Institute of Medical Research at St James's, University of Leeds, Leeds, UK
| | - M Nix
- Department of Medical Physics and Engineering, Leeds Cancer Centre, St James's University Hospital, Leeds, UK
| |
Collapse
|
59
|
Faes L, Sim DA, van Smeden M, Held U, Bossuyt PM, Bachmann LM. Artificial Intelligence and Statistics: Just the Old Wine in New Wineskins? Front Digit Health 2022; 4:833912. [PMID: 35156082 PMCID: PMC8825497 DOI: 10.3389/fdgth.2022.833912] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/14/2021] [Accepted: 01/03/2022] [Indexed: 01/03/2023] Open
Affiliation(s)
- Livia Faes
- Medical Retina Department, Moorfields Eye Hospital NHS Foundation Trust, London, United Kingdom
- Medignition Inc., Research Consultants, Zurich, Switzerland
- *Correspondence: Livia Faes
| | - Dawn A. Sim
- Medical Retina Department, Moorfields Eye Hospital NHS Foundation Trust, London, United Kingdom
- Health Data Research UK, London, United Kingdom
- National Institute for Health Research (NIHR) Biomedical Research Centre for Ophthalmology, Moorfields Eye Hospital National Health Service (NHS) Foundation Trust and University College London (UCL) Institute of Ophthalmology, London, United Kingdom
| | - Maarten van Smeden
- Julius Center for Health Science and Primary Care, University Medical Center Utrecht, University of Utrecht, Utrecht, Netherlands
| | - Ulrike Held
- Department of Biostatistics, Epidemiology, Biostatistics and Prevention Institute, University of Zurich, Zurich, Switzerland
| | - Patrick M. Bossuyt
- Department of Clinical Epidemiology, Biostatistics and Bioinformatics, Amsterdam Public Health Research Institute, Amsterdam University Medical Centers, Amsterdam, Netherlands
| | | |
Collapse
|
60
|
Nijman S, Leeuwenberg AM, Beekers I, Verkouter I, Jacobs J, Bots ML, Asselbergs FW, Moons K, Debray T. Missing data is poorly handled and reported in prediction model studies using machine learning: a literature review. J Clin Epidemiol 2021; 142:218-229. [PMID: 34798287 DOI: 10.1016/j.jclinepi.2021.11.023] [Citation(s) in RCA: 60] [Impact Index Per Article: 15.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/01/2021] [Revised: 11/01/2021] [Accepted: 11/10/2021] [Indexed: 12/23/2022]
Abstract
OBJECTIVES Missing data is a common problem during the development, evaluation, and implementation of prediction models. Although machine learning (ML) methods are often said to be capable of circumventing missing data, it is unclear how these methods are used in medical research. We aim to find out if and how well prediction model studies using machine learning report on their handling of missing data. STUDY DESIGN AND SETTING We systematically searched the literature on published papers between 2018 and 2019 about primary studies developing and/or validating clinical prediction models using any supervised ML methodology across medical fields. From the retrieved studies information about the amount and nature (e.g. missing completely at random, potential reasons for missingness) of missing data and the way they were handled were extracted. RESULTS We identified 152 machine learning-based clinical prediction model studies. A substantial amount of these 152 papers did not report anything on missing data (n = 56/152). A majority (n = 96/152) reported details on the handling of missing data (e.g., methods used), though many of these (n = 46/96) did not report the amount of the missingness in the data. In these 96 papers the authors only sometimes reported possible reasons for missingness (n = 7/96) and information about missing data mechanisms (n = 8/96). The most common approach for handling missing data was deletion (n = 65/96), mostly via complete-case analysis (CCA) (n = 43/96). Very few studies used multiple imputation (n = 8/96) or built-in mechanisms such as surrogate splits (n = 7/96) that directly address missing data during the development, validation, or implementation of the prediction model. CONCLUSION Though missing values are highly common in any type of medical research and certainly in the research based on routine healthcare data, a majority of the prediction model studies using machine learning does not report sufficient information on the presence and handling of missing data. Strategies in which patient data are simply omitted are unfortunately the most often used methods, even though it is generally advised against and well known that it likely causes bias and loss of analytical power in prediction model development and in the predictive accuracy estimates. Prediction model researchers should be much more aware of alternative methodologies to address missing data.
Collapse
Affiliation(s)
- Swj Nijman
- Julius Center for Health Sciences and Primary Care, University Medical Center Utrecht, Utrecht University, Heidelberglaan 100, Utrecht, 3584 CX , The Netherlands.
| | - A M Leeuwenberg
- Julius Center for Health Sciences and Primary Care, University Medical Center Utrecht, Utrecht University, Heidelberglaan 100, Utrecht, 3584 CX , The Netherlands
| | - I Beekers
- Department of Health, Ortec B.V. Zoetermeer, The Netherlands
| | - I Verkouter
- Department of Health, Ortec B.V. Zoetermeer, The Netherlands
| | - Jjl Jacobs
- Department of Health, Ortec B.V. Zoetermeer, The Netherlands
| | - M L Bots
- Julius Center for Health Sciences and Primary Care, University Medical Center Utrecht, Utrecht University, Heidelberglaan 100, Utrecht, 3584 CX , The Netherlands
| | - F W Asselbergs
- Department of Cardiology, University Medical Center Utrecht, Utrecht University, The Netherlands; Institute of Cardiovascular Science, Population Health Sciences, University College London, London, UK; Health Data Research UK, Institute of Health Informatics, University College London, London, UK
| | - Kgm Moons
- Julius Center for Health Sciences and Primary Care, University Medical Center Utrecht, Utrecht University, Heidelberglaan 100, Utrecht, 3584 CX , The Netherlands
| | - Tpa Debray
- Julius Center for Health Sciences and Primary Care, University Medical Center Utrecht, Utrecht University, Heidelberglaan 100, Utrecht, 3584 CX , The Netherlands; Health Data Research UK, Institute of Health Informatics, University College London, London, UK
| |
Collapse
|