1
|
Vlontzou ME, Athanasiou M, Dalakleidi KV, Skampardoni I, Davatzikos C, Nikita K. A comprehensive interpretable machine learning framework for mild cognitive impairment and Alzheimer's disease diagnosis. Sci Rep 2025; 15:8410. [PMID: 40069342 PMCID: PMC11897299 DOI: 10.1038/s41598-025-92577-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/18/2024] [Accepted: 02/28/2025] [Indexed: 03/15/2025] Open
Abstract
An interpretable machine learning (ML) framework is introduced to enhance the diagnosis of Mild Cognitive Impairment (MCI) and Alzheimer's disease (AD) by ensuring robustness of the ML models' interpretations. The dataset used comprises volumetric measurements from brain MRI and genetic data from healthy individuals and patients with MCI/AD, obtained through the Alzheimer's Disease Neuroimaging Initiative. The existing class imbalance is addressed by an ensemble learning approach, while various attribution-based and counterfactual-based interpretability methods are leveraged towards producing diverse explanations related to the pathophysiology of MCI/AD. A unification method combining SHAP with counterfactual explanations assesses the interpretability techniques' robustness. The best performing model yielded 87.5% balanced accuracy and 90.8% F1-score. The attribution-based interpretability methods highlighted significant volumetric and genetic features related to MCI/AD risk. The unification method provided useful insights regarding those features' necessity and sufficiency, further showcasing their significance in MCI/AD diagnosis.
Collapse
Affiliation(s)
- Maria Eleftheria Vlontzou
- Faculty of Electrical and Computer Engineering, National Technical University of Athens, Athens, 15773, Greece.
| | - Maria Athanasiou
- Faculty of Electrical and Computer Engineering, National Technical University of Athens, Athens, 15773, Greece
| | - Kalliopi V Dalakleidi
- Faculty of Electrical and Computer Engineering, National Technical University of Athens, Athens, 15773, Greece
| | - Ioanna Skampardoni
- Faculty of Electrical and Computer Engineering, National Technical University of Athens, Athens, 15773, Greece
| | - Christos Davatzikos
- Center for Biomedical Image Computing and Analytics, University of Pennsylvania, Philadelphia, PA, USA
- Department ofRadiology, University of Pennsylvania, Philadelphia, PA, USA
| | - Konstantina Nikita
- Faculty of Electrical and Computer Engineering, National Technical University of Athens, Athens, 15773, Greece
| |
Collapse
|
2
|
Deng Y, Tang M, Liu Q, Fan X, Shu J, Chen J, Chen M, Yang L. Prediction of cardiac remodeling and myocardial fibrosis in athletes based on IVIM-DWI images. iScience 2025; 28:111567. [PMID: 39829680 PMCID: PMC11742616 DOI: 10.1016/j.isci.2024.111567] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2024] [Revised: 09/29/2024] [Accepted: 12/06/2024] [Indexed: 01/22/2025] Open
Abstract
Myocardial microcirculation in athletes and its relationship with cardiac remodeling (CR) and myocardial fibrosis (MF) are not fully understood. We prospectively enrolled 174 athletes and 54 healthy sedentary controls for intravoxel incoherent motion (IVIM) diffusion-weighted imaging of cardiac magnetic resonance imaging. Athletes exhibited significantly lower fast apparent diffusion coefficient (ADCfast) and perfusion fraction (f) in 16 myocardial segments and each blood supply area compared to controls (p < 0.05). Athletes with CR and/or MF had lower myocardial slow apparent diffusion coefficient (ADCslow) values than those without (p < 0.05). A gradient boosting machine (GBM) effectively predicted CR and/or MF based on these hypoperfusion parameters, with an area under the receiver operating characteristic curve of 0.947 in the training set and 0.841 in the test set. The GBM model, leveraging IVIM parameters, could predict the occurrence of CR and/or MF, offering a potential tool for monitoring and managing the athletes' health.
Collapse
Affiliation(s)
- Yujiao Deng
- Department of Radiology, The Affiliated Hospital of Southwest Medical University, 25# Tai Ping Street, Luzhou, Sichuan 646000, China
- Department of Nuclear Medicine, Sichuan Provincial People’s Hospital, University of Electronic Science and Technology of China, 32#, West Second Section of First Ring Road, Chengdu, Sichuan 610000, China
| | - Min Tang
- Department of Radiology, The Affiliated Hospital of Southwest Medical University, 25# Tai Ping Street, Luzhou, Sichuan 646000, China
| | - Qian Liu
- Department of Radiology, The Affiliated Hospital of Southwest Medical University, 25# Tai Ping Street, Luzhou, Sichuan 646000, China
| | - Xinrong Fan
- Department of Cardiology, the Affiliated Hospital of Southwest Medical University, 25# Tai Ping Street, Luzhou, Sichuan 646000, China
| | - Jian Shu
- Department of Radiology, The Affiliated Hospital of Southwest Medical University, 25# Tai Ping Street, Luzhou, Sichuan 646000, China
| | - Jing Chen
- Department of Radiology, The Affiliated Hospital of Southwest Medical University, 25# Tai Ping Street, Luzhou, Sichuan 646000, China
| | - Meining Chen
- MR Research Collaboration, Siemens Healthineers Ltd., Chengdu, China
| | - Lu Yang
- Department of Radiology, The Affiliated Hospital of Southwest Medical University, 25# Tai Ping Street, Luzhou, Sichuan 646000, China
| |
Collapse
|
3
|
Hafeez Y, Memon K, AL-Quraishi MS, Yahya N, Elferik S, Ali SSA. Explainable AI in Diagnostic Radiology for Neurological Disorders: A Systematic Review, and What Doctors Think About It. Diagnostics (Basel) 2025; 15:168. [PMID: 39857052 PMCID: PMC11764244 DOI: 10.3390/diagnostics15020168] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/18/2024] [Revised: 12/22/2024] [Accepted: 01/08/2025] [Indexed: 01/27/2025] Open
Abstract
Background: Artificial intelligence (AI) has recently made unprecedented contributions in every walk of life, but it has not been able to work its way into diagnostic medicine and standard clinical practice yet. Although data scientists, researchers, and medical experts have been working in the direction of designing and developing computer aided diagnosis (CAD) tools to serve as assistants to doctors, their large-scale adoption and integration into the healthcare system still seems far-fetched. Diagnostic radiology is no exception. Imagining techniques like magnetic resonance imaging (MRI), computed tomography (CT), and positron emission tomography (PET) scans have been widely and very effectively employed by radiologists and neurologists for the differential diagnoses of neurological disorders for decades, yet no AI-powered systems to analyze such scans have been incorporated into the standard operating procedures of healthcare systems. Why? It is absolutely understandable that in diagnostic medicine, precious human lives are on the line, and hence there is no room even for the tiniest of mistakes. Nevertheless, with the advent of explainable artificial intelligence (XAI), the old-school black boxes of deep learning (DL) systems have been unraveled. Would XAI be the turning point for medical experts to finally embrace AI in diagnostic radiology? This review is a humble endeavor to find the answers to these questions. Methods: In this review, we present the journey and contributions of AI in developing systems to recognize, preprocess, and analyze brain MRI scans for differential diagnoses of various neurological disorders, with special emphasis on CAD systems embedded with explainability. A comprehensive review of the literature from 2017 to 2024 was conducted using host databases. We also present medical domain experts' opinions and summarize the challenges up ahead that need to be addressed in order to fully exploit the tremendous potential of XAI in its application to medical diagnostics and serve humanity. Results: Forty-seven studies were summarized and tabulated with information about the XAI technology and datasets employed, along with performance accuracies. The strengths and weaknesses of the studies have also been discussed. In addition, the opinions of seven medical experts from around the world have been presented to guide engineers and data scientists in developing such CAD tools. Conclusions: Current CAD research was observed to be focused on the enhancement of the performance accuracies of the DL regimens, with less attention being paid to the authenticity and usefulness of explanations. A shortage of ground truth data for explainability was also observed. Visual explanation methods were found to dominate; however, they might not be enough, and more thorough and human professor-like explanations would be required to build the trust of healthcare professionals. Special attention to these factors along with the legal, ethical, safety, and security issues can bridge the current gap between XAI and routine clinical practice.
Collapse
Affiliation(s)
- Yasir Hafeez
- Faculty of Science and Engineering, University of Nottingham, Jalan Broga, Semenyih 43500, Selangor Darul Ehsan, Malaysia;
| | - Khuhed Memon
- Centre for Intelligent Signal and Imaging Research, Department of Electrical and Electronic Engineering, Universiti Teknologi PETRONAS, Seri Iskandar 32610, Perak Darul Ridzuan, Malaysia; (K.M.); (N.Y.)
| | - Maged S. AL-Quraishi
- Interdisciplinary Research Center for Smart Mobility and Logistics, King Fahd University of Petroleum and Minerals, Dhahran 31261, Saudi Arabia; (M.S.A.-Q.); (S.E.)
| | - Norashikin Yahya
- Centre for Intelligent Signal and Imaging Research, Department of Electrical and Electronic Engineering, Universiti Teknologi PETRONAS, Seri Iskandar 32610, Perak Darul Ridzuan, Malaysia; (K.M.); (N.Y.)
| | - Sami Elferik
- Interdisciplinary Research Center for Smart Mobility and Logistics, King Fahd University of Petroleum and Minerals, Dhahran 31261, Saudi Arabia; (M.S.A.-Q.); (S.E.)
| | - Syed Saad Azhar Ali
- Aerospace Engineering Department and Interdisciplinary Research Center for Smart Mobility and Logistics, and Interdisciplinary Research Center Aviation and Space Exploration, King Fahd University of Petroleum and Minerals, Dhahran 31261, Saudi Arabia
| |
Collapse
|
4
|
Salih AM, Galazzo IB, Raisi-Estabragh Z, Petersen SE, Menegaz G, Radeva P. Characterizing the Contribution of Dependent Features in XAI Methods. IEEE J Biomed Health Inform 2024; 28:6466-6473. [PMID: 38696291 DOI: 10.1109/jbhi.2024.3395289] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/04/2024]
Abstract
Explainable Artificial Intelligence (XAI) provides tools to help understanding how AI models work and reach a particular decision or outcome. It helps to increase the interpretability of models and makes them more trustworthy and transparent. In this context, many XAI methods have been proposed to make black-box and complex models more digestible from a human perspective. However, one of the main issues that XAI methods have to face especially when dealing with a high number of features is the presence of multicollinearity, which casts shadows on the robustness of the XAI outcomes, such as the ranking of informative features. Most of the current XAI methods either do not consider the collinearity or assume the features are independent which, in general, is not necessarily true. Here, we propose a simple, yet useful, proxy that modifies the outcome of any XAI feature ranking method allowing to account for the dependency among the features, and to reveal their impact on the outcome. The proposed method was applied to SHAP, as an example of XAI method which assume that the features are independent. For this purpose, several models were exploited for a well-known classification task (males versus females) using nine cardiac phenotypes extracted from cardiac magnetic resonance imaging as features. Principal component analysis and biological plausibility were employed to validate the proposed method. Our results showed that the proposed proxy could lead to a more robust list of informative features compared to the original SHAP in presence of collinearity.
Collapse
|
5
|
Ponce‐Bobadilla AV, Schmitt V, Maier CS, Mensing S, Stodtmann S. Practical guide to SHAP analysis: Explaining supervised machine learning model predictions in drug development. Clin Transl Sci 2024; 17:e70056. [PMID: 39463176 PMCID: PMC11513550 DOI: 10.1111/cts.70056] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/14/2024] [Revised: 10/01/2024] [Accepted: 10/13/2024] [Indexed: 10/29/2024] Open
Abstract
Despite increasing interest in using Artificial Intelligence (AI) and Machine Learning (ML) models for drug development, effectively interpreting their predictions remains a challenge, which limits their impact on clinical decisions. We address this issue by providing a practical guide to SHapley Additive exPlanations (SHAP), a popular feature-based interpretability method, which can be seamlessly integrated into supervised ML models to gain a deeper understanding of their predictions, thereby enhancing their transparency and trustworthiness. This tutorial focuses on the application of SHAP analysis to standard ML black-box models for regression and classification problems. We provide an overview of various visualization plots and their interpretation, available software for implementing SHAP, and highlight best practices, as well as special considerations, when dealing with binary endpoints and time-series models. To enhance the reader's understanding for the method, we also apply it to inherently explainable regression models. Finally, we discuss the limitations and ongoing advancements aimed at tackling the current drawbacks of the method.
Collapse
Affiliation(s)
| | | | | | - Sven Mensing
- AbbVie Deutschland GmbH & Co. KGLudwigshafenGermany
| | | |
Collapse
|
6
|
Salih AM, Galazzo IB, Gkontra P, Rauseo E, Lee AM, Lekadir K, Radeva P, Petersen SE, Menegaz G. A review of evaluation approaches for explainable AI with applications in cardiology. Artif Intell Rev 2024; 57:240. [PMID: 39132011 PMCID: PMC11315784 DOI: 10.1007/s10462-024-10852-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 07/03/2024] [Indexed: 08/13/2024]
Abstract
Explainable artificial intelligence (XAI) elucidates the decision-making process of complex AI models and is important in building trust in model predictions. XAI explanations themselves require evaluation as to accuracy and reasonableness and in the context of use of the underlying AI model. This review details the evaluation of XAI in cardiac AI applications and has found that, of the studies examined, 37% evaluated XAI quality using literature results, 11% used clinicians as domain-experts, 11% used proxies or statistical analysis, with the remaining 43% not assessing the XAI used at all. We aim to inspire additional studies within healthcare, urging researchers not only to apply XAI methods but to systematically assess the resulting explanations, as a step towards developing trustworthy and safe models. Supplementary Information The online version contains supplementary material available at 10.1007/s10462-024-10852-w.
Collapse
Affiliation(s)
- Ahmed M. Salih
- William Harvey Research Institute, NIHR Barts Biomedical Research Centre, Queen Mary University of London, Charterhouse Square, London, EC1M 6BQ UK
- Department of Population Health Sciences, University of Leicester, University Rd, Leicester, LE1 7RH UK
- Department of Computer Science, University of Zakho, Duhok road, Zakho, Kurdistan Iraq
| | - Ilaria Boscolo Galazzo
- Department of Engineering for Innovative Medicine, University of Verona, S. Francesco, 22, 37129 Verona, Italy
| | - Polyxeni Gkontra
- Artificial Intelligence in Medicine Lab (BCN-AIM), Departament de Matemàtiques i Informàtica, Universitat de Barcelona, Gran Via de les Corts Catalanes, 585, 08007 Barcelona, Spain
| | - Elisa Rauseo
- William Harvey Research Institute, NIHR Barts Biomedical Research Centre, Queen Mary University of London, Charterhouse Square, London, EC1M 6BQ UK
| | - Aaron Mark Lee
- William Harvey Research Institute, NIHR Barts Biomedical Research Centre, Queen Mary University of London, Charterhouse Square, London, EC1M 6BQ UK
| | - Karim Lekadir
- Artificial Intelligence in Medicine Lab (BCN-AIM), Departament de Matemàtiques i Informàtica, Universitat de Barcelona, Gran Via de les Corts Catalanes, 585, 08007 Barcelona, Spain
- Institució Catalana de Recerca i Estudis Avançats (ICREA), Passeig Lluís Companys 23, Barcelona, Spain
| | - Petia Radeva
- Departament de Matemàtiques i Informàtica, Universitat de Barcelona, Gran Via de les Corts Catalanes, 585, 08007 Barcelona, Spain
| | - Steffen E. Petersen
- William Harvey Research Institute, NIHR Barts Biomedical Research Centre, Queen Mary University of London, Charterhouse Square, London, EC1M 6BQ UK
- Barts Heart Centre, St Bartholomew’s Hospital, Barts Health NHS Trust, West Smithfield, London, UK
- Health Data Research, London, UK
- Alan Turing Institute, London, UK
| | - Gloria Menegaz
- Department of Engineering for Innovative Medicine, University of Verona, S. Francesco, 22, 37129 Verona, Italy
| |
Collapse
|
7
|
Brusini L, Cruciani F, Dall’Aglio G, Zajac T, Boscolo Galazzo I, Zucchelli M, Menegaz G. XAI-Based Assessment of the AMURA Model for Detecting Amyloid-β and Tau Microstructural Signatures in Alzheimer's Disease. IEEE JOURNAL OF TRANSLATIONAL ENGINEERING IN HEALTH AND MEDICINE 2024; 12:569-579. [PMID: 39155922 PMCID: PMC11329216 DOI: 10.1109/jtehm.2024.3430035] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/28/2023] [Revised: 05/24/2024] [Accepted: 07/08/2024] [Indexed: 08/20/2024]
Abstract
Brain microstructural changes already occur in the earliest phases of Alzheimer's disease (AD) as evidenced in diffusion magnetic resonance imaging (dMRI) literature. This study investigates the potential of the novel dMRI Apparent Measures Using Reduced Acquisitions (AMURA) as imaging markers for capturing such tissue modifications.Tract-based spatial statistics (TBSS) and support vector machines (SVMs) based on different measures were exploited to distinguish between amyloid-beta/tau negative (A[Formula: see text]-/tau-) and A[Formula: see text]+/tau+ or A[Formula: see text]+/tau- subjects. Moreover, eXplainable Artificial Intelligence (XAI) was used to highlight the most influential features in the SVMs classifications and to validate the results by seeing the explanations' recurrence across different methods.TBSS analysis revealed significant differences between A[Formula: see text]-/tau- and other groups in line with the literature. The best SVM classification performance reached an accuracy of 0.73 by using advanced measures compared to more standard ones. Moreover, the explainability analysis suggested the results' stability and the central role of the cingulum to show early sign of AD.By relying on SVM classification and XAI interpretation of the outcomes, AMURA indices can be considered viable markers for amyloid and tau pathology. Clinical impact: This pre-clinical research revealed AMURA indices as viable imaging markers for timely AD diagnosis by acquiring clinically feasible dMR images, with advantages compared to more invasive methods employed nowadays.
Collapse
Affiliation(s)
- Lorenza Brusini
- Department of Engineering for Innovation MedicineUniversity of VeronaVerona37134Italy
| | - Federica Cruciani
- Department of Engineering for Innovation MedicineUniversity of VeronaVerona37134Italy
| | | | - Tommaso Zajac
- Department of Computer ScienceUniversity of VeronaVerona37134Italy
| | | | - Mauro Zucchelli
- Department of Research and Development Advanced ApplicationsOlea MedicalLa Ciotat13600France
| | - Gloria Menegaz
- Department of Engineering for Innovation MedicineUniversity of VeronaVerona37134Italy
| |
Collapse
|
8
|
Vimbi V, Shaffi N, Mahmud M. Interpreting artificial intelligence models: a systematic review on the application of LIME and SHAP in Alzheimer's disease detection. Brain Inform 2024; 11:10. [PMID: 38578524 PMCID: PMC10997568 DOI: 10.1186/s40708-024-00222-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/02/2023] [Accepted: 03/04/2024] [Indexed: 04/06/2024] Open
Abstract
Explainable artificial intelligence (XAI) has gained much interest in recent years for its ability to explain the complex decision-making process of machine learning (ML) and deep learning (DL) models. The Local Interpretable Model-agnostic Explanations (LIME) and Shaply Additive exPlanation (SHAP) frameworks have grown as popular interpretive tools for ML and DL models. This article provides a systematic review of the application of LIME and SHAP in interpreting the detection of Alzheimer's disease (AD). Adhering to PRISMA and Kitchenham's guidelines, we identified 23 relevant articles and investigated these frameworks' prospective capabilities, benefits, and challenges in depth. The results emphasise XAI's crucial role in strengthening the trustworthiness of AI-based AD predictions. This review aims to provide fundamental capabilities of LIME and SHAP XAI frameworks in enhancing fidelity within clinical decision support systems for AD prognosis.
Collapse
Affiliation(s)
- Viswan Vimbi
- College of Computing and Information Sciences, University of Technology and Applied Sciences, OM 311, Sohar, Sultanate of Oman
| | - Noushath Shaffi
- College of Computing and Information Sciences, University of Technology and Applied Sciences, OM 311, Sohar, Sultanate of Oman
| | - Mufti Mahmud
- Department of Computer Science, Nottingham Trent University, Nottingham, NG11 8NS, UK.
- Medical Technologies Innovation Facility, Nottingham Trent University, Nottingham, NG11 8NS, UK.
- Computing and Informatics Research Centre, Nottingham Trent University, Nottingham, NG11 8NS, UK.
| |
Collapse
|