1
|
Salvi M, Seoni S, Campagner A, Gertych A, Acharya UR, Molinari F, Cabitza F. Explainability and uncertainty: Two sides of the same coin for enhancing the interpretability of deep learning models in healthcare. Int J Med Inform 2025; 197:105846. [PMID: 39993336 DOI: 10.1016/j.ijmedinf.2025.105846] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/14/2024] [Revised: 02/19/2025] [Accepted: 02/19/2025] [Indexed: 02/26/2025]
Abstract
BACKGROUND The increasing use of Deep Learning (DL) in healthcare has highlighted the critical need for improved transparency and interpretability. While Explainable Artificial Intelligence (XAI) methods provide insights into model predictions, reliability cannot be guaranteed by simply relying on explanations. OBJECTIVES This position paper proposes the integration of Uncertainty Quantification (UQ) with XAI methods to improve model reliability and trustworthiness in healthcare applications. METHODS We examine state-of-the-art XAI and UQ techniques, discuss implementation challenges, and suggest solutions to combine UQ with XAI methods. We propose a framework for estimating both aleatoric and epistemic uncertainty in the XAI context, providing illustrative examples of their potential application. RESULTS Our analysis indicates that integrating UQ with XAI could significantly enhance the reliability of DL models in practice. This approach has the potential to reduce interpretation biases and over-reliance, leading to more cautious and conscious use of AI in healthcare.
Collapse
Affiliation(s)
- Massimo Salvi
- Biolab, PoliToBIOMed Lab, Department of Electronics and Telecommunications, Politecnico di Torino, Corso Duca degli Abruzzi 24, 10129 Turin, Italy.
| | - Silvia Seoni
- Biolab, PoliToBIOMed Lab, Department of Electronics and Telecommunications, Politecnico di Torino, Corso Duca degli Abruzzi 24, 10129 Turin, Italy
| | | | - Arkadiusz Gertych
- Faculty of Biomedical Engineering, Silesian University of Technology, Zabrze, Poland; Department of Surgery, Cedars-Sinai Medical Center, Los Angeles, CA, United States; Department of Pathology and Laboratory Medicine, Cedars-Sinai Medical Center, Los Angeles, CA, United States
| | - U Rajendra Acharya
- School of Mathematics, Physics and Computing, University of Southern Queensland, Springfield, Australia; Centre for Health Research, University of Southern Queensland, Springfield, QLD 4300, Australia
| | - Filippo Molinari
- Biolab, PoliToBIOMed Lab, Department of Electronics and Telecommunications, Politecnico di Torino, Corso Duca degli Abruzzi 24, 10129 Turin, Italy
| | - Federico Cabitza
- IRCCS Ospedale Galeazzi - Sant'Ambrogio, Milan, Italy; Department of Computer Science, Systems and Communication, University of Milano-Bicocca, Milan, Italy
| |
Collapse
|
2
|
Ghazi R, Ibrahim TK, Nasir JA, Gai S, Ali G, Boukhris I, Rehman Z. Iron oxide based magnetic nanoparticles for hyperthermia, MRI and drug delivery applications: a review. RSC Adv 2025; 15:11587-11616. [PMID: 40230636 PMCID: PMC11995399 DOI: 10.1039/d5ra00728c] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2025] [Accepted: 03/24/2025] [Indexed: 04/16/2025] Open
Abstract
Iron-oxide nanoparticles (IONPs) have garnered substantial attention in both research and technological domains due to their exceptional chemical and physical properties. These nanoparticles have mitigated the adverse effects of conventional treatment procedures by facilitating advanced theranostic approaches in integration with biomedicine. These IONPs have been extensively utilized in MRI (as contrast agents in diagnosis), drug delivery (as drug carriers), and hyperthermia (treatment), demonstrating promising results with potential for further enhancement. This study elucidates the operational principles of these NPs during diagnosis, drug delivery, and treatment, and emphasizes their precision and efficacy in transporting therapeutic agents to targeted sites without drug loss. It also analyses various challenges associated with the application of these IONPs in this field, such as biocompatibility, agglomeration, and toxicity. Furthermore, diverse strategies have been delineated to address these challenges. Overall, this review provides a comprehensive overview of the applications of IONPs in the field of biomedicine and treatment, along with the associated challenges. It offers significant assistance to researchers, professionals, and clinicians in the field of biomedicine.
Collapse
Affiliation(s)
- Rizwana Ghazi
- Department of Chemistry, Quaid-i-Azam University Islamabad 45320 Pakistan +92-(051)90642241 +92-(051)90642245
| | - Talib K Ibrahim
- Department of Petroleum Engineering, College of Engineering, Knowledge University Erbil Iraq
- Department of Petroleum Engineering, Al-Kitab University Altun Kupri Iraq
| | - Jamal Abdul Nasir
- Department of Chemistry, Quaid-i-Azam University Islamabad 45320 Pakistan +92-(051)90642241 +92-(051)90642245
| | - Shili Gai
- Key Laboratory of Superlight Materials and Surface Technology, Ministry of Education, College of Materials Science and Chemical Engineering, Harbin Engineering University Harbin 150001 P. R. China
| | - Ghafar Ali
- Nanomaterials Research Group (NRG), Physics Division, PINSTECH Nilore Islamabad Pakistan
| | - Imed Boukhris
- Department of Physics, Faculty of Science, King Khalid University P. O. Box 9004 Abha Saudi Arabia
| | - Ziaur Rehman
- Department of Chemistry, Quaid-i-Azam University Islamabad 45320 Pakistan +92-(051)90642241 +92-(051)90642245
| |
Collapse
|
3
|
Qi W, Xu X, Qian K, Schuller BW, Fortino G, Aliverti A. A Review of AIoT-Based Human Activity Recognition: From Application to Technique. IEEE J Biomed Health Inform 2025; 29:2425-2438. [PMID: 38809724 DOI: 10.1109/jbhi.2024.3406737] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/31/2024]
Abstract
This scoping review paper redefines the Artificial Intelligence-based Internet of Things (AIoT) driven Human Activity Recognition (HAR) field by systematically extrapolating from various application domains to deduce potential techniques and algorithms. We distill a general model with adaptive learning and optimization mechanisms by conducting a detailed analysis of human activity types and utilizing contact or non-contact devices. It presents various system integration mathematical paradigms driven by multimodal data fusion, covering predictions of complex behaviors and redefining valuable methods, devices, and systems for HAR. Additionally, this paper establishes benchmarks for behavior recognition across different application requirements, from simple localized actions to group activities. It summarizes open research directions, including data diversity and volume, computational limitations, interoperability, real-time recognition, data security, and privacy concerns. Finally, we aim to serve as a comprehensive and foundational resource for researchers delving into the complex and burgeoning realm of AIoT-enhanced HAR, providing insights and guidance for future innovations and developments.
Collapse
|
4
|
da Costa Nascimento JJ, Marques AG, do Nascimento Souza L, de Mattos Dourado Junior CMJ, da Silva Barros AC, de Albuquerque VHC, de Freitas Sousa LF. A novel generative model for brain tumor detection using magnetic resonance imaging. Comput Med Imaging Graph 2025; 121:102498. [PMID: 39985841 DOI: 10.1016/j.compmedimag.2025.102498] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/26/2024] [Revised: 12/27/2024] [Accepted: 01/22/2025] [Indexed: 02/24/2025]
Abstract
Brain tumors are a disease that kills thousands of people worldwide each year. Early identification through diagnosis is essential for monitoring and treating patients. The proposed study brings a new method through intelligent computational cells that are capable of segmenting the tumor region with high precision. The method uses deep learning to detect brain tumors with the "You only look once" (Yolov8) framework, and a fine-tuning process at the end of the network layer using intelligent computational cells capable of traversing the detected region, segmenting the edges of the brain tumor. In addition, the method uses a classification pipeline that combines a set of classifiers and extractors combined with grid search, to find the best combination and the best parameters for the dataset. The method obtained satisfactory results above 98% accuracy for region detection, and above 99% for brain tumor segmentation and accuracies above 98% for binary classification of brain tumor, and segmentation time obtaining less than 1 s, surpassing the state of the art compared to the same database, demonstrating the effectiveness of the proposed method. The new approach proposes the classification of different databases through data fusion to classify the presence of tumor in MRI images, as well as the patient's life span. The segmentation and classification steps are validated by comparing them with the literature, with comparisons between works that used the same dataset. The method addresses a new generative AI for brain tumor capable of generating a pre-diagnosis through input data through Large Language Model (LLM), and can be used in systems to aid medical imaging diagnosis. As a contribution, this study employs new detection models combined with innovative methods based on digital image processing to improve segmentation metrics, as well as the use of Data Fusion, combining two tumor datasets to enhance classification performance. The study also utilizes LLM models to refine the pre-diagnosis obtained post-classification. Thus, this study proposes a Computer-Aided Diagnosis (CAD) method through AI with PDI, CNN, and LLM.
Collapse
Affiliation(s)
| | - Adriell Gomes Marques
- Instituto Federal de Educação, Ciência e Tecnologia do Ceará - Campus Fortaleza, Fortaleza, 60040-531, CE, Brazil.
| | | | | | | | | | | |
Collapse
|
5
|
Vellmer S, Aydogan DB, Roine T, Cacciola A, Picht T, Fekonja LS. Diffusion MRI GAN synthesizing fibre orientation distribution data using generative adversarial networks. Commun Biol 2025; 8:512. [PMID: 40155540 PMCID: PMC11953217 DOI: 10.1038/s42003-025-07936-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/15/2025] [Accepted: 03/13/2025] [Indexed: 04/01/2025] Open
Abstract
Machine learning may enhance clinical data analysis but requires large amounts of training data, which are scarce for rare pathologies. While generative neural network models can create realistic synthetic data such as 3D MRI volumes and, thus, augment training datasets, the generation of complex data remains challenging. Fibre orientation distributions (FODs) represent one such complex data type, modelling diffusion as spherical harmonics with stored weights as multiple three-dimensional volumes. We successfully trained an α-WGAN combining a generative adversarial network and a variational autoencoder to generate synthetic FODs, using the Human Connectome Project (HCP) data. Our resulting synthetic FODs produce anatomically accurate fibre bundles and connectomes, with properties matching those from our validation dataset. Our approach extends beyond FODs and could be adapted for generating various types of complex medical imaging data, particularly valuable for augmenting limited clinical datasets.
Collapse
Affiliation(s)
- Sebastian Vellmer
- Department of Neurosurgery, Charité Universitätsmedizin Berlin, Berlin, Germany.
- Cluster of Excellence, Matters of Activity, Image Space Material, Berlin, Germany.
| | - Dogu Baran Aydogan
- A.I. Virtanen Institute for Molecular Sciences, University of Eastern Finland, Kuopio, Finland
- Department of Neuroscience and Biomedical Engineering, Aalto University School of Science, Espoo, Finland
| | - Timo Roine
- Department of Neuroscience and Biomedical Engineering, Aalto University School of Science, Espoo, Finland
| | - Alberto Cacciola
- Brain Mapping Lab, Department of Biomedical, Dental Sciences and Morphological and Functional Images, University of Messina, Messina, Italy
| | - Thomas Picht
- Department of Neurosurgery, Charité Universitätsmedizin Berlin, Berlin, Germany
- Cluster of Excellence, Matters of Activity, Image Space Material, Berlin, Germany
| | - Lucius S Fekonja
- Department of Neurosurgery, Charité Universitätsmedizin Berlin, Berlin, Germany.
- Cluster of Excellence, Matters of Activity, Image Space Material, Berlin, Germany.
| |
Collapse
|
6
|
Rujas M, Martín Gómez Del Moral Herranz R, Fico G, Merino-Barbancho B. Synthetic data generation in healthcare: A scoping review of reviews on domains, motivations, and future applications. Int J Med Inform 2025; 195:105763. [PMID: 39719743 DOI: 10.1016/j.ijmedinf.2024.105763] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/19/2024] [Revised: 12/11/2024] [Accepted: 12/14/2024] [Indexed: 12/26/2024]
Abstract
BACKGROUND The development of Artificial Intelligence in the healthcare sector is generating a great impact. However, one of the primary challenges for the implementation of this technology is the access to high-quality data due to issues in data collection and regulatory constraints, for which synthetic data is an emerging alternative. While previous research has reviewed synthetic data generation techniques, there is limited focus on their applications and the motivations driving their synthesis. A comprehensive review is needed to expand the potential of synthetic data into less explored healthcare areas. OBJECTIVE This review aims to identify the healthcare domains where synthetic data are currently generated, the motivations behind their creation, their future uses, limitations, and types of data. MATERIALS AND METHODS Following the PRISMA-ScR framework, this review analysed literature from the last 10 years within PubMed, Scopus, and Web of Science. Reviews containing information on synthetic data generation in healthcare were screened and analysed. Key healthcare domains, motivations, future uses, and gaps in the literature were identified through a structured data extraction process. RESULTS Of the 346 reviews identified, 42 were included for data extraction. Thirteen main domains were identified, with Oncology, Neurology, and Cardiology being the most frequently mentioned. Five primary motivations for synthetic data generation and three major categories of future applications were highlighted. Additionally, unstructured data, particularly images, were found to be the predominant type of synthetic data generated. DISCUSSION AND CONCLUSION Synthetic data are currently being generated across diverse healthcare domains, showcasing their adaptability and potential. Despite their early stage, synthetic data technologies hold significant promise for future applications. Expanding their use into new domains and less common data types (e.g., video and text) could further enhance their impact. Future work should focus on developing evaluation benchmarks and standardized generative models tailored to specific healthcare domains.
Collapse
Affiliation(s)
- Miguel Rujas
- Life Supporting Technologies Research Group, Universidad Politécnica de Madrid, Avda Complutense 30, 28040 Madrid, Spain.
| | | | - Giuseppe Fico
- Life Supporting Technologies Research Group, Universidad Politécnica de Madrid, Avda Complutense 30, 28040 Madrid, Spain
| | - Beatriz Merino-Barbancho
- Life Supporting Technologies Research Group, Universidad Politécnica de Madrid, Avda Complutense 30, 28040 Madrid, Spain
| |
Collapse
|
7
|
Talukder MA, Layek MA, Hossain MA, Islam MA, Nur-e-Alam M, Kazi M. ACU-Net: Attention-based convolutional U-Net model for segmenting brain tumors in fMRI images. Digit Health 2025; 11:20552076251320288. [PMID: 39968528 PMCID: PMC11833834 DOI: 10.1177/20552076251320288] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/20/2025] Open
Abstract
Objective Accurate segmentation of brain tumors in medical imaging is essential for diagnosis and treatment planning. Current techniques often struggle with capturing complex tumor features and are computationally demanding, limiting their clinical application. This study introduces the attention-based convolutional U-Net (ACU-Net) model, designed to improve segmentation accuracy and efficiency in fMRI images by incorporating attention mechanisms that selectively highlight critical features while preserving spatial context. Methods The ACU-Net model combines convolutional neural networks (CNNs) with attention mechanisms to enhance feature extraction and spatial coherence. We evaluated ACU-Net on the BraTS 2018 and BraTS 2020 fMRI datasets using rigorous data splitting for training, validation, and testing. Performance metrics, particularly Dice scores, were used to assess segmentation accuracy across different tumor regions, including whole tumor (WT), tumor core (TC), and enhancing tumor (ET) classes. Results ACU-Net demonstrated high segmentation accuracy, achieving Dice scores of 99.23%, 99.27%, and 96.99% for WT, TC, and ET, respectively, on the BraTS 2018 dataset, and 98.72%, 98.40%, and 97.66% for WT, TC, and ET on the BraTS 2020 dataset. These results indicate that ACU-Net effectively captures tumor boundaries and subregions with precision, surpassing traditional segmentation approaches. Conclusion The ACU-Net model shows significant potential to enhance clinical diagnosis and treatment planning by providing precise and efficient brain tumor segmentation in fMRI images. The integration of attention mechanisms within a CNN architecture proves beneficial for identifying complex tumor structures, suggesting that ACU-Net can be a valuable tool in medical imaging applications.
Collapse
Affiliation(s)
- Md Alamin Talukder
- Department of Computer Science and Engineering, International University of Business Agriculture and Technology, Dhaka, Bangladesh
| | - Md Abu Layek
- Department of Computer Science and Engineering, Jagannath University, Dhaka, Bangladesh
| | - Md Aslam Hossain
- Department of Computer Science and Engineering, Jagannath University, Dhaka, Bangladesh
| | - Md Aminul Islam
- Department of Computer Science and Engineering, Jagannath University, Dhaka, Bangladesh
- Department of Computer Science, Tennessee Technological University, Cookeville, USA
| | - Mohammad Nur-e-Alam
- Department of Pharmacognosy, College of Pharmacy, King Saud University, Riyadh, Saudi Arabia
| | - Mohsin Kazi
- Department of Pharmaceutics, College of Pharmacy, King Saud University, Riyadh, Saudi Arabia
| |
Collapse
|
8
|
Berghout T. The Neural Frontier of Future Medical Imaging: A Review of Deep Learning for Brain Tumor Detection. J Imaging 2024; 11:2. [PMID: 39852315 PMCID: PMC11766058 DOI: 10.3390/jimaging11010002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/05/2024] [Revised: 12/21/2024] [Accepted: 12/23/2024] [Indexed: 01/26/2025] Open
Abstract
Brain tumor detection is crucial in medical research due to high mortality rates and treatment challenges. Early and accurate diagnosis is vital for improving patient outcomes, however, traditional methods, such as manual Magnetic Resonance Imaging (MRI) analysis, are often time-consuming and error-prone. The rise of deep learning has led to advanced models for automated brain tumor feature extraction, segmentation, and classification. Despite these advancements, comprehensive reviews synthesizing recent findings remain scarce. By analyzing over 100 research papers over past half-decade (2019-2024), this review fills that gap, exploring the latest methods and paradigms, summarizing key concepts, challenges, datasets, and offering insights into future directions for brain tumor detection using deep learning. This review also incorporates an analysis of previous reviews and targets three main aspects: feature extraction, segmentation, and classification. The results revealed that research primarily focuses on Convolutional Neural Networks (CNNs) and their variants, with a strong emphasis on transfer learning using pre-trained models. Other methods, such as Generative Adversarial Networks (GANs) and Autoencoders, are used for feature extraction, while Recurrent Neural Networks (RNNs) are employed for time-sequence modeling. Some models integrate with Internet of Things (IoT) frameworks or federated learning for real-time diagnostics and privacy, often paired with optimization algorithms. However, the adoption of eXplainable AI (XAI) remains limited, despite its importance in building trust in medical diagnostics. Finally, this review outlines future opportunities, focusing on image quality, underexplored deep learning techniques, expanding datasets, and exploring deeper learning representations and model behavior such as recurrent expansion to advance medical imaging diagnostics.
Collapse
Affiliation(s)
- Tarek Berghout
- Laboratory of Automation and Manufacturing Engineering, Department of Industrial Engineering, Batna 2 University, Batna 05000, Algeria
| |
Collapse
|
9
|
Muhammad D, Bendechache M. Unveiling the black box: A systematic review of Explainable Artificial Intelligence in medical image analysis. Comput Struct Biotechnol J 2024; 24:542-560. [PMID: 39252818 PMCID: PMC11382209 DOI: 10.1016/j.csbj.2024.08.005] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/05/2024] [Revised: 08/07/2024] [Accepted: 08/07/2024] [Indexed: 09/11/2024] Open
Abstract
This systematic literature review examines state-of-the-art Explainable Artificial Intelligence (XAI) methods applied to medical image analysis, discussing current challenges and future research directions, and exploring evaluation metrics used to assess XAI approaches. With the growing efficiency of Machine Learning (ML) and Deep Learning (DL) in medical applications, there's a critical need for adoption in healthcare. However, their "black-box" nature, where decisions are made without clear explanations, hinders acceptance in clinical settings where decisions have significant medicolegal consequences. Our review highlights the advanced XAI methods, identifying how they address the need for transparency and trust in ML/DL decisions. We also outline the challenges faced by these methods and propose future research directions to improve XAI in healthcare. This paper aims to bridge the gap between cutting-edge computational techniques and their practical application in healthcare, nurturing a more transparent, trustworthy, and effective use of AI in medical settings. The insights guide both research and industry, promoting innovation and standardisation in XAI implementation in healthcare.
Collapse
Affiliation(s)
- Dost Muhammad
- ADAPT Research Centre, School of Computer Science, University of Galway, Galway, Ireland
| | - Malika Bendechache
- ADAPT Research Centre, School of Computer Science, University of Galway, Galway, Ireland
| |
Collapse
|
10
|
Zhuang Y, Liu H, Fang W, Ma G, Sun S, Zhu Y, Zhang X, Ge C, Chen W, Long J, Song E. A 3D hierarchical cross-modality interaction network using transformers and convolutions for brain glioma segmentation in MR images. Med Phys 2024; 51:8371-8389. [PMID: 39137295 DOI: 10.1002/mp.17354] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2024] [Revised: 06/20/2024] [Accepted: 08/02/2024] [Indexed: 08/15/2024] Open
Abstract
BACKGROUND Precise glioma segmentation from multi-parametric magnetic resonance (MR) images is essential for brain glioma diagnosis. However, due to the indistinct boundaries between tumor sub-regions and the heterogeneous appearances of gliomas in volumetric MR scans, designing a reliable and automated glioma segmentation method is still challenging. Although existing 3D Transformer-based or convolution-based segmentation networks have obtained promising results via multi-modal feature fusion strategies or contextual learning methods, they widely lack the capability of hierarchical interactions between different modalities and cannot effectively learn comprehensive feature representations related to all glioma sub-regions. PURPOSE To overcome these problems, in this paper, we propose a 3D hierarchical cross-modality interaction network (HCMINet) using Transformers and convolutions for accurate multi-modal glioma segmentation, which leverages an effective hierarchical cross-modality interaction strategy to sufficiently learn modality-specific and modality-shared knowledge correlated to glioma sub-region segmentation from multi-parametric MR images. METHODS In the HCMINet, we first design a hierarchical cross-modality interaction Transformer (HCMITrans) encoder to hierarchically encode and fuse heterogeneous multi-modal features by Transformer-based intra-modal embeddings and inter-modal interactions in multiple encoding stages, which effectively captures complex cross-modality correlations while modeling global contexts. Then, we collaborate an HCMITrans encoder with a modality-shared convolutional encoder to construct the dual-encoder architecture in the encoding stage, which can learn the abundant contextual information from global and local perspectives. Finally, in the decoding stage, we present a progressive hybrid context fusion (PHCF) decoder to progressively fuse local and global features extracted by the dual-encoder architecture, which utilizes the local-global context fusion (LGCF) module to efficiently alleviate the contextual discrepancy among the decoding features. RESULTS Extensive experiments are conducted on two public and competitive glioma benchmark datasets, including the BraTS2020 dataset with 494 patients and the BraTS2021 dataset with 1251 patients. Results show that our proposed method outperforms existing Transformer-based and CNN-based methods using other multi-modal fusion strategies in our experiments. Specifically, the proposed HCMINet achieves state-of-the-art mean DSC values of 85.33% and 91.09% on the BraTS2020 online validation dataset and the BraTS2021 local testing dataset, respectively. CONCLUSIONS Our proposed method can accurately and automatically segment glioma regions from multi-parametric MR images, which is beneficial for the quantitative analysis of brain gliomas and helpful for reducing the annotation burden of neuroradiologists.
Collapse
Affiliation(s)
- Yuzhou Zhuang
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, China
| | - Hong Liu
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, China
| | - Wei Fang
- Wuhan Zhongke Industrial Research Institute of Medical Science Co., Ltd, Wuhan, China
| | - Guangzhi Ma
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, China
| | - Sisi Sun
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, China
| | - Yunfeng Zhu
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, China
| | - Xu Zhang
- Wuhan United Imaging Healthcare Surgical Technology Co., Ltd, Wuhan, China
| | - Chuanbin Ge
- Wuhan United Imaging Healthcare Surgical Technology Co., Ltd, Wuhan, China
| | - Wenyang Chen
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, China
| | - Jiaosong Long
- School of Art and Design, Hubei University of Technology, Wuhan, China
| | - Enmin Song
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, China
| |
Collapse
|
11
|
Ullah MS, Khan MA, Albarakati HM, Damaševičius R, Alsenan S. Multimodal brain tumor segmentation and classification from MRI scans based on optimized DeepLabV3+ and interpreted networks information fusion empowered with explainable AI. Comput Biol Med 2024; 182:109183. [PMID: 39357134 DOI: 10.1016/j.compbiomed.2024.109183] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/12/2024] [Revised: 09/03/2024] [Accepted: 09/20/2024] [Indexed: 10/04/2024]
Abstract
Explainable artificial intelligence (XAI) aims to offer machine learning (ML) methods that enable people to comprehend, properly trust, and create more explainable models. In medical imaging, XAI has been adopted to interpret deep learning black box models to demonstrate the trustworthiness of machine decisions and predictions. In this work, we proposed a deep learning and explainable AI-based framework for segmenting and classifying brain tumors. The proposed framework consists of two parts. The first part, encoder-decoder-based DeepLabv3+ architecture, is implemented with Bayesian Optimization (BO) based hyperparameter initialization. The different scales are performed, and features are extracted through the Atrous Spatial Pyramid Pooling (ASPP) technique. The extracted features are passed to the output layer for tumor segmentation. In the second part of the proposed framework, two customized models have been proposed named Inverted Residual Bottleneck 96 layers (IRB-96) and Inverted Residual Bottleneck Self-Attention (IRB-Self). Both models are trained on the selected brain tumor datasets and extracted features from the global average pooling and self-attention layers. Features are fused using a serial approach, and classification is performed. The BO-based hyperparameters optimization of the neural network classifiers is performed and the classification results have been optimized. An XAI method named LIME is implemented to check the interpretability of the proposed models. The experimental process of the proposed framework was performed on the Figshare dataset, and an average segmentation accuracy of 92.68 % and classification accuracy of 95.42 % were obtained, respectively. Compared with state-of-the-art techniques, the proposed framework shows improved accuracy.
Collapse
Affiliation(s)
| | - Muhammad Attique Khan
- Department of Artificial Intelligence, College of Computer Engineering and Science, Prince Mohammad Bin Fahd University, Al Khobar, Saudi Arabia.
| | - Hussain Mobarak Albarakati
- Computer and Network Engineering Department, College of Computing, Umm Al-Qura University, Makkah, 24382, Saudi Arabia
| | - Robertas Damaševičius
- Faculty of Applied Mathematics, Silesian University of Technology, 44-100, Gliwice, Poland
| | - Shrooq Alsenan
- Information Systems Department, College of Computer and Information Sciences, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, Riyadh, 11671, Saudi Arabia.
| |
Collapse
|
12
|
Liu S, Wang H, Li S, Zhang C. Mixture-of-experts and semantic-guided network for brain tumor segmentation with missing MRI modalities. Med Biol Eng Comput 2024; 62:3179-3191. [PMID: 38789839 DOI: 10.1007/s11517-024-03130-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/05/2024] [Accepted: 05/14/2024] [Indexed: 05/26/2024]
Abstract
Accurate brain tumor segmentation with multi-modal MRI images is crucial, but missing modalities in clinical practice often reduce accuracy. The aim of this study is to propose a mixture-of-experts and semantic-guided network to tackle the issue of missing modalities in brain tumor segmentation. We introduce a transformer-based encoder with novel mixture-of-experts blocks. In each block, four modality experts aim for modality-specific feature learning. Learnable modality embeddings are employed to alleviate the negative effect of missing modalities. We also introduce a decoder guided by semantic information, designed to pay higher attention to various tumor regions. Finally, we conduct extensive comparison experiments with other models as well as ablation experiments to validate the performance of the proposed model on the BraTS2018 dataset. The proposed model can accurately segment brain tumor sub-regions even with missing modalities. It achieves an average Dice score of 0.81 for the whole tumor, 0.66 for the tumor core, and 0.52 for the enhanced tumor across the 15 modality combinations, achieving top or near-top results in most cases, while also exhibiting a lower computational cost. Our mixture-of-experts and sematic-guided network achieves accurate and reliable brain tumor segmentation results with missing modalities, indicating its significant potential for clinical applications. Our source code is already available at https://github.com/MaggieLSY/MESG-Net .
Collapse
Affiliation(s)
- Siyu Liu
- Digital Medical Research Center, School of Basic Medical Sciences, Fudan University, Shanghai, 200032, China
- Shanghai Key Lab of Medical Image Computing and Computer Assisted Intervention, Shanghai, 200032, China
| | - Haoran Wang
- Digital Medical Research Center, School of Basic Medical Sciences, Fudan University, Shanghai, 200032, China
- Shanghai Key Lab of Medical Image Computing and Computer Assisted Intervention, Shanghai, 200032, China
| | - Shiman Li
- Digital Medical Research Center, School of Basic Medical Sciences, Fudan University, Shanghai, 200032, China
- Shanghai Key Lab of Medical Image Computing and Computer Assisted Intervention, Shanghai, 200032, China
| | - Chenxi Zhang
- Digital Medical Research Center, School of Basic Medical Sciences, Fudan University, Shanghai, 200032, China.
- Shanghai Key Lab of Medical Image Computing and Computer Assisted Intervention, Shanghai, 200032, China.
| |
Collapse
|
13
|
Bian C, Hu C, Cao N. Exploiting K-Space in Magnetic Resonance Imaging Diagnosis: Dual-Path Attention Fusion for K-Space Global and Image Local Features. Bioengineering (Basel) 2024; 11:958. [PMID: 39451334 PMCID: PMC11504126 DOI: 10.3390/bioengineering11100958] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/05/2024] [Revised: 09/04/2024] [Accepted: 09/21/2024] [Indexed: 10/26/2024] Open
Abstract
Magnetic resonance imaging (MRI) diagnosis, enhanced by deep learning methods, plays a crucial role in medical image processing, facilitating precise clinical diagnosis and optimal treatment planning. Current methodologies predominantly focus on feature extraction from the image domain, which often results in the loss of global features during down-sampling processes. However, the unique global representational capacity of MRI K-space is often overlooked. In this paper, we present a novel MRI K-space-based global feature extraction and dual-path attention fusion network. Our proposed method extracts global features from MRI K-space data and fuses them with local features from the image domain using a dual-path attention mechanism, thereby achieving accurate MRI segmentation for diagnosis. Specifically, our method consists of four main components: an image-domain feature extraction module, a K-space domain feature extraction module, a dual-path attention feature fusion module, and a decoder. We conducted ablation studies and comprehensive comparisons on the Brain Tumor Segmentation (BraTS) MRI dataset to validate the effectiveness of each module. The results demonstrate that our method exhibits superior performance in segmentation diagnostics, outperforming state-of-the-art methods with improvements of up to 63.82% in the HD95 distance evaluation metric. Furthermore, we performed generalization testing and complexity analysis on the Automated Cardiac Diagnosis Challenge (ACDC) MRI cardiac segmentation dataset. The findings indicate robust performance across different datasets, highlighting strong generalizability and favorable algorithmic complexity. Collectively, these results suggest that our proposed method holds significant potential for practical clinical applications.
Collapse
Affiliation(s)
- Congchao Bian
- College of Information Science and Engineering, Hohai University, Nanjing 210098, China;
| | - Can Hu
- College of Computer Science and Software Engineering, Hohai University, Nanjing 210098, China;
| | - Ning Cao
- College of Information Science and Engineering, Hohai University, Nanjing 210098, China;
| |
Collapse
|
14
|
Zhou T. M2GCNet: Multi-Modal Graph Convolution Network for Precise Brain Tumor Segmentation Across Multiple MRI Sequences. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2024; 33:4896-4910. [PMID: 39236123 DOI: 10.1109/tip.2024.3451936] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/07/2024]
Abstract
Accurate segmentation of brain tumors across multiple MRI sequences is essential for diagnosis, treatment planning, and clinical decision-making. In this paper, I propose a cutting-edge framework, named multi-modal graph convolution network (M2GCNet), to explore the relationships across different MR modalities, and address the challenge of brain tumor segmentation. The core of M2GCNet is the multi-modal graph convolution module (M2GCM), a pivotal component that represents MR modalities as graphs, with nodes corresponding to image pixels and edges capturing latent relationships between pixels. This graph-based representation enables the effective utilization of both local and global contextual information. Notably, M2GCM comprises two important modules: the spatial-wise graph convolution module (SGCM), adept at capturing extensive spatial dependencies among distinct regions within an image, and the channel-wise graph convolution module (CGCM), dedicated to modelling intricate contextual dependencies among different channels within the image. Additionally, acknowledging the intrinsic correlation present among different MR modalities, a multi-modal correlation loss function is introduced. This novel loss function aims to capture specific nonlinear relationships between correlated modality pairs, enhancing the model's ability to achieve accurate segmentation results. The experimental evaluation on two brain tumor datasets demonstrates the superiority of the proposed M2GCNet over other state-of-the-art segmentation methods. Furthermore, the proposed method paves the way for improved tumor diagnosis, multi-modal information fusion, and a deeper understanding of brain tumor pathology.
Collapse
|
15
|
Wang T, Wen Y, Wang Z. nnU-Net based segmentation and 3D reconstruction of uterine fibroids with MRI images for HIFU surgery planning. BMC Med Imaging 2024; 24:233. [PMID: 39243001 PMCID: PMC11380377 DOI: 10.1186/s12880-024-01385-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/18/2024] [Accepted: 08/01/2024] [Indexed: 09/09/2024] Open
Abstract
High-Intensity Focused Ultrasound (HIFU) ablation represents a rapidly advancing non-invasive treatment modality that has achieved considerable success in addressing uterine fibroids, which constitute over 50% of benign gynecological tumors. Preoperative Magnetic Resonance Imaging (MRI) plays a pivotal role in the planning and guidance of HIFU surgery for uterine fibroids, wherein the segmentation of tumors holds critical significance. The segmentation process was previously manually executed by medical experts, entailing a time-consuming and labor-intensive procedure heavily reliant on clinical expertise. This study introduced deep learning-based nnU-Net models, offering a cost-effective approach for their application in the segmentation of uterine fibroids utilizing preoperative MRI images. Furthermore, 3D reconstruction of the segmented targets was implemented to guide HIFU surgery. The evaluation of segmentation and 3D reconstruction performance was conducted with a focus on enhancing the safety and effectiveness of HIFU surgery. Results demonstrated the nnU-Net's commendable performance in the segmentation of uterine fibroids and their surrounding organs. Specifically, 3D nnU-Net achieved Dice Similarity Coefficients (DSC) of 92.55% for the uterus, 95.63% for fibroids, 92.69% for the spine, 89.63% for the endometrium, 97.75% for the bladder, and 90.45% for the urethral orifice. Compared to other state-of-the-art methods such as HIFUNet, U-Net, R2U-Net, ConvUNeXt and 2D nnU-Net, 3D nnU-Net demonstrated significantly higher DSC values, highlighting its superior accuracy and robustness. In conclusion, the efficacy of the 3D nnU-Net model for automated segmentation of the uterus and its surrounding organs was robustly validated. When integrated with intra-operative ultrasound imaging, this segmentation method and 3D reconstruction hold substantial potential to enhance the safety and efficiency of HIFU surgery in the clinical treatment of uterine fibroids.
Collapse
Affiliation(s)
- Ting Wang
- State Key Laboratory of Ultrasound in Medicine and Engineering, College of Biomedical Engineering, Chongqing Medical University, Chongqing, 400016, China
| | - Yingang Wen
- National Engineering Research Center of Ultrasonic Medicine, Chongqing, 401121, China
| | - Zhibiao Wang
- State Key Laboratory of Ultrasound in Medicine and Engineering, College of Biomedical Engineering, Chongqing Medical University, Chongqing, 400016, China.
| |
Collapse
|
16
|
Chen Z, Liang N, Li H, Zhang H, Li H, Yan L, Hu Z, Chen Y, Zhang Y, Wang Y, Ke D, Shi N. Exploring explainable AI features in the vocal biomarkers of lung disease. Comput Biol Med 2024; 179:108844. [PMID: 38981214 DOI: 10.1016/j.compbiomed.2024.108844] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/02/2024] [Revised: 05/15/2024] [Accepted: 06/04/2024] [Indexed: 07/11/2024]
Abstract
This review delves into the burgeoning field of explainable artificial intelligence (XAI) in the detection and analysis of lung diseases through vocal biomarkers. Lung diseases, often elusive in their early stages, pose a significant public health challenge. Recent advancements in AI have ushered in innovative methods for early detection, yet the black-box nature of many AI models limits their clinical applicability. XAI emerges as a pivotal tool, enhancing transparency and interpretability in AI-driven diagnostics. This review synthesizes current research on the application of XAI in analyzing vocal biomarkers for lung diseases, highlighting how these techniques elucidate the connections between specific vocal features and lung pathology. We critically examine the methodologies employed, the types of lung diseases studied, and the performance of various XAI models. The potential for XAI to aid in early detection, monitor disease progression, and personalize treatment strategies in pulmonary medicine is emphasized. Furthermore, this review identifies current challenges, including data heterogeneity and model generalizability, and proposes future directions for research. By offering a comprehensive analysis of explainable AI features in the context of lung disease detection, this review aims to bridge the gap between advanced computational approaches and clinical practice, paving the way for more transparent, reliable, and effective diagnostic tools.
Collapse
Affiliation(s)
- Zhao Chen
- Institute of Basic Research in Clinical Medicine, China Academy of Chinese Medical Sciences, Beijing, China
| | - Ning Liang
- Institute of Basic Research in Clinical Medicine, China Academy of Chinese Medical Sciences, Beijing, China
| | - Haoyuan Li
- Institute of Basic Research in Clinical Medicine, China Academy of Chinese Medical Sciences, Beijing, China
| | - Haili Zhang
- Institute of Basic Research in Clinical Medicine, China Academy of Chinese Medical Sciences, Beijing, China
| | - Huizhen Li
- Institute of Basic Research in Clinical Medicine, China Academy of Chinese Medical Sciences, Beijing, China
| | - Lijiao Yan
- Institute of Basic Research in Clinical Medicine, China Academy of Chinese Medical Sciences, Beijing, China
| | - Ziteng Hu
- Institute of Basic Research in Clinical Medicine, China Academy of Chinese Medical Sciences, Beijing, China
| | - Yaxin Chen
- Institute of Basic Research in Clinical Medicine, China Academy of Chinese Medical Sciences, Beijing, China
| | - Yujing Zhang
- Institute of Basic Research in Clinical Medicine, China Academy of Chinese Medical Sciences, Beijing, China
| | - Yanping Wang
- Institute of Basic Research in Clinical Medicine, China Academy of Chinese Medical Sciences, Beijing, China
| | - Dandan Ke
- Special Disease Clinic, Huaishuling Branch of Beijing Fengtai Hospital of Integrated Traditional Chinese and Western Medicine, Beijing, China.
| | - Nannan Shi
- State Key Laboratory of Cognitive Neuroscience and Learning, Beijing Normal University, Beijing, China; Institute of Basic Research in Clinical Medicine, China Academy of Chinese Medical Sciences, Beijing, China.
| |
Collapse
|
17
|
Huang L, Zhang N, Yi Y, Zhou W, Zhou B, Dai J, Wang J. SAMCF: Adaptive global style alignment and multi-color spaces fusion for joint optic cup and disc segmentation. Comput Biol Med 2024; 178:108639. [PMID: 38878394 DOI: 10.1016/j.compbiomed.2024.108639] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/26/2024] [Revised: 04/21/2024] [Accepted: 05/18/2024] [Indexed: 07/24/2024]
Abstract
The optic cup (OC) and optic disc (OD) are two critical structures in retinal fundus images, and their relative positions and sizes are essential for effectively diagnosing eye diseases. With the success of deep learning in computer vision, deep learning-based segmentation models have been widely used for joint optic cup and disc segmentation. However, there are three prominent issues that impact the segmentation performance. First, significant differences among datasets collecting from various institutions, protocols, and devices lead to performance degradation of models. Second, we find that images with only RGB information struggle to counteract the interference caused by brightness variations, affecting color representation capability. Finally, existing methods typically ignored the edge perception, facing the challenges in obtaining clear and smooth edge segmentation results. To address these drawbacks, we propose a novel framework based on Style Alignment and Multi-Color Fusion (SAMCF) for joint OC and OD segmentation. Initially, we introduce a domain generalization method to generate uniformly styled images without damaged image content for mitigating domain shift issues. Next, based on multiple color spaces, we propose a feature extraction and fusion network aiming to handle brightness variation interference and improve color representation capability. Lastly, an edge aware loss is designed to generate fine edge segmentation results. Our experiments conducted on three public datasets, DGS, RIM, and REFUGE, demonstrate that our proposed SAMCF achieves superior performance to existing state-of-the-art methods. Moreover, SAMCF exhibits remarkable generalization ability across multiple retinal fundus image datasets, showcasing its outstanding generality.
Collapse
Affiliation(s)
- Longjun Huang
- School of Software, Nanchang Key Laboratory for Blindness and Visual Impairment Prevention Technology and Equipment, Jiangxi Normal University, Nanchang, 330022, China
| | - Ningyi Zhang
- School of Software, Nanchang Key Laboratory for Blindness and Visual Impairment Prevention Technology and Equipment, Jiangxi Normal University, Nanchang, 330022, China
| | - Yugen Yi
- School of Software, Nanchang Key Laboratory for Blindness and Visual Impairment Prevention Technology and Equipment, Jiangxi Normal University, Nanchang, 330022, China.
| | - Wei Zhou
- College of Computer Science, Shenyang Aerospace University, Shenyang, 110136, China
| | - Bin Zhou
- School of Software, Nanchang Key Laboratory for Blindness and Visual Impairment Prevention Technology and Equipment, Jiangxi Normal University, Nanchang, 330022, China
| | - Jiangyan Dai
- School of Computer Engineering, Weifang University, 261061, China.
| | - Jianzhong Wang
- College of Information Science and Technology, Northeast Normal University, Changchun, 130117, China
| |
Collapse
|
18
|
Teng S, Wang B, Yang F, Yi X, Zhang X, Sun Y. MediDRNet: Tackling category imbalance in diabetic retinopathy classification with dual-branch learning and prototypical contrastive learning. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 253:108230. [PMID: 38810377 DOI: 10.1016/j.cmpb.2024.108230] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/09/2024] [Revised: 04/17/2024] [Accepted: 05/14/2024] [Indexed: 05/31/2024]
Abstract
BACKGROUND AND OBJECTIVE The classification of diabetic retinopathy (DR) aims to utilize the implicit information in images for early diagnosis, to prevent and mitigate the further worsening of the condition. However, existing methods are often limited by the need to operate within large, annotated datasets to show significant advantages. Additionally, the number of samples for different categories within the dataset needs to be evenly distributed, because the characteristic of sample imbalance distribution can lead to an excessive focus on high-frequency disease categories, while neglecting the less common but equally important disease categories. Therefore, there is an urgent need to develop a new classification method that can effectively alleviate the issue of sample distribution imbalance, thereby enhancing the accuracy of diabetic retinopathy classification. METHODS In this work, we propose MediDRNet, a dual-branch network model based on prototypical contrastive learning. This model adopts prototype contrastive learning, creating prototypes for different levels of lesions, ensuring they represent the core features of each lesion level. It classifies by comparing the similarity between data points and their category prototypes. Our dual-branch network structure effectively resolves the issue of category imbalance and improves classification accuracy by emphasizing subtle differences in retinal lesions. Moreover, our approach combines a dual-branch network with specific lesion-level prototypes for core feature representation and incorporates the convolutional block attention module for enhanced lesion feature identification. RESULTS Our experiments using both the Kaggle and UWF classification datasets have demonstrated that MediDRNet exhibits exceptional performance compared to other advanced models in the industry, especially on the UWF DR classification dataset where it achieved state-of-the-art performance across all metrics. On the Kaggle DR classification dataset, it achieved the highest average classification accuracy (0.6327) and Macro-F1 score (0.6361). Particularly in the classification tasks for minority categories of diabetic retinopathy on the Kaggle dataset (Grades 1, 2, 3, and 4), the model reached high classification accuracies of 58.08%, 55.32%, 69.73%, and 90.21%, respectively. In the ablation study, the MediDRNet model proved to be more effective in feature extraction from diabetic retinal fundus images compared to other feature extraction methods. CONCLUSIONS This study employed prototype contrastive learning and bidirectional branch learning strategies, successfully constructing a grading system for diabetic retinopathy lesions within imbalanced diabetic retinopathy datasets. Through a dual-branch network, the feature learning branch effectively facilitated a smooth transition of features from the grading network to the classification learning branch, accurately identifying minority sample categories. This method not only effectively resolved the issue of sample imbalance but also provided strong support for the precise grading and early diagnosis of diabetic retinopathy in clinical applications, showcasing exceptional performance in handling complex diabetic retinopathy datasets. Moreover, this research significantly improved the efficiency of prevention and management of disease progression in diabetic retinopathy patients within medical practice. We encourage the use and modification of our code, which is publicly accessible on GitHub: https://github.com/ReinforceLove/MediDRNet.
Collapse
Affiliation(s)
- Siying Teng
- Department of Ophthalmology, the First Hospital of Jilin University, Changchun, 130021, Jilin, China
| | - Bo Wang
- University of Minho, Braga, 4710-057, Braga District, Portugal
| | - Feiyang Yang
- College of Computer Science and Technology, Jilin University, Changchun, 130012, Jilin, China
| | - Xingcheng Yi
- Laboratory of Cancer Precision Medicine, the First Hospital of Jilin University, Changchun, 130013, Jilin, China
| | - Xinmin Zhang
- Department of Regenerative Medicine, School of Pharmaceutical Science, Jilin University, Changchun, 130021, Jilin, China
| | - Yabin Sun
- Department of Ophthalmology, the First Hospital of Jilin University, Changchun, 130021, Jilin, China.
| |
Collapse
|
19
|
Abidin ZU, Naqvi RA, Haider A, Kim HS, Jeong D, Lee SW. Recent deep learning-based brain tumor segmentation models using multi-modality magnetic resonance imaging: a prospective survey. Front Bioeng Biotechnol 2024; 12:1392807. [PMID: 39104626 PMCID: PMC11298476 DOI: 10.3389/fbioe.2024.1392807] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2024] [Accepted: 06/14/2024] [Indexed: 08/07/2024] Open
Abstract
Radiologists encounter significant challenges when segmenting and determining brain tumors in patients because this information assists in treatment planning. The utilization of artificial intelligence (AI), especially deep learning (DL), has emerged as a useful tool in healthcare, aiding radiologists in their diagnostic processes. This empowers radiologists to understand the biology of tumors better and provide personalized care to patients with brain tumors. The segmentation of brain tumors using multi-modal magnetic resonance imaging (MRI) images has received considerable attention. In this survey, we first discuss multi-modal and available magnetic resonance imaging modalities and their properties. Subsequently, we discuss the most recent DL-based models for brain tumor segmentation using multi-modal MRI. We divide this section into three parts based on the architecture: the first is for models that use the backbone of convolutional neural networks (CNN), the second is for vision transformer-based models, and the third is for hybrid models that use both convolutional neural networks and transformer in the architecture. In addition, in-depth statistical analysis is performed of the recent publication, frequently used datasets, and evaluation metrics for segmentation tasks. Finally, open research challenges are identified and suggested promising future directions for brain tumor segmentation to improve diagnostic accuracy and treatment outcomes for patients with brain tumors. This aligns with public health goals to use health technologies for better healthcare delivery and population health management.
Collapse
Affiliation(s)
- Zain Ul Abidin
- Department of Intelligent Mechatronics Engineering, Sejong University, Seoul, Republic of Korea
| | - Rizwan Ali Naqvi
- Department of Intelligent Mechatronics Engineering, Sejong University, Seoul, Republic of Korea
| | - Amir Haider
- Department of Intelligent Mechatronics Engineering, Sejong University, Seoul, Republic of Korea
| | - Hyung Seok Kim
- Department of Intelligent Mechatronics Engineering, Sejong University, Seoul, Republic of Korea
| | - Daesik Jeong
- College of Convergence Engineering, Sangmyung University, Seoul, Republic of Korea
| | - Seung Won Lee
- School of Medicine, Sungkyunkwan University, Suwon, Republic of Korea
| |
Collapse
|
20
|
Yu H, Yang Z, Zhang Z, Wang T, Ran M, Wang Z, Liu L, Liu Y, Zhang Y. Multiple organ segmentation framework for brain metastasis radiotherapy. Comput Biol Med 2024; 177:108637. [PMID: 38824789 DOI: 10.1016/j.compbiomed.2024.108637] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/20/2024] [Revised: 04/24/2024] [Accepted: 05/18/2024] [Indexed: 06/04/2024]
Abstract
Radiotherapy is a preferred treatment for brain metastases, which kills cancer cells via high doses of radiation meanwhile hardly avoiding damage to surrounding healthy cells. Therefore, the delineation of organs-at-risk (OARs) is vital in treatment planning to minimize radiation-induced toxicity. However, the following aspects make OAR delineation a challenging task: extremely imbalanced organ sizes, ambiguous boundaries, and complex anatomical structures. To alleviate these challenges, we imitate how specialized clinicians delineate OARs and present a novel cascaded multi-OAR segmentation framework, called OAR-SegNet. OAR-SegNet comprises two distinct levels of segmentation networks: an Anatomical-Prior-Guided network (APG-Net) and a Point-Cloud-Guided network (PCG-Net). Specifically, APG-Net handles segmentation for all organs, where multi-view segmentation modules and a deep prior loss are designed under the guidance of prior knowledge. After APG-Net, PCG-Net refines small organs through the mini-segmentation and the point-cloud alignment heads. The mini-segmentation head is further equipped with the deep prior feature. Extensive experiments were conducted to demonstrate the superior performance of the proposed method compared to other state-of-the-art medical segmentation methods.
Collapse
Affiliation(s)
- Hui Yu
- College of Computer Science, Sichuan University, China
| | - Ziyuan Yang
- College of Computer Science, Sichuan University, China
| | | | - Tao Wang
- College of Computer Science, Sichuan University, China
| | - Maoson Ran
- College of Computer Science, Sichuan University, China
| | - Zhiwen Wang
- College of Computer Science, Sichuan University, China
| | - Lunxin Liu
- Department of Neurosurgery, West China Hospital of Sichuan University, China
| | - Yan Liu
- College of Electrical Engineering, Sichuan University, China.
| | - Yi Zhang
- School of Cyber Science and Engineering, Sichuan University, China
| |
Collapse
|
21
|
Xuan P, Chu X, Cui H, Nakaguchi T, Wang L, Ning Z, Ning Z, Li C, Zhang T. Multi-view attribute learning and context relationship encoding enhanced segmentation of lung tumors from CT images. Comput Biol Med 2024; 177:108640. [PMID: 38833798 DOI: 10.1016/j.compbiomed.2024.108640] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/17/2023] [Revised: 04/25/2024] [Accepted: 05/18/2024] [Indexed: 06/06/2024]
Abstract
Graph convolutional neural networks (GCN) have shown the promise in medical image segmentation due to the flexibility of representing diverse range of image regions using graph nodes and propagating knowledge via graph edges. However, existing methods did not fully exploit the various attributes of image nodes and the context relationship among their attributes. We propose a new segmentation method with multi-similarity view enhancement and node attribute context learning (MNSeg). First, multiple views were formed by measuring the similarities among the image nodes, and MNSeg has a GCN based multi-view image node attribute learning (MAL) module to integrate various node attributes learnt from multiple similarity views. Each similarity view contains the specific similarities among all the image nodes, and it was integrated with the node attributes from all the channels to form the enhanced attributes of image nodes. Second, the context relationships among the attributes of image nodes are formulated by a transformer-based context relationship encoding (CRE) strategy to propagate these relationships across all the image nodes. During the transformer-based learning, the relationships were estimated based on the self-attention on all the image nodes, and then they were encoded into the learned node features. Finally, we design an attention at attribute category level (ACA) to discriminate and fuse the learnt diverse information from MAL, CRE, and the original node attributes. ACA identifies the more informative attribute categories by adaptively learn their importance. We validate the performance of MNSeg on a public lung tumor CT dataset and an in-house non-small cell lung cancer (NSCLC) dataset collected from the hospital. The segmentation results show that MNSeg outperformed the compared segmentation methods in terms of spatial overlap and the shape similarities. The ablation studies demonstrated the effectiveness of MAL, CRE, and ACA. The generalization ability of MNSeg was proved by the consistent improved segmentation performances using different 3D segmentation backbones.
Collapse
Affiliation(s)
- Ping Xuan
- Department of Computer Science and Technology, Shantou University, Shantou, China; School of Computer Science and Technology, Heilongjiang University, Harbin, China
| | - Xiuqiang Chu
- School of Computer Science and Technology, Heilongjiang University, Harbin, China
| | - Hui Cui
- Department of Computer Science and Information Technology, La Trobe University, Melbourne, Australia
| | - Toshiya Nakaguchi
- Center for Frontier Medical Engineering, Chiba University, Chiba, Japan
| | - Linlin Wang
- Department of Radiation Oncology, Shandong First Medical University and Shandong Academy of Medical Sciences, Jinan, China
| | - Zhiyuan Ning
- School of Electrical and Information Engineering, The University of Sydney, Sydney, Australia
| | - Zhiyu Ning
- School of Electrical and Information Engineering, The University of Sydney, Sydney, Australia
| | | | - Tiangang Zhang
- School of Computer Science and Technology, Heilongjiang University, Harbin, China; School of Mathematical Science, Heilongjiang University, Harbin, China.
| |
Collapse
|
22
|
Al-Kadi OS, Al-Emaryeen R, Al-Nahhas S, Almallahi I, Braik R, Mahafza W. Empowering brain cancer diagnosis: harnessing artificial intelligence for advanced imaging insights. Rev Neurosci 2024; 35:399-419. [PMID: 38291768 DOI: 10.1515/revneuro-2023-0115] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/19/2023] [Accepted: 12/10/2023] [Indexed: 02/01/2024]
Abstract
Artificial intelligence (AI) is increasingly being used in the medical field, specifically for brain cancer imaging. In this review, we explore how AI-powered medical imaging can impact the diagnosis, prognosis, and treatment of brain cancer. We discuss various AI techniques, including deep learning and causality learning, and their relevance. Additionally, we examine current applications that provide practical solutions for detecting, classifying, segmenting, and registering brain tumors. Although challenges such as data quality, availability, interpretability, transparency, and ethics persist, we emphasise the enormous potential of intelligent applications in standardising procedures and enhancing personalised treatment, leading to improved patient outcomes. Innovative AI solutions have the power to revolutionise neuro-oncology by enhancing the quality of routine clinical practice.
Collapse
Affiliation(s)
- Omar S Al-Kadi
- King Abdullah II School for Information Technology, University of Jordan, Amman, 11942, Jordan
| | - Roa'a Al-Emaryeen
- King Abdullah II School for Information Technology, University of Jordan, Amman, 11942, Jordan
| | - Sara Al-Nahhas
- King Abdullah II School for Information Technology, University of Jordan, Amman, 11942, Jordan
| | - Isra'a Almallahi
- Department of Diagnostic Radiology, Jordan University Hospital, Amman, 11942, Jordan
| | - Ruba Braik
- Department of Diagnostic Radiology, Jordan University Hospital, Amman, 11942, Jordan
| | - Waleed Mahafza
- Department of Diagnostic Radiology, Jordan University Hospital, Amman, 11942, Jordan
| |
Collapse
|
23
|
Mokhtarpour K, Akbarzadehmoallemkolaei M, Rezaei N. A viral attack on brain tumors: the potential of oncolytic virus therapy. J Neurovirol 2024; 30:229-250. [PMID: 38806994 DOI: 10.1007/s13365-024-01209-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2023] [Revised: 04/17/2024] [Accepted: 04/18/2024] [Indexed: 05/30/2024]
Abstract
Managing malignant brain tumors remains a significant therapeutic hurdle that necessitates further research to comprehend their treatment potential fully. Oncolytic viruses (OVs) offer many opportunities for predicting and combating tumors through several mechanisms, with both preclinical and clinical studies demonstrating potential. OV therapy has emerged as a potent and effective method with a dual mechanism. Developing innovative and effective strategies for virus transduction, coupled with immune checkpoint inhibitors or chemotherapy drugs, strengthens this new technique. Furthermore, the discovery and creation of new OVs that can seamlessly integrate gene therapy strategies, such as cytotoxic, anti-angiogenic, and immunostimulatory, are promising advancements. This review presents an overview of the latest advancements in OVs transduction for brain cancer, focusing on the safety and effectiveness of G207, G47Δ, M032, rQNestin34.5v.2, C134, DNX-2401, Ad-TD-nsIL12, NSC-CRAd-S-p7, TG6002, and PVSRIPO. These are evaluated in both preclinical and clinical models of various brain tumors.
Collapse
Affiliation(s)
- Kasra Mokhtarpour
- Animal Model Integrated Network (AMIN), Universal Scientific Education and Research Network (USERN), Tehran, 1419733151, Iran
| | - Milad Akbarzadehmoallemkolaei
- Animal Model Integrated Network (AMIN), Universal Scientific Education and Research Network (USERN), Tehran, 1419733151, Iran
- Research Center for Immunodeficiencies, Children's Medical Center, Tehran University of Medical Sciences, Dr. Gharib St, Keshavarz Blvd, Tehran, 1419733151, Iran
| | - Nima Rezaei
- Animal Model Integrated Network (AMIN), Universal Scientific Education and Research Network (USERN), Tehran, 1419733151, Iran.
- Research Center for Immunodeficiencies, Children's Medical Center, Tehran University of Medical Sciences, Dr. Gharib St, Keshavarz Blvd, Tehran, 1419733151, Iran.
- Department of Immunology, School of Medicine, Tehran University of Medical Sciences, Tehran, 1417653761, Iran.
| |
Collapse
|
24
|
Peng Y, Yi X, Zhang D, Zhang L, Tian Y, Zhou Z. ConvMedSegNet: A multi-receptive field depthwise convolutional neural network for medical image segmentation. Comput Biol Med 2024; 176:108559. [PMID: 38759586 DOI: 10.1016/j.compbiomed.2024.108559] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/21/2024] [Revised: 04/11/2024] [Accepted: 05/05/2024] [Indexed: 05/19/2024]
Abstract
In order to achieve highly precise medical image segmentation, this paper presents ConvMedSegNet, a novel convolutional neural network designed with a U-shaped architecture that seamlessly integrates two crucial modules: the multi-receptive field depthwise convolution module (MRDC) and the guided fusion module (GF). The MRDC module's primary function is to capture texture information of varying sizes through multi-scale convolutional layers. This information is subsequently utilized to enhance the correlation of global feature data by expanding the network's width. This strategy adeptly preserves the inherent inductive biases of convolution while concurrently amplifying the network's ability to establish dependencies on global information. Conversely, the GF module assumes responsibility for implementing multi-scale feature fusion by connecting the encoder and decoder components. It facilitates the transfer of information between features that are separated over substantial distances through guided fusion, effectively minimizing the loss of critical data. In experiments conducted on public medical image datasets such as BUSI and ISIC2018, ConvMedSegNet outperforms several advanced competing methods, yielding superior results. Additionally, the code can be accessed at https://github.com/csust-yixin/ConvMedSegNet.
Collapse
Affiliation(s)
- Yuxu Peng
- School of Computer and Communication Engineering, Changsha University of Science and Technology, Changsha, 410114, China
| | - Xin Yi
- School of Computer and Communication Engineering, Changsha University of Science and Technology, Changsha, 410114, China
| | - Dengyong Zhang
- School of Computer and Communication Engineering, Changsha University of Science and Technology, Changsha, 410114, China
| | | | - Yuehong Tian
- Changkuangao Beijing Technology Co., Ltd, Beijing 101100, China
| | - Zhifeng Zhou
- Wenzhou University Library, Wenzhou, 325035, China.
| |
Collapse
|
25
|
Chen B, Sun Q, Han Y, Liu B, Zhang J, Zhang Q. Adaptive cascaded transformer U-Net for MRI brain tumor segmentation. Phys Med Biol 2024; 69:115036. [PMID: 38636503 DOI: 10.1088/1361-6560/ad4081] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2023] [Accepted: 04/18/2024] [Indexed: 04/20/2024]
Abstract
Objective.Brain tumor segmentation on magnetic resonance imaging (MRI) plays an important role in assisting the diagnosis and treatment of cancer patients. Recently, cascaded U-Net models have achieved excellent performance via conducting coarse-to-fine segmentation of MRI brain tumors. However, they are still restricted by obvious global and local differences among various brain tumors, which are difficult to solve with conventional convolutions.Approach.To address the issue, this study proposes a novel Adaptive Cascaded Transformer U-Net (ACTransU-Net) for MRI brain tumor segmentation, which simultaneously integrates Transformer and dynamic convolution into a single cascaded U-Net architecture to adaptively capture global information and local details of brain tumors. ACTransU-Net first cascades two 3D U-Nets into a two-stage network to segment brain tumors from coarse to fine. Subsequently, it integrates omni-dimensional dynamic convolution modules into the second-stage shallow encoder and decoder, thereby enhancing the local detail representation of various brain tumors through dynamically adjusting convolution kernel parameters. Moreover, 3D Swin-Transformer modules are introduced into the second-stage deep encoder and decoder to capture image long-range dependencies, which helps adapt the global representation of brain tumors.Main results.Extensive experiment results evaluated on the public BraTS 2020 and BraTS 2021 brain tumor data sets demonstrate the effectiveness of ACTransU-Net, with average DSC of 84.96% and 91.37%, and HD95 of 10.81 and 7.31 mm, proving competitiveness with the state-of-the-art methods.Significance.The proposed method focuses on adaptively capturing both global information and local details of brain tumors, aiding physicians in their accurate diagnosis. In addition, it has the potential to extend ACTransU-Net for segmenting other types of lesions. The source code is available at:https://github.com/chenbn266/ACTransUnet.
Collapse
Affiliation(s)
- Bonian Chen
- School of Computer Science and Engineering, Dalian Minzu University, Dalian, 116600, People's Republic of China
| | - Qiule Sun
- Faculty of Electronic and Information Engineering, Dalian University of Technology, Dalian, 116024, People's Republic of China
| | - Yutong Han
- School of Computer Science and Engineering, Dalian Minzu University, Dalian, 116600, People's Republic of China
| | - Bin Liu
- International School of Information Science and Engineering (DUT-RUISE), Dalian University of Technology, Dalian, 116620, People's Republic of China
| | - Jianxin Zhang
- School of Computer Science and Engineering, Dalian Minzu University, Dalian, 116600, People's Republic of China
| | - Qiang Zhang
- Faculty of Electronic and Information Engineering, Dalian University of Technology, Dalian, 116024, People's Republic of China
| |
Collapse
|
26
|
Delgado-López PD, Montalvo-Afonso A, Martín-Alonso J, Martín-Velasco V, Diana-Martín R, Castilla-Díez JM. Predicting histological grade in symptomatic meningioma by an objective estimation of the tumoral surface irregularity. NEUROCIRUGIA (ENGLISH EDITION) 2024; 35:113-121. [PMID: 38244923 DOI: 10.1016/j.neucie.2023.10.002] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/01/2023] [Accepted: 10/03/2023] [Indexed: 01/22/2024]
Abstract
INTRODUCTION Predicting the histopathologic grade of meningioma is relevant because local recurrence is significantly greater in WHO grade II-III compared to WHO grade I tumours, which would ideally benefit from a more aggressive surgical strategy. It has been suggested that higher WHO grade tumours are more irregularly-shaped. However, irregularity is a subjective and observer-dependent feature. In this study, the tumour surface irregularity of a large series of meningiomas, measured upon preoperative MRI, is quantified and correlated with the WHO grade. METHODS Unicentric retrospective observational study of a cohort of symptomatic meningiomas surgically removed in the time period between January 2015 and December 2022. Using specific segmentation software, the Surface Factor (SF) was calculated for each meningioma. SF is an objective parameter that compares the surface of a sphere (minimum surface area for a given volume) with the same volume of the tumour against the actual surface of the tumour. This ratio varies from 0 to 1, being 1 the maximum sphericity. Since irregularly-shaped meningiomas present proportionally greater surface area, the SF tends to decrease as irregularity increases. SF was correlated with WHO grade and its predictive power was estimated with ROC curve analysis. RESULTS A total of 176 patients (64.7% females) were included in the study; 120 WHO grade I (71.9%), 43 WHO grade II (25.7%) and 4 WHO grade III (2.4%). A statistically significant difference was found between the mean SF of WHO grade I and WHO grade II-III tumours (0.8651 ± 0.049 versus 0.7081 ± 0.105, p < 0.0001). Globally, the SF correctly classified more than 90% of cases (area under ROC curve 0.940) with 93.3% sensibility and 80.9% specificity. A cutoff value of 0.79 yielded the maximum precision, with positive and negative predictive powers of 82.6% and 92.6%, respectively. Multivariate analysis yielded SF as an independent prognostic factor of WHO grade. CONCLUSION The Surface Factor is an objective and quantitative parameter that helps to identify aggressive meningiomas preoperatively. A cutoff value of 0.79 allowed differentiation between WHO grade I and WHO grade II-III with high precision.
Collapse
Affiliation(s)
| | | | | | | | - Rubén Diana-Martín
- Servicio de Neurocirugía, Hospital Universitario de Burgos, Burgos, Spain
| | | |
Collapse
|
27
|
Kang N, Wang M, Pang C, Lan R, Li B, Guan J, Wang H. Cross-patch feature interactive net with edge refinement for retinal vessel segmentation. Comput Biol Med 2024; 174:108443. [PMID: 38608328 DOI: 10.1016/j.compbiomed.2024.108443] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/03/2024] [Revised: 03/14/2024] [Accepted: 04/07/2024] [Indexed: 04/14/2024]
Abstract
Retinal vessel segmentation based on deep learning is an important auxiliary method for assisting clinical doctors in diagnosing retinal diseases. However, existing methods often produce mis-segmentation when dealing with low contrast images and thin blood vessels, which affects the continuity and integrity of the vessel skeleton. In addition, existing deep learning methods tend to lose a lot of detailed information during training, which affects the accuracy of segmentation. To address these issues, we propose a novel dual-decoder based Cross-patch Feature Interactive Net with Edge Refinement (CFI-Net) for end-to-end retinal vessel segmentation. In the encoder part, a joint refinement down-sampling method (JRDM) is proposed to compress feature information in the process of reducing image size, so as to reduce the loss of thin vessels and vessel edge information during the encoding process. In the decoder part, we adopt a dual-path model based on edge detection, and propose a Cross-patch Interactive Attention Mechanism (CIAM) in the main path to enhancing multi-scale spatial channel features and transferring cross-spatial information. Consequently, it improve the network's ability to segment complete and continuous vessel skeletons, reducing vessel segmentation fractures. Finally, the Adaptive Spatial Context Guide Method (ASCGM) is proposed to fuse the prediction results of the two decoder paths, which enhances segmentation details while removing part of the background noise. We evaluated our model on two retinal image datasets and one coronary angiography dataset, achieving outstanding performance in segmentation comprehensive assessment metrics such as AUC and CAL. The experimental results showed that the proposed CFI-Net has superior segmentation performance compared with other existing methods, especially for thin vessels and vessel edges. The code is available at https://github.com/kita0420/CFI-Net.
Collapse
Affiliation(s)
- Ning Kang
- School of Computer Science and Information Security, Guilin University of Electronic Technology, Guilin, 541004, China
| | - Maofa Wang
- School of Computer Science and Information Security, Guilin University of Electronic Technology, Guilin, 541004, China
| | - Cheng Pang
- School of Computer Science and Information Security, Guilin University of Electronic Technology, Guilin, 541004, China; Guangxi Key Laboratory of Image and Graphic Intelligent Processing, Guilin, 541004, China
| | - Rushi Lan
- School of Computer Science and Information Security, Guilin University of Electronic Technology, Guilin, 541004, China; Guangxi Key Laboratory of Image and Graphic Intelligent Processing, Guilin, 541004, China
| | - Bingbing Li
- Department of Pathology, Ganzhou Municipal Hospital, Ganzhou, 341000, China
| | - Junlin Guan
- School of Computer Science and Information Security, Guilin University of Electronic Technology, Guilin, 541004, China.
| | - Huadeng Wang
- School of Computer Science and Information Security, Guilin University of Electronic Technology, Guilin, 541004, China; Guangxi Key Laboratory of Image and Graphic Intelligent Processing, Guilin, 541004, China
| |
Collapse
|
28
|
Grignaffini F, Barbuto F, Troiano M, Piazzo L, Simeoni P, Mangini F, De Stefanis C, Onetti Muda A, Frezza F, Alisi A. The Use of Artificial Intelligence in the Liver Histopathology Field: A Systematic Review. Diagnostics (Basel) 2024; 14:388. [PMID: 38396427 PMCID: PMC10887838 DOI: 10.3390/diagnostics14040388] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2023] [Revised: 02/02/2024] [Accepted: 02/06/2024] [Indexed: 02/25/2024] Open
Abstract
Digital pathology (DP) has begun to play a key role in the evaluation of liver specimens. Recent studies have shown that a workflow that combines DP and artificial intelligence (AI) applied to histopathology has potential value in supporting the diagnosis, treatment evaluation, and prognosis prediction of liver diseases. Here, we provide a systematic review of the use of this workflow in the field of hepatology. Based on the PRISMA 2020 criteria, a search of the PubMed, SCOPUS, and Embase electronic databases was conducted, applying inclusion/exclusion filters. The articles were evaluated by two independent reviewers, who extracted the specifications and objectives of each study, the AI tools used, and the results obtained. From the 266 initial records identified, 25 eligible studies were selected, mainly conducted on human liver tissues. Most of the studies were performed using whole-slide imaging systems for imaging acquisition and applying different machine learning and deep learning methods for image pre-processing, segmentation, feature extractions, and classification. Of note, most of the studies selected demonstrated good performance as classifiers of liver histological images compared to pathologist annotations. Promising results to date bode well for the not-too-distant inclusion of these techniques in clinical practice.
Collapse
Affiliation(s)
- Flavia Grignaffini
- Department of Information Engineering, Electronics and Telecommunications (DIET), “La Sapienza”, University of Rome, 00184 Rome, Italy; (F.G.); (F.B.); (L.P.); (F.M.); (F.F.)
| | - Francesco Barbuto
- Department of Information Engineering, Electronics and Telecommunications (DIET), “La Sapienza”, University of Rome, 00184 Rome, Italy; (F.G.); (F.B.); (L.P.); (F.M.); (F.F.)
| | - Maurizio Troiano
- Research Unit of Genetics of Complex Phenotypes, Bambino Gesù Children’s Hospital, IRCCS, 00165 Rome, Italy; (M.T.); (C.D.S.)
| | - Lorenzo Piazzo
- Department of Information Engineering, Electronics and Telecommunications (DIET), “La Sapienza”, University of Rome, 00184 Rome, Italy; (F.G.); (F.B.); (L.P.); (F.M.); (F.F.)
| | - Patrizio Simeoni
- National Transport Authority (NTA), D02 WT20 Dublin, Ireland;
- Faculty of Lifelong Learning, South East Technological University (SETU), R93 V960 Carlow, Ireland
| | - Fabio Mangini
- Department of Information Engineering, Electronics and Telecommunications (DIET), “La Sapienza”, University of Rome, 00184 Rome, Italy; (F.G.); (F.B.); (L.P.); (F.M.); (F.F.)
| | - Cristiano De Stefanis
- Research Unit of Genetics of Complex Phenotypes, Bambino Gesù Children’s Hospital, IRCCS, 00165 Rome, Italy; (M.T.); (C.D.S.)
| | | | - Fabrizio Frezza
- Department of Information Engineering, Electronics and Telecommunications (DIET), “La Sapienza”, University of Rome, 00184 Rome, Italy; (F.G.); (F.B.); (L.P.); (F.M.); (F.F.)
| | - Anna Alisi
- Research Unit of Genetics of Complex Phenotypes, Bambino Gesù Children’s Hospital, IRCCS, 00165 Rome, Italy; (M.T.); (C.D.S.)
| |
Collapse
|
29
|
Ranjbarzadeh R, Zarbakhsh P, Caputo A, Tirkolaee EB, Bendechache M. Brain tumor segmentation based on optimized convolutional neural network and improved chimp optimization algorithm. Comput Biol Med 2024; 168:107723. [PMID: 38000242 DOI: 10.1016/j.compbiomed.2023.107723] [Citation(s) in RCA: 14] [Impact Index Per Article: 14.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2023] [Revised: 10/21/2023] [Accepted: 11/15/2023] [Indexed: 11/26/2023]
Abstract
Reliable and accurate brain tumor segmentation is a challenging task even with the appropriate acquisition of brain images. Tumor grading and segmentation utilizing Magnetic Resonance Imaging (MRI) are necessary steps for correct diagnosis and treatment planning. There are different MRI sequence images (T1, Flair, T1ce, T2, etc.) for identifying different parts of the tumor. Due to the diversity in the illumination of each brain imaging modality, different information and details can be obtained from each input modality. Therefore, by using various MRI modalities, the diagnosis system is capable of finding more unique details that lead to a better segmentation result, especially in fuzzy borders. In this study, to achieve an automatic and robust brain tumor segmentation framework using four MRI sequence images, an optimized Convolutional Neural Network (CNN) is proposed. All weight and bias values of the CNN model are adjusted using an Improved Chimp Optimization Algorithm (IChOA). In the first step, all four input images are normalized to find some potential areas of the existing tumor. Next, by employing the IChOA, the best features are selected using a Support Vector Machine (SVM) classifier. Finally, the best-extracted features are fed to the optimized CNN model to classify each object for brain tumor segmentation. Accordingly, the proposed IChOA is utilized for feature selection and optimizing Hyperparameters in the CNN model. The experimental outcomes conducted on the BRATS 2018 dataset demonstrate superior performance (Precision of 97.41 %, Recall of 95.78 %, and Dice Score of 97.04 %) compared to the existing frameworks.
Collapse
Affiliation(s)
- Ramin Ranjbarzadeh
- School of Computing, Faculty of Engineering and Computing, Dublin City University, Ireland.
| | - Payam Zarbakhsh
- Electrical and Electronic Engineering Department, Cyprus International University, Via Mersin 10, Nicosia, Northern Cyprus, Turkey.
| | - Annalina Caputo
- School of Computing, Faculty of Engineering and Computing, Dublin City University, Ireland.
| | - Erfan Babaee Tirkolaee
- Department of Industrial Engineering, Istinye University, Istanbul, Turkey; Department of Industrial Engineering and Management, Yuan Ze University, Taoyuan, Taiwan; Department of Industrial and Mechanical Engineering, Lebanese American University, Byblos, Lebanon.
| | - Malika Bendechache
- Lero & ADAPT Research Centres, School of Computer Science, University of Galway, Ireland.
| |
Collapse
|
30
|
Kumar A, Pandey SK, Varshney N, Singh KU, Singh T, Shah MA. Distinctive approach in brain tumor detection and feature extraction using biologically inspired DWT method and SVM. Sci Rep 2023; 13:22735. [PMID: 38123666 PMCID: PMC10733354 DOI: 10.1038/s41598-023-50073-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/27/2023] [Accepted: 12/14/2023] [Indexed: 12/23/2023] Open
Abstract
Brain tumors result from uncontrolled cell growth, potentially leading to fatal consequences if left untreated. While significant efforts have been made with some promising results, the segmentation and classification of brain tumors remain challenging due to their diverse locations, shapes, and sizes. In this study, we employ a combination of Discrete Wavelet Transform (DWT) and Principal Component Analysis (PCA) to enhance performance and streamline the medical image segmentation process. Proposed method using Otsu's segmentation method followed by PCA to identify the most informative features. Leveraging the grey-level co-occurrence matrix, we extract numerous valuable texture features. Subsequently, we apply a Support Vector Machine (SVM) with various kernels for classification. We evaluate the proposed method's performance using metrics such as accuracy, sensitivity, specificity, and the Dice Similarity Index coefficient. The experimental results validate the effectiveness of our approach, with recall rates of 86.9%, precision of 95.2%, F-measure of 90.9%, and overall accuracy. Simulation of the results shows improvements in both quality and accuracy compared to existing techniques. In results section, experimental Dice Similarity Index coefficient of 0.82 indicates a strong overlap between the machine-extracted tumor region and the manually delineated tumor region.
Collapse
Affiliation(s)
- Ankit Kumar
- Department of Information Technology, Guru Ghasidas Vishwavidyalaya, Bilaspur, India
| | - Saroj Kumar Pandey
- Department of Computer Engineering & Applications, GLA University, Mathura, Uttar Pradesh, India
| | - Neeraj Varshney
- Department of Computer Engineering & Applications, GLA University, Mathura, Uttar Pradesh, India
| | - Kamred Udham Singh
- School of Computer Science and Engineering, Graphic Hill Era University, Dehradun, 248002, India
| | - Teekam Singh
- Department of Computer Science and Engineering, Graphic Era Deemed to be University, Dehradun, 248002, India
| | - Mohd Asif Shah
- Kebri Dehar University, Kebri Dehar, Somali, 250, Ethiopia.
- Centre of Research Impact and Outcome, Chitkara University Institute of Engineering and Technology, Chitkara University, Rajpura, 140401, Punjab, India.
- Division of Research and Development, Lovely Professional University, Phagwara, Punjab, 144001, India.
| |
Collapse
|
31
|
Chen X, Chen F, Liang C, He G, Chen H, Wu Y, Chen Y, Shuai J, Yang Y, Dai C, Cao L, Wang X, Cai E, Wang J, Wu M, Zeng L, Zhu J, Hai D, Pan W, Pan S, Zhang C, Quan S, Su F. MRI advances in the imaging diagnosis of tuberculous meningitis: opportunities and innovations. Front Microbiol 2023; 14:1308149. [PMID: 38149270 PMCID: PMC10750405 DOI: 10.3389/fmicb.2023.1308149] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2023] [Accepted: 11/14/2023] [Indexed: 12/28/2023] Open
Abstract
Tuberculous meningitis (TBM) is not only one of the most fatal forms of tuberculosis, but also a major public health concern worldwide, presenting grave clinical challenges due to its nonspecific symptoms and the urgent need for timely intervention. The severity and the rapid progression of TBM underscore the necessity of early and accurate diagnosis to prevent irreversible neurological deficits and reduce mortality rates. Traditional diagnostic methods, reliant primarily on clinical findings and cerebrospinal fluid analysis, often falter in delivering timely and conclusive results. Moreover, such methods struggle to distinguish TBM from other forms of neuroinfections, making it critical to seek advanced diagnostic solutions. Against this backdrop, magnetic resonance imaging (MRI) has emerged as an indispensable modality in diagnostics, owing to its unique advantages. This review provides an overview of the advancements in MRI technology, specifically emphasizing its crucial applications in the early detection and identification of complex pathological changes in TBM. The integration of artificial intelligence (AI) has further enhanced the transformative impact of MRI on TBM diagnostic imaging. When these cutting-edge technologies synergize with deep learning algorithms, they substantially improve diagnostic precision and efficiency. Currently, the field of TBM imaging diagnosis is undergoing a phase of technological amalgamation. The melding of MRI and AI technologies unquestionably signals new opportunities in this specialized area.
Collapse
Affiliation(s)
- Xingyu Chen
- Department of Infectious Diseases, Wenzhou Central Hospital, Wenzhou, China
- The First School of Medicine, Wenzhou Medical University, Wenzhou, China
| | - Fanxuan Chen
- School of Biomedical Engineering, School of Ophthalmology and Optometry, Eye Hospital, Wenzhou Medical University, Wenzhou, China
| | - Chenglong Liang
- The First School of Medicine, Wenzhou Medical University, Wenzhou, China
| | - Guoqiang He
- Postgraduate Training Base Alliance of Wenzhou Medical University, Wenzhou, China
- Wenzhou Institute, University of Chinese Academy of Sciences, Wenzhou, China
| | - Hao Chen
- The First School of Medicine, Wenzhou Medical University, Wenzhou, China
| | - Yanchan Wu
- School of Electrical and Information Engineering, Quzhou University, Quzhou, China
| | - Yinda Chen
- School of Electrical and Information Engineering, Quzhou University, Quzhou, China
| | - Jincen Shuai
- Baskin Engineering, University of California, Santa Cruz, CA, United States
| | - Yilei Yang
- The First School of Medicine, Wenzhou Medical University, Wenzhou, China
| | | | - Luhuan Cao
- Wenzhou Medical University, Wenzhou, China
| | - Xian Wang
- Wenzhou Medical University, Wenzhou, China
| | - Enna Cai
- Wenzhou Medical University, Wenzhou, China
| | | | | | - Li Zeng
- Wenzhou Medical University, Wenzhou, China
| | | | - Darong Hai
- Wenzhou Medical University, Wenzhou, China
| | - Wangzheng Pan
- Renji College of Wenzhou Medical University, Wenzhou, China
| | - Shuo Pan
- The First School of Medicine, Wenzhou Medical University, Wenzhou, China
| | - Chengxi Zhang
- School of Materials Science and Engineering, Shandong Jianzhu University, Jinan, China
| | - Shichao Quan
- Department of Big Data in Health Science, The First Affiliated Hospital of Wenzhou Medical University, Wenzhou, China
- Key Laboratory of Intelligent Treatment and Life Support for Critical Diseases of Zhejiang Province, Wenzhou, China
- Zhejiang Engineering Research Center for Hospital Emergency and Process Digitization, Wenzhou, China
| | - Feifei Su
- Department of Infectious Diseases, Wenzhou Central Hospital, Wenzhou, China
- Wenzhou Key Laboratory of Diagnosis and Treatment of Emerging and Recurrent Infectious Diseases, Wenzhou, China
- Department of Infectious Diseases, Wenzhou Sixth People’s Hospital, Wenzhou, China
| |
Collapse
|
32
|
Chen H, An J, Jiang B, Xia L, Bai Y, Gao Z. WS-MTST: Weakly Supervised Multi-Label Brain Tumor Segmentation With Transformers. IEEE J Biomed Health Inform 2023; 27:5914-5925. [PMID: 37788198 DOI: 10.1109/jbhi.2023.3321602] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/05/2023]
Abstract
Brain tumor segmentation is a key step in brain cancer diagnosis. Segmentation of brain tumor sub-regions, including necrotic, enhancing, and edematous regions, can provide more detailed guidance for clinical diagnosis. Weakly supervised brain tumor segmentation methods have received much attention because they do not require time-consuming pixel-level annotations. However, existing weakly supervised methods focus on the segmentation of the entire tumor region while ignoring the challenging task of multi-label segmentation for the tumor sub-regions. In this article, we propose a weakly supervised approach to solve the multi-label brain tumor segmentation problem. To the best of our knowledge, it's the first end-to-end multi-label weakly supervised segmentation model applied to brain tumor segmentation. With well-designed loss functions and a contrastive learning pre-training process, our proposed Transformer-based segmentation method (WS-MTST) has the ability to perform segmentation of brain tumor sub-regions. We conduct comprehensive experiments and demonstrate that our method reaches the state-of-the-art on the popular brain tumor dataset BraTS (from 2018 to 2020).
Collapse
|
33
|
Ahamed MF, Hossain MM, Nahiduzzaman M, Islam MR, Islam MR, Ahsan M, Haider J. A review on brain tumor segmentation based on deep learning methods with federated learning techniques. Comput Med Imaging Graph 2023; 110:102313. [PMID: 38011781 DOI: 10.1016/j.compmedimag.2023.102313] [Citation(s) in RCA: 9] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/13/2023] [Revised: 11/13/2023] [Accepted: 11/13/2023] [Indexed: 11/29/2023]
Abstract
Brain tumors have become a severe medical complication in recent years due to their high fatality rate. Radiologists segment the tumor manually, which is time-consuming, error-prone, and expensive. In recent years, automated segmentation based on deep learning has demonstrated promising results in solving computer vision problems such as image classification and segmentation. Brain tumor segmentation has recently become a prevalent task in medical imaging to determine the tumor location, size, and shape using automated methods. Many researchers have worked on various machine and deep learning approaches to determine the most optimal solution using the convolutional methodology. In this review paper, we discuss the most effective segmentation techniques based on the datasets that are widely used and publicly available. We also proposed a survey of federated learning methodologies to enhance global segmentation performance and ensure privacy. A comprehensive literature review is suggested after studying more than 100 papers to generalize the most recent techniques in segmentation and multi-modality information. Finally, we concentrated on unsolved problems in brain tumor segmentation and a client-based federated model training strategy. Based on this review, future researchers will understand the optimal solution path to solve these issues.
Collapse
Affiliation(s)
- Md Faysal Ahamed
- Department of Computer Science & Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Md Munawar Hossain
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Md Nahiduzzaman
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Md Rabiul Islam
- Department of Computer Science & Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Md Robiul Islam
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Mominul Ahsan
- Department of Computer Science, University of York, Deramore Lane, Heslington, York YO10 5GH, UK
| | - Julfikar Haider
- Department of Engineering, Manchester Metropolitan University, Chester St, Manchester M1 5GD, UK.
| |
Collapse
|
34
|
Sun W, Wang C, Tian C, Li X, Hu X, Liu S. Nanotechnology for brain tumor imaging and therapy based on π-conjugated materials: state-of-the-art advances and prospects. Front Chem 2023; 11:1301496. [PMID: 38025074 PMCID: PMC10663370 DOI: 10.3389/fchem.2023.1301496] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2023] [Accepted: 10/26/2023] [Indexed: 12/01/2023] Open
Abstract
In contemporary biomedical research, the development of nanotechnology has brought forth numerous possibilities for brain tumor imaging and therapy. Among these, π-conjugated materials have garnered significant attention as a special class of nanomaterials in brain tumor-related studies. With their excellent optical and electronic properties, π-conjugated materials can be tailored in structure and nature to facilitate applications in multimodal imaging, nano-drug delivery, photothermal therapy, and other related fields. This review focuses on presenting the cutting-edge advances and application prospects of π-conjugated materials in brain tumor imaging and therapeutic nanotechnology.
Collapse
Affiliation(s)
- Wenshe Sun
- Department of Interventional Medical Center, Affiliated Hospital of Qingdao University, Qingdao, Shandong, China
- Qingdao Cancer Institute, Qingdao University, Qingdao, China
| | - Congxiao Wang
- Department of Interventional Medical Center, Affiliated Hospital of Qingdao University, Qingdao, Shandong, China
| | - Chuan Tian
- Department of Interventional Medical Center, Affiliated Hospital of Qingdao University, Qingdao, Shandong, China
| | - Xueda Li
- Department of Interventional Medical Center, Affiliated Hospital of Qingdao University, Qingdao, Shandong, China
| | - Xiaokun Hu
- Department of Interventional Medical Center, Affiliated Hospital of Qingdao University, Qingdao, Shandong, China
| | - Shifeng Liu
- Department of Interventional Medical Center, Affiliated Hospital of Qingdao University, Qingdao, Shandong, China
| |
Collapse
|
35
|
Kim S, Yoon H, Lee J, Yoo S. Facial wrinkle segmentation using weighted deep supervision and semi-automatic labeling. Artif Intell Med 2023; 145:102679. [PMID: 37925209 DOI: 10.1016/j.artmed.2023.102679] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/15/2023] [Revised: 07/28/2023] [Accepted: 10/03/2023] [Indexed: 11/06/2023]
Abstract
Facial wrinkles are important indicators of human aging. Recently, a method using deep learning and a semi-automatic labeling was proposed to segment facial wrinkles, which showed much better performance than conventional image-processing-based methods. However, the difficulty of wrinkle segmentation remains challenging due to the thinness of wrinkles and their small proportion in the entire image. Therefore, performance improvement in wrinkle segmentation is still necessary. To address this issue, we propose a novel loss function that takes into account the thickness of wrinkles based on the semi-automatic labeling approach. First, considering the different spatial dimensions of the decoder in the U-Net architecture, we generated weighted wrinkle maps from ground truth. These weighted wrinkle maps were used to calculate the training losses more accurately than the existing deep supervision approach. This new loss computation approach is defined as weighted deep supervision in our study. The proposed method was evaluated using an image dataset obtained from a professional skin analysis device and labeled using semi-automatic labeling. In our experiment, the proposed weighted deep supervision showed higher Jaccard Similarity Index (JSI) performance for wrinkle segmentation compared to conventional deep supervision and traditional image processing methods. Additionally, we conducted experiments on the labeling using a semi-automatic labeling approach, which had not been explored in previous research, and compared it with human labeling. The semi-automatic labeling technology showed more consistent wrinkle labels than human-made labels. Furthermore, to assess the scalability of the proposed method to other domains, we applied it to retinal vessel segmentation. The results demonstrated superior performance of the proposed method compared to existing retinal vessel segmentation approaches. In conclusion, the proposed method offers high performance and can be easily applied to various biomedical domains and U-Net-based architectures. Therefore, the proposed approach will be beneficial for various biomedical imaging approaches. To facilitate this, we have made the source code of the proposed method publicly available at: https://github.com/resemin/WeightedDeepSupervision.
Collapse
Affiliation(s)
- Semin Kim
- AI R&D Center, Lululab Inc., 318, Dosan-daero, Gangnam-gu, Seoul, Republic of Korea.
| | - Huisu Yoon
- AI R&D Center, Lululab Inc., 318, Dosan-daero, Gangnam-gu, Seoul, Republic of Korea.
| | - Jongha Lee
- AI R&D Center, Lululab Inc., 318, Dosan-daero, Gangnam-gu, Seoul, Republic of Korea.
| | - Sangwook Yoo
- AI R&D Center, Lululab Inc., 318, Dosan-daero, Gangnam-gu, Seoul, Republic of Korea.
| |
Collapse
|
36
|
Valente J, António J, Mora C, Jardim S. Developments in Image Processing Using Deep Learning and Reinforcement Learning. J Imaging 2023; 9:207. [PMID: 37888314 PMCID: PMC10607786 DOI: 10.3390/jimaging9100207] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2023] [Revised: 09/24/2023] [Accepted: 09/28/2023] [Indexed: 10/28/2023] Open
Abstract
The growth in the volume of data generated, consumed, and stored, which is estimated to exceed 180 zettabytes in 2025, represents a major challenge both for organizations and for society in general. In addition to being larger, datasets are increasingly complex, bringing new theoretical and computational challenges. Alongside this evolution, data science tools have exploded in popularity over the past two decades due to their myriad of applications when dealing with complex data, their high accuracy, flexible customization, and excellent adaptability. When it comes to images, data analysis presents additional challenges because as the quality of an image increases, which is desirable, so does the volume of data to be processed. Although classic machine learning (ML) techniques are still widely used in different research fields and industries, there has been great interest from the scientific community in the development of new artificial intelligence (AI) techniques. The resurgence of neural networks has boosted remarkable advances in areas such as the understanding and processing of images. In this study, we conducted a comprehensive survey regarding advances in AI design and the optimization solutions proposed to deal with image processing challenges. Despite the good results that have been achieved, there are still many challenges to face in this field of study. In this work, we discuss the main and more recent improvements, applications, and developments when targeting image processing applications, and we propose future research directions in this field of constant and fast evolution.
Collapse
Affiliation(s)
- Jorge Valente
- Techframe-Information Systems, SA, 2785-338 São Domingos de Rana, Portugal; (J.V.); (J.A.)
| | - João António
- Techframe-Information Systems, SA, 2785-338 São Domingos de Rana, Portugal; (J.V.); (J.A.)
| | - Carlos Mora
- Smart Cities Research Center, Polytechnic Institute of Tomar, 2300-313 Tomar, Portugal;
| | - Sandra Jardim
- Smart Cities Research Center, Polytechnic Institute of Tomar, 2300-313 Tomar, Portugal;
| |
Collapse
|
37
|
Krishnan G, Singh S, Pathania M, Gosavi S, Abhishek S, Parchani A, Dhar M. Artificial intelligence in clinical medicine: catalyzing a sustainable global healthcare paradigm. Front Artif Intell 2023; 6:1227091. [PMID: 37705603 PMCID: PMC10497111 DOI: 10.3389/frai.2023.1227091] [Citation(s) in RCA: 78] [Impact Index Per Article: 39.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2023] [Accepted: 08/09/2023] [Indexed: 09/15/2023] Open
Abstract
As the demand for quality healthcare increases, healthcare systems worldwide are grappling with time constraints and excessive workloads, which can compromise the quality of patient care. Artificial intelligence (AI) has emerged as a powerful tool in clinical medicine, revolutionizing various aspects of patient care and medical research. The integration of AI in clinical medicine has not only improved diagnostic accuracy and treatment outcomes, but also contributed to more efficient healthcare delivery, reduced costs, and facilitated better patient experiences. This review article provides an extensive overview of AI applications in history taking, clinical examination, imaging, therapeutics, prognosis and research. Furthermore, it highlights the critical role AI has played in transforming healthcare in developing nations.
Collapse
Affiliation(s)
- Gokul Krishnan
- Department of Internal Medicine, Kasturba Medical College, Manipal, India
| | - Shiana Singh
- Department of Emergency Medicine, All India Institute of Medical Sciences, Rishikesh, India
| | - Monika Pathania
- Department of Geriatric Medicine, All India Institute of Medical Sciences, Rishikesh, India
| | - Siddharth Gosavi
- Department of Internal Medicine, Kasturba Medical College, Manipal, India
| | - Shuchi Abhishek
- Department of Internal Medicine, Kasturba Medical College, Manipal, India
| | - Ashwin Parchani
- Department of Geriatric Medicine, All India Institute of Medical Sciences, Rishikesh, India
| | - Minakshi Dhar
- Department of Geriatric Medicine, All India Institute of Medical Sciences, Rishikesh, India
| |
Collapse
|
38
|
Darwish SM, Abu Shaheen LJ, Elzoghabi AA. A New Medical Analytical Framework for Automated Detection of MRI Brain Tumor Using Evolutionary Quantum Inspired Level Set Technique. Bioengineering (Basel) 2023; 10:819. [PMID: 37508846 PMCID: PMC10376225 DOI: 10.3390/bioengineering10070819] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/05/2023] [Revised: 06/29/2023] [Accepted: 07/04/2023] [Indexed: 07/30/2023] Open
Abstract
Segmenting brain tumors in 3D magnetic resonance imaging (3D-MRI) accurately is critical for easing the diagnostic and treatment processes. In the field of energy functional theory-based methods for image segmentation and analysis, level set methods have emerged as a potent computational approach that has greatly aided in the advancement of the geometric active contour model. An important factor in reducing segmentation error and the number of required iterations when using the level set technique is the choice of the initial contour points, both of which are important when dealing with the wide range of sizes, shapes, and structures that brain tumors may take. To define the velocity function, conventional methods simply use the image gradient, edge strength, and region intensity. This article suggests a clustering method influenced by the Quantum Inspired Dragonfly Algorithm (QDA), a metaheuristic optimizer inspired by the swarming behaviors of dragonflies, to accurately extract initial contour points. The proposed model employs a quantum-inspired computing paradigm to stabilize the trade-off between exploitation and exploration, thereby compensating for any shortcomings of the conventional DA-based clustering method, such as slow convergence or falling into a local optimum. To begin, the quantum rotation gate concept can be used to relocate a colony of agents to a location where they can better achieve the optimum value. The main technique is then given a robust local search capacity by adopting a mutation procedure to enhance the swarm's mutation and realize its variety. After a preliminary phase in which the cranium is disembodied from the brain, tumor contours (edges) are determined with the help of QDA. An initial contour for the MRI series will be derived from these extracted edges. The final step is to use a level set segmentation technique to isolate the tumor area across all volume segments. When applied to 3D-MRI images from the BraTS' 2019 dataset, the proposed technique outperformed state-of-the-art approaches to brain tumor segmentation, as shown by the obtained results.
Collapse
Affiliation(s)
- Saad M Darwish
- Department of Information Technology, Institute of Graduate Studies and Research, Alexandria University, 163 Horreya Avenue, El Shatby, Alexandria 21526, Egypt
| | - Lina J Abu Shaheen
- Department of Computer Information Systems, College of Technology and Applied Sciences, Al-Quds Open University, Deir AL Balah P920, Palestine
| | - Adel A Elzoghabi
- Department of Information Technology, Institute of Graduate Studies and Research, Alexandria University, 163 Horreya Avenue, El Shatby, Alexandria 21526, Egypt
| |
Collapse
|
39
|
Eslami M, Tabarestani S, Adjouadi M. A unique color-coded visualization system with multimodal information fusion and deep learning in a longitudinal study of Alzheimer's disease. Artif Intell Med 2023; 140:102543. [PMID: 37210151 PMCID: PMC10204620 DOI: 10.1016/j.artmed.2023.102543] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/06/2022] [Revised: 03/28/2023] [Accepted: 04/02/2023] [Indexed: 05/22/2023]
Abstract
PURPOSE Automated diagnosis and prognosis of Alzheimer's Disease remain a challenging problem that machine learning (ML) techniques have attempted to resolve in the last decade. This study introduces a first-of-its-kind color-coded visualization mechanism driven by an integrated ML model to predict disease trajectory in a 2-year longitudinal study. The main aim of this study is to help capture visually in 2D and 3D renderings the diagnosis and prognosis of AD, therefore augmenting our understanding of the processes of multiclass classification and regression analysis. METHOD The proposed method, Machine Learning for Visualizing AD (ML4VisAD), is designed to predict disease progression through a visual output. This newly developed model takes baseline measurements as input to generate a color-coded visual image that reflects disease progression at different time points. The architecture of the network relies on convolutional neural networks. With 1123 subjects selected from the ADNI QT-PAD dataset, we use a 10-fold cross-validation process to evaluate the method. Multimodal inputs* include neuroimaging data (MRI, PET), neuropsychological test scores (excluding MMSE, CDR-SB, and ADAS to avoid bias), cerebrospinal fluid (CSF) biomarkers with measures of amyloid beta (ABETA), phosphorylated tau protein (PTAU), total tau protein (TAU), and risk factors that include age, gender, years of education, and ApoE4 gene. FINDINGS/RESULTS Based on subjective scores reached by three raters, the results showed an accuracy of 0.82 ± 0.03 for a 3-way classification and 0.68 ± 0.05 for a 5-way classification. The visual renderings were generated in 0.08 msec for a 23 × 23 output image and in 0.17 ms for a 45 × 45 output image. Through visualization, this study (1) demonstrates that the ML visual output augments the prospects for a more accurate diagnosis and (2) highlights why multiclass classification and regression analysis are incredibly challenging. An online survey was conducted to gauge this visualization platform's merits and obtain valuable feedback from users. All implementation codes are shared online on GitHub. CONCLUSION This approach makes it possible to visualize the many nuances that lead to a specific classification or prediction in the disease trajectory, all in context to multimodal measurements taken at baseline. This ML model can serve as a multiclass classification and prediction model while reinforcing the diagnosis and prognosis capabilities by including a visualization platform.
Collapse
Affiliation(s)
- Mohammad Eslami
- Harvard Ophthalmology AI lab, Schepens Eye Research Institute of Massachusetts Eye and Ear, Harvard Medical School, Boston, MA, USA; Center for Advanced Technology and Education, Florida International University, Miami, FL, United States.
| | - Solale Tabarestani
- Center for Advanced Technology and Education, Florida International University, Miami, FL, United States.
| | - Malek Adjouadi
- Center for Advanced Technology and Education, Florida International University, Miami, FL, United States.
| |
Collapse
|
40
|
Chaudhury S, Sau K, Khan MA, Shabaz M. Deep transfer learning for IDC breast cancer detection using fast AI technique and Sqeezenet architecture. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2023; 20:10404-10427. [PMID: 37322939 DOI: 10.3934/mbe.2023457] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/17/2023]
Abstract
One of the most effective approaches for identifying breast cancer is histology, which is the meticulous inspection of tissues under a microscope. The kind of cancer cells, or whether they are cancerous (malignant) or non-cancerous, is typically determined by the type of tissue that is analyzed by the test performed by the technician (benign). The goal of this study was to automate IDC classification within breast cancer histology samples using a transfer learning technique. To improve our outcomes, we combined a Gradient Color Activation Mapping (Grad CAM) and image coloring mechanism with a discriminative fine-tuning methodology employing a one-cycle strategy using FastAI techniques. There have been lots of research studies related to deep transfer learning which use the same mechanism, but this report uses a transfer learning mechanism based on lightweight Squeeze Net architecture, a variant of CNN (Convolution neural network). This strategy demonstrates that fine-tuning on Squeeze Net makes it possible to achieve satisfactory results when transitioning generic features from natural images to medical images.
Collapse
Affiliation(s)
- Sushovan Chaudhury
- University of Engineering and Management, Kolkata, Department of Computer Science and Engineering, University Area, Plot No. III, B/5, New Town Rd, Action Area III, Newtown, Kolkata, West Bengal 700160, India
| | - Kartik Sau
- University of Engineering and Management, Kolkata, Department of Computer Science and Engineering, University Area, Plot No. III, B/5, New Town Rd, Action Area III, Newtown, Kolkata, West Bengal 700160, India
| | | | - Mohammad Shabaz
- Model Institute of Engineering and Technology Jammu, J&K, India
| |
Collapse
|
41
|
ME-CCNN: Multi-encoded images and a cascade convolutional neural network for breast tumor segmentation and recognition. Artif Intell Rev 2023. [DOI: 10.1007/s10462-023-10426-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/21/2023]
|