1
|
Jia Y, Dong L, Jiao Y. Medical image classification based on contour processing attention mechanism. Comput Biol Med 2025; 191:110102. [PMID: 40203738 DOI: 10.1016/j.compbiomed.2025.110102] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/19/2024] [Revised: 03/11/2025] [Accepted: 03/26/2025] [Indexed: 04/11/2025]
Abstract
Medical diagnosis, often constrained by a doctor's experience and capabilities, remains a critical challenge. In recent years, intelligent algorithms have emerged as promising tools to assist in improving diagnostic accuracy. Among these, medical image classification plays a pivotal role in enhancing diagnostic precision. This paper proposes a flexible and concise medical image classification method based on a contour processing attention mechanism, designed to improve accuracy by emphasizing target regions during image processing. First, the training images undergo sequential grayscale and binarization processes, after which binary images are subjected to opening and closing operations to generate two distinct contour maps. These contour maps are then concatenated with the grayscale image along the channel dimension to produce a feature map, which is subsequently convolved. Next, pixel-wise multiplication is performed between the resulting image containing contour information and the original training image, thereby enhancing the contour and positional information of the target regions. The enhanced image is then fed into a residual network for classification training, forming a model based on the contour processing attention mechanism. Finally, classification experiments using three different types of medical image datasets were conducted. The experimental results demonstrate that the contour processing attention mechanism significantly improves the performance of residual networks in medical image classification, achieving the 0.0368 increase in classification accuracy, the 0.0413 improvement in the F1 score and the 0.0821 improvement in the Kappa score. Furthermore, the proposed model demonstrates versatility, with potential applications not only in medical image classification but also in other domains, such as remote sensing, urban landscape analysis, and transportation vehicle image classification.
Collapse
Affiliation(s)
- Yongnan Jia
- School of Automation and Electrical Engineering, University of Science and Technology Beijing, Beijing, 100083, PR China; Key Laboratory of Knowledge Automation for Industrial Processes of Ministry of Education, University of Science and Technology Beijing, Beijing, 100083, PR China.
| | - Linjie Dong
- School of Automation and Electrical Engineering, University of Science and Technology Beijing, Beijing, 100083, PR China
| | - Yuhang Jiao
- School of Automation and Electrical Engineering, University of Science and Technology Beijing, Beijing, 100083, PR China
| |
Collapse
|
2
|
Kumar S, Bhowmik B. ADConv-Net: Advanced Deep Convolution Neural Network for COVID-19 Diagnostics Using Chest X-Ray and CT Images. SN COMPUTER SCIENCE 2025; 6:423. [DOI: https:/doi.org/10.1007/s42979-025-03923-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/31/2024] [Accepted: 03/22/2025] [Indexed: 04/30/2025]
|
3
|
Kumar S, Bhowmik B. EffiCOVID-net: A highly efficient convolutional neural network for COVID-19 diagnosis using chest X-ray imaging. Methods 2025; 240:81-100. [PMID: 40252941 DOI: 10.1016/j.ymeth.2025.04.008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/25/2024] [Revised: 04/04/2025] [Accepted: 04/14/2025] [Indexed: 04/21/2025] Open
Abstract
The global COVID-19 pandemic has drastically affected daily life, emphasizing the urgent need for early and accurate detection to provide adequate medical treatment, especially with limited antiviral options. Chest X-ray imaging has proven crucial for distinguishing COVID-19 from other respiratory conditions, providing an essential diagnostic tool. Deep learning (DL)-based models have proven highly effective in image diagnostics in recent years. Many of these models are computationally intensive and prone to overfitting, especially when trained on limited datasets. Additionally, conventional models often fail to capture multi-scale features, reducing diagnostic accuracy. This paper proposed a highly efficient convolutional neural network (CNN) called EffiCOVID-Net, incorporating diverse feature learning units. The proposed model consists of a bunch of EffiCOVID blocks that incorporate several layers of convolution containing (3×3) filters and recurrent connections to extract complex features while preserving spatial integrity. The performance of EffiCOVID-Net is rigorously evaluated using standard performance metrics on two publicly available COVID-19 chest X-ray datasets. Experimental results demonstrate that EffiCOVID-Net outperforms existing models, achieving 98.68% accuracy on the COVID-19 radiography dataset (D1), 98.55% on the curated chest X-ray dataset (D2), and 98.87% on the mixed dataset (DMix) in multi-class classification (COVID-19 vs. Normal vs. Pneumonia). For binary classification (COVID-19 vs. Normal), the model attains 99.06%, 99.78%, and 99.07% accuracy, respectively. Integrating Grad-CAM-based visualizations further enhances interpretability by highlighting critical regions influencing model predictions. EffiCOVID-Net's lightweight architecture ensures low computational overhead, making it suitable for deployment in resource-constrained clinical settings. A comparative analysis with existing methods highlights its superior accuracy, efficiency, and robustness performance. However, while the model enhances diagnostic workflows, it is best utilized as an assistive tool rather than a standalone diagnostic method.
Collapse
Affiliation(s)
- Sunil Kumar
- Maharshi Patanjali CPS Lab, BRICS Laboratory, Department of Computer Science and Engineering, National Institute of Technology Karnataka, Mangalore 575025, Karnataka, India.
| | - Biswajit Bhowmik
- Maharshi Patanjali CPS Lab, BRICS Laboratory, Department of Computer Science and Engineering, National Institute of Technology Karnataka, Mangalore 575025, Karnataka, India
| |
Collapse
|
4
|
Quintero-Rincón A, Di-Pasquale R, Quintero-Rodríguez K, Batatia H. Computer-based quantitative image texture analysis using multi-collinearity diagnosis in chest X-ray images. PLoS One 2025; 20:e0320706. [PMID: 40228193 PMCID: PMC11996224 DOI: 10.1371/journal.pone.0320706] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2024] [Accepted: 02/23/2025] [Indexed: 04/16/2025] Open
Abstract
Despite tremendous efforts devoted to the area, image texture analysis is still an open research field. This paper presents an algorithm and experimental results demonstrating the feasibility of developing automated tools to detect abnormal X-ray images based on tissue attenuation. Specifically, this work proposes using the variability characterised by singular values and conditional indices extracted from the singular value decomposition (SVD) as image texture features. In addition, the paper introduces a "tuning weight" parameter to consider the variability of the X-ray attenuation in tissues affected by pathologies. This weight is estimated using the coefficient of variation of the minimum covariance determinant from the bandwidth yielded by the non-parametric distribution of variance-decomposition proportions of the SVD. When multiplied by the two features (singular values and conditional indices), this single parameter acts as a tuning weight, reducing misclassification and improving the classic performance metrics, such as true positive rate, false negative rate, positive predictive values, false discovery rate, area-under-curve, accuracy rate, and total cost. The proposed method implements an ensemble bagged trees classification model to classify X-ray chest images as COVID-19, viral pneumonia, lung opacity, or normal. It was tested using a challenging, imbalanced chest X-ray public dataset. The results show an accuracy of 88% without applying the tuning weight and 99% with its application. The proposed method outperforms state-of-the-art methods, as attested by all performance metrics.
Collapse
Affiliation(s)
- Antonio Quintero-Rincón
- Department of Data Science, Data Science and AI Laboratory, Catholic University of Argentina (UCA), Buenos Aires, Argentina
- Department of Computer Sciences, Catholic University of Argentina (UCA), Buenos Aires, Argentina
| | - Ricardo Di-Pasquale
- Department of Data Science, Data Science and AI Laboratory, Catholic University of Argentina (UCA), Buenos Aires, Argentina
| | | | - Hadj Batatia
- MACS School, Heriot-Watt University, Dubai, United Arab Emirates
| |
Collapse
|
5
|
Zhang F, Zhai D, Bai G, Jiang J, Ye Q, Ji X, Liu X. Towards fairness-aware and privacy-preserving enhanced collaborative learning for healthcare. Nat Commun 2025; 16:2852. [PMID: 40122892 PMCID: PMC11930927 DOI: 10.1038/s41467-025-58055-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/26/2024] [Accepted: 03/11/2025] [Indexed: 03/25/2025] Open
Abstract
The widespread integration of AI algorithms in healthcare has sparked ethical concerns, particularly regarding privacy and fairness. Federated Learning (FL) offers a promising solution to learn from a broad spectrum of patient data without directly accessing individual records, enhancing privacy while facilitating knowledge sharing across distributed data sources. However, healthcare institutions face significant variations in access to crucial computing resources, with resource budgets often linked to demographic and socio-economic factors, exacerbating unfairness in participation. While heterogeneous federated learning methods allow healthcare institutions with varying computational capacities to collaborate, they fail to address the performance gap between resource-limited and resource-rich institutions. As a result, resource-limited institutions may receive suboptimal models, further reinforcing disparities in AI-driven healthcare outcomes. Here, we propose a resource-adaptive framework for collaborative learning that dynamically adjusts to varying computational capacities, ensuring fair participation. Our approach enhances model accuracy, safeguards patient privacy, and promotes equitable access to trustworthy and efficient AI-driven healthcare solutions.
Collapse
Affiliation(s)
- Feilong Zhang
- Faculty of Computing, Harbin Institute of Technology, Harbin, China
| | - Deming Zhai
- Faculty of Computing, Harbin Institute of Technology, Harbin, China
| | - Guo Bai
- Shanghai Ninth People's Hospital, School of Medicine, Shanghai Jiao Tong University, Shanghai, China
| | - Junjun Jiang
- Faculty of Computing, Harbin Institute of Technology, Harbin, China
| | - Qixiang Ye
- School of Electronic Electrical and Communication Engineering, University of Chinese Academy of Sciences, Beijing, China
| | - Xiangyang Ji
- Department of Automation, Tsinghua University, Beijing, China.
| | - Xianming Liu
- Faculty of Computing, Harbin Institute of Technology, Harbin, China.
| |
Collapse
|
6
|
Agha S, Nazir S, Kaleem M, Najeeb F, Talat R. Performance evaluation of reduced complexity deep neural networks. PLoS One 2025; 20:e0319859. [PMID: 40112278 PMCID: PMC11925470 DOI: 10.1371/journal.pone.0319859] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2024] [Accepted: 02/10/2025] [Indexed: 03/22/2025] Open
Abstract
Deep Neural Networks (DNN) have achieved state-of-the-art performance in medical image classification and are increasingly being used for disease diagnosis. However, these models are quite complex and that necessitates the need to reduce the model complexity for their use in low-power edge applications that are becoming common. The model complexity reduction techniques in most cases comprise of time-consuming operations and are often associated with a loss of model performance in proportion to the model size reduction. In this paper, we propose a simplified model complexity reduction technique based on reducing the number of channels for any DNN and demonstrate the complexity reduction approaches for the ResNet-50 model integration in low-power devices. The model performance of the proposed models was evaluated for multiclass classification of CXR images, as normal, pneumonia, and COVID-19 classes. We demonstrate successive size reductions down to 75%, 87%, and 93% reduction with an acceptable classification performance reduction of 0.5%, 0.5%, and 0.8% respectively. We also provide the results for the model generalization, and visualization with Grad-CAM at an acceptable performance and interpretable level. In addition, a theoretical VLSI architecture for the best performing architecture has been presented.
Collapse
Affiliation(s)
- Shahrukh Agha
- Department of Electrical and Computer Engineering, COMSATS University, Islamabad, Pakistan
| | - Sajid Nazir
- School of Computing, Engineering and Built Environment, Glasgow Caledonian University, Glasgow, Scotland, United Kingdom
| | - Mohammad Kaleem
- Department of Electrical and Computer Engineering, COMSATS University, Islamabad, Pakistan
| | - Faisal Najeeb
- Department of Electrical and Computer Engineering, COMSATS University, Islamabad, Pakistan
| | - Rehab Talat
- Islamic International Medical College, Riphah International University, Rawalpindi, Pakistan
| |
Collapse
|
7
|
Oltu B, Güney S, Yuksel SE, Dengiz B. Automated classification of chest X-rays: a deep learning approach with attention mechanisms. BMC Med Imaging 2025; 25:71. [PMID: 40038588 DOI: 10.1186/s12880-025-01604-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2024] [Accepted: 02/18/2025] [Indexed: 03/06/2025] Open
Abstract
BACKGROUND Pulmonary diseases such as COVID-19 and pneumonia, are life-threatening conditions, that require prompt and accurate diagnosis for effective treatment. Chest X-ray (CXR) has become the most common alternative method for detecting pulmonary diseases such as COVID-19, pneumonia, and lung opacity due to their availability, cost-effectiveness, and ability to facilitate comparative analysis. However, the interpretation of CXRs is a challenging task. METHODS This study presents an automated deep learning (DL) model that outperforms multiple state-of-the-art methods in diagnosing COVID-19, Lung Opacity, and Viral Pneumonia. Using a dataset of 21,165 CXRs, the proposed framework introduces a seamless combination of the Vision Transformer (ViT) for capturing long-range dependencies, DenseNet201 for powerful feature extraction, and global average pooling (GAP) for retaining critical spatial details. This combination results in a robust classification system, achieving remarkable accuracy. RESULTS The proposed methodology delivers outstanding results across all categories: achieving 99.4% accuracy and an F1-score of 98.43% for COVID-19, 96.45% accuracy and an F1-score of 93.64% for Lung Opacity, 99.63% accuracy and an F1-score of 97.05% for Viral Pneumonia, and 95.97% accuracy with an F1-score of 95.87% for Normal subjects. CONCLUSION The proposed framework achieves a remarkable overall accuracy of 97.87%, surpassing several state-of-the-art methods with reproducible and objective outcomes. To ensure robustness and minimize variability in train-test splits, our study employs five-fold cross-validation, providing reliable and consistent performance evaluation. For transparency and to facilitate future comparisons, the specific training and testing splits have been made publicly accessible. Furthermore, Grad-CAM-based visualizations are integrated to enhance the interpretability of the model, offering valuable insights into its decision-making process. This innovative framework not only boosts classification accuracy but also sets a new benchmark in CXR-based disease diagnosis.
Collapse
Affiliation(s)
- Burcu Oltu
- Department of Biomedical Engineering, Baskent University, Etimesgut, Ankara, 06790, Türkiye.
| | - Selda Güney
- Department of Electrical and Electronics Engineering, Baskent University, Etimesgut, Ankara, 06790, Türkiye
| | - Seniha Esen Yuksel
- Department of Electrical and Electronics Engineering, Hacettepe University, Beytepe, Ankara, 06800, Türkiye
| | - Berna Dengiz
- Department of Industrial Engineering, Baskent University, Etimesgut, Ankara, 06790, Türkiye
| |
Collapse
|
8
|
Chu S, Ren X, Ji G, Zhao J, Shi J, Wei Y, Pei B, Qiang Y. Learning Consistent Semantic Representation for Chest X-ray via Anatomical Localization in Self-Supervised Pre-Training. IEEE J Biomed Health Inform 2025; 29:2100-2112. [PMID: 40030350 DOI: 10.1109/jbhi.2024.3505303] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/08/2025]
Abstract
Despite the similar global structures in Chest X-ray (CXR) images, the same anatomy exhibits varying appearances across images, including differences in local textures, shapes, colors, etc. Learning consistent representations for anatomical semantics through these diverse appearances poses a great challenge for self-supervised pre-training in CXR images. To address this challenge, we propose two new pre-training tasks: inner-image anatomy localization (IIAL) and cross-image anatomy localization (CIAL). Leveraging the relatively stable positions of identical anatomy across images, we utilize position information directly as supervision to learn consistent semantic representations. Specifically, IIAL adopts a coarse-to-fine heatmap localization approach to correlate anatomical semantics with positions, while CIAL leverages feature affine alignment and heatmap localization to establish a correspondence between identical anatomical semantics across varying images, despite their appearance diversity. Furthermore, we introduce a unified end-to-end pre-training framework, anatomy-aware representation learning (AARL), integrating IIAL, CIAL, and a pixel restoration task. The advantages of AARL are: 1) preserving the appearance diversity and 2) training in a simple end-to-end way avoiding complicated preprocessing. Extensive experiments on six downstream tasks, including classification and segmentation tasks in various application scenarios, demonstrate that our AARL: 1) has more powerful representation and transferring ability; 2) is annotation-efficient, reducing the demand for labeled data and 3) improves the sensitivity to detecting various pathological and anatomical patterns.
Collapse
|
9
|
Prince R, Niu Z, Khan ZY, Chambua J, Yousif A, Patrick N, Jennifer B. Interpretable COVID-19 chest X-ray detection based on handcrafted feature analysis and sequential neural network. Comput Biol Med 2025; 186:109659. [PMID: 39847942 DOI: 10.1016/j.compbiomed.2025.109659] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/14/2024] [Revised: 12/24/2024] [Accepted: 01/06/2025] [Indexed: 01/25/2025]
Abstract
Deep learning methods have significantly improved medical image analysis, particularly in detecting COVID-19 chest X-rays. Nonetheless, these methodologies frequently inhibit some drawbacks, such as limited interpretability, extensive computational resources, and the need for extensive datasets. To tackle these issues, we introduced two novel algorithms: the Dynamic Co-Occurrence Grey Level Matrix (DC-GLM) and the Contextual Adaptation Multiscale Gabor Network (CAMSGNeT). Ground-glass opacity, consolidation, and fibrosis are key indicators of COVID-19 that are effectively captured by DC-GLM, which is designed to adaptively respond to diverse texture sizes and orientations. It emphasizes coarse texture patterns, adeptly catching significant structural alterations in the texture of chest X-rays, enhancing diagnostic precision by documenting the spatial correlations among pixel intensities and facilitating the detection of both significant and minor irregularities. To enhance coarse feature extraction, we introduced CAMSGNeT, which emphasizes fine features via Contextual Adaptive Diffusion. In contrast to conventional multiscale Gabor filtering, CAMSGNeT improves feature extraction by modifying the diffusion process according to both gradients and local texture complexity. The Contextual Adaptation Diffusion approach adjusts the diffusion coefficient by incorporating both gradient and local variance, enabling intricate texture areas to preserve finer details while smoothing regions are diffused to decrease noise. Air bronchograms and crazy-paving patterns are maintained by this adaptive method, which enhances edge identification and texture characteristics while preserving essential tiny details. Finally, a simple optimized sequential neural network analyzes these refined features, resulting in enhanced classification accuracy. Feature importance analysis improves the model's interpretability by revealing the contributions of individual features to its decisions. Our methodology outperforms numerous state-of-the-art models, achieving 98.27% and 100% accuracy on two datasets, providing a more interpretable, precise, and resource-efficient solution for COVID-19 detection.
Collapse
|
10
|
Altan G, Narli SS. DeepCOVIDNet-CXR: deep learning strategies for identifying COVID-19 on enhanced chest X-rays. BIOMED ENG-BIOMED TE 2025; 70:21-35. [PMID: 39370946 DOI: 10.1515/bmt-2021-0272] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/03/2021] [Accepted: 09/10/2024] [Indexed: 10/08/2024]
Abstract
OBJECTIVES COVID-19 is one of the recent major epidemics, which accelerates its mortality and prevalence worldwide. Most literature on chest X-ray-based COVID-19 analysis has focused on multi-case classification (COVID-19, pneumonia, and normal) by the advantages of Deep Learning. However, the limited number of chest X-rays with COVID-19 is a prominent deficiency for clinical relevance. This study aims at evaluating COVID-19 identification performances using adaptive histogram equalization (AHE) to feed the ConvNet architectures with reliable lung anatomy of airways. METHODS We experimented with balanced small- and large-scale COVID-19 databases using left lung, right lung, and complete chest X-rays with various AHE parameters. On multiple strategies, we applied transfer learning on four ConvNet architectures (MobileNet, DarkNet19, VGG16, and AlexNet). RESULTS Whereas DarkNet19 reached the highest multi-case identification performance with an accuracy rate of 98.26 % on the small-scale dataset, VGG16 achieved the best generalization performance with an accuracy rate of 95.04 % on the large-scale dataset. CONCLUSIONS Our study is one of the pioneering approaches that analyses 3615 COVID-19 cases and specifies the most responsible AHE parameters for ConvNet architectures in the multi-case classification.
Collapse
Affiliation(s)
- Gokhan Altan
- Computer Engineering Department, Iskenderun Technical University, Hatay, Türkiye
| | | |
Collapse
|
11
|
Gangwar S, Devi R, Mat Isa NA. Optimized exposer region-based modified adaptive histogram equalization method for contrast enhancement in CXR imaging. Sci Rep 2025; 15:6693. [PMID: 40000697 PMCID: PMC11862226 DOI: 10.1038/s41598-025-90876-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2024] [Accepted: 02/17/2025] [Indexed: 02/27/2025] Open
Abstract
In medical imaging, low-contrast chest X-ray (CXR) images may fail to provide adequate information for accurate visual interpretation and disease diagnosis. Conventional contrast enhancement techniques, such as histogram equalization, often introduce intensity shifts and loss of fine details. This study presents an advanced Exposure Region-Based Modified Adaptive Histogram Equalization (ERBMAHE) method, further optimized using Particle Swarm Optimization (PSO) to enhance contrast, preserve brightness, and strengthen fine details. The ERBMAHE method segments CXR images into underexposed, well-exposed, and overexposed regions using the 9IEC algorithm. The well-exposed region is further divided, generating five histograms. Each region undergoes adaptive contrast enhancement via a novel weighted probability density function (PDF) and power-law transformation to ensure balanced enhancement across different exposure levels. The PSO algorithm is then employed to optimize power-law parameters, further refining contrast enhancement and illumination uniformity while maintaining the natural appearance of medical images. The PSO-ERBMAHE method was tested on 600 Kaggle CXR images and compared against six state-of-the-art techniques. It achieved a superior peak signal-to-noise ratio (PSNR = 31.10 dB), entropy (7.48), feature similarity index (FSIM = 0.98), tenengrad function (TEN = 0.19), quality-aware relative contrast measure (QRCM = 0.10), and contrast ratio, while maintaining a low absolute mean brightness error (AMBE = 0.10). The method effectively enhanced image contrast while preserving brightness and visual quality, as confirmed by medical expert evaluations. The proposed PSO-ERBMAHE method delivers high-quality contrast enhancement in medical imaging, ensuring better visibility of critical anatomical features. By strengthening fine details, maintaining mean brightness, and improving computational efficiency, this technique enhances disease examination and diagnosis, reducing misinterpretation risks and improving clinical decision-making.
Collapse
Affiliation(s)
- Shivam Gangwar
- Department of Electronics and Communication Engineering, University Institute of Engineering & Technology, Kurukshetra University, Kurukshetra, Haryana, 136119, India
| | - Reeta Devi
- Department of Electronics and Communication Engineering, University Institute of Engineering & Technology, Kurukshetra University, Kurukshetra, Haryana, 136119, India.
| | - Nor Ashidi Mat Isa
- School of Electrical and Electronic Engineering, Universiti Sains Malaysia, Engineering Campus, Nibong Tebal, Penang, 14300, Malaysia
| |
Collapse
|
12
|
Wang K, Zhu M, Chen Z, Weng J, Li M, Yiu SM, Ding W, Gu T. A Statistical Physics Perspective: Understanding the Causality Behind Convolutional Neural Network Adversarial Vulnerability. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2025; 36:2118-2132. [PMID: 38324429 DOI: 10.1109/tnnls.2024.3359269] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/09/2024]
Abstract
The adversarial vulnerability of convolutional neural networks (CNNs) refers to the performance degradation of CNNs under adversarial attacks, leading to incorrect decisions. However, the causes of adversarial vulnerability in CNNs remain unknown. To address this issue, we propose a unique cross-scale analytical approach from a statistical physics perspective. It reveals that the huge amount of nonlinear effects inherent in CNNs is the fundamental cause for the formation and evolution of system vulnerability. Vulnerability is spontaneously formed on the macroscopic level after the symmetry of the system is broken through the nonlinear interaction between microscopic state order parameters. We develop a cascade failure algorithm, visualizing how micro perturbations on neurons' activation can cascade and influence macro decision paths. Our empirical results demonstrate the interplay between microlevel activation maps and macrolevel decision-making and provide a statistical physics perspective to understand the causality behind CNN vulnerability. Our work will help subsequent research to improve the adversarial robustness of CNNs.
Collapse
|
13
|
Chen C, Mat Isa NA, Liu X. A review of convolutional neural network based methods for medical image classification. Comput Biol Med 2025; 185:109507. [PMID: 39631108 DOI: 10.1016/j.compbiomed.2024.109507] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/12/2024] [Revised: 11/20/2024] [Accepted: 11/28/2024] [Indexed: 12/07/2024]
Abstract
This study systematically reviews CNN-based medical image classification methods. We surveyed 149 of the latest and most important papers published to date and conducted an in-depth analysis of the methods used therein. Based on the selected literature, we organized this review systematically. First, the development and evolution of CNN in the field of medical image classification are analyzed. Subsequently, we provide an in-depth overview of the main techniques of CNN applied to medical image classification, which is also the current research focus in this field, including data preprocessing, transfer learning, CNN architectures, and explainability, and their role in improving classification accuracy and efficiency. In addition, this overview summarizes the main public datasets for various diseases. Although CNN has great potential in medical image classification tasks and has achieved good results, clinical application is still difficult. Therefore, we conclude by discussing the main challenges faced by CNNs in medical image analysis and pointing out future research directions to address these challenges. This review will help researchers with their future studies and can promote the successful integration of deep learning into clinical practice and smart medical systems.
Collapse
Affiliation(s)
- Chao Chen
- School of Electrical and Electronic Engineering, Engineering Campus, Universiti Sains Malaysia, 14300, Nibong Tebal, Pulau Pinang, Malaysia; School of Automation and Information Engineering, Sichuan University of Science and Engineering, Yibin, 644000, China
| | - Nor Ashidi Mat Isa
- School of Electrical and Electronic Engineering, Engineering Campus, Universiti Sains Malaysia, 14300, Nibong Tebal, Pulau Pinang, Malaysia.
| | - Xin Liu
- School of Electrical and Electronic Engineering, Engineering Campus, Universiti Sains Malaysia, 14300, Nibong Tebal, Pulau Pinang, Malaysia
| |
Collapse
|
14
|
Kabir S, Sarmun R, Ramírez-Velázquez E, Takvani A, Ali M, Chowdhury MEH, Abbas TO. Automated detection of posterior urethral valves in voiding cystourethrography images: A novel AI-Based pipeline for enhanced diagnosis and classification. Comput Biol Med 2025; 185:109509. [PMID: 39705793 DOI: 10.1016/j.compbiomed.2024.109509] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/05/2024] [Revised: 11/27/2024] [Accepted: 11/29/2024] [Indexed: 12/23/2024]
Abstract
INTRODUCTION Posterior Urethral Valves (PUV) are rare congenital anomalies of the male urinary tract that can lead to urethral obstruction and increased risk of kidney disease. Traditional diagnosis relies on subjective interpretation of imaging techniques. This study aimed to automate and increase accuracy of PUV detection in voiding cystourethrography (VCUG) images using an AI-based pipeline. The main objective was to detect presence of PUV based on urethral ratio calculated automatically from segmented urethra region. METHODS A total of 181 VCUG images were evaluated by 9 clinicians to determine presence of PUV. Various different encoders (DenseNet, MobileNet, ResNet and VGG) were combined with Unet and Unet++ architectures to segment the urethra region. Some preprocessing and postprocessing steps were investigated to improve segmentation performance. Urethral ratios were automatically calculated with image processing and morphological operations. Finally, samples were classified between PUV or non PUV based on urethral ratio. RESULTS An overall classification accuracy of 81.52 % was achieved between PUV and non PUV cases. DenseNet201 combined with Unet achieved the best overall segmentation performance (Dice score coefficient 66.15 %). Optimal cut-off value of urethral ratio for PUV detection was determined as 2.01. CONCLUSION PUV detection from VCUG images through automated segmentation and processing can reduce subjectivity and decrease physician workloads. The proposed approach can serve as a foundation for future efforts to fully automate PUV diagnosis and follow-up.
Collapse
Affiliation(s)
- Saidul Kabir
- Department of Electrical and Electronic Engineering, University of Dhaka, Dhaka, Bangladesh
| | - Rusab Sarmun
- Department of Electrical and Electronic Engineering, University of Dhaka, Dhaka, Bangladesh
| | - Elias Ramírez-Velázquez
- Pediatric Urology Department, Hospital Infantil de México Federico Gómez, México City, Mexico
| | - Anil Takvani
- Takvani Kidney Hospital, Junagadh, Gujarat, India
| | - Mansour Ali
- Department of Surgery, Sidra Medicine, Doha, Qatar
| | | | - Tariq O Abbas
- Urology Division, Surgery Department, Sidra Medicine, Qatar; College of Medicine, Qatar University, Doha, Qatar; Weill Cornell Medicine Qatar, Doha, Qatar.
| |
Collapse
|
15
|
Liu H, Zhao M, She C, Peng H, Liu M, Li B. Classification of CT scan and X-ray dataset based on deep learning and particle swarm optimization. PLoS One 2025; 20:e0317450. [PMID: 39869555 PMCID: PMC11771893 DOI: 10.1371/journal.pone.0317450] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/12/2024] [Accepted: 12/25/2024] [Indexed: 01/29/2025] Open
Abstract
In 2019, the novel coronavirus swept the world, exposing the monitoring and early warning problems of the medical system. Computer-aided diagnosis models based on deep learning have good universality and can well alleviate these problems. However, traditional image processing methods may lead to high false positive rates, which is unacceptable in disease monitoring and early warning. This paper proposes a low false positive rate disease detection method based on COVID-19 lung images and establishes a two-stage optimization model. In the first stage, the model is trained using classical gradient descent, and relevant features are extracted; in the second stage, an objective function that minimizes the false positive rate is constructed to obtain a network model with high accuracy and low false positive rate. Therefore, the proposed method has the potential to effectively classify medical images. The proposed model was verified using a public COVID-19 radiology dataset and a public COVID-19 lung CT scan dataset. The results show that the model has made significant progress, with the false positive rate reduced to 11.3% and 7.5%, and the area under the ROC curve increased to 92.8% and 97.01%.
Collapse
Affiliation(s)
- Honghua Liu
- Hunan University of Chinese Medicine, Changsha, PR China
| | | | - Chang She
- Changsha Hospital of Traditional Chinese Medicine(Changsha Eighth Hospital), Changsha, PR China
| | - Han Peng
- Hunan University of Chinese Medicine, Changsha, PR China
| | - Mailan Liu
- Hunan University of Chinese Medicine, Changsha, PR China
| | - Bo Li
- The First Hospital of Hunan University of Chinese Medicine, Changsha, PR China
| |
Collapse
|
16
|
Siddiqui AA, Tirunagari S, Zia T, Windridge D. A latent diffusion approach to visual attribution in medical imaging. Sci Rep 2025; 15:962. [PMID: 39762275 PMCID: PMC11704132 DOI: 10.1038/s41598-024-81646-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/11/2023] [Accepted: 11/28/2024] [Indexed: 01/11/2025] Open
Abstract
Visual attribution in medical imaging seeks to make evident the diagnostically-relevant components of a medical image, in contrast to the more common detection of diseased tissue deployed in standard machine vision pipelines (which are less straightforwardly interpretable/explainable to clinicians). We here present a novel generative visual attribution technique, one that leverages latent diffusion models in combination with domain-specific large language models, in order to generate normal counterparts of abnormal images. The discrepancy between the two hence gives rise to a mapping indicating the diagnostically-relevant image components. To achieve this, we deploy image priors in conjunction with appropriate conditioning mechanisms in order to control the image generative process, including natural language text prompts acquired from medical science and applied radiology. We perform experiments and quantitatively evaluate our results on the COVID-19 Radiography Database containing labelled chest X-rays with differing pathologies via the Frechet Inception Distance (FID), Structural Similarity (SSIM) and Multi Scale Structural Similarity Metric (MS-SSIM) metrics obtained between real and generated images. The resulting system also exhibits a range of latent capabilities including zero-shot localized disease induction, which are evaluated with real examples from the cheXpert dataset.
Collapse
|
17
|
Anter JM, Yakimovich A. Artificial Intelligence Methods in Infection Biology Research. Methods Mol Biol 2025; 2890:291-333. [PMID: 39890733 DOI: 10.1007/978-1-0716-4326-6_15] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/03/2025]
Abstract
Despite unprecedented achievements, the domain-specific application of artificial intelligence (AI) in the realm of infection biology was still in its infancy just a couple of years ago. This is largely attributable to the proneness of the infection biology community to shirk quantitative techniques. The so-called "sorting machine" paradigm was prevailing at that time, meaning that AI applications were primarily confined to the automation of tedious laboratory tasks. However, fueled by the severe acute respiratory syndrome coronavirus 2 pandemic, AI-driven applications in infection biology made giant leaps beyond mere automation. Instead, increasingly sophisticated tasks were successfully tackled, thereby ushering in the transition to the "Swiss army knife" paradigm. Incentivized by the urgent need to subdue a raging pandemic, AI achieved maturity in infection biology and became a versatile tool. In this chapter, the maturation of AI in the field of infection biology from the "sorting machine" paradigm to the "Swiss army knife" paradigm is outlined. Successful applications are illustrated for the three data modalities in the domain, that is, images, molecular data, and language data, with a particular emphasis on disentangling host-pathogen interactions. Along the way, fundamental terminology mentioned in the same breath as AI is elaborated on, and relationships between the subfields these terms represent are established. Notably, in order to dispel the fears of infection biologists toward quantitative methodologies and lower the initial hurdle, this chapter features a hands-on guide on software installation, virtual environment setup, data preparation, and utilization of pretrained models at its very end.
Collapse
Affiliation(s)
- Jacob Marcel Anter
- Center for Advanced Systems Understanding (CASUS), Görlitz, Germany
- Helmholtz-Zentrum Dresden-Rossendorf e. V. (HZDR), Dresden, Germany
| | - Artur Yakimovich
- Center for Advanced Systems Understanding (CASUS), Görlitz, Germany.
- Helmholtz-Zentrum Dresden-Rossendorf e. V. (HZDR), Dresden, Germany.
- Institute of Computer Science, University of Wrocław, Wrocław, Poland.
| |
Collapse
|
18
|
Khan AY, Luque-Nieto MA, Saleem MI, Nava-Baro E. X-Ray Image-Based Real-Time COVID-19 Diagnosis Using Deep Neural Networks (CXR-DNNs). J Imaging 2024; 10:328. [PMID: 39728225 PMCID: PMC11728291 DOI: 10.3390/jimaging10120328] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/25/2024] [Accepted: 12/12/2024] [Indexed: 12/28/2024] Open
Abstract
On 11 February 2020, the prevalent outbreak of COVID-19, a coronavirus illness, was declared a global pandemic. Since then, nearly seven million people have died and over 765 million confirmed cases of COVID-19 have been reported. The goal of this study is to develop a diagnostic tool for detecting COVID-19 infections more efficiently. Currently, the most widely used method is Reverse Transcription Polymerase Chain Reaction (RT-PCR), a clinical technique for infection identification. However, RT-PCR is expensive, has limited sensitivity, and requires specialized medical expertise. One of the major challenges in the rapid diagnosis of COVID-19 is the need for reliable imaging, particularly X-ray imaging. This work takes advantage of artificial intelligence (AI) techniques to enhance diagnostic accuracy by automating the detection of COVID-19 infections from chest X-ray (CXR) images. We obtained and analyzed CXR images from the Kaggle public database (4035 images in total), including cases of COVID-19, viral pneumonia, pulmonary opacity, and healthy controls. By integrating advanced techniques with transfer learning from pre-trained convolutional neural networks (CNNs), specifically InceptionV3, ResNet50, and Xception, we achieved an accuracy of 95%, significantly higher than the 85.5% achieved with ResNet50 alone. Additionally, our proposed method, CXR-DNNs, can accurately distinguish between three different types of chest X-ray images for the first time. This computer-assisted diagnostic tool has the potential to significantly enhance the speed and accuracy of COVID-19 diagnoses.
Collapse
Affiliation(s)
- Ali Yousuf Khan
- Telecommunications Engineering School, University of Malaga, 29010 Malaga, Spain;
| | | | - Muhammad Imran Saleem
- Department of Software Engineering, Sir Syed University of Engineering & Technology, Karachi 75300, Pakistan;
| | - Enrique Nava-Baro
- Institute of Oceanic Engineering Research, University of Malaga, 29010 Malaga, Spain;
| |
Collapse
|
19
|
Hadhoud Y, Mekhaznia T, Bennour A, Amroune M, Kurdi NA, Aborujilah AH, Al-Sarem M. From Binary to Multi-Class Classification: A Two-Step Hybrid CNN-ViT Model for Chest Disease Classification Based on X-Ray Images. Diagnostics (Basel) 2024; 14:2754. [PMID: 39682662 DOI: 10.3390/diagnostics14232754] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/11/2024] [Revised: 12/03/2024] [Accepted: 12/04/2024] [Indexed: 12/18/2024] Open
Abstract
BACKGROUND/OBJECTIVES Chest disease identification for Tuberculosis and Pneumonia diseases presents diagnostic challenges due to overlapping radiographic features and the limited availability of expert radiologists, especially in developing countries. The present study aims to address these challenges by developing a Computer-Aided Diagnosis (CAD) system to provide consistent and objective analyses of chest X-ray images, thereby reducing potential human error. By leveraging the complementary strengths of convolutional neural networks (CNNs) and vision transformers (ViTs), we propose a hybrid model for the accurate detection of Tuberculosis and for distinguishing between Tuberculosis and Pneumonia. METHODS We designed a two-step hybrid model that integrates the ResNet-50 CNN with the ViT-b16 architecture. It uses the transfer learning on datasets from Guangzhou Women's and Children's Medical Center for Pneumonia cases and datasets from Qatar and Dhaka (Bangladesh) universities for Tuberculosis cases. CNNs capture hierarchical structures in images, while ViTs, with their self-attention mechanisms, excel at identifying relationships between features. Combining these approaches enhances the model's performance on binary and multi-class classification tasks. RESULTS Our hybrid CNN-ViT model achieved a binary classification accuracy of 98.97% for Tuberculosis detection. For multi-class classification, distinguishing between Tuberculosis, viral Pneumonia, and bacterial Pneumonia, the model achieved an accuracy of 96.18%. These results underscore the model's potential in improving diagnostic accuracy and reliability for chest disease classification based on X-ray images. CONCLUSIONS The proposed hybrid CNN-ViT model demonstrates substantial potential in advancing the accuracy and robustness of CAD systems for chest disease diagnosis. By integrating CNN and ViT architectures, our approach enhances the diagnostic precision, which may help to alleviate the burden on healthcare systems in resource-limited settings and improve patient outcomes in chest disease diagnosis.
Collapse
Affiliation(s)
- Yousra Hadhoud
- LAMIS Laboratory, Larbi Tebessi University, Tebessa 12002, Algeria
| | - Tahar Mekhaznia
- LAMIS Laboratory, Larbi Tebessi University, Tebessa 12002, Algeria
| | - Akram Bennour
- LAMIS Laboratory, Larbi Tebessi University, Tebessa 12002, Algeria
| | - Mohamed Amroune
- LAMIS Laboratory, Larbi Tebessi University, Tebessa 12002, Algeria
| | - Neesrin Ali Kurdi
- College of Computer Science and Engineering, Taibah University, Medina 41477, Saudi Arabia
| | - Abdulaziz Hadi Aborujilah
- Department of Management Information Systems, College of Commerce & Business Administration, Dhofar University, Salalaha 211, Oman
| | - Mohammed Al-Sarem
- Department of Information Technology, Aylol University College, Yarim 547, Yemen
| |
Collapse
|
20
|
Matta S, Lamard M, Zhang P, Le Guilcher A, Borderie L, Cochener B, Quellec G. A systematic review of generalization research in medical image classification. Comput Biol Med 2024; 183:109256. [PMID: 39427426 DOI: 10.1016/j.compbiomed.2024.109256] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2024] [Revised: 09/17/2024] [Accepted: 10/06/2024] [Indexed: 10/22/2024]
Abstract
Numerous Deep Learning (DL) classification models have been developed for a large spectrum of medical image analysis applications, which promises to reshape various facets of medical practice. Despite early advances in DL model validation and implementation, which encourage healthcare institutions to adopt them, a fundamental questions remain: how can these models effectively handle domain shift? This question is crucial to limit DL models performance degradation. Medical data are dynamic and prone to domain shift, due to multiple factors. Two main shift types can occur over time: (1) covariate shift mainly arising due to updates to medical equipment and (2) concept shift caused by inter-grader variability. To mitigate the problem of domain shift, existing surveys mainly focus on domain adaptation techniques, with an emphasis on covariate shift. More generally, no work has reviewed the state-of-the-art solutions while focusing on the shift types. This paper aims to explore existing domain generalization methods for DL-based classification models through a systematic review of literature. It proposes a taxonomy based on the shift type they aim to solve. Papers were searched and gathered on Scopus till 10 April 2023, and after the eligibility screening and quality evaluation, 77 articles were identified. Exclusion criteria included: lack of methodological novelty (e.g., reviews, benchmarks), experiments conducted on a single mono-center dataset, or articles not written in English. The results of this paper show that learning based methods are emerging, for both shift types. Finally, we discuss future challenges, including the need for improved evaluation protocols and benchmarks, and envisioned future developments to achieve robust, generalized models for medical image classification.
Collapse
Affiliation(s)
- Sarah Matta
- Université de Bretagne Occidentale, Brest, Bretagne, 29200, France; Inserm, UMR 1101, Brest, F-29200, France.
| | - Mathieu Lamard
- Université de Bretagne Occidentale, Brest, Bretagne, 29200, France; Inserm, UMR 1101, Brest, F-29200, France
| | - Philippe Zhang
- Université de Bretagne Occidentale, Brest, Bretagne, 29200, France; Inserm, UMR 1101, Brest, F-29200, France; Evolucare Technologies, Villers-Bretonneux, F-80800, France
| | | | | | - Béatrice Cochener
- Université de Bretagne Occidentale, Brest, Bretagne, 29200, France; Inserm, UMR 1101, Brest, F-29200, France; Service d'Ophtalmologie, CHRU Brest, Brest, F-29200, France
| | | |
Collapse
|
21
|
Yen CT, Tsao CY. Lightweight convolutional neural network for chest X-ray images classification. Sci Rep 2024; 14:29759. [PMID: 39613790 DOI: 10.1038/s41598-024-80826-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/04/2024] [Accepted: 11/21/2024] [Indexed: 12/01/2024] Open
Abstract
In this study, we developed a lightweight and rapid convolutional neural network (CNN) architecture for chest X-ray images; it primarily consists of a redesigned feature extraction (FE) module and multiscale feature (MF) module and validated using publicly available COVID-19 datasets. Experiments were conducted on multiple updated versions of the COVID-19 Radiography Database, a publicly accessible dataset on Kaggle. The database contained images categorized into three classes: COVID-19 coronavirus, viral or bacterial pneumonia, and normal. The results revealed that the proposed method achieved a training accuracy of 99.85% and a validation accuracy of 96.28% when detecting the three classes. In the test set, the optimal results were 96.03% accuracy for COVID-19, 97.10% accuracy for viral or bacterial pneumonia, and 97.86% accuracy for normal individuals. By reducing the computational requirements and improving the speed of the model, the proposed method can achieve real-time, low-error performance to help medical professionals with rapid diagnosis of COVID-19.
Collapse
Affiliation(s)
- Chih-Ta Yen
- Department of Electrical Engineering, National Taiwan Ocean University, Keelung City, 202301, Taiwan.
| | - Chia-Yu Tsao
- Department of Electrical Engineering, National Taiwan Ocean University, Keelung City, 202301, Taiwan
| |
Collapse
|
22
|
Kuş Z, Aydin M. MedSegBench: A comprehensive benchmark for medical image segmentation in diverse data modalities. Sci Data 2024; 11:1283. [PMID: 39587124 PMCID: PMC11589128 DOI: 10.1038/s41597-024-04159-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2024] [Accepted: 11/19/2024] [Indexed: 11/27/2024] Open
Abstract
MedSegBench is a comprehensive benchmark designed to evaluate deep learning models for medical image segmentation across a wide range of modalities. It covers a wide range of modalities, including 35 datasets with over 60,000 images from ultrasound, MRI, and X-ray. The benchmark addresses challenges in medical imaging by providing standardized datasets with train/validation/test splits, considering variability in image quality and dataset imbalances. The benchmark supports binary and multi-class segmentation tasks with up to 19 classes and uses the U-Net architecture with various encoder/decoder networks such as ResNets, EfficientNet, and DenseNet for evaluations. MedSegBench is a valuable resource for developing robust and flexible segmentation algorithms and allows for fair comparisons across different models, promoting the development of universal models for medical tasks. It is the most comprehensive study among medical segmentation datasets. The datasets and source code are publicly available, encouraging further research and development in medical image analysis.
Collapse
Affiliation(s)
- Zeki Kuş
- Fatih Sultan Mehmet Vakif University, Computer Engineering, İstanbul, 34445, Türkiye.
| | - Musa Aydin
- Fatih Sultan Mehmet Vakif University, Computer Engineering, İstanbul, 34445, Türkiye
| |
Collapse
|
23
|
Zheng Q, Zhao W, Wu C, Zhang X, Dai L, Guan H, Li Y, Zhang Y, Wang Y, Xie W. Large-scale long-tailed disease diagnosis on radiology images. Nat Commun 2024; 15:10147. [PMID: 39578456 PMCID: PMC11584732 DOI: 10.1038/s41467-024-54424-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/29/2024] [Accepted: 11/08/2024] [Indexed: 11/24/2024] Open
Abstract
Developing a generalist radiology diagnosis system can greatly enhance clinical diagnostics. In this paper, we introduce RadDiag, a foundational model supporting 2D and 3D inputs across various modalities and anatomies, using a transformer-based fusion module for comprehensive disease diagnosis. Due to patient privacy concerns and the lack of large-scale radiology diagnosis datasets, we utilize high-quality, clinician-reviewed radiological images available online with diagnosis labels. Our dataset, RP3D-DiagDS, contains 40,936 cases with 195,010 scans covering 5568 disorders (930 unique ICD-10-CM codes). Experimentally, our RadDiag achieves 95.14% AUC on internal evaluation with the knowledge-enhancement strategy. Additionally, RadDiag can be zero-shot applied or fine-tuned to external diagnosis datasets sourced from various medical centers, demonstrating state-of-the-art results. In conclusion, we show that publicly shared medical data on the Internet is a tremendous and valuable resource that can potentially support building strong models for image understanding in healthcare.
Collapse
Affiliation(s)
- Qiaoyu Zheng
- Shanghai Jiao Tong University, Shanghai, China
- Shanghai Artificial Intelligence Laboratory, Shanghai, China
| | - Weike Zhao
- Shanghai Jiao Tong University, Shanghai, China
- Shanghai Artificial Intelligence Laboratory, Shanghai, China
| | - Chaoyi Wu
- Shanghai Jiao Tong University, Shanghai, China
- Shanghai Artificial Intelligence Laboratory, Shanghai, China
| | - Xiaoman Zhang
- Shanghai Jiao Tong University, Shanghai, China
- Shanghai Artificial Intelligence Laboratory, Shanghai, China
| | - Lisong Dai
- Shanghai Jiao Tong University, Shanghai, China
- Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University, Shanghai, China
| | - Hengyu Guan
- Department of Reproductive Medicine, Ren Ji Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
- Shanghai Key Laboratory for Assisted Reproduction and Reproductive Genetics, Shanghai, China
| | - Yuehua Li
- Shanghai Jiao Tong University, Shanghai, China
- Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University, Shanghai, China
| | - Ya Zhang
- Shanghai Jiao Tong University, Shanghai, China
- Shanghai Artificial Intelligence Laboratory, Shanghai, China
| | - Yanfeng Wang
- Shanghai Jiao Tong University, Shanghai, China.
- Shanghai Artificial Intelligence Laboratory, Shanghai, China.
| | - Weidi Xie
- Shanghai Jiao Tong University, Shanghai, China.
- Shanghai Artificial Intelligence Laboratory, Shanghai, China.
| |
Collapse
|
24
|
Kamalakannan N, Macharla SR, Kanimozhi M, Sudhakar MS. Exponential Pixelating Integral transform with dual fractal features for enhanced chest X-ray abnormality detection. Comput Biol Med 2024; 182:109093. [PMID: 39232407 DOI: 10.1016/j.compbiomed.2024.109093] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/09/2024] [Revised: 08/25/2024] [Accepted: 08/29/2024] [Indexed: 09/06/2024]
Abstract
The heightened prevalence of respiratory disorders, particularly exacerbated by a significant upswing in fatalities due to the novel coronavirus, underscores the critical need for early detection and timely intervention. This imperative is paramount, possessing the potential to profoundly impact and safeguard numerous lives. Medically, chest radiography stands out as an essential and economically viable medical imaging approach for diagnosing and assessing the severity of diverse Respiratory Disorders. However, their detection in Chest X-Rays is a cumbersome task even for well-trained radiologists owing to low contrast issues, overlapping of the tissue structures, subjective variability, and the presence of noise. To address these issues, a novel analytical model termed Exponential Pixelating Integral is introduced for the automatic detection of infections in Chest X-Rays in this work. Initially, the presented Exponential Pixelating Integral enhances the pixel intensities to overcome the low-contrast issues that are then polar-transformed followed by their representation using the locally invariant Mandelbrot and Julia fractal geometries for effective distinction of structural features. The collated features labeled Exponential Pixelating Integral with dually characterized fractal features are then classified by the non-parametric multivariate adaptive regression splines to establish an ensemble model between each pair of classes for effective diagnosis of diverse diseases. Rigorous analysis of the proposed classification framework on large medical benchmarked datasets showcases its superiority over its peers by registering a higher classification accuracy and F1 scores ranging from 98.46 to 99.45 % and 96.53-98.10 % respectively, making it a precise and interpretable automated system for diagnosing respiratory disorders.
Collapse
Affiliation(s)
| | | | - M Kanimozhi
- School of Electrical & Electronics, Sathyabama Institute of Science and Technology, Chennai, Tamilnadu, India
| | - M S Sudhakar
- School of Electronics Engineering, Vellore Institute of Technology, Vellore, Tamilnadu, India.
| |
Collapse
|
25
|
Huang Q, Li G. Knowledge graph based reasoning in medical image analysis: A scoping review. Comput Biol Med 2024; 182:109100. [PMID: 39244959 DOI: 10.1016/j.compbiomed.2024.109100] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/01/2024] [Revised: 08/04/2024] [Accepted: 08/31/2024] [Indexed: 09/10/2024]
Abstract
Automated computer-aided diagnosis (CAD) is becoming more significant in the field of medicine due to advancements in computer hardware performance and the progress of artificial intelligence. The knowledge graph is a structure for visually representing knowledge facts. In the last decade, a large body of work based on knowledge graphs has effectively improved the organization and interpretability of large-scale complex knowledge. Introducing knowledge graph inference into CAD is a research direction with significant potential. In this review, we briefly review the basic principles and application methods of knowledge graphs firstly. Then, we systematically organize and analyze the research and application of knowledge graphs in medical imaging-assisted diagnosis. We also summarize the shortcomings of the current research, such as medical data barriers and deficiencies, low utilization of multimodal information, and weak interpretability. Finally, we propose future research directions with possibilities and potentials to address the shortcomings of current approaches.
Collapse
Affiliation(s)
- Qinghua Huang
- School of Artificial Intelligence, OPtics and ElectroNics (iOPEN), Northwestern Polytechnical University, 127 West Youyi Road, Beilin District, Xi'an, 710072, Shaanxi, China.
| | - Guanghui Li
- School of Artificial Intelligence, OPtics and ElectroNics (iOPEN), Northwestern Polytechnical University, 127 West Youyi Road, Beilin District, Xi'an, 710072, Shaanxi, China; School of Computer Science, Northwestern Polytechnical University, 1 Dongxiang Road, Chang'an District, Xi'an, 710129, Shaanxi, China.
| |
Collapse
|
26
|
Zeng X, Abdullah N, Sumari P. Self-supervised learning framework application for medical image analysis: a review and summary. Biomed Eng Online 2024; 23:107. [PMID: 39465395 PMCID: PMC11514943 DOI: 10.1186/s12938-024-01299-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/10/2024] [Accepted: 10/17/2024] [Indexed: 10/29/2024] Open
Abstract
Manual annotation of medical image datasets is labor-intensive and prone to biases. Moreover, the rate at which image data accumulates significantly outpaces the speed of manual annotation, posing a challenge to the advancement of machine learning, particularly in the realm of supervised learning. Self-supervised learning is an emerging field that capitalizes on unlabeled data for training, thereby circumventing the need for extensive manual labeling. This learning paradigm generates synthetic pseudo-labels through pretext tasks, compelling the network to acquire image representations in a pseudo-supervised manner and subsequently fine-tuning with a limited set of annotated data to achieve enhanced performance. This review begins with an overview of prevalent types and advancements in self-supervised learning, followed by an exhaustive and systematic examination of methodologies within the medical imaging domain from 2018 to September 2024. The review encompasses a range of medical image modalities, including CT, MRI, X-ray, Histology, and Ultrasound. It addresses specific tasks, such as Classification, Localization, Segmentation, Reduction of False Positives, Improvement of Model Performance, and Enhancement of Image Quality. The analysis reveals a descending order in the volume of related studies, with CT and MRI leading the list, followed by X-ray, Histology, and Ultrasound. Except for CT and MRI, there is a greater prevalence of studies focusing on contrastive learning methods over generative learning approaches. The performance of MRI/Ultrasound classification and all image types segmentation still has room for further exploration. Generally, this review can provide conceptual guidance for medical professionals to combine self-supervised learning with their research.
Collapse
Affiliation(s)
- Xiangrui Zeng
- School of Computer Sciences, Universiti Sains Malaysia, USM, 11800, Pulau Pinang, Malaysia.
| | - Nibras Abdullah
- Faculty of Computer Studies, Arab Open University, Jeddah, Saudi Arabia.
| | - Putra Sumari
- School of Computer Sciences, Universiti Sains Malaysia, USM, 11800, Pulau Pinang, Malaysia
| |
Collapse
|
27
|
Buriboev AS, Muhamediyeva D, Primova H, Sultanov D, Tashev K, Jeon HS. Concatenated CNN-Based Pneumonia Detection Using a Fuzzy-Enhanced Dataset. SENSORS (BASEL, SWITZERLAND) 2024; 24:6750. [PMID: 39460230 PMCID: PMC11510836 DOI: 10.3390/s24206750] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/19/2024] [Revised: 10/12/2024] [Accepted: 10/17/2024] [Indexed: 10/28/2024]
Abstract
Pneumonia is a form of acute respiratory infection affecting the lungs. Symptoms of viral and bacterial pneumonia are similar. Rapid diagnosis of the disease is difficult, since polymerase chain reaction-based methods, which have the greatest reliability, provide results in a few hours, while ensuring high requirements for compliance with the analysis technology and professionalism of the personnel. This study proposed a Concatenated CNN model for pneumonia detection combined with a fuzzy logic-based image improvement method. The fuzzy logic-based image enhancement process is based on a new fuzzification refinement algorithm, with significantly improved image quality and feature extraction for the CCNN model. Four datasets, original and upgraded images utilizing fuzzy entropy, standard deviation, and histogram equalization, were utilized to train the algorithm. The CCNN's performance was demonstrated to be significantly improved by the upgraded datasets, with the fuzzy entropy-added dataset producing the best results. The suggested CCNN attained remarkable classification metrics, including 98.9% accuracy, 99.3% precision, 99.8% F1-score, and 99.6% recall. Experimental comparisons showed that the fuzzy logic-based enhancement worked significantly better than traditional image enhancement methods, resulting in higher diagnostic precision. This study demonstrates how well deep learning models and sophisticated image enhancement techniques work together to analyze medical images.
Collapse
Affiliation(s)
- Abror Shavkatovich Buriboev
- School of Computing, Department of AI-Software, Gachon University, Seongnam-si 13306, Republic of Korea;
- Department of Infocommunication Engineering, Tashkent University of Information Technologies, Tashkent 100084, Uzbekistan; (D.S.); (K.T.)
| | - Dilnoz Muhamediyeva
- Tashkent Institute of Irrigation and Agricultural Mechanization Engineers, National Research University, Tashkent 100000, Uzbekistan;
| | - Holida Primova
- Department of IT, Samarkand Branch of Tashkent University of Information Technologies, Samarkand 140100, Uzbekistan;
| | - Djamshid Sultanov
- Department of Infocommunication Engineering, Tashkent University of Information Technologies, Tashkent 100084, Uzbekistan; (D.S.); (K.T.)
| | - Komil Tashev
- Department of Infocommunication Engineering, Tashkent University of Information Technologies, Tashkent 100084, Uzbekistan; (D.S.); (K.T.)
| | - Heung Seok Jeon
- Department of Computer Engineering, Konkuk University, Chungju 27478, Republic of Korea
| |
Collapse
|
28
|
Bouarroudj R, Bellala FZ, Souami F. High Capacity and Reversible Fragile Watermarking Method for Medical Image Authentication and Patient Data Hiding. J Med Syst 2024; 48:98. [PMID: 39425804 DOI: 10.1007/s10916-024-02110-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/20/2024] [Accepted: 09/17/2024] [Indexed: 10/21/2024]
Abstract
The exchange of medical images and patient data over the internet has attracted considerable attention in the past decade, driven by advancements in communication and health services. However, transferring confidential data through insecure channels, such as the internet, exposes it to potential manipulations and attacks. To ensure the authenticity of medical images while concealing patient data within them, this paper introduces a high-capacity and reversible fragile watermarking model in which an authentication watermark is initially generated from the cover image and merged with the patient's information, photo, and medical report to form the global watermark. This watermark is subsequently encrypted using the chaotic Chen system technique, enhancing the model's security and ensuring patient data confidentiality. The cover image then undergoes a Discrete Fourier Transform (DFT) and the encrypted watermark is inserted into the frequency coefficients using a new embedding technique. The experimental results demonstrate that the proposed method achieves great watermarked image quality, with a PSNR exceeding 113 dB and an SSIM close to 1, while maintaining a high embedding capacity of 3 BPP (Bits Per Pixel) and offering perfect reversibility. Furthermore, the proposed model demonstrates high sensitivity to attacks, successfully detecting tampering in all 18 tested attacks, and achieves nearly perfect watermark extraction accuracy, with a Bit Error Rate (BER) of 0.0004%. This high watermark extraction accuracy is crucial in our situation where patient data need to be retrieved from the watermarked images with almost no alteration.
Collapse
Affiliation(s)
- Riadh Bouarroudj
- LRIA, Department of Computer Science, University of Science and Technology Houari Boumediene (USTHB), B.P. 32, El Alia, 16111 Bab Ezzouar, Algiers, Algeria.
| | - Fatma Zohra Bellala
- LRIA, Department of Computer Science, University of Science and Technology Houari Boumediene (USTHB), B.P. 32, El Alia, 16111 Bab Ezzouar, Algiers, Algeria
| | - Feryel Souami
- LRIA, Department of Computer Science, University of Science and Technology Houari Boumediene (USTHB), B.P. 32, El Alia, 16111 Bab Ezzouar, Algiers, Algeria
- Department of Computer Science, University of Algiers 1, Algiers, Algeria
| |
Collapse
|
29
|
Yi S, Chen Z. MIDC: Medical image dataset cleaning framework based on deep learning. Heliyon 2024; 10:e38910. [PMID: 39444398 PMCID: PMC11497395 DOI: 10.1016/j.heliyon.2024.e38910] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/25/2024] [Revised: 09/17/2024] [Accepted: 10/02/2024] [Indexed: 10/25/2024] Open
Abstract
Deep learning technology is widely used in the field of medical imaging. Among them, Convolutional Neural Networks (CNNs) are the most widely used, and the quality of the dataset is crucial for the training of CNN diagnostic models, as mislabeled data can easily affect the accuracy of the diagnostic models. However, due to medical specialization, it is difficult for non-professional physicians to judge mislabeled medical image data. In this paper, we proposed a new framework named medical image dataset cleaning (MIDC), whose main contribution is to improve the quality of public datasets by automatically cleaning up mislabeled data. The main innovations of MIDC are: firstly, the framework innovatively utilizes multiple public datasets of the same disease, relying on different CNNs to automatically recognize images and remove mislabeled data to complete the data cleaning process. This process does not rely on annotations from professional physicians and does not require additional datasets with more reliable labels; Secondly, a novel grading rule is designed to divide the datasets into high-accuracy datasets and low-accuracy datasets, based on which the data cleaning process can be performed; Thirdly, a novel data cleaning module based on CNN is designed to identify and clean low-accuracy datasets by using high-accuracy datasets. In the experiments, the validity of the proposed framework was verified by using four kinds of datasets diabetic retinal, viral pneumonia, breast tumor, and skin cancer, with results showing an increase in the average diagnostic accuracy from 71.18 % to 85.13 %, 82.50 %to 93.79 %, 85.59 %to 93.45 %, and 84.55 %to 94.21 %, respectively. The proposed data cleaning framework MIDC could better help physicians diagnose diseases based on the dataset with mislabeled data.
Collapse
Affiliation(s)
- Sanli Yi
- School of Information Engineering and Automation, Kunming University of Science and Technology, Kunming, 650500, China
- Key Laboratory of Computer Technology Application of Yunnan Province, Kunming, 650500, China
| | - Ziyan Chen
- School of Information Engineering and Automation, Kunming University of Science and Technology, Kunming, 650500, China
- Key Laboratory of Computer Technology Application of Yunnan Province, Kunming, 650500, China
| |
Collapse
|
30
|
Veras Florentino PT, Araújo VDO, Zatti H, Luis CV, Cavalcanti CRS, de Oliveira MHC, Leão AHFF, Bertoldo Junior J, Barbosa GGC, Ravera E, Cebukin A, David RB, de Melo DBV, Machado TM, Bellei NCJ, Boaventura V, Barral-Netto M, Smaili SS. Text mining method to unravel long COVID's clinical condition in hospitalized patients. Cell Death Dis 2024; 15:671. [PMID: 39271699 PMCID: PMC11399332 DOI: 10.1038/s41419-024-07043-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/13/2024] [Revised: 08/28/2024] [Accepted: 08/29/2024] [Indexed: 09/15/2024]
Abstract
Long COVID is characterized by persistent that extends symptoms beyond established timeframes. Its varied presentation across different populations and healthcare systems poses significant challenges in understanding its clinical manifestations and implications. In this study, we present a novel application of text mining technique to automatically extract unstructured data from a long COVID survey conducted at a prominent university hospital in São Paulo, Brazil. Our phonetic text clustering (PTC) method enables the exploration of unstructured Electronic Healthcare Records (EHR) data to unify different written forms of similar terms into a single phonemic representation. We used n-gram text analysis to detect compound words and negated terms in Portuguese-BR, focusing on medical conditions and symptoms related to long COVID. By leveraging text mining, we aim to contribute to a deeper understanding of this chronic condition and its implications for healthcare systems globally. The model developed in this study has the potential for scalability and applicability in other healthcare settings, thereby supporting broader research efforts and informing clinical decision-making for long COVID patients.
Collapse
Affiliation(s)
- Pilar Tavares Veras Florentino
- Laboratório de Medicina e Saúde Pública de Precisão (MeSP2), Instituto Gonçalo Moniz, Fundação Oswaldo Cruz, Salvador, Brazil
- Centro de Integração de Dados e Conhecimentos para a Saúde (CIDACS), Instituto Gonçalo Moniz, Fundação Oswaldo Cruz, Salvador, Brazil
| | - Vinícius de Oliveira Araújo
- Centro de Integração de Dados e Conhecimentos para a Saúde (CIDACS), Instituto Gonçalo Moniz, Fundação Oswaldo Cruz, Salvador, Brazil
- Faculdade de Medicina da Bahia, Universidade Federal da Bahia, Salvador, Brazil
| | - Henrique Zatti
- Centro de Integração de Dados e Conhecimentos para a Saúde (CIDACS), Instituto Gonçalo Moniz, Fundação Oswaldo Cruz, Salvador, Brazil
| | - Caio Vinícius Luis
- Departamento de Farmacologia, Escola Paulista de Medicina, Universidade Federal de São Paulo, São Paulo, Brazil
| | | | | | | | - Juracy Bertoldo Junior
- Centro de Integração de Dados e Conhecimentos para a Saúde (CIDACS), Instituto Gonçalo Moniz, Fundação Oswaldo Cruz, Salvador, Brazil
| | - George G Caique Barbosa
- Centro de Integração de Dados e Conhecimentos para a Saúde (CIDACS), Instituto Gonçalo Moniz, Fundação Oswaldo Cruz, Salvador, Brazil
| | - Ernesto Ravera
- Departamento de Farmacologia, Escola Paulista de Medicina, Universidade Federal de São Paulo, São Paulo, Brazil
| | - Alberto Cebukin
- Departamento de Farmacologia, Escola Paulista de Medicina, Universidade Federal de São Paulo, São Paulo, Brazil
| | - Renata Bernardes David
- Departamento de Farmacologia, Escola Paulista de Medicina, Universidade Federal de São Paulo, São Paulo, Brazil
| | | | - Tales Mota Machado
- Centro de Integração de Dados e Conhecimentos para a Saúde (CIDACS), Instituto Gonçalo Moniz, Fundação Oswaldo Cruz, Salvador, Brazil
- Diretoria de Tecnologia da Informação, Universidade Federal de Ouro Preto, Ouro Preto, Brazil
| | - Nancy C J Bellei
- Disciplina de Moléstias Infecciosas, Escola Paulista de Medicina, Universidade Federal de São Paulo, São Paulo, Brazil
| | - Viviane Boaventura
- Laboratório de Medicina e Saúde Pública de Precisão (MeSP2), Instituto Gonçalo Moniz, Fundação Oswaldo Cruz, Salvador, Brazil
- Faculdade de Medicina da Bahia, Universidade Federal da Bahia, Salvador, Brazil
| | - Manoel Barral-Netto
- Laboratório de Medicina e Saúde Pública de Precisão (MeSP2), Instituto Gonçalo Moniz, Fundação Oswaldo Cruz, Salvador, Brazil.
- Faculdade de Medicina da Bahia, Universidade Federal da Bahia, Salvador, Brazil.
| | - Soraya S Smaili
- Departamento de Farmacologia, Escola Paulista de Medicina, Universidade Federal de São Paulo, São Paulo, Brazil.
| |
Collapse
|
31
|
Lee SH, Fox S, Smith R, Skrobarcek KA, Keyserling H, Phares CR, Lee D, Posey DL. Development and validation of a deep learning model for detecting signs of tuberculosis on chest radiographs among US-bound immigrants and refugees. PLOS DIGITAL HEALTH 2024; 3:e0000612. [PMID: 39348377 PMCID: PMC11441656 DOI: 10.1371/journal.pdig.0000612] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/08/2024] [Accepted: 08/12/2024] [Indexed: 10/02/2024]
Abstract
Immigrants and refugees seeking admission to the United States must first undergo an overseas medical exam, overseen by the US Centers for Disease Control and Prevention (CDC), during which all persons ≥15 years old receive a chest x-ray to look for signs of tuberculosis. Although individual screening sites often implement quality control (QC) programs to ensure radiographs are interpreted correctly, the CDC does not currently have a method for conducting similar QC reviews at scale. We obtained digitized chest radiographs collected as part of the overseas immigration medical exam. Using radiographs from applicants 15 years old and older, we trained deep learning models to perform three tasks: identifying abnormal radiographs; identifying abnormal radiographs suggestive of tuberculosis; and identifying the specific findings (e.g., cavities or infiltrates) in abnormal radiographs. We then evaluated the models on both internal and external testing datasets, focusing on two classes of performance metrics: individual-level metrics, like sensitivity and specificity, and sample-level metrics, like accuracy in predicting the prevalence of abnormal radiographs. A total of 152,012 images (one image per applicant; mean applicant age 39 years) were used for model training. On our internal test dataset, our models performed well both in identifying abnormalities suggestive of TB (area under the curve [AUC] of 0.97; 95% confidence interval [CI]: 0.95, 0.98) and in estimating sample-level counts of the same (-2% absolute percentage error; 95% CIC: -8%, 6%). On the external test datasets, our models performed similarly well in identifying both generic abnormalities (AUCs ranging from 0.89 to 0.92) and those suggestive of TB (AUCs from 0.94 to 0.99). This performance was consistent across metrics, including those based on thresholded class predictions, like sensitivity, specificity, and F1 score. Strong performance relative to high-quality radiological reference standards across a variety of datasets suggests our models may make reliable tools for supporting chest radiography QC activities at CDC.
Collapse
Affiliation(s)
- Scott H. Lee
- National Center for Emerging and Zoonotic Infectious Diseases, US Centers for Disease Control and Prevention, Atlanta, Georgia, United States of America
| | - Shannon Fox
- National Center for Emerging and Zoonotic Infectious Diseases, US Centers for Disease Control and Prevention, Atlanta, Georgia, United States of America
| | - Raheem Smith
- National Center for Emerging and Zoonotic Infectious Diseases, US Centers for Disease Control and Prevention, Atlanta, Georgia, United States of America
| | - Kimberly A. Skrobarcek
- National Center for Emerging and Zoonotic Infectious Diseases, US Centers for Disease Control and Prevention, Atlanta, Georgia, United States of America
| | | | - Christina R. Phares
- National Center for Emerging and Zoonotic Infectious Diseases, US Centers for Disease Control and Prevention, Atlanta, Georgia, United States of America
| | - Deborah Lee
- National Center for Emerging and Zoonotic Infectious Diseases, US Centers for Disease Control and Prevention, Atlanta, Georgia, United States of America
| | - Drew L. Posey
- National Center for Emerging and Zoonotic Infectious Diseases, US Centers for Disease Control and Prevention, Atlanta, Georgia, United States of America
| |
Collapse
|
32
|
Senthil R, Anand T, Somala CS, Saravanan KM. Bibliometric analysis of artificial intelligence in healthcare research: Trends and future directions. Future Healthc J 2024; 11:100182. [PMID: 39310219 PMCID: PMC11414662 DOI: 10.1016/j.fhj.2024.100182] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/03/2024] [Revised: 08/06/2024] [Accepted: 08/30/2024] [Indexed: 09/25/2024]
Abstract
Objective The presence of artificial intelligence (AI) in healthcare is a powerful and game-changing force that is completely transforming the industry as a whole. Using sophisticated algorithms and data analytics, AI has unparalleled prospects for improving patient care, streamlining operational efficiency, and fostering innovation across the healthcare ecosystem. This study conducts a comprehensive bibliometric analysis of research on AI in healthcare, utilising the SCOPUS database as the primary data source. Methods Preliminary findings from 2013 identified 153 publications on AI and healthcare. Between 2019 and 2023, the number of publications increased exponentially, indicating significant growth and development in the field. The analysis employs various bibliometric indicators to assess research production performance, science mapping techniques, and thematic mapping analysis. Results The study reveals insights into research hotspots, thematic focus, and emerging trends in AI and healthcare research. Based on an extensive examination of the Scopus database provides a brief overview and suggests potential avenues for further investigation. Conclusion This article provides valuable contributions to understanding the current landscape of AI in healthcare, offering insights for future research directions and informing strategic decision making in the field.
Collapse
Affiliation(s)
- Renganathan Senthil
- Department of Bioinformatics, School of Lifesciences, Vels Institute of Science Technology and Advanced Studies (VISTAS), Pallavaram, Chennai 600117, Tamil Nadu, India
| | - Thirunavukarasou Anand
- SRIIC Lab, Faculty of Clinical Research, Sri Ramachandra Institute of Higher Education and Research, Chennai 600116, Tamil Nadu, India
- B Aatral Biosciences Private Limited, Bangalore 560091, Karnataka, India
| | | | - Konda Mani Saravanan
- B Aatral Biosciences Private Limited, Bangalore 560091, Karnataka, India
- Department of Biotechnology, Bharath Institute of Higher Education and Research, Chennai 600073, Tamil Nadu, India
| |
Collapse
|
33
|
Ejiyi CJ, Qin Z, Ejiyi MB, Ukwuoma C, Ejiyi TU, Muoka GW, Gyarteng ESA, Bamisile OO. MACCoM: A multiple attention and convolutional cross-mixer framework for detailed 2D biomedical image segmentation. Comput Biol Med 2024; 179:108847. [PMID: 39004046 DOI: 10.1016/j.compbiomed.2024.108847] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/19/2024] [Revised: 06/28/2024] [Accepted: 07/02/2024] [Indexed: 07/16/2024]
Abstract
The UNet architecture, which is widely used for biomedical image segmentation, has limitations like blurred feature maps and over- or under-segmented regions. To overcome these limitations, we propose a novel network architecture called MACCoM (Multiple Attention and Convolutional Cross-Mixer) - an end-to-end depthwise encoder-decoder fully convolutional network designed for binary and multi-class biomedical image segmentation built upon deeperUNet. We proposed a multi-scope attention module (MSAM) that allows the model to attend to diverse scale features, preserving fine details and high-level semantic information thus useful at the encoder-decoder connection. As the depth increases, our proposed spatial multi-head attention (SMA) is added to facilitate inter-layer communication and information exchange, enabling the network to effectively capture long-range dependencies and global context. MACCoM is also equipped with a convolutional cross-mixer we proposed to enhance the feature extraction capability of the model. By incorporating these modules, we effectively combine semantically similar features and reduce artifacts during the early stages of training. Experimental results on 4 biomedical datasets crafted from 3 datasets of varying modalities consistently demonstrate that MACCoM outperforms or matches state-of-the-art baselines in the segmentation tasks. With Breast Ultrasound Image (BUSI), MACCoM recorded 99.06 % Jaccard, 77.58 % Dice, and 93.92 % Accuracy, while recording 99.50 %, 98.44 %, and 99.29 % respectively for Jaccard, Dice, and Accuracy on the Chest X-ray (CXR) images used. The Jaccard, Dice, and Accuracy for the High-Resolution Fundus (HRF) images are 95.77 %, 74.35 %, and 95.95 % respectively. The findings here highlight MACCoM's effectiveness in improving segmentation performance and its valuable potential in image analysis.
Collapse
Affiliation(s)
- Chukwuebuka Joseph Ejiyi
- Network and Data Security Key Laboratory of Sichuan Province, University of Electronic Science and Technology of China, Chengdu, Sichuan, China; Sichuan Industrial Internet Intelligent Monitoring and Application Engineering Research Center, Chengdu University of Technology, Chengdu, China
| | - Zhen Qin
- Network and Data Security Key Laboratory of Sichuan Province, University of Electronic Science and Technology of China, Chengdu, Sichuan, China.
| | | | - Chiagoziem Ukwuoma
- Network and Data Security Key Laboratory of Sichuan Province, University of Electronic Science and Technology of China, Chengdu, Sichuan, China
| | - Thomas Ugochukwu Ejiyi
- Department of Pure and Industrial Chemistry, University of Nigeria Nsukka, Nsukka, Enugu State, Nigeria
| | - Gladys Wavinya Muoka
- Network and Data Security Key Laboratory of Sichuan Province, University of Electronic Science and Technology of China, Chengdu, Sichuan, China
| | - Emmanuel S A Gyarteng
- School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, China
| | - Olusola O Bamisile
- Sichuan Industrial Internet Intelligent Monitoring and Application Engineering Research Center, Chengdu University of Technology, Chengdu, China
| |
Collapse
|
34
|
Yan C, Yan H, Liang W, Yin M, Luo H, Luo J. DP-SSLoRA: A privacy-preserving medical classification model combining differential privacy with self-supervised low-rank adaptation. Comput Biol Med 2024; 179:108792. [PMID: 38964242 DOI: 10.1016/j.compbiomed.2024.108792] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2023] [Revised: 05/31/2024] [Accepted: 06/18/2024] [Indexed: 07/06/2024]
Abstract
BACKGROUND AND OBJECTIVE Concerns about patient privacy issues have limited the application of medical deep learning models in certain real-world scenarios. Differential privacy (DP) can alleviate this problem by injecting random noise into the model. However, naively applying DP to medical models will not achieve a satisfactory balance between privacy and utility due to the high dimensionality of medical models and the limited labeled samples. METHODS This work proposed the DP-SSLoRA model, a privacy-preserving classification model for medical images combining differential privacy with self-supervised low-rank adaptation. In this work, a self-supervised pre-training method is used to obtain enhanced representations from unlabeled publicly available medical data. Then, a low-rank decomposition method is employed to mitigate the impact of differentially private noise and combined with pre-trained features to conduct the classification task on private datasets. RESULTS In the classification experiments using three real chest-X ray datasets, DP-SSLoRA achieves good performance with strong privacy guarantees. Under the premise of ɛ=2, with the AUC of 0.942 in RSNA, the AUC of 0.9658 in Covid-QU-mini, and the AUC of 0.9886 in Chest X-ray 15k. CONCLUSION Extensive experiments on real chest X-ray datasets show that DP-SSLoRA can achieve satisfactory performance with stronger privacy guarantees. This study provides guidance for studying privacy-preserving in the medical field. Source code is publicly available online. https://github.com/oneheartforone/DP-SSLoRA.
Collapse
Affiliation(s)
- Chaokun Yan
- School of Computer and Information Engineering, Henan University, Kaifeng, 475004, Henan, China; Academy for Advanced Interdisciplinary Studies, Henan University, Kaifeng, 475004, Henan, China; Henan Engineering Research Center of Intelligent Technology and Application, Henan University, Kaifeng, 475004, Henan, China
| | - Haicao Yan
- School of Computer and Information Engineering, Henan University, Kaifeng, 475004, Henan, China
| | - Wenjuan Liang
- School of Computer and Information Engineering, Henan University, Kaifeng, 475004, Henan, China; Academy for Advanced Interdisciplinary Studies, Henan University, Kaifeng, 475004, Henan, China; Henan Engineering Research Center of Intelligent Technology and Application, Henan University, Kaifeng, 475004, Henan, China.
| | - Menghan Yin
- School of Computer and Information Engineering, Henan University, Kaifeng, 475004, Henan, China
| | - Huimin Luo
- School of Computer and Information Engineering, Henan University, Kaifeng, 475004, Henan, China; Academy for Advanced Interdisciplinary Studies, Henan University, Kaifeng, 475004, Henan, China; Henan Engineering Research Center of Intelligent Technology and Application, Henan University, Kaifeng, 475004, Henan, China
| | - Junwei Luo
- School of Software, Henan Polytecgnic University, Jiaozuo, 454000, Henan, China
| |
Collapse
|
35
|
Abdel-Salam M, Hu G, Çelik E, Gharehchopogh FS, El-Hasnony IM. Chaotic RIME optimization algorithm with adaptive mutualism for feature selection problems. Comput Biol Med 2024; 179:108803. [PMID: 38955125 DOI: 10.1016/j.compbiomed.2024.108803] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2024] [Revised: 05/17/2024] [Accepted: 06/24/2024] [Indexed: 07/04/2024]
Abstract
The RIME optimization algorithm is a newly developed physics-based optimization algorithm used for solving optimization problems. The RIME algorithm proved high-performing in various fields and domains, providing a high-performance solution. Nevertheless, like many swarm-based optimization algorithms, RIME suffers from many limitations, including the exploration-exploitation balance not being well balanced. In addition, the likelihood of falling into local optimal solutions is high, and the convergence speed still needs some work. Hence, there is room for enhancement in the search mechanism so that various search agents can discover new solutions. The authors suggest an adaptive chaotic version of the RIME algorithm named ACRIME, which incorporates four main improvements, including an intelligent population initialization using chaotic maps, a novel adaptive modified Symbiotic Organism Search (SOS) mutualism phase, a novel mixed mutation strategy, and the utilization of restart strategy. The main goal of these improvements is to improve the variety of the population, achieve a better balance between exploration and exploitation, and improve RIME's local and global search abilities. The study assesses the effectiveness of ACRIME by using the standard benchmark functions of the CEC2005 and CEC2019 benchmarks. The proposed ACRIME is also applied as a feature selection to fourteen various datasets to test its applicability to real-world problems. Besides, the ACRIME algorithm is applied to the COVID-19 classification real problem to test its applicability and performance further. The suggested algorithm is compared to other sophisticated classical and advanced metaheuristics, and its performance is assessed using statistical tests such as Wilcoxon rank-sum and Friedman rank tests. The study demonstrates that ACRIME exhibits a high level of competitiveness and often outperforms competing algorithms. It discovers the optimal subset of features, enhancing the accuracy of classification and minimizing the number of features employed. This study primarily focuses on enhancing the equilibrium between exploration and exploitation, extending the scope of local search.
Collapse
Affiliation(s)
- Mahmoud Abdel-Salam
- Faculty of Computer and Information Science, Mansoura University, Mansoura, 35516, Egypt.
| | - Gang Hu
- Department of Applied Mathematics, Xi'an University of Technology, Xi'an, 710054, PR China
| | - Emre Çelik
- Department of Electrical and Electronics Engineering, Faculty of Engineering, Düzce University, Düzce, Turkey
| | | | - Ibrahim M El-Hasnony
- Faculty of Computer and Information Science, Mansoura University, Mansoura, 35516, Egypt
| |
Collapse
|
36
|
Ding H, Fan L, Zhang J, Gao G. Deep Learning-Based System Combining Chest X-Ray and Computerized Tomography Images for COVID-19 Diagnosis. Br J Hosp Med (Lond) 2024; 85:1-15. [PMID: 39212565 DOI: 10.12968/hmed.2024.0244] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/04/2024]
Abstract
Aims/Background: The coronavirus disease 2019 (COVID-19) pandemic has highlighted the need for accurate and efficient diagnostic methods. This study aims to improve COVID-19 detection by integrating chest X-ray (CXR) and computerized tomography (CT) images using deep learning techniques, further improving diagnostic accuracy by using a combined imaging approach. Methods: The study used two publicly accessible databases, COVID-19 Questionnaires for Understanding the Exposure (COVID-QU-Ex) and Integrated Clinical and Translational Cancer Foundation (iCTCF), containing CXR and CT images, respectively. The proposed system employed convolutional neural networks (CNNs) for classification, specifically EfficientNet and ResNet architectures. The data underwent preprocessing steps, including image resizing, Gaussian noise addition, and data augmentation. The dataset was divided into training, validation, and test sets. Gradient-weighted Class Activation Mapping (Grad-CAM) was used for model interpretability. Results: The EfficientNet-based models outperformed the ResNet-based models across all metrics. The highest accuracy achieved was 99.44% for CXR images and 99.81% for CT images with EfficientNetB5. The models also demonstrated high precision, recall, and F1 scores. For statistical significance, the p-values were less than 0.05, indicating that the results are significant. Conclusion: Integrating CXR and CT images using deep learning significantly improves the accuracy of COVID-19 diagnosis. The EfficientNet-based models, with their superior feature extraction capabilities, show better performance than ResNet models. Grad-CAM Visualizations provide insights into the model's decision-making process, potentially reducing diagnostic errors and accelerating diagnosis processes. This approach can improve patient care and support healthcare systems in managing the pandemic more effectively.
Collapse
Affiliation(s)
- Hui Ding
- Department of Radiology, Ningbo No.2 Hospital, Ningbo, Zhejiang, China
| | - Lingyan Fan
- Department of Acute Infectious Diseases, Ningbo No.2 Hospital, Ningbo, Zhejiang, China
| | - Jingfeng Zhang
- Department of Radiology, Ningbo No.2 Hospital, Ningbo, Zhejiang, China
| | - Guosheng Gao
- Department of Clinical Laboratory, Ningbo No.2 Hospital, Ningbo, Zhejiang, China
| |
Collapse
|
37
|
Singh T, Mishra S, Kalra R, Satakshi, Kumar M, Kim T. COVID-19 severity detection using chest X-ray segmentation and deep learning. Sci Rep 2024; 14:19846. [PMID: 39191941 PMCID: PMC11349901 DOI: 10.1038/s41598-024-70801-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/24/2024] [Accepted: 08/21/2024] [Indexed: 08/29/2024] Open
Abstract
COVID-19 has resulted in a significant global impact on health, the economy, education, and daily life. The disease can range from mild to severe, with individuals over 65 or those with underlying medical conditions being more susceptible to severe illness. Early testing and isolation are vital due to the virus's variable incubation period. Chest radiographs (CXR) have gained importance as a diagnostic tool due to their efficiency and reduced radiation exposure compared to CT scans. However, the sensitivity of CXR in detecting COVID-19 may be lower. This paper introduces a deep learning framework for accurate COVID-19 classification and severity prediction using CXR images. U-Net is used for lung segmentation, achieving a precision of 0.9924. Classification is performed using a Convulation-capsule network, with high true positive rates of 86% for COVID-19, 93% for pneumonia, and 85% for normal cases. Severity assessment employs ResNet50, VGG-16, and DenseNet201, with DenseNet201 showing superior accuracy. Empirical results, validated with 95% confidence intervals, confirm the framework's reliability and robustness. This integration of advanced deep learning techniques with radiological imaging enhances early detection and severity assessment, improving patient management and resource allocation in clinical settings.
Collapse
Affiliation(s)
- Tinku Singh
- School of Information and Communication Engineering, Chungbuk National University, Cheongju, South Korea
| | - Suryanshi Mishra
- Department of Mathematics & Statistics, SHUATS, Prayagraj, Uttar Pradesh, India
| | - Riya Kalra
- Indian Institute of Information Technology Allahabad, Prayagraj, Uttar Pradesh, India
| | - Satakshi
- Department of Mathematics & Statistics, SHUATS, Prayagraj, Uttar Pradesh, India
| | - Manish Kumar
- Indian Institute of Information Technology Allahabad, Prayagraj, Uttar Pradesh, India
| | - Taehong Kim
- School of Information and Communication Engineering, Chungbuk National University, Cheongju, South Korea.
| |
Collapse
|
38
|
Sannasi Chakravarthy SR, Bharanidharan N, Vinothini C, Vinoth Kumar V, Mahesh TR, Guluwadi S. Adaptive Mish activation and ranger optimizer-based SEA-ResNet50 model with explainable AI for multiclass classification of COVID-19 chest X-ray images. BMC Med Imaging 2024; 24:206. [PMID: 39123118 PMCID: PMC11313131 DOI: 10.1186/s12880-024-01394-2] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/04/2024] [Accepted: 08/06/2024] [Indexed: 08/12/2024] Open
Abstract
A recent global health crisis, COVID-19 is a significant global health crisis that has profoundly affected lifestyles. The detection of such diseases from similar thoracic anomalies using medical images is a challenging task. Thus, the requirement of an end-to-end automated system is vastly necessary in clinical treatments. In this way, the work proposes a Squeeze-and-Excitation Attention-based ResNet50 (SEA-ResNet50) model for detecting COVID-19 utilizing chest X-ray data. Here, the idea lies in improving the residual units of ResNet50 using the squeeze-and-excitation attention mechanism. For further enhancement, the Ranger optimizer and adaptive Mish activation function are employed to improve the feature learning of the SEA-ResNet50 model. For evaluation, two publicly available COVID-19 radiographic datasets are utilized. The chest X-ray input images are augmented during experimentation for robust evaluation against four output classes namely normal, pneumonia, lung opacity, and COVID-19. Then a comparative study is done for the SEA-ResNet50 model against VGG-16, Xception, ResNet18, ResNet50, and DenseNet121 architectures. The proposed framework of SEA-ResNet50 together with the Ranger optimizer and adaptive Mish activation provided maximum classification accuracies of 98.38% (multiclass) and 99.29% (binary classification) as compared with the existing CNN architectures. The proposed method achieved the highest Kappa validation scores of 0.975 (multiclass) and 0.98 (binary classification) over others. Furthermore, the visualization of the saliency maps of the abnormal regions is represented using the explainable artificial intelligence (XAI) model, thereby enhancing interpretability in disease diagnosis.
Collapse
Affiliation(s)
- S R Sannasi Chakravarthy
- Department of Electronics and Communication Engineering, Bannari Amman Institute of Technology, Sathyamangalam, India
| | - N Bharanidharan
- School of Computer Science Engineering and Information Systems, Vellore Institute of Technology, Vellore, 632014, India
| | - C Vinothini
- Department of Computer Science and Engineering, Dayananda Sagar College of Engineering, Bangalore, India
| | - Venkatesan Vinoth Kumar
- School of Computer Science Engineering and Information Systems, Vellore Institute of Technology, Vellore, 632014, India
| | - T R Mahesh
- Department of Computer Science and Engineering, JAIN (Deemed-to-Be University), Bengaluru, 562112, India
| | - Suresh Guluwadi
- Adama Science and Technology University, Adama, 302120, Ethiopia.
| |
Collapse
|
39
|
Rahman MF, Tseng TL(B, Pokojovy M, McCaffrey P, Walser E, Moen S, Vo A, Ho JC. Machine-Learning-Enabled Diagnostics with Improved Visualization of Disease Lesions in Chest X-ray Images. Diagnostics (Basel) 2024; 14:1699. [PMID: 39202188 PMCID: PMC11353848 DOI: 10.3390/diagnostics14161699] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2024] [Revised: 07/31/2024] [Accepted: 08/02/2024] [Indexed: 09/03/2024] Open
Abstract
The class activation map (CAM) represents the neural-network-derived region of interest, which can help clarify the mechanism of the convolutional neural network's determination of any class of interest. In medical imaging, it can help medical practitioners diagnose diseases like COVID-19 or pneumonia by highlighting the suspicious regions in Computational Tomography (CT) or chest X-ray (CXR) film. Many contemporary deep learning techniques only focus on COVID-19 classification tasks using CXRs, while few attempt to make it explainable with a saliency map. To fill this research gap, we first propose a VGG-16-architecture-based deep learning approach in combination with image enhancement, segmentation-based region of interest (ROI) cropping, and data augmentation steps to enhance classification accuracy. Later, a multi-layer Gradient CAM (ML-Grad-CAM) algorithm is integrated to generate a class-specific saliency map for improved visualization in CXR images. We also define and calculate a Severity Assessment Index (SAI) from the saliency map to quantitatively measure infection severity. The trained model achieved an accuracy score of 96.44% for the three-class CXR classification task, i.e., COVID-19, pneumonia, and normal (healthy patients), outperforming many existing techniques in the literature. The saliency maps generated from the proposed ML-GRAD-CAM algorithm are compared with the original Gran-CAM algorithm.
Collapse
Affiliation(s)
- Md Fashiar Rahman
- Department of Industrial, Manufacturing and Systems Engineering, The University of Texas, El Paso, TX 79968, USA
| | - Tzu-Liang (Bill) Tseng
- Department of Industrial, Manufacturing and Systems Engineering, The University of Texas, El Paso, TX 79968, USA
| | - Michael Pokojovy
- Department of Mathematics and Statistics, Old Dominion University, Norfolk, VA 23529, USA;
| | - Peter McCaffrey
- Department of Radiology, The University of Texas Medical Branch, Galveston, TX 77550, USA; (P.M.); (E.W.); (S.M.); (A.V.)
| | - Eric Walser
- Department of Radiology, The University of Texas Medical Branch, Galveston, TX 77550, USA; (P.M.); (E.W.); (S.M.); (A.V.)
| | - Scott Moen
- Department of Radiology, The University of Texas Medical Branch, Galveston, TX 77550, USA; (P.M.); (E.W.); (S.M.); (A.V.)
| | - Alex Vo
- Department of Radiology, The University of Texas Medical Branch, Galveston, TX 77550, USA; (P.M.); (E.W.); (S.M.); (A.V.)
| | - Johnny C. Ho
- Department of Management and Marketing, Turner College of Business, Columbus State University, Columbus, GA 31907, USA;
| |
Collapse
|
40
|
Liu Z, Xu J, Yin C, Han G, Che Y, Fan G, Li X, Xie L, Bao L, Peng Z, Wang J, Chen Y, Zhang F, Ouyang W, Wang S, Guo J, Ma Y, Meng X, Fan T, Zhi A, Dawaciren, Yi K, You T, Yang Y, Liu J, Shi Y, Huang Y, Pan X. Development and External Validation of an Artificial Intelligence-Based Method for Scalable Chest Radiograph Diagnosis: A Multi-Country Cross-Sectional Study. RESEARCH (WASHINGTON, D.C.) 2024; 7:0426. [PMID: 39109248 PMCID: PMC11301699 DOI: 10.34133/research.0426] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/26/2024] [Accepted: 06/22/2024] [Indexed: 03/30/2025]
Abstract
Problem: Chest radiography is a crucial tool for diagnosing thoracic disorders, but interpretation errors and a lack of qualified practitioners can cause delays in treatment. Aim: This study aimed to develop a reliable multi-classification artificial intelligence (AI) tool to improve the accuracy and efficiency of chest radiograph diagnosis. Methods: We developed a convolutional neural network (CNN) capable of distinguishing among 26 thoracic diagnoses. The model was trained and externally validated using 795,055 chest radiographs from 13 datasets across 4 countries. Results: The CNN model achieved an average area under the curve (AUC) of 0.961 across all 26 diagnoses in the testing set. COVID-19 detection achieved perfect accuracy (AUC 1.000, [95% confidence interval {CI}, 1.000 to 1.000]), while effusion or pleural effusion detection showed the lowest accuracy (AUC 0.8453, [95% CI, 0.8417 to 0.8489]). In external validation, the model demonstrated strong reproducibility and generalizability within the local dataset, achieving an AUC of 0.9634 for lung opacity detection (95% CI, 0.9423 to 0.9702). The CNN outperformed both radiologists and nonradiological physicians, particularly in trans-device image recognition. Even for diseases not specifically trained on, such as aortic dissection, the AI model showed considerable scalability and enhanced diagnostic accuracy for physicians of varying experience levels (all P < 0.05). Additionally, our model exhibited no gender bias (P > 0.05). Conclusion: The developed AI algorithm, now available as professional web-based software, substantively improves chest radiograph interpretation. This research advances medical imaging and offers substantial diagnostic support in clinical settings.
Collapse
Affiliation(s)
- Zeye Liu
- Department of Cardiac Surgery,
Peking University People’s Hospital, Peking University, Xicheng District, Beijing, China
- Department of Structural Heart Disease, National Center for Cardiovascular Disease, China & Fuwai Hospital,
Chinese Academy of Medical Sciences & Peking Union Medical College, Beijing 100037, China
- National Health Commission Key Laboratory of Cardiovascular Regeneration Medicine, Beijing 100037, China
- Key Laboratory of Innovative Cardiovascular Devices,
Chinese Academy of Medical Sciences, Beijing 100037, China
- National Clinical Research Center for Cardiovascular Diseases, Fuwai Hospital,
Chinese Academy of Medical Sciences, Beijing 100037, China
| | - Jing Xu
- State Key Laboratory of Cardiovascular Disease, Fuwai Hospital, National Center for Cardiovascular Diseases, Fuwai Hospital, Chinese Academy of Medical Sciences,
and Peking Union Medical College, Beijing, China
| | - Chengliang Yin
- Medical Big Data Research Center, Medical Innovation Research Division of Chinese PLA General Hospital, Beijing, China
- National Engineering Research Center for Medical Big Data Application Technology, Chinese People’s Liberation Army (PLA) General Hospital, Beijing, China
| | - Guojing Han
- College of Pulmonary & Critical Care Medicine, Chinese PLA General Hospital, Beijing, China
| | - Yue Che
- Center for Health Policy Research and Evaluation,
Renmin University of China, Beijing, China
- School of Public Administration and Policy,
Renmin University of China, Beijing, China
| | - Ge Fan
- Lightspeed & Quantum Studios, Tencent Inc., Shenzhen, China
| | - Xiaofei Li
- Department of Cardiology, Fuwai Hospital,
Chinese Academy of Medical Sciences & Peking Union Medical College, Beijing, China
| | - Lixin Xie
- College of Pulmonary & Critical Care Medicine, Chinese PLA General Hospital, Beijing, China
| | - Lei Bao
- Shenzhen Benevolence Medical Sci&Tech Co. Ltd., Shenzhen, China
| | - Zimin Peng
- Shenzhen Benevolence Medical Sci&Tech Co. Ltd., Shenzhen, China
| | - Jinduo Wang
- University of Science and Technology of China, School of Cyber Science and Technology, Hefei 230000, China
| | - Yan Chen
- University of Science and Technology of China, School of Cyber Science and Technology, Hefei 230000, China
| | - Fengwen Zhang
- Department of Structural Heart Disease, National Center for Cardiovascular Disease, China & Fuwai Hospital,
Chinese Academy of Medical Sciences & Peking Union Medical College, Beijing 100037, China
- National Health Commission Key Laboratory of Cardiovascular Regeneration Medicine, Beijing 100037, China
- Key Laboratory of Innovative Cardiovascular Devices,
Chinese Academy of Medical Sciences, Beijing 100037, China
- National Clinical Research Center for Cardiovascular Diseases, Fuwai Hospital,
Chinese Academy of Medical Sciences, Beijing 100037, China
| | - Wenbin Ouyang
- Department of Structural Heart Disease, National Center for Cardiovascular Disease, China & Fuwai Hospital,
Chinese Academy of Medical Sciences & Peking Union Medical College, Beijing 100037, China
- National Health Commission Key Laboratory of Cardiovascular Regeneration Medicine, Beijing 100037, China
- Key Laboratory of Innovative Cardiovascular Devices,
Chinese Academy of Medical Sciences, Beijing 100037, China
- National Clinical Research Center for Cardiovascular Diseases, Fuwai Hospital,
Chinese Academy of Medical Sciences, Beijing 100037, China
| | - Shouzheng Wang
- Department of Structural Heart Disease, National Center for Cardiovascular Disease, China & Fuwai Hospital,
Chinese Academy of Medical Sciences & Peking Union Medical College, Beijing 100037, China
- National Health Commission Key Laboratory of Cardiovascular Regeneration Medicine, Beijing 100037, China
- Key Laboratory of Innovative Cardiovascular Devices,
Chinese Academy of Medical Sciences, Beijing 100037, China
- National Clinical Research Center for Cardiovascular Diseases, Fuwai Hospital,
Chinese Academy of Medical Sciences, Beijing 100037, China
| | - Junwei Guo
- Department of Respiratory and Critical Care Medicine, Peking Union Medical College Hospital,
Chinese Academy of Medical Sciences & Peking Union Medical College, Beijing, China
| | - Yanqiu Ma
- Peking University Third Hospital, Beijing, China
| | - Xiangzhi Meng
- Department of Thoracic Surgical Oncology, National Cancer Center/Cancer Hospital,
Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing 100021, China
| | - Taibing Fan
- Department of Pediatric Cardiac Surgery,
Zhengzhou University Fuwai Central China Cardiovascular Hospital, Zhengzhou, Henan 450000, China
| | - Aihua Zhi
- Fuwai Yunnan Cardiovascular Hospital, Department of Medical Imaging, Kunming 650000, China
| | - Dawaciren
- The Autonomous Region People’s Hospital, Xizang, China
| | - Kang Yi
- Department of Cardiovascular Surgery, Gansu Provincial Hospital, Lanzhou, China
- Gansu International Scientific and Technological Cooperation Base of Diagnosis and Treatment of Congenital Heart Disease, Lanzhou, China
| | - Tao You
- Department of Cardiovascular Surgery, Gansu Provincial Hospital, Lanzhou, China
- Gansu International Scientific and Technological Cooperation Base of Diagnosis and Treatment of Congenital Heart Disease, Lanzhou, China
| | - Yuejin Yang
- State Key Laboratory of Cardiovascular Disease, Fuwai Hospital, National Center for Cardiovascular Diseases, Fuwai Hospital, Chinese Academy of Medical Sciences,
and Peking Union Medical College, Beijing, China
| | - Jue Liu
- National Health Commission Key Laboratory of Cardiovascular Regeneration Medicine, Beijing 100037, China
- National Clinical Research Center for Cardiovascular Diseases, Fuwai Hospital,
Chinese Academy of Medical Sciences, Beijing 100037, China
| | - Yi Shi
- Department of Cardiac Surgery,
Peking University People’s Hospital, Peking University, Xicheng District, Beijing, China
| | - Yuan Huang
- State Key Laboratory of Cardiovascular Disease, Fuwai Hospital, National Center for Cardiovascular Diseases, Fuwai Hospital, Chinese Academy of Medical Sciences,
and Peking Union Medical College, Beijing, China
| | - Xiangbin Pan
- Department of Structural Heart Disease, National Center for Cardiovascular Disease, China & Fuwai Hospital,
Chinese Academy of Medical Sciences & Peking Union Medical College, Beijing 100037, China
- National Health Commission Key Laboratory of Cardiovascular Regeneration Medicine, Beijing 100037, China
- Key Laboratory of Innovative Cardiovascular Devices,
Chinese Academy of Medical Sciences, Beijing 100037, China
- National Clinical Research Center for Cardiovascular Diseases, Fuwai Hospital,
Chinese Academy of Medical Sciences, Beijing 100037, China
| |
Collapse
|
41
|
Lamprou V, Kallipolitis A, Maglogiannis I. On the evaluation of deep learning interpretability methods for medical images under the scope of faithfulness. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 253:108238. [PMID: 38823117 DOI: 10.1016/j.cmpb.2024.108238] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/05/2023] [Revised: 05/01/2024] [Accepted: 05/21/2024] [Indexed: 06/03/2024]
Abstract
BACKGROUND AND OBJECTIVE Evaluating the interpretability of Deep Learning models is crucial for building trust and gaining insights into their decision-making processes. In this work, we employ class activation map based attribution methods in a setting where only High-Resolution Class Activation Mapping (HiResCAM) is known to produce faithful explanations. The objective is to evaluate the quality of the attribution maps using quantitative metrics and investigate whether faithfulness aligns with the metrics results. METHODS We fine-tune pre-trained deep learning architectures over four medical image datasets in order to calculate attribution maps. The maps are evaluated on a threefold metrics basis utilizing well-established evaluation scores. RESULTS Our experimental findings suggest that the Area Over Perturbation Curve (AOPC) and Max-Sensitivity scores favor the HiResCAM maps. On the other hand, the Heatmap Assisted Accuracy Score (HAAS) does not provide insights to our comparison as it evaluates almost all maps as inaccurate. To this purpose we further compare our calculated values against values obtained over a diverse group of models which are trained on non-medical benchmark datasets, to eventually achieve more responsive results. CONCLUSION This study develops a series of experiments to discuss the connection between faithfulness and quantitative metrics over medical attribution maps. HiResCAM preserves the gradient effect on a pixel level ultimately producing high-resolution, informative and resilient mappings. In turn, this is depicted in the results of AOPC and Max-Sensitivity metrics, successfully identifying the faithful algorithm. In regards to HAAS, our experiments yield that it is sensitive over complex medical patterns, commonly characterized by strong color dependency and multiple attention areas.
Collapse
Affiliation(s)
- Vangelis Lamprou
- Department of Digital Systems, University of Piraeus, 80, M. Karaoli & A. Dimitriou St, Piraeus 18534, Greece
| | - Athanasios Kallipolitis
- Department of Digital Systems, University of Piraeus, 80, M. Karaoli & A. Dimitriou St, Piraeus 18534, Greece.
| | - Ilias Maglogiannis
- Department of Digital Systems, University of Piraeus, 80, M. Karaoli & A. Dimitriou St, Piraeus 18534, Greece
| |
Collapse
|
42
|
Ye RZ, Lipatov K, Diedrich D, Bhattacharyya A, Erickson BJ, Pickering BW, Herasevich V. Automatic ARDS surveillance with chest X-ray recognition using convolutional neural networks. J Crit Care 2024; 82:154794. [PMID: 38552452 DOI: 10.1016/j.jcrc.2024.154794] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2023] [Revised: 11/20/2023] [Accepted: 12/01/2023] [Indexed: 06/01/2024]
Abstract
OBJECTIVE This study aims to design, validate and assess the accuracy a deep learning model capable of differentiation Chest X-Rays between pneumonia, acute respiratory distress syndrome (ARDS) and normal lungs. MATERIALS AND METHODS A diagnostic performance study was conducted using Chest X-Ray images from adult patients admitted to a medical intensive care unit between January 2003 and November 2014. X-ray images from 15,899 patients were assigned one of three prespecified categories: "ARDS", "Pneumonia", or "Normal". RESULTS A two-step convolutional neural network (CNN) pipeline was developed and tested to distinguish between the three patterns with sensitivity ranging from 91.8% to 97.8% and specificity ranging from 96.6% to 98.8%. The CNN model was validated with a sensitivity of 96.3% and specificity of 96.6% using a previous dataset of patients with Acute Lung Injury (ALI)/ARDS. DISCUSSION The results suggest that a deep learning model based on chest x-ray pattern recognition can be a useful tool in distinguishing patients with ARDS from patients with normal lungs, providing faster results than digital surveillance tools based on text reports. CONCLUSION A CNN-based deep learning model showed clinically significant performance, providing potential for faster ARDS identification. Future research should prospectively evaluate these tools in a clinical setting.
Collapse
Affiliation(s)
- Run Zhou Ye
- Department of Anesthesiology and Perioperative Medicine, Mayo Clinic, 200 First Street Southwest, Rochester, MN 55905, USA.; Division of Endocrinology, Department of Medicine, Centre de Recherche du CHUS, Sherbrooke QC J1H 5N4, Canada
| | - Kirill Lipatov
- Critical Care Medicine, Mayo Clinic, Eau Claire, WI, United States
| | - Daniel Diedrich
- Department of Anesthesiology and Perioperative Medicine, Mayo Clinic, 200 First Street Southwest, Rochester, MN 55905, USA
| | | | - Bradley J Erickson
- Department of Diagnostic Radiology, Mayo Clinic, 200 First Street Southwest, Rochester, MN 55905, USA
| | - Brian W Pickering
- Department of Anesthesiology and Perioperative Medicine, Mayo Clinic, 200 First Street Southwest, Rochester, MN 55905, USA
| | - Vitaly Herasevich
- Department of Anesthesiology and Perioperative Medicine, Mayo Clinic, 200 First Street Southwest, Rochester, MN 55905, USA..
| |
Collapse
|
43
|
Kinger S, Kulkarni V. Transparent and trustworthy interpretation of COVID-19 features in chest X-rays using explainable AI. MULTIMEDIA TOOLS AND APPLICATIONS 2024. [DOI: 10.1007/s11042-024-19755-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/12/2024] [Revised: 05/14/2024] [Accepted: 06/23/2024] [Indexed: 01/03/2025]
|
44
|
Li Y, Xin Y, Li X, Zhang Y, Liu C, Cao Z, Du S, Wang L. Omni-dimensional dynamic convolution feature coordinate attention network for pneumonia classification. Vis Comput Ind Biomed Art 2024; 7:17. [PMID: 38976189 PMCID: PMC11231110 DOI: 10.1186/s42492-024-00168-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/08/2024] [Accepted: 06/22/2024] [Indexed: 07/09/2024] Open
Abstract
Pneumonia is a serious disease that can be fatal, particularly among children and the elderly. The accuracy of pneumonia diagnosis can be improved by combining artificial-intelligence technology with X-ray imaging. This study proposes X-ODFCANet, which addresses the issues of low accuracy and excessive parameters in existing deep-learning-based pneumonia-classification methods. This network incorporates a feature coordination attention module and an omni-dimensional dynamic convolution (ODConv) module, leveraging the residual module for feature extraction from X-ray images. The feature coordination attention module utilizes two one-dimensional feature encoding processes to aggregate feature information from different spatial directions. Additionally, the ODConv module extracts and fuses feature information in four dimensions: the spatial dimension of the convolution kernel, input and output channel quantities, and convolution kernel quantity. The experimental results demonstrate that the proposed method can effectively improve the accuracy of pneumonia classification, which is 3.77% higher than that of ResNet18. The model parameters are 4.45M, which was reduced by approximately 2.5 times. The code is available at https://github.com/limuni/X-ODFCANET .
Collapse
Affiliation(s)
- Yufei Li
- School of Information Science and Technology, Northwest University, Xi'an, 710127, Shaanxi Province, China
| | - Yufei Xin
- School of Information Science and Technology, Northwest University, Xi'an, 710127, Shaanxi Province, China
| | - Xinni Li
- School of Information Science and Technology, Northwest University, Xi'an, 710127, Shaanxi Province, China
| | - Yinrui Zhang
- School of Information Science and Technology, Northwest University, Xi'an, 710127, Shaanxi Province, China
| | - Cheng Liu
- School of Information Science and Technology, Northwest University, Xi'an, 710127, Shaanxi Province, China
| | - Zhengwen Cao
- School of Information Science and Technology, Northwest University, Xi'an, 710127, Shaanxi Province, China
| | - Shaoyi Du
- Department of Ultrasound, the Second Affiliated Hospital of Xi'an Jiaotong University, Xi'an, Shaanxi Province, 710004, China.
- National Key Laboratory of Human-Machine Hybrid Augmented Intelligence, National Engineering Research Center for Visual Information and Applications, and Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, Shaanxi Province, 710049, China.
| | - Lin Wang
- School of Information Science and Technology, Northwest University, Xi'an, 710127, Shaanxi Province, China.
| |
Collapse
|
45
|
Azevedo KS, de Souza LC, Coutinho MGF, de M Barbosa R, Fernandes MAC. Deepvirusclassifier: a deep learning tool for classifying SARS-CoV-2 based on viral subtypes within the coronaviridae family. BMC Bioinformatics 2024; 25:231. [PMID: 38969970 PMCID: PMC11225326 DOI: 10.1186/s12859-024-05754-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2023] [Accepted: 03/19/2024] [Indexed: 07/07/2024] Open
Abstract
PURPOSE In this study, we present DeepVirusClassifier, a tool capable of accurately classifying Severe Acute Respiratory Syndrome Coronavirus 2 (SARS-CoV-2) viral sequences among other subtypes of the coronaviridae family. This classification is achieved through a deep neural network model that relies on convolutional neural networks (CNNs). Since viruses within the same family share similar genetic and structural characteristics, the classification process becomes more challenging, necessitating more robust models. With the rapid evolution of viral genomes and the increasing need for timely classification, we aimed to provide a robust and efficient tool that could increase the accuracy of viral identification and classification processes. Contribute to advancing research in viral genomics and assist in surveilling emerging viral strains. METHODS Based on a one-dimensional deep CNN, the proposed tool is capable of training and testing on the Coronaviridae family, including SARS-CoV-2. Our model's performance was assessed using various metrics, including F1-score and AUROC. Additionally, artificial mutation tests were conducted to evaluate the model's generalization ability across sequence variations. We also used the BLAST algorithm and conducted comprehensive processing time analyses for comparison. RESULTS DeepVirusClassifier demonstrated exceptional performance across several evaluation metrics in the training and testing phases. Indicating its robust learning capacity. Notably, during testing on more than 10,000 viral sequences, the model exhibited a more than 99% sensitivity for sequences with fewer than 2000 mutations. The tool achieves superior accuracy and significantly reduced processing times compared to the Basic Local Alignment Search Tool algorithm. Furthermore, the results appear more reliable than the work discussed in the text, indicating that the tool has great potential to revolutionize viral genomic research. CONCLUSION DeepVirusClassifier is a powerful tool for accurately classifying viral sequences, specifically focusing on SARS-CoV-2 and other subtypes within the Coronaviridae family. The superiority of our model becomes evident through rigorous evaluation and comparison with existing methods. Introducing artificial mutations into the sequences demonstrates the tool's ability to identify variations and significantly contributes to viral classification and genomic research. As viral surveillance becomes increasingly critical, our model holds promise in aiding rapid and accurate identification of emerging viral strains.
Collapse
Affiliation(s)
- Karolayne S Azevedo
- InovAI Lab, nPITI/IMD, Federal University of Rio Grande do Norte, Natal, RN, 59078-970, Brazil
| | - Luísa C de Souza
- InovAI Lab, nPITI/IMD, Federal University of Rio Grande do Norte, Natal, RN, 59078-970, Brazil
| | - Maria G F Coutinho
- InovAI Lab, nPITI/IMD, Federal University of Rio Grande do Norte, Natal, RN, 59078-970, Brazil
| | - Raquel de M Barbosa
- InovAI Lab, nPITI/IMD, Federal University of Rio Grande do Norte, Natal, RN, 59078-970, Brazil.
- Department of Pharmacy and Pharmaceutical Technology, University of Seville, 41012, Seville, Spain.
| | - Marcelo A C Fernandes
- InovAI Lab, nPITI/IMD, Federal University of Rio Grande do Norte, Natal, RN, 59078-970, Brazil.
- Bioinformatics Multidisciplinary Environment (BioME), Federal University of Rio Grande do Norte, Natal, RN, 59078-970, Brazil.
- Department of Computer Engineering and Automation (DCA), Federal University of Rio Grande do Norte, Natal, RN, 59078-970, Brazil.
| |
Collapse
|
46
|
Guan H, Yap PT, Bozoki A, Liu M. Federated learning for medical image analysis: A survey. PATTERN RECOGNITION 2024; 151:110424. [PMID: 38559674 PMCID: PMC10976951 DOI: 10.1016/j.patcog.2024.110424] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/04/2024]
Abstract
Machine learning in medical imaging often faces a fundamental dilemma, namely, the small sample size problem. Many recent studies suggest using multi-domain data pooled from different acquisition sites/centers to improve statistical power. However, medical images from different sites cannot be easily shared to build large datasets for model training due to privacy protection reasons. As a promising solution, federated learning, which enables collaborative training of machine learning models based on data from different sites without cross-site data sharing, has attracted considerable attention recently. In this paper, we conduct a comprehensive survey of the recent development of federated learning methods in medical image analysis. We have systematically gathered research papers on federated learning and its applications in medical image analysis published between 2017 and 2023. Our search and compilation were conducted using databases from IEEE Xplore, ACM Digital Library, Science Direct, Springer Link, Web of Science, Google Scholar, and PubMed. In this survey, we first introduce the background of federated learning for dealing with privacy protection and collaborative learning issues. We then present a comprehensive review of recent advances in federated learning methods for medical image analysis. Specifically, existing methods are categorized based on three critical aspects of a federated learning system, including client end, server end, and communication techniques. In each category, we summarize the existing federated learning methods according to specific research problems in medical image analysis and also provide insights into the motivations of different approaches. In addition, we provide a review of existing benchmark medical imaging datasets and software platforms for current federated learning research. We also conduct an experimental study to empirically evaluate typical federated learning methods for medical image analysis. This survey can help to better understand the current research status, challenges, and potential research opportunities in this promising research field.
Collapse
Affiliation(s)
- Hao Guan
- Department of Radiology and Biomedical Research Imaging Center, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
| | - Pew-Thian Yap
- Department of Radiology and Biomedical Research Imaging Center, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
| | - Andrea Bozoki
- Department of Neurology, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
| | - Mingxia Liu
- Department of Radiology and Biomedical Research Imaging Center, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
| |
Collapse
|
47
|
Lee SB. Development of a chest X-ray machine learning convolutional neural network model on a budget and using artificial intelligence explainability techniques to analyze patterns of machine learning inference. JAMIA Open 2024; 7:ooae035. [PMID: 38699648 PMCID: PMC11064095 DOI: 10.1093/jamiaopen/ooae035] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/23/2024] [Revised: 04/03/2024] [Accepted: 04/10/2024] [Indexed: 05/05/2024] Open
Abstract
Objective Machine learning (ML) will have a large impact on medicine and accessibility is important. This study's model was used to explore various concepts including how varying features of a model impacted behavior. Materials and Methods This study built an ML model that classified chest X-rays as normal or abnormal by using ResNet50 as a base with transfer learning. A contrast enhancement mechanism was implemented to improve performance. After training with a dataset of publicly available chest radiographs, performance metrics were determined with a test set. The ResNet50 base was substituted with deeper architectures (ResNet101/152) and visualization methods used to help determine patterns of inference. Results Performance metrics were an accuracy of 79%, recall 69%, precision 96%, and area under the curve of 0.9023. Accuracy improved to 82% and recall to 74% with contrast enhancement. When visualization methods were applied and the ratio of pixels used for inference measured, deeper architectures resulted in the model using larger portions of the image for inference as compared to ResNet50. Discussion The model performed on par with many existing models despite consumer-grade hardware and smaller datasets. Individual models vary thus a single model's explainability may not be generalizable. Therefore, this study varied architecture and studied patterns of inference. With deeper ResNet architectures, the machine used larger portions of the image to make decisions. Conclusion An example using a custom model showed that AI (Artificial Intelligence) can be accessible on consumer-grade hardware, and it also demonstrated an example of studying themes of ML explainability by varying ResNet architectures.
Collapse
Affiliation(s)
- Stephen B Lee
- Division of Infectious Diseases, Department of Medicine, College of Medicine, University of Saskatchewan, Regina, S4P 0W5, Canada
| |
Collapse
|
48
|
Kanwal K, Asif M, Khalid SG, Liu H, Qurashi AG, Abdullah S. Current Diagnostic Techniques for Pneumonia: A Scoping Review. SENSORS (BASEL, SWITZERLAND) 2024; 24:4291. [PMID: 39001069 PMCID: PMC11244398 DOI: 10.3390/s24134291] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/29/2024] [Revised: 06/22/2024] [Accepted: 06/28/2024] [Indexed: 07/16/2024]
Abstract
Community-acquired pneumonia is one of the most lethal infectious diseases, especially for infants and the elderly. Given the variety of causative agents, the accurate early detection of pneumonia is an active research area. To the best of our knowledge, scoping reviews on diagnostic techniques for pneumonia are lacking. In this scoping review, three major electronic databases were searched and the resulting research was screened. We categorized these diagnostic techniques into four classes (i.e., lab-based methods, imaging-based techniques, acoustic-based techniques, and physiological-measurement-based techniques) and summarized their recent applications. Major research has been skewed towards imaging-based techniques, especially after COVID-19. Currently, chest X-rays and blood tests are the most common tools in the clinical setting to establish a diagnosis; however, there is a need to look for safe, non-invasive, and more rapid techniques for diagnosis. Recently, some non-invasive techniques based on wearable sensors achieved reasonable diagnostic accuracy that could open a new chapter for future applications. Consequently, further research and technology development are still needed for pneumonia diagnosis using non-invasive physiological parameters to attain a better point of care for pneumonia patients.
Collapse
Affiliation(s)
- Kehkashan Kanwal
- College of Speech, Language, and Hearing Sciences, Ziauddin University, Karachi 75000, Pakistan
| | - Muhammad Asif
- Faculty of Computing and Applied Sciences, Sir Syed University of Engineering and Technology, Karachi 75300, Pakistan;
| | - Syed Ghufran Khalid
- Department of Engineering, Faculty of Science and Technology, Nottingham Trent University, Nottingham B15 3TN, UK
| | - Haipeng Liu
- Research Centre for Intelligent Healthcare, Coventry University, Coventry CV1 5FB, UK;
| | | | - Saad Abdullah
- School of Innovation, Design and Engineering, Mälardalen University, 721 23 Västerås, Sweden
| |
Collapse
|
49
|
Brima Y, Atemkeng M. Saliency-driven explainable deep learning in medical imaging: bridging visual explainability and statistical quantitative analysis. BioData Min 2024; 17:18. [PMID: 38909228 PMCID: PMC11193223 DOI: 10.1186/s13040-024-00370-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/07/2023] [Accepted: 06/10/2024] [Indexed: 06/24/2024] Open
Abstract
Deep learning shows great promise for medical image analysis but often lacks explainability, hindering its adoption in healthcare. Attribution techniques that explain model reasoning can potentially increase trust in deep learning among clinical stakeholders. In the literature, much of the research on attribution in medical imaging focuses on visual inspection rather than statistical quantitative analysis.In this paper, we proposed an image-based saliency framework to enhance the explainability of deep learning models in medical image analysis. We use adaptive path-based gradient integration, gradient-free techniques, and class activation mapping along with its derivatives to attribute predictions from brain tumor MRI and COVID-19 chest X-ray datasets made by recent deep convolutional neural network models.The proposed framework integrates qualitative and statistical quantitative assessments, employing Accuracy Information Curves (AICs) and Softmax Information Curves (SICs) to measure the effectiveness of saliency methods in retaining critical image information and their correlation with model predictions. Visual inspections indicate that methods such as ScoreCAM, XRAI, GradCAM, and GradCAM++ consistently produce focused and clinically interpretable attribution maps. These methods highlighted possible biomarkers, exposed model biases, and offered insights into the links between input features and predictions, demonstrating their ability to elucidate model reasoning on these datasets. Empirical evaluations reveal that ScoreCAM and XRAI are particularly effective in retaining relevant image regions, as reflected in their higher AUC values. However, SICs highlight variability, with instances of random saliency masks outperforming established methods, emphasizing the need for combining visual and empirical metrics for a comprehensive evaluation.The results underscore the importance of selecting appropriate saliency methods for specific medical imaging tasks and suggest that combining qualitative and quantitative approaches can enhance the transparency, trustworthiness, and clinical adoption of deep learning models in healthcare. This study advances model explainability to increase trust in deep learning among healthcare stakeholders by revealing the rationale behind predictions. Future research should refine empirical metrics for stability and reliability, include more diverse imaging modalities, and focus on improving model explainability to support clinical decision-making.
Collapse
Affiliation(s)
- Yusuf Brima
- Computer Vision, Institute of Cognitive Science, Osnabrück University, Osnabrueck, D-49090, Lower Saxony, Germany.
| | - Marcellin Atemkeng
- Department of Mathematics, Rhodes University, Grahamstown, 6140, Eastern Cape, South Africa.
| |
Collapse
|
50
|
Qiu Y, Liu Y, Li S, Xu J. MiniSeg: An Extremely Minimum Network Based on Lightweight Multiscale Learning for Efficient COVID-19 Segmentation. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:8570-8584. [PMID: 37015641 DOI: 10.1109/tnnls.2022.3230821] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
The rapid spread of the new pandemic, i.e., coronavirus disease 2019 (COVID-19), has severely threatened global health. Deep-learning-based computer-aided screening, e.g., COVID-19 infected area segmentation from computed tomography (CT) image, has attracted much attention by serving as an adjunct to increase the accuracy of COVID-19 screening and clinical diagnosis. Although lesion segmentation is a hot topic, traditional deep learning methods are usually data-hungry with millions of parameters, easy to overfit under limited available COVID-19 training data. On the other hand, fast training/testing and low computational cost are also necessary for quick deployment and development of COVID-19 screening systems, but traditional methods are usually computationally intensive. To address the above two problems, we propose MiniSeg, a lightweight model for efficient COVID-19 segmentation from CT images. Our efforts start with the design of an attentive hierarchical spatial pyramid (AHSP) module for lightweight, efficient, effective multiscale learning that is essential for image segmentation. Then, we build a two-path (TP) encoder for deep feature extraction, where one path uses AHSP modules for learning multiscale contextual features and the other is a shallow convolutional path for capturing fine details. The two paths interact with each other for learning effective representations. Based on the extracted features, a simple decoder is added for COVID-19 segmentation. For comparing MiniSeg to previous methods, we build a comprehensive COVID-19 segmentation benchmark. Extensive experiments demonstrate that the proposed MiniSeg achieves better accuracy because its only 83k parameters make it less prone to overfitting. Its high efficiency also makes it easy to deploy and develop. The code has been released at https://github.com/yun-liu/MiniSeg.
Collapse
|