51
|
Ullah F, Srivastava G, Xiao H, Ullah S, Lin JCW, Zhao Y. A Scalable Federated Learning Approach for Collaborative Smart Healthcare Systems With Intermittent Clients Using Medical Imaging. IEEE J Biomed Health Inform 2024; 28:3293-3304. [PMID: 37279135 DOI: 10.1109/jbhi.2023.3282955] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
The healthcare industry is one of the most vulnerable to cybercrime and privacy violations because health data is very sensitive and spread out in many places. Recent confidentiality trends and a rising number of infringements in different sectors make it crucial to implement new methods that protect data privacy while maintaining accuracy and sustainability. Moreover, the intermittent nature of remote clients with imbalanced datasets poses a significant obstacle for decentralized healthcare systems. Federated learning (FL) is a decentralized and privacy-protecting approach to deep learning and machine learning models. In this article, we implement a scalable FL framework for interactive smart healthcare systems with intermittent clients using chest X-ray images. Remote hospitals may have imbalanced datasets with intermittent clients communicating with the FL global server. The data augmentation method is used to balance datasets for local model training. In practice, some clients may leave the training process while others join due to technical or connectivity issues. The proposed method is tested with five to eighteen clients and different testing data sizes to evaluate performance in various situations. The experiments show that the proposed FL approach produces competitive results when dealing with two distinct problems, such as intermittent clients and imbalanced data. These findings would encourage medical institutions to collaborate and use rich private data to quickly develop a powerful patient diagnostic model.
Collapse
|
52
|
Zhou Z, Islam MT, Xing L. Multibranch CNN With MLP-Mixer-Based Feature Exploration for High-Performance Disease Diagnosis. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:7351-7362. [PMID: 37028335 PMCID: PMC11779602 DOI: 10.1109/tnnls.2023.3250490] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Deep learning-based diagnosis is becoming an indispensable part of modern healthcare. For high-performance diagnosis, the optimal design of deep neural networks (DNNs) is a prerequisite. Despite its success in image analysis, existing supervised DNNs based on convolutional layers often suffer from their rudimentary feature exploration ability caused by the limited receptive field and biased feature extraction of conventional convolutional neural networks (CNNs), which compromises the network performance. Here, we propose a novel feature exploration network named manifold embedded multilayer perceptron (MLP) mixer (ME-Mixer), which utilizes both supervised and unsupervised features for disease diagnosis. In the proposed approach, a manifold embedding network is employed to extract class-discriminative features; then, two MLP-Mixer-based feature projectors are adopted to encode the extracted features with the global reception field. Our ME-Mixer network is quite general and can be added as a plugin to any existing CNN. Comprehensive evaluations on two medical datasets are performed. The results demonstrate that their approach greatly enhances the classification accuracy in comparison with different configurations of DNNs with acceptable computational complexity.
Collapse
|
53
|
Aasem M, Javed Iqbal M. Toward explainable AI in radiology: Ensemble-CAM for effective thoracic disease localization in chest X-ray images using weak supervised learning. Front Big Data 2024; 7:1366415. [PMID: 38756502 PMCID: PMC11096460 DOI: 10.3389/fdata.2024.1366415] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/06/2024] [Accepted: 04/08/2024] [Indexed: 05/18/2024] Open
Abstract
Chest X-ray (CXR) imaging is widely employed by radiologists to diagnose thoracic diseases. Recently, many deep learning techniques have been proposed as computer-aided diagnostic (CAD) tools to assist radiologists in minimizing the risk of incorrect diagnosis. From an application perspective, these models have exhibited two major challenges: (1) They require large volumes of annotated data at the training stage and (2) They lack explainable factors to justify their outcomes at the prediction stage. In the present study, we developed a class activation mapping (CAM)-based ensemble model, called Ensemble-CAM, to address both of these challenges via weakly supervised learning by employing explainable AI (XAI) functions. Ensemble-CAM utilizes class labels to predict the location of disease in association with interpretable features. The proposed work leverages ensemble and transfer learning with class activation functions to achieve three objectives: (1) minimizing the dependency on strongly annotated data when locating thoracic diseases, (2) enhancing confidence in predicted outcomes by visualizing their interpretable features, and (3) optimizing cumulative performance via fusion functions. Ensemble-CAM was trained on three CXR image datasets and evaluated through qualitative and quantitative measures via heatmaps and Jaccard indices. The results reflect the enhanced performance and reliability in comparison to existing standalone and ensembled models.
Collapse
Affiliation(s)
- Muhammad Aasem
- Department of Computer Science, University of Engineering and Technology, Taxila, Pakistan
| | | |
Collapse
|
54
|
Liu Z, Shen L. CECT: Controllable ensemble CNN and transformer for COVID-19 image classification. Comput Biol Med 2024; 173:108388. [PMID: 38569235 DOI: 10.1016/j.compbiomed.2024.108388] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/16/2023] [Revised: 03/15/2024] [Accepted: 03/25/2024] [Indexed: 04/05/2024]
Abstract
The COVID-19 pandemic has resulted in hundreds of million cases and numerous deaths worldwide. Here, we develop a novel classification network CECT by controllable ensemble convolutional neural network and transformer to provide a timely and accurate COVID-19 diagnosis. The CECT is composed of a parallel convolutional encoder block, an aggregate transposed-convolutional decoder block, and a windowed attention classification block. Each block captures features at different scales from 28 × 28 to 224 × 224 from the input, composing enriched and comprehensive information. Different from existing methods, our CECT can capture features at both multi-local and global scales without any sophisticated module design. Moreover, the contribution of local features at different scales can be controlled with the proposed ensemble coefficients. We evaluate CECT on two public COVID-19 datasets and it reaches the highest accuracy of 98.1% in the intra-dataset evaluation, outperforming existing state-of-the-art methods. Moreover, the developed CECT achieves an accuracy of 90.9% on the unseen dataset in the inter-dataset evaluation, showing extraordinary generalization ability. With remarkable feature capture ability and generalization ability, we believe CECT can be extended to other medical scenarios as a powerful diagnosis tool. Code is available at https://github.com/NUS-Tim/CECT.
Collapse
Affiliation(s)
- Zhaoshan Liu
- Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore, 117575, Singapore.
| | - Lei Shen
- Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore, 117575, Singapore.
| |
Collapse
|
55
|
Rana S, Hosen MJ, Tonni TJ, Rony MAH, Fatema K, Hasan MZ, Rahman MT, Khan RT, Jan T, Whaiduzzaman M. DeepChestGNN: A Comprehensive Framework for Enhanced Lung Disease Identification through Advanced Graphical Deep Features. SENSORS (BASEL, SWITZERLAND) 2024; 24:2830. [PMID: 38732936 PMCID: PMC11086108 DOI: 10.3390/s24092830] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/02/2024] [Revised: 04/06/2024] [Accepted: 04/16/2024] [Indexed: 05/13/2024]
Abstract
Lung diseases are the third-leading cause of mortality in the world. Due to compromised lung function, respiratory difficulties, and physiological complications, lung disease brought on by toxic substances, pollution, infections, or smoking results in millions of deaths every year. Chest X-ray images pose a challenge for classification due to their visual similarity, leading to confusion among radiologists. To imitate those issues, we created an automated system with a large data hub that contains 17 datasets of chest X-ray images for a total of 71,096, and we aim to classify ten different disease classes. For combining various resources, our large datasets contain noise and annotations, class imbalances, data redundancy, etc. We conducted several image pre-processing techniques to eliminate noise and artifacts from images, such as resizing, de-annotation, CLAHE, and filtering. The elastic deformation augmentation technique also generates a balanced dataset. Then, we developed DeepChestGNN, a novel medical image classification model utilizing a deep convolutional neural network (DCNN) to extract 100 significant deep features indicative of various lung diseases. This model, incorporating Batch Normalization, MaxPooling, and Dropout layers, achieved a remarkable 99.74% accuracy in extensive trials. By combining graph neural networks (GNNs) with feedforward layers, the architecture is very flexible when it comes to working with graph data for accurate lung disease classification. This study highlights the significant impact of combining advanced research with clinical application potential in diagnosing lung diseases, providing an optimal framework for precise and efficient disease identification and classification.
Collapse
Affiliation(s)
- Shakil Rana
- Health Informatics Research Laboratory (HIRL), Department of Computer Science and Engineering, Daffodil International University, Dhaka 1207, Bangladesh; (S.R.); (M.J.H.); (T.J.T.); (M.A.H.R.); (K.F.); (M.Z.H.)
| | - Md Jabed Hosen
- Health Informatics Research Laboratory (HIRL), Department of Computer Science and Engineering, Daffodil International University, Dhaka 1207, Bangladesh; (S.R.); (M.J.H.); (T.J.T.); (M.A.H.R.); (K.F.); (M.Z.H.)
| | - Tasnim Jahan Tonni
- Health Informatics Research Laboratory (HIRL), Department of Computer Science and Engineering, Daffodil International University, Dhaka 1207, Bangladesh; (S.R.); (M.J.H.); (T.J.T.); (M.A.H.R.); (K.F.); (M.Z.H.)
| | - Md. Awlad Hossen Rony
- Health Informatics Research Laboratory (HIRL), Department of Computer Science and Engineering, Daffodil International University, Dhaka 1207, Bangladesh; (S.R.); (M.J.H.); (T.J.T.); (M.A.H.R.); (K.F.); (M.Z.H.)
| | - Kaniz Fatema
- Health Informatics Research Laboratory (HIRL), Department of Computer Science and Engineering, Daffodil International University, Dhaka 1207, Bangladesh; (S.R.); (M.J.H.); (T.J.T.); (M.A.H.R.); (K.F.); (M.Z.H.)
| | - Md. Zahid Hasan
- Health Informatics Research Laboratory (HIRL), Department of Computer Science and Engineering, Daffodil International University, Dhaka 1207, Bangladesh; (S.R.); (M.J.H.); (T.J.T.); (M.A.H.R.); (K.F.); (M.Z.H.)
| | - Md. Tanvir Rahman
- School of Health and Rehabilitation Sciences, The University of Queensland, St. Lucia, QLD 4072, Australia
- Department of Information and Communication Technology, Mawlana Bhashani Science and Technology University, Tangail 1902, Bangladesh
| | - Risala Tasin Khan
- Institute of Information Technology, Jahangirnagar University, Dhaka 1342, Bangladesh;
| | - Tony Jan
- Centre for Artificial Intelligence Research and Optimisation (AIRO), Torrens University, Ultimo, NSW 2007, Australia;
| | - Md Whaiduzzaman
- Centre for Artificial Intelligence Research and Optimisation (AIRO), Torrens University, Ultimo, NSW 2007, Australia;
- School of Information Systems, Queensland University of Technology, Brisbane, QLD 4000, Australia
| |
Collapse
|
56
|
Althenayan AS, AlSalamah SA, Aly S, Nouh T, Mahboub B, Salameh L, Alkubeyyer M, Mirza A. COVID-19 Hierarchical Classification Using a Deep Learning Multi-Modal. SENSORS (BASEL, SWITZERLAND) 2024; 24:2641. [PMID: 38676257 PMCID: PMC11053684 DOI: 10.3390/s24082641] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/05/2024] [Revised: 04/15/2024] [Accepted: 04/18/2024] [Indexed: 04/28/2024]
Abstract
Coronavirus disease 2019 (COVID-19), originating in China, has rapidly spread worldwide. Physicians must examine infected patients and make timely decisions to isolate them. However, completing these processes is difficult due to limited time and availability of expert radiologists, as well as limitations of the reverse-transcription polymerase chain reaction (RT-PCR) method. Deep learning, a sophisticated machine learning technique, leverages radiological imaging modalities for disease diagnosis and image classification tasks. Previous research on COVID-19 classification has encountered several limitations, including binary classification methods, single-feature modalities, small public datasets, and reliance on CT diagnostic processes. Additionally, studies have often utilized a flat structure, disregarding the hierarchical structure of pneumonia classification. This study aims to overcome these limitations by identifying pneumonia caused by COVID-19, distinguishing it from other types of pneumonia and healthy lungs using chest X-ray (CXR) images and related tabular medical data, and demonstrate the value of incorporating tabular medical data in achieving more accurate diagnoses. Resnet-based and VGG-based pre-trained convolutional neural network (CNN) models were employed to extract features, which were then combined using early fusion for the classification of eight distinct classes. We leveraged the hierarchal structure of pneumonia classification within our approach to achieve improved classification outcomes. Since an imbalanced dataset is common in this field, a variety of versions of generative adversarial networks (GANs) were used to generate synthetic data. The proposed approach tested in our private datasets of 4523 patients achieved a macro-avg F1-score of 95.9% and an F1-score of 87.5% for COVID-19 identification using a Resnet-based structure. In conclusion, in this study, we were able to create an accurate deep learning multi-modal to diagnose COVID-19 and differentiate it from other kinds of pneumonia and normal lungs, which will enhance the radiological diagnostic process.
Collapse
Affiliation(s)
- Albatoul S. Althenayan
- Information Systems Department, College of Computer and Information Sciences, King Saud University, Riyadh 11543, Saudi Arabia; (S.A.A.); (A.M.)
- Information Systems Department, College of Computer and Information Sciences, Imam Mohammed Bin Saud Islamic University, Riyadh 11432, Saudi Arabia
| | - Shada A. AlSalamah
- Information Systems Department, College of Computer and Information Sciences, King Saud University, Riyadh 11543, Saudi Arabia; (S.A.A.); (A.M.)
- National Health Information Center, Saudi Health Council, Riyadh 13315, Saudi Arabia
- Digital Health and Innovation Department, Science Division, World Health Organization, 1211 Geneva, Switzerland
| | - Sherin Aly
- Institute of Graduate Studies and Research, Alexandria University, Alexandria 21526, Egypt;
| | - Thamer Nouh
- Trauma and Acute Care Surgery Unit, College of Medicine, King Saud University, Riyadh 12271, Saudi Arabia;
| | - Bassam Mahboub
- Clinical Sciences Department, College of Medicine, University of Sharjah, Sharjah P.O. Box 27272, United Arab Emirates;
| | - Laila Salameh
- Sharjah Institute for Medical Research, University of Sharjah, Sharjah P.O. Box 27272, United Arab Emirates;
| | - Metab Alkubeyyer
- Department of Radiology and Medical Imaging, King Khalid University Hospital, King Saud University, Riyadh 12372, Saudi Arabia;
| | - Abdulrahman Mirza
- Information Systems Department, College of Computer and Information Sciences, King Saud University, Riyadh 11543, Saudi Arabia; (S.A.A.); (A.M.)
| |
Collapse
|
57
|
Arefin MS, Rahman MM, Hasan MT, Mahmud M. A Topical Review on Enabling Technologies for the Internet of Medical Things: Sensors, Devices, Platforms, and Applications. MICROMACHINES 2024; 15:479. [PMID: 38675290 PMCID: PMC11051832 DOI: 10.3390/mi15040479] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/24/2023] [Revised: 03/17/2024] [Accepted: 03/22/2024] [Indexed: 04/28/2024]
Abstract
The Internet of Things (IoT) is still a relatively new field of research, and its potential to be used in the healthcare and medical sectors is enormous. In the last five years, IoT has been a go-to option for various applications such as using sensors for different features, machine-to-machine communication, etc., but precisely in the medical sector, it is still lagging far behind compared to other sectors. Hence, this study emphasises IoT applications in medical fields, Medical IoT sensors and devices, IoT platforms for data visualisation, and artificial intelligence in medical applications. A systematic review considering PRISMA guidelines on research articles as well as the websites on IoMT sensors and devices has been carried out. After the year 2001, an integrated outcome of 986 articles was initially selected, and by applying the inclusion-exclusion criterion, a total of 597 articles were identified. 23 new studies have been finally found, including records from websites and citations. This review then analyses different sensor monitoring circuits in detail, considering an Intensive Care Unit (ICU) scenario, device applications, and the data management system, including IoT platforms for the patients. Lastly, detailed discussion and challenges have been outlined, and possible prospects have been presented.
Collapse
Affiliation(s)
- Md. Shamsul Arefin
- Department of Electrical and Electronic Engineering (EEE), Bangladesh University of Business & Technology, Dhaka 1216, Bangladesh;
| | | | - Md. Tanvir Hasan
- Department of Electrical and Electronic Engineering (EEE), Jashore University of Science & Technology, Jashore 7408, Bangladesh;
- Department of Electrical Engineering, University of South Carolina, Columbia, SC 29208, USA
| | - Mufti Mahmud
- Department of Computer Science, Nottingham Trent University, Nottingham NG11 8NS, UK
- Computing and Informatics Research Centre, Nottingham Trent University, Nottingham NG11 8NS, UK
- Medical Technologies Innovation Facility, Nottingham Trent University, Nottingham NG11 8NS, UK
| |
Collapse
|
58
|
Kabir MM, Mridha M, Rahman A, Hamid MA, Monowar MM. Detection of COVID-19, pneumonia, and tuberculosis from radiographs using AI-driven knowledge distillation. Heliyon 2024; 10:e26801. [PMID: 38444490 PMCID: PMC10912466 DOI: 10.1016/j.heliyon.2024.e26801] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/10/2023] [Revised: 01/30/2024] [Accepted: 02/20/2024] [Indexed: 03/07/2024] Open
Abstract
Chest radiography is an essential diagnostic tool for respiratory diseases such as COVID-19, pneumonia, and tuberculosis because it accurately depicts the structures of the chest. However, accurate detection of these diseases from radiographs is a complex task that requires the availability of medical imaging equipment and trained personnel. Conventional deep learning models offer a viable automated solution for this task. However, the high complexity of these models often poses a significant obstacle to their practical deployment within automated medical applications, including mobile apps, web apps, and cloud-based platforms. This study addresses and resolves this dilemma by reducing the complexity of neural networks using knowledge distillation techniques (KDT). The proposed technique trains a neural network on an extensive collection of chest X-ray images and propagates the knowledge to a smaller network capable of real-time detection. To create a comprehensive dataset, we have integrated three popular chest radiograph datasets with chest radiographs for COVID-19, pneumonia, and tuberculosis. Our experiments show that this knowledge distillation approach outperforms conventional deep learning methods in terms of computational complexity and performance for real-time respiratory disease detection. Specifically, our system achieves an impressive average accuracy of 0.97, precision of 0.94, and recall of 0.97.
Collapse
Affiliation(s)
- Md Mohsin Kabir
- Department of Computer Science & Engineering, Bangladesh University of Business & Technology, Dhaka-1216, Bangladesh
| | - M.F. Mridha
- Department of Computer Science, American International University-Bangladesh, Dhaka-1229, Bangladesh
| | - Ashifur Rahman
- Department of Computer Science & Engineering, Bangladesh University of Business & Technology, Dhaka-1216, Bangladesh
| | - Md. Abdul Hamid
- Department of Information Technology, Faculty of Computing & Information Technology, King Abdulaziz University, Jeddah-21589, Kingdom of Saudi Arabia
| | - Muhammad Mostafa Monowar
- Department of Information Technology, Faculty of Computing & Information Technology, King Abdulaziz University, Jeddah-21589, Kingdom of Saudi Arabia
| |
Collapse
|
59
|
Rainio O, Teuho J, Klén R. Evaluation metrics and statistical tests for machine learning. Sci Rep 2024; 14:6086. [PMID: 38480847 PMCID: PMC10937649 DOI: 10.1038/s41598-024-56706-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2023] [Accepted: 03/09/2024] [Indexed: 03/17/2024] Open
Abstract
Research on different machine learning (ML) has become incredibly popular during the past few decades. However, for some researchers not familiar with statistics, it might be difficult to understand how to evaluate the performance of ML models and compare them with each other. Here, we introduce the most common evaluation metrics used for the typical supervised ML tasks including binary, multi-class, and multi-label classification, regression, image segmentation, object detection, and information retrieval. We explain how to choose a suitable statistical test for comparing models, how to obtain enough values of the metric for testing, and how to perform the test and interpret its results. We also present a few practical examples about comparing convolutional neural networks used to classify X-rays with different lung infections and detect cancer tumors in positron emission tomography images.
Collapse
Affiliation(s)
- Oona Rainio
- Turku PET Centre, University of Turku and Turku University Hospital, Turku, Finland.
| | - Jarmo Teuho
- Turku PET Centre, University of Turku and Turku University Hospital, Turku, Finland
| | - Riku Klén
- Turku PET Centre, University of Turku and Turku University Hospital, Turku, Finland
| |
Collapse
|
60
|
Malik H, Anees T. Multi-modal deep learning methods for classification of chest diseases using different medical imaging and cough sounds. PLoS One 2024; 19:e0296352. [PMID: 38470893 DOI: 10.1371/journal.pone.0296352] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/16/2023] [Accepted: 12/11/2023] [Indexed: 03/14/2024] Open
Abstract
Chest disease refers to a wide range of conditions affecting the lungs, such as COVID-19, lung cancer (LC), consolidation lung (COL), and many more. When diagnosing chest disorders medical professionals may be thrown off by the overlapping symptoms (such as fever, cough, sore throat, etc.). Additionally, researchers and medical professionals make use of chest X-rays (CXR), cough sounds, and computed tomography (CT) scans to diagnose chest disorders. The present study aims to classify the nine different conditions of chest disorders, including COVID-19, LC, COL, atelectasis (ATE), tuberculosis (TB), pneumothorax (PNEUTH), edema (EDE), pneumonia (PNEU). Thus, we suggested four novel convolutional neural network (CNN) models that train distinct image-level representations for nine different chest disease classifications by extracting features from images. Furthermore, the proposed CNN employed several new approaches such as a max-pooling layer, batch normalization layers (BANL), dropout, rank-based average pooling (RBAP), and multiple-way data generation (MWDG). The scalogram method is utilized to transform the sounds of coughing into a visual representation. Before beginning to train the model that has been developed, the SMOTE approach is used to calibrate the CXR and CT scans as well as the cough sound images (CSI) of nine different chest disorders. The CXR, CT scan, and CSI used for training and evaluating the proposed model come from 24 publicly available benchmark chest illness datasets. The classification performance of the proposed model is compared with that of seven baseline models, namely Vgg-19, ResNet-101, ResNet-50, DenseNet-121, EfficientNetB0, DenseNet-201, and Inception-V3, in addition to state-of-the-art (SOTA) classifiers. The effectiveness of the proposed model is further demonstrated by the results of the ablation experiments. The proposed model was successful in achieving an accuracy of 99.01%, making it superior to both the baseline models and the SOTA classifiers. As a result, the proposed approach is capable of offering significant support to radiologists and other medical professionals.
Collapse
Affiliation(s)
- Hassaan Malik
- Department of Computer Science, School of Systems and Technology, University of Management and Technology, Lahore, Pakistan
| | - Tayyaba Anees
- Department of Software Engineering, School of Systems and Technology, University of Management and Technology, Lahore, Pakistan
| |
Collapse
|
61
|
Abad M, Casas-Roma J, Prados F. Generalizable disease detection using model ensemble on chest X-ray images. Sci Rep 2024; 14:5890. [PMID: 38467705 PMCID: PMC10928229 DOI: 10.1038/s41598-024-56171-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/04/2023] [Accepted: 03/03/2024] [Indexed: 03/13/2024] Open
Abstract
In the realm of healthcare, the demand for swift and precise diagnostic tools has been steadily increasing. This study delves into a comprehensive performance analysis of three pre-trained convolutional neural network (CNN) architectures: ResNet50, DenseNet121, and Inception-ResNet-v2. To ensure the broad applicability of our approach, we curated a large-scale dataset comprising a diverse collection of chest X-ray images, that included both positive and negative cases of COVID-19. The models' performance was evaluated using separate datasets for internal validation (from the same source as the training images) and external validation (from different sources). Our examination uncovered a significant drop in network efficacy, registering a 10.66% reduction for ResNet50, a 36.33% decline for DenseNet121, and a 19.55% decrease for Inception-ResNet-v2 in terms of accuracy. Best results were obtained with DenseNet121 achieving the highest accuracy at 96.71% in internal validation and Inception-ResNet-v2 attaining 76.70% accuracy in external validation. Furthermore, we introduced a model ensemble approach aimed at improving network performance when making inferences on images from diverse sources beyond their training data. The proposed method uses uncertainty-based weighting by calculating the entropy in order to assign appropriate weights to the outputs of each network. Our results showcase the effectiveness of the ensemble method in enhancing accuracy up to 97.38% for internal validation and 81.18% for external validation, while maintaining a balanced ability to detect both positive and negative cases.
Collapse
Affiliation(s)
- Maider Abad
- Universitat Oberta de Catalunya, e-Health Center, Barcelona, Spain.
| | - Jordi Casas-Roma
- Universitat Oberta de Catalunya, e-Health Center, Barcelona, Spain
- Department of Computer Science, Universitat Autònoma de Barcelona (UAB), Bellaterra, Spain
- Computer Vision Center (CVC), Universitat Autònoma de Barcelona (UAB), Bellaterra, Spain
| | - Ferran Prados
- Universitat Oberta de Catalunya, e-Health Center, Barcelona, Spain
- Queen Square MS Centre, Department of Neuroinflammation, UCL Queen Square Institute of Neurology, Faculty of Brain Science, University College of London, London, WC1N 3BG, UK
- Centre for Medical Image Computing (CMIC), Department of Medical Physics and Bioengineering, University College London, London, WC1V 6LJ, UK
| |
Collapse
|
62
|
Ju H, Cui Y, Su Q, Juan L, Manavalan B. CODENET: A deep learning model for COVID-19 detection. Comput Biol Med 2024; 171:108229. [PMID: 38447500 DOI: 10.1016/j.compbiomed.2024.108229] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2023] [Revised: 02/20/2024] [Accepted: 02/25/2024] [Indexed: 03/08/2024]
Abstract
Conventional COVID-19 testing methods have some flaws: they are expensive and time-consuming. Chest X-ray (CXR) diagnostic approaches can alleviate these flaws to some extent. However, there is no accurate and practical automatic diagnostic framework with good interpretability. The application of artificial intelligence (AI) technology to medical radiography can help to accurately detect the disease, reduce the burden on healthcare organizations, and provide good interpretability. Therefore, this study proposes a new deep neural network (CNN) based on CXR for COVID-19 diagnosis - CodeNet. This method uses contrastive learning to make full use of latent image data to enhance the model's ability to extract features and generalize across different data domains. On the evaluation dataset, the proposed method achieves an accuracy as high as 94.20%, outperforming several other existing methods used for comparison. Ablation studies validate the efficacy of the proposed method, while interpretability analysis shows that the method can effectively guide clinical professionals. This work demonstrates the superior detection performance of a CNN using contrastive learning techniques on CXR images, paving the way for computer vision and artificial intelligence technologies to leverage massive medical data for disease diagnosis.
Collapse
Affiliation(s)
- Hong Ju
- Heilongjiang Agricultural Engineering Vocational College, China
| | - Yanyan Cui
- Beidahuang Industry Group General Hospital, Harbin, China
| | - Qiaosen Su
- Department of Integrative Biotechnology, College of Biotechnology and Bioengineering, Sungkyunkwan University, Suwon, 16419, Gyeonggi-do, Republic of Korea
| | - Liran Juan
- School of Life Science and Technology, Harbin Institute of Technology, Harbin, 150001, China.
| | - Balachandran Manavalan
- Department of Integrative Biotechnology, College of Biotechnology and Bioengineering, Sungkyunkwan University, Suwon, 16419, Gyeonggi-do, Republic of Korea.
| |
Collapse
|
63
|
Ali MU, Zafar A, Tanveer J, Khan MA, Kim SH, Alsulami MM, Lee SW. Deep learning network selection and optimized information fusion for enhanced COVID‐19 detection. INTERNATIONAL JOURNAL OF IMAGING SYSTEMS AND TECHNOLOGY 2024; 34. [DOI: 10.1002/ima.23001] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/09/2023] [Accepted: 11/12/2023] [Indexed: 08/25/2024]
Abstract
AbstractThis study proposes a wrapper‐based technique to improve the classification performance of chest infection (including COVID‐19) detection using X‐rays. Deep features were extracted using pretrained deep learning models. Ten optimization techniques, including poor and rich optimization, path finder algorithm, Henry gas solubility optimization, Harris hawks optimization, atom search optimization, manta‐ray foraging optimization, equilibrium optimizer, slime mold algorithm, generalized normal distribution optimization, and marine predator algorithm, were used to determine the optimal features using a support vector machine. Moreover, a network selection technique was used to select the deep learning models. An online chest infection detection X‐ray scan dataset was used to validate the proposed approach. The results suggest that the proposed wrapper‐based automatic deep learning network selection and feature optimization framework has a high classification rate of 97.7%. The comparative analysis further validates the credibility of the framework in COVID‐19 and other chest infection classifications, suggesting that the proposed approach can help doctors in clinical practice.
Collapse
Affiliation(s)
- Muhammad Umair Ali
- Department of Intelligent Mechatronics Engineering Sejong University Seoul Republic of Korea
| | - Amad Zafar
- Department of Intelligent Mechatronics Engineering Sejong University Seoul Republic of Korea
| | - Jawad Tanveer
- Department of Computer Science and Engineering Sejong University Seoul Republic of Korea
| | | | - Seong Han Kim
- Department of Intelligent Mechatronics Engineering Sejong University Seoul Republic of Korea
| | - Mashael M. Alsulami
- Department of Information Technology, College of Computers and Information Technology Taif University Taif Saudi Arabia
| | - Seung Won Lee
- Department of Precision Medicine Sungkyunkwan University School of Medicine Suwon Republic of Korea
| |
Collapse
|
64
|
Liao H, Li Y. LFC-UNet: learned lossless medical image fast compression with U-Net. PeerJ Comput Sci 2024; 10:e1924. [PMID: 38435602 PMCID: PMC10909184 DOI: 10.7717/peerj-cs.1924] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/07/2023] [Accepted: 02/13/2024] [Indexed: 03/05/2024]
Abstract
In the field of medicine, the rapid advancement of medical technology has significantly increased the speed of medical image generation, compelling us to seek efficient methods for image compression. Neural networks, owing to their outstanding image estimation capabilities, have provided new avenues for lossless compression. In recent years, learning-based lossless image compression methods, combining neural network predictions with residuals, have achieved performance comparable to traditional non-learning algorithms. However, existing methods have not taken into account that residuals often concentrate excessively, hindering the neural network's ability to learn accurate residual probability estimation. To address this issue, this study employs a weighted cross-entropy method to handle the imbalance in residual categories. In terms of network architecture, we introduce skip connections from U-Net to better capture image features, thereby obtaining accurate probability estimates. Furthermore, our framework boasts excellent encoding speed, as the model is able to acquire all residuals and residual probabilities in a single inference pass. The experimental results demonstrate that the proposed method achieves state-of-the-art performance on medical datasets while also offering the fastest processing speed. As illustrated by an instance using head CT data, our approach achieves a compression efficiency of 2.30 bits per pixel, with a processing time of only 0.320 seconds per image.
Collapse
Affiliation(s)
- Hengrui Liao
- School of Computer, University of South China, Hengyang, Hunan, China
| | - Yue Li
- School of Computer, University of South China, Hengyang, Hunan, China
| |
Collapse
|
65
|
Kumar S, Kumar H, Kumar G, Singh SP, Bijalwan A, Diwakar M. A methodical exploration of imaging modalities from dataset to detection through machine learning paradigms in prominent lung disease diagnosis: a review. BMC Med Imaging 2024; 24:30. [PMID: 38302883 PMCID: PMC10832080 DOI: 10.1186/s12880-024-01192-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2023] [Accepted: 01/03/2024] [Indexed: 02/03/2024] Open
Abstract
BACKGROUND Lung diseases, both infectious and non-infectious, are the most prevalent cause of mortality overall in the world. Medical research has identified pneumonia, lung cancer, and Corona Virus Disease 2019 (COVID-19) as prominent lung diseases prioritized over others. Imaging modalities, including X-rays, computer tomography (CT) scans, magnetic resonance imaging (MRIs), positron emission tomography (PET) scans, and others, are primarily employed in medical assessments because they provide computed data that can be utilized as input datasets for computer-assisted diagnostic systems. Imaging datasets are used to develop and evaluate machine learning (ML) methods to analyze and predict prominent lung diseases. OBJECTIVE This review analyzes ML paradigms, imaging modalities' utilization, and recent developments for prominent lung diseases. Furthermore, the research also explores various datasets available publically that are being used for prominent lung diseases. METHODS The well-known databases of academic studies that have been subjected to peer review, namely ScienceDirect, arXiv, IEEE Xplore, MDPI, and many more, were used for the search of relevant articles. Applied keywords and combinations used to search procedures with primary considerations for review, such as pneumonia, lung cancer, COVID-19, various imaging modalities, ML, convolutional neural networks (CNNs), transfer learning, and ensemble learning. RESULTS This research finding indicates that X-ray datasets are preferred for detecting pneumonia, while CT scan datasets are predominantly favored for detecting lung cancer. Furthermore, in COVID-19 detection, X-ray datasets are prioritized over CT scan datasets. The analysis reveals that X-rays and CT scans have surpassed all other imaging techniques. It has been observed that using CNNs yields a high degree of accuracy and practicability in identifying prominent lung diseases. Transfer learning and ensemble learning are complementary techniques to CNNs to facilitate analysis. Furthermore, accuracy is the most favored metric for assessment.
Collapse
Affiliation(s)
- Sunil Kumar
- Department of Computer Engineering, J. C. Bose University of Science and Technology, YMCA, Faridabad, India
- Department of Information Technology, School of Engineering and Technology (UIET), CSJM University, Kanpur, India
| | - Harish Kumar
- Department of Computer Engineering, J. C. Bose University of Science and Technology, YMCA, Faridabad, India
| | - Gyanendra Kumar
- Department of Computer and Communication Engineering, Manipal University Jaipur, Jaipur, India
| | | | - Anchit Bijalwan
- Faculty of Electrical and Computer Engineering, Arba Minch University, Arba Minch, Ethiopia.
| | - Manoj Diwakar
- Department of Computer Science and Engineering, Graphic Era Deemed to Be University, Dehradun, India
| |
Collapse
|
66
|
Chen S, Ren S, Wang G, Huang M, Xue C. Interpretable CNN-Multilevel Attention Transformer for Rapid Recognition of Pneumonia From Chest X-Ray Images. IEEE J Biomed Health Inform 2024; 28:753-764. [PMID: 37027681 DOI: 10.1109/jbhi.2023.3247949] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/24/2023]
Abstract
Chest imaging plays an essential role in diagnosing and predicting patients with COVID-19 with evidence of worsening respiratory status. Many deep learning-based approaches for pneumonia recognition have been developed to enable computer-aided diagnosis. However, the long training and inference time makes them inflexible, and the lack of interpretability reduces their credibility in clinical medical practice. This paper aims to develop a pneumonia recognition framework with interpretability, which can understand the complex relationship between lung features and related diseases in chest X-ray (CXR) images to provide high-speed analytics support for medical practice. To reduce the computational complexity to accelerate the recognition process, a novel multi-level self-attention mechanism within Transformer has been proposed to accelerate convergence and emphasize the task-related feature regions. Moreover, a practical CXR image data augmentation has been adopted to address the scarcity of medical image data problems to boost the model's performance. The effectiveness of the proposed method has been demonstrated on the classic COVID-19 recognition task using the widespread pneumonia CXR image dataset. In addition, abundant ablation experiments validate the effectiveness and necessity of all of the components of the proposed method.
Collapse
|
67
|
Prince R, Niu Z, Khan ZY, Emmanuel M, Patrick N. COVID-19 detection from chest X-ray images using CLAHE-YCrCb, LBP, and machine learning algorithms. BMC Bioinformatics 2024; 25:28. [PMID: 38233764 PMCID: PMC10792799 DOI: 10.1186/s12859-023-05427-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/11/2023] [Accepted: 07/20/2023] [Indexed: 01/19/2024] Open
Abstract
BACKGROUND COVID-19 is a disease that caused a contagious respiratory ailment that killed and infected hundreds of millions. It is necessary to develop a computer-based tool that is fast, precise, and inexpensive to detect COVID-19 efficiently. Recent studies revealed that machine learning and deep learning models accurately detect COVID-19 using chest X-ray (CXR) images. However, they exhibit notable limitations, such as a large amount of data to train, larger feature vector sizes, enormous trainable parameters, expensive computational resources (GPUs), and longer run-time. RESULTS In this study, we proposed a new approach to address some of the above-mentioned limitations. The proposed model involves the following steps: First, we use contrast limited adaptive histogram equalization (CLAHE) to enhance the contrast of CXR images. The resulting images are converted from CLAHE to YCrCb color space. We estimate reflectance from chrominance using the Illumination-Reflectance model. Finally, we use a normalized local binary patterns histogram generated from reflectance (Cr) and YCb as the classification feature vector. Decision tree, Naive Bayes, support vector machine, K-nearest neighbor, and logistic regression were used as the classification algorithms. The performance evaluation on the test set indicates that the proposed approach is superior, with accuracy rates of 99.01%, 100%, and 98.46% across three different datasets, respectively. Naive Bayes, a probabilistic machine learning algorithm, emerged as the most resilient. CONCLUSION Our proposed method uses fewer handcrafted features, affordable computational resources, and less runtime than existing state-of-the-art approaches. Emerging nations where radiologists are in short supply can adopt this prototype. We made both coding materials and datasets accessible to the general public for further improvement. Check the manuscript's availability of the data and materials under the declaration section for access.
Collapse
Affiliation(s)
- Rukundo Prince
- Department of Computer Science and Technology, Beijing Institute of Technology, Beijing, China
| | - Zhendong Niu
- Department of Computer Science and Technology, Beijing Institute of Technology, Beijing, China.
| | - Zahid Younas Khan
- Computer Science and Information Technology, University of Azad Jammu and Kashmir, Kashmir, Pakistan
| | - Masabo Emmanuel
- Software Engineering, African Center of Excellence in Data Science(ACE-DS), and the African Center of Excellence in Internet of Things(ACEIoT), University of Rwanda, Kigali, Rwanda
| | - Niyishaka Patrick
- Computer and Information Sciences, University of Hyderabad, Hyderabad, India
| |
Collapse
|
68
|
Kaur BP, Singh H, Hans R, Sharma SK, Kaushal C, Hassan MM, Shah MA. An augmentation aided concise CNN based architecture for COVID-19 diagnosis in real time. Sci Rep 2024; 14:1136. [PMID: 38212647 PMCID: PMC10784465 DOI: 10.1038/s41598-024-51317-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/28/2023] [Accepted: 01/03/2024] [Indexed: 01/13/2024] Open
Abstract
Over 6.5 million people around the world have lost their lives due to the highly contagious COVID 19 virus. The virus increases the danger of fatal health effects by damaging the lungs severely. The only method to reduce mortality and contain the spread of this disease is by promptly detecting it. Recently, deep learning has become one of the most prominent approaches to CAD, helping surgeons make more informed decisions. But deep learning models are computation hungry and devices with TPUs and GPUs are needed to run these models. The current focus of machine learning research is on developing models that can be deployed on mobile and edge devices. To this end, this research aims to develop a concise convolutional neural network-based computer-aided diagnostic system for detecting the COVID 19 virus in X-ray images, which may be deployed on devices with limited processing resources, such as mobile phones and tablets. The proposed architecture aspires to use the image enhancement in first phase and data augmentation in the second phase for image pre-processing, additionally hyperparameters are also optimized to obtain the optimal parameter settings in the third phase that provide the best results. The experimental analysis has provided empirical evidence of the impact of image enhancement, data augmentation, and hyperparameter tuning on the proposed convolutional neural network model, which increased accuracy from 94 to 98%. Results from the evaluation show that the suggested method gives an accuracy of 98%, which is better than popular transfer learning models like Xception, Resnet50, and Inception.
Collapse
Affiliation(s)
- Balraj Preet Kaur
- Department of Computer Science and Engineering, DAV University, Jalandhar, India
| | - Harpreet Singh
- Department of Computer Science and Engineering, Thapar Institute of Engineering and Technology, Patiala, India
| | - Rahul Hans
- Department of Computer Science and Engineering, DAV University, Jalandhar, India
| | - Sanjeev Kumar Sharma
- Department of Computer Science and Applications, DAV University, Jalandhar, India
| | - Chetna Kaushal
- Chitkara University Institute of Engineering and Technology, Chitkara University, Punjab, 140401, India
| | - Md Mehedi Hassan
- Computer Science and Engineering Discipline, Khulna University, Khulna, 9208, Bangladesh
| | - Mohd Asif Shah
- Department of Economics, Kebri Dehar University, Kebri Dehar, 250, Ethiopia.
- Centre of Research Impact and Outcome, Chitkara University Institute of Engineering and Technology, Chitkara University, Rajpura, 140401, Punjab, India.
- Division of Research and Development, Lovely Professional University, Phagwara, 144001, Punjab, India.
| |
Collapse
|
69
|
Chauhan S, Edla DR, Boddu V, Rao MJ, Cheruku R, Nayak SR, Martha S, Lavanya K, Nigat TD. Detection of COVID-19 using edge devices by a light-weight convolutional neural network from chest X-ray images. BMC Med Imaging 2024; 24:1. [PMID: 38166813 PMCID: PMC10759384 DOI: 10.1186/s12880-023-01155-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/21/2023] [Accepted: 11/14/2023] [Indexed: 01/05/2024] Open
Abstract
Deep learning is a highly significant technology in clinical treatment and diagnostics nowadays. Convolutional Neural Network (CNN) is a new idea in deep learning that is being used in the area of computer vision. The COVID-19 detection is the subject of our medical study. Researchers attempted to increase the detection accuracy but at the cost of high model complexity. In this paper, we desire to achieve better accuracy with little training space and time so that this model easily deployed in edge devices. In this paper, a new CNN design is proposed that has three stages: pre-processing, which removes the black padding on the side initially; convolution, which employs filter banks; and feature extraction, which makes use of deep convolutional layers with skip connections. In order to train the model, chest X-ray images are partitioned into three sets: learning(0.7), validation(0.1), and testing(0.2). The models are then evaluated using the test and training data. The LMNet, CoroNet, CVDNet, and Deep GRU-CNN models are the other four models used in the same experiment. The propose model achieved 99.47% & 98.91% accuracy on training and testing respectively. Additionally, it achieved 97.54%, 98.19%, 99.49%, and 97.86% scores for precision, recall, specificity, and f1-score respectively. The proposed model obtained nearly equivalent accuracy and other similar metrics when compared with other models but greatly reduced the model complexity. Moreover, it is found that proposed model is less prone to over fitting as compared to other models.
Collapse
Affiliation(s)
- Sohamkumar Chauhan
- Department of Computer Science and Engineering, National Institute of Technology Goa, Ponda, 403401, Goa, India
| | - Damoder Reddy Edla
- Department of Computer Science and Engineering, National Institute of Technology Goa, Ponda, 403401, Goa, India
| | - Vijayasree Boddu
- Department of Electronics and Communication Engineering, National Institute of Technology Warangal, Hanamkonda, 506004, Telangana, India
| | - M Jayanthi Rao
- Department of CSE, Aditya Institute of Technology and Management, Kotturu, Tekkali, Andhra Pradesh, India
| | - Ramalingaswamy Cheruku
- Department of Computer Science and Engineering, National Institute of Technology Warangal, Hanamkonda, 506004, Telangana, India
| | - Soumya Ranjan Nayak
- School of Computer Engineering, KIIT Deemed to be University, Bhubaneswar, 751024, Odisha, India
| | - Sheshikala Martha
- School of Computer Science and Artificial Intelligence, SR University, Warangal, 506004, Telangana, India
| | - Kamppa Lavanya
- University College of Sciences, Acharya Nagarjuna Univesity, Guntur, Andhra Pradesh, India
| | - Tsedenya Debebe Nigat
- Faculty of Computing and Informatics, Jimma Institute of Technology, Jimma, Oromia, Ethiopia.
| |
Collapse
|
70
|
Ouis MY, A Akhloufi M. Deep learning for report generation on chest X-ray images. Comput Med Imaging Graph 2024; 111:102320. [PMID: 38134726 DOI: 10.1016/j.compmedimag.2023.102320] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/18/2023] [Revised: 11/13/2023] [Accepted: 11/29/2023] [Indexed: 12/24/2023]
Abstract
Medical imaging, specifically chest X-ray image analysis, is a crucial component of early disease detection and screening in healthcare. Deep learning techniques, such as convolutional neural networks (CNNs), have emerged as powerful tools for computer-aided diagnosis (CAD) in chest X-ray image analysis. These techniques have shown promising results in automating tasks such as classification, detection, and segmentation of abnormalities in chest X-ray images, with the potential to surpass human radiologists. In this review, we provide an overview of the importance of chest X-ray image analysis, historical developments, impact of deep learning techniques, and availability of labeled databases. We specifically focus on advancements and challenges in radiology report generation using deep learning, highlighting potential future advancements in this area. The use of deep learning for report generation has the potential to reduce the burden on radiologists, improve patient care, and enhance the accuracy and efficiency of chest X-ray image analysis in medical imaging.
Collapse
Affiliation(s)
- Mohammed Yasser Ouis
- Perception, Robotics and Intelligent Machines Lab(PRIME), Department of Computer Science, Université de Moncton, Moncton, NB E1C 3E9, Canada.
| | - Moulay A Akhloufi
- Perception, Robotics and Intelligent Machines Lab(PRIME), Department of Computer Science, Université de Moncton, Moncton, NB E1C 3E9, Canada.
| |
Collapse
|
71
|
Azad R, Kazerouni A, Heidari M, Aghdam EK, Molaei A, Jia Y, Jose A, Roy R, Merhof D. Advances in medical image analysis with vision Transformers: A comprehensive review. Med Image Anal 2024; 91:103000. [PMID: 37883822 DOI: 10.1016/j.media.2023.103000] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/09/2023] [Revised: 09/30/2023] [Accepted: 10/11/2023] [Indexed: 10/28/2023]
Abstract
The remarkable performance of the Transformer architecture in natural language processing has recently also triggered broad interest in Computer Vision. Among other merits, Transformers are witnessed as capable of learning long-range dependencies and spatial correlations, which is a clear advantage over convolutional neural networks (CNNs), which have been the de facto standard in Computer Vision problems so far. Thus, Transformers have become an integral part of modern medical image analysis. In this review, we provide an encyclopedic review of the applications of Transformers in medical imaging. Specifically, we present a systematic and thorough review of relevant recent Transformer literature for different medical image analysis tasks, including classification, segmentation, detection, registration, synthesis, and clinical report generation. For each of these applications, we investigate the novelty, strengths and weaknesses of the different proposed strategies and develop taxonomies highlighting key properties and contributions. Further, if applicable, we outline current benchmarks on different datasets. Finally, we summarize key challenges and discuss different future research directions. In addition, we have provided cited papers with their corresponding implementations in https://github.com/mindflow-institue/Awesome-Transformer.
Collapse
Affiliation(s)
- Reza Azad
- Faculty of Electrical Engineering and Information Technology, RWTH Aachen University, Aachen, Germany
| | - Amirhossein Kazerouni
- School of Electrical Engineering, Iran University of Science and Technology, Tehran, Iran
| | - Moein Heidari
- School of Electrical Engineering, Iran University of Science and Technology, Tehran, Iran
| | | | - Amirali Molaei
- School of Computer Engineering, Iran University of Science and Technology, Tehran, Iran
| | - Yiwei Jia
- Faculty of Electrical Engineering and Information Technology, RWTH Aachen University, Aachen, Germany
| | - Abin Jose
- Faculty of Electrical Engineering and Information Technology, RWTH Aachen University, Aachen, Germany
| | - Rijo Roy
- Faculty of Electrical Engineering and Information Technology, RWTH Aachen University, Aachen, Germany
| | - Dorit Merhof
- Faculty of Informatics and Data Science, University of Regensburg, Regensburg, Germany; Fraunhofer Institute for Digital Medicine MEVIS, Bremen, Germany.
| |
Collapse
|
72
|
Mujahid M, Rustam F, Chakrabarti P, Mallampati B, de la Torre Diez I, Gali P, Chunduri V, Ashraf I. Pneumonia detection on chest X-rays from Xception-based transfer learning and logistic regression. Technol Health Care 2024; 32:3847-3870. [PMID: 39520166 PMCID: PMC11612971 DOI: 10.3233/thc-230313] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2023] [Accepted: 03/31/2024] [Indexed: 11/16/2024]
Abstract
Pneumonia is a dangerous disease that kills millions of children and elderly patients worldwide every year. The detection of pneumonia from a chest x-ray is perpetrated by expert radiologists. The chest x-ray is cheaper and is most often used to diagnose pneumonia. However, chest x-ray-based diagnosis requires expert radiologists which is time-consuming and laborious. Moreover, COVID-19 and pneumonia have similar symptoms which leads to false positives. Machine learning-based solutions have been proposed for the automatic prediction of pneumonia from chest X-rays, however, such approaches lack robustness and high accuracy due to data imbalance and generalization errors. This study focuses on elevating the performance of machine learning models by dealing with data imbalanced problems using data augmentation. Contrary to traditional machine learning models that required hand-crafted features, this study uses transfer learning for automatic feature extraction using Xception and VGG-16 to train classifiers like support vector machine, logistic regression, K nearest neighbor, stochastic gradient descent, extra tree classifier, and gradient boosting machine. Experiments involve the use of hand-crafted features, as well as, transfer learning-based feature extraction for pneumonia detection. Performance comparison using Xception and VGG-16 features suggest that transfer learning-based features tend to show better performance than hand-crafted features and an accuracy of 99.23% can be obtained for pneumonia using chest X-rays.
Collapse
Affiliation(s)
- Muhammad Mujahid
- Artificial Intelligence & Data Analytics Lab, CCIS, Prince Sultan University, Riyadh 11586, Saudi Arabia
| | - Furqan Rustam
- School of Systems and Technology, University of Management and Technology, Lahore, Pakistan
| | | | | | - Isabel de la Torre Diez
- Department of Signal Theory and Communications and Telematic Engineering, Unviersity of Valladolid, Paseo de Belén, Spain
| | - Pradeep Gali
- University of North Texas, North Texas, Denton, TX, USA
| | | | - Imran Ashraf
- Department of Information and Communication Engineering, Yeungnam University, Gyeongsan-si, Korea
| |
Collapse
|
73
|
Talukder MA, Layek MA, Kazi M, Uddin MA, Aryal S. Empowering COVID-19 detection: Optimizing performance through fine-tuned EfficientNet deep learning architecture. Comput Biol Med 2024; 168:107789. [PMID: 38042105 DOI: 10.1016/j.compbiomed.2023.107789] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2023] [Revised: 11/21/2023] [Accepted: 11/28/2023] [Indexed: 12/04/2023]
Abstract
The worldwide COVID-19 pandemic has profoundly influenced the health and everyday experiences of individuals across the planet. It is a highly contagious respiratory disease requiring early and accurate detection to curb its rapid transmission. Initial testing methods primarily revolved around identifying the genetic composition of the coronavirus, exhibiting a relatively low detection rate and requiring a time-intensive procedure. To address this challenge, experts have suggested using radiological imagery, particularly chest X-rays, as a valuable approach within the diagnostic protocol. This study investigates the potential of leveraging radiographic imaging (X-rays) with deep learning algorithms to swiftly and precisely identify COVID-19 patients. The proposed approach elevates the detection accuracy by fine-tuning with appropriate layers on various established transfer learning models. The experimentation was conducted on a COVID-19 X-ray dataset containing 2000 images. The accuracy rates achieved were impressive of 99.55%, 97.32%, 99.11%, 99.55%, 99.11% and 100% for Xception, InceptionResNetV2, ResNet50 , ResNet50V2, EfficientNetB0 and EfficientNetB4 respectively. The fine-tuned EfficientNetB4 achieved an excellent accuracy score, showcasing its potential as a robust COVID-19 detection model. Furthermore, EfficientNetB4 excelled in identifying Lung disease using Chest X-ray dataset containing 4,350 Images, achieving remarkable performance with an accuracy of 99.17%, precision of 99.13%, recall of 99.16%, and f1-score of 99.14%. These results highlight the promise of fine-tuned transfer learning for efficient lung detection through medical imaging, especially with X-ray images. This research offers radiologists an effective means of aiding rapid and precise COVID-19 diagnosis and contributes valuable assistance for healthcare professionals in accurately identifying affected patients.
Collapse
Affiliation(s)
- Md Alamin Talukder
- Department of Computer Science and Engineering, Jagannath University, Dhaka, Bangladesh.
| | - Md Abu Layek
- Department of Computer Science and Engineering, Jagannath University, Dhaka, Bangladesh.
| | - Mohsin Kazi
- Department of Pharmaceutics, College of Pharmacy, King Saud University, P.O. Box-2457, Riyadh 11451, Saudi Arabia.
| | - Md Ashraf Uddin
- School of Information Technology, Deakin University, Waurn Ponds Campus, Geelong, Australia.
| | - Sunil Aryal
- School of Information Technology, Deakin University, Waurn Ponds Campus, Geelong, Australia.
| |
Collapse
|
74
|
Gopatoti A, Jayakumar R, Billa P, Patteeswaran V. DDA-SSNets: Dual decoder attention-based semantic segmentation networks for COVID-19 infection segmentation and classification using chest X-Ray images. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2024; 32:623-649. [PMID: 38607728 DOI: 10.3233/xst-230421] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/14/2024]
Abstract
BACKGROUND COVID-19 needs to be diagnosed and staged to be treated accurately. However, prior studies' diagnostic and staging abilities for COVID-19 infection needed to be improved. Therefore, new deep learning-based approaches are required to aid radiologists in detecting and quantifying COVID-19-related lung infections. OBJECTIVE To develop deep learning-based models to classify and quantify COVID-19-related lung infections. METHODS Initially, Dual Decoder Attention-based Semantic Segmentation Networks (DDA-SSNets) such as Dual Decoder Attention-UNet (DDA-UNet) and Dual Decoder Attention-SegNet (DDA-SegNet) are proposed to facilitate the dual segmentation tasks such as lung lobes and infection segmentation in chest X-ray (CXR) images. The lung lobe and infection segmentations are mapped to grade the severity of COVID-19 infection in both the lungs of CXRs. Later, a Genetic algorithm-based Deep Convolutional Neural Network classifier with the optimum number of layers, namely GADCNet, is proposed to classify the extracted regions of interest (ROI) from the CXR lung lobes into COVID-19 and non-COVID-19. RESULTS The DDA-SegNet shows better segmentation with an average BCSSDC of 99.53% and 99.97% for lung lobes and infection segmentations, respectively, compared with DDA-UNet with an average BCSSDC of 99.14% and 99.92%. The proposed DDA-SegNet with GADCNet classifier offered excellent classification results with an average BCCAC of 99.98%, followed by the GADCNet with DDA-UNet with an average BCCAC of 99.92% after extensive testing and analysis. CONCLUSIONS The results show that the proposed DDA-SegNet has superior performance in the segmentation of lung lobes and COVID-19-infected regions in CXRs, along with improved severity grading compared to the DDA-UNet and improved accuracy of the GADCNet classifier in classifying the CXRs into COVID-19, and non-COVID-19.
Collapse
Affiliation(s)
- Anandbabu Gopatoti
- Department of Electronics and Communication Engineering, Hindusthan College of Engineering and Technology, Coimbatore, Tamil Nadu, India
| | - Ramya Jayakumar
- Department of Electronics and Communication Engineering, Hindusthan College of Engineering and Technology, Coimbatore, Tamil Nadu, India
| | - Poornaiah Billa
- Department of Electronics and Communication Engineering, Lakireddy Bali Reddy College of Engineering, Mylavaram, Andhra Pradesh, India
| | - Vijayalakshmi Patteeswaran
- Department of Electronics and Communication Engineering, Hindusthan College of Engineering and Technology, Coimbatore, Tamil Nadu, India
| |
Collapse
|
75
|
Islam MT, Zhou Z, Ren H, Khuzani MB, Kapp D, Zou J, Tian L, Liao JC, Xing L. Revealing hidden patterns in deep neural network feature space continuum via manifold learning. Nat Commun 2023; 14:8506. [PMID: 38129376 PMCID: PMC10739971 DOI: 10.1038/s41467-023-43958-w] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/17/2023] [Accepted: 11/24/2023] [Indexed: 12/23/2023] Open
Abstract
Deep neural networks (DNNs) extract thousands to millions of task-specific features during model training for inference and decision-making. While visualizing these features is critical for comprehending the learning process and improving the performance of the DNNs, existing visualization techniques work only for classification tasks. For regressions, the feature points lie on a high dimensional continuum having an inherently complex shape, making a meaningful visualization of the features intractable. Given that the majority of deep learning applications are regression-oriented, developing a conceptual framework and computational method to reliably visualize the regression features is of great significance. Here, we introduce a manifold discovery and analysis (MDA) method for DNN feature visualization, which involves learning the manifold topology associated with the output and target labels of a DNN. MDA leverages the acquired topological information to preserve the local geometry of the feature space manifold and provides insightful visualizations of the DNN features, highlighting the appropriateness, generalizability, and adversarial robustness of a DNN. The performance and advantages of the MDA approach compared to the existing methods are demonstrated in different deep learning applications.
Collapse
Affiliation(s)
- Md Tauhidul Islam
- Department of Radiation Oncology, Stanford University, Stanford, CA, 94305, USA
| | - Zixia Zhou
- Department of Radiation Oncology, Stanford University, Stanford, CA, 94305, USA
| | - Hongyi Ren
- Department of Radiation Oncology, Stanford University, Stanford, CA, 94305, USA
| | | | - Daniel Kapp
- Department of Radiation Oncology, Stanford University, Stanford, CA, 94305, USA
| | - James Zou
- Department of Biomedical Data Science, Stanford University, Stanford, CA, 94305, USA
| | - Lu Tian
- Department of Biomedical Data Science, Stanford University, Stanford, CA, 94305, USA
| | - Joseph C Liao
- Department of Urology, Stanford University, Stanford, CA, 94305, USA.
| | - Lei Xing
- Department of Radiation Oncology, Stanford University, Stanford, CA, 94305, USA.
| |
Collapse
|
76
|
Phumkuea T, Wongsirichot T, Damkliang K, Navasakulpong A, Andritsch J. MSTAC: A Multi-Stage Automated Classification of COVID-19 Chest X-ray Images Using Stacked CNN Models. Tomography 2023; 9:2233-2246. [PMID: 38133077 PMCID: PMC10747997 DOI: 10.3390/tomography9060173] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/02/2023] [Revised: 12/08/2023] [Accepted: 12/08/2023] [Indexed: 12/23/2023] Open
Abstract
This study introduces a Multi-Stage Automated Classification (MSTAC) system for COVID-19 chest X-ray (CXR) images, utilizing stacked Convolutional Neural Network (CNN) models. Suspected COVID-19 patients often undergo CXR imaging, making it valuable for disease classification. The study collected CXR images from public datasets and aimed to differentiate between COVID-19, non-COVID-19, and healthy cases. MSTAC employs two classification stages: the first distinguishes healthy from unhealthy cases, and the second further classifies COVID-19 and non-COVID-19 cases. Compared to a single CNN-Multiclass model, MSTAC demonstrated superior classification performance, achieving 97.30% accuracy and sensitivity. In contrast, the CNN-Multiclass model showed 94.76% accuracy and sensitivity. MSTAC's effectiveness is highlighted in its promising results over the CNN-Multiclass model, suggesting its potential to assist healthcare professionals in efficiently diagnosing COVID-19 cases. The system outperformed similar techniques, emphasizing its accuracy and efficiency in COVID-19 diagnosis. This research underscores MSTAC as a valuable tool in medical image analysis for enhanced disease classification.
Collapse
Affiliation(s)
- Thanakorn Phumkuea
- College of Digital Science, Prince of Songkla University, Songkhla 90110, Thailand
| | - Thakerng Wongsirichot
- Division of Computational Science, Faculty of Science, Prince of Songkla University, Songkhla 90110, Thailand;
| | - Kasikrit Damkliang
- Division of Computational Science, Faculty of Science, Prince of Songkla University, Songkhla 90110, Thailand;
| | - Asma Navasakulpong
- Division of Respiratory and Respiratory Critical Care Medicine, Prince of Songkla University, Songkhla 90110, Thailand;
| | - Jarutas Andritsch
- Faculty of Business, Law and Digital Technologies, Solent University, Southampton SO14 0YN, UK;
| |
Collapse
|
77
|
Wu JCH, Yu HW, Tsai TH, Lu HHS. Dynamically Synthetic Images for Federated Learning of medical images. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2023; 242:107845. [PMID: 37852147 DOI: 10.1016/j.cmpb.2023.107845] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/01/2023] [Revised: 09/28/2023] [Accepted: 10/03/2023] [Indexed: 10/20/2023]
Abstract
BACKGROUND To develop deep learning models for medical diagnosis, it is important to collect more medical data from several medical institutions. Due to the regulations for privacy concerns, it is infeasible to collect data from various medical institutions to one institution for centralized learning. Federated Learning (FL) provides a feasible approach to jointly train the deep learning model with data stored in various medical institutions instead of collected together. However, the resulting FL models could be biased towards institutions with larger training datasets. METHODOLOGY In this study, we propose the applicable method of Dynamically Synthetic Images for Federated Learning (DSIFL) that aims to integrate the information of local institutions with heterogeneous types of data. The main technique of DSIFL is to develop a synthetic method that can dynamically adjust the number of synthetic images similar to local data that are misclassified by the current model. The resulting global model can handle the diversity in heterogeneous types of data collected in local medical institutions by including the training of synthetic images similar to misclassified cases in local collections. RESULTS In model performance evaluation metrics, we focus on the accuracy of each client's dataset. Finally, the accuracy of the model of DSIFL in the experiments can achieve the higher accuracy of the FL approach. CONCLUSION In this study, we propose the framework of DSIFL that achieves improvements over the conventional FL approach. We conduct empirical studies with two kinds of medical images. We compare the performance by variants of FL vs. DSIFL approaches. The performance by individual training is used as the baseline, whereas the performance by centralized learning is used as the target for the comparison studies. The empirical findings suggest that the DSIFL has improved performance over the FL via the technique of dynamically synthetic images in training.
Collapse
Affiliation(s)
- Jacky Chung-Hao Wu
- Institute of Statistics, National Yang Ming Chiao Tung University, Hsinchu, Taiwan, ROC
| | - Hsuan-Wen Yu
- Institute of Statistics, National Yang Ming Chiao Tung University, Hsinchu, Taiwan, ROC
| | - Tsung-Hung Tsai
- Institute of Statistics, National Yang Ming Chiao Tung University, Hsinchu, Taiwan, ROC
| | - Henry Horng-Shing Lu
- Institute of Statistics, National Yang Ming Chiao Tung University, Hsinchu, Taiwan, ROC; Department of Statistics and Data Science, Cornell University, New York, USA.
| |
Collapse
|
78
|
Zhang Z, Wu H, Zhao H, Shi Y, Wang J, Bai H, Sun B. A Novel Deep Learning Model for Medical Image Segmentation with Convolutional Neural Network and Transformer. Interdiscip Sci 2023; 15:663-677. [PMID: 37665496 DOI: 10.1007/s12539-023-00585-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/09/2023] [Revised: 07/26/2023] [Accepted: 08/01/2023] [Indexed: 09/05/2023]
Abstract
Accurate segmentation of medical images is essential for clinical decision-making, and deep learning techniques have shown remarkable results in this area. However, existing segmentation models that combine transformer and convolutional neural networks often use skip connections in U-shaped networks, which may limit their ability to capture contextual information in medical images. To address this limitation, we propose a coordinated mobile and residual transformer UNet (MRC-TransUNet) that combines the strengths of transformer and UNet architectures. Our approach uses a lightweight MR-ViT to address the semantic gap and a reciprocal attention module to compensate for the potential loss of details. To better explore long-range contextual information, we use skip connections only in the first layer and add MR-ViT and RPA modules in the subsequent downsampling layers. In our study, we evaluated the effectiveness of our proposed method on three different medical image segmentation datasets, namely, breast, brain, and lung. Our proposed method outperformed state-of-the-art methods in terms of various evaluation metrics, including the Dice coefficient and Hausdorff distance. These results demonstrate that our proposed method can significantly improve the accuracy of medical image segmentation and has the potential for clinical applications. Illustration of the proposed MRC-TransUNet. For the input medical images, we first subject them to an intrinsic downsampling operation and then replace the original jump connection structure using MR-ViT. The output feature representations at different scales are fused by the RPA module. Finally, an upsampling operation is performed to fuse the features to restore them to the same resolution as the input image.
Collapse
Affiliation(s)
- Zhuo Zhang
- Tianjin Key Laboratory of Optoelectronic Detection Technology and Systems, School of Electronic and Information Engineering, Tiangong University, Tianjin, 300387, China
| | - Hongbing Wu
- School of Computer Science and Technology, Tiangong University, Tianjin, 300387, China
| | - Huan Zhao
- Tianjin Key Laboratory of Optoelectronic Detection Technology and Systems, School of Electronic and Information Engineering, Tiangong University, Tianjin, 300387, China
| | - Yicheng Shi
- College of Management and Economics, Tianjin University, Tianjin, 300072, China
| | - Jifang Wang
- Tianjin Key Laboratory of Optoelectronic Detection Technology and Systems, School of Electronic and Information Engineering, Tiangong University, Tianjin, 300387, China
| | - Hua Bai
- Tianjin Key Laboratory of Optoelectronic Detection Technology and Systems, School of Electronic and Information Engineering, Tiangong University, Tianjin, 300387, China.
| | - Baoshan Sun
- School of Computer Science and Technology, Tiangong University, Tianjin, 300387, China.
| |
Collapse
|
79
|
Fan R, Bowd C, Brye N, Christopher M, Weinreb RN, Kriegman DJ, Zangwill LM. One-Vote Veto: Semi-Supervised Learning for Low-Shot Glaucoma Diagnosis. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:3764-3778. [PMID: 37610903 PMCID: PMC11214580 DOI: 10.1109/tmi.2023.3307689] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/25/2023]
Abstract
Convolutional neural networks (CNNs) are a promising technique for automated glaucoma diagnosis from images of the fundus, and these images are routinely acquired as part of an ophthalmic exam. Nevertheless, CNNs typically require a large amount of well-labeled data for training, which may not be available in many biomedical image classification applications, especially when diseases are rare and where labeling by experts is costly. This article makes two contributions to address this issue: 1) It extends the conventional Siamese network and introduces a training method for low-shot learning when labeled data are limited and imbalanced, and 2) it introduces a novel semi-supervised learning strategy that uses additional unlabeled training data to achieve greater accuracy. Our proposed multi-task Siamese network (MTSN) can employ any backbone CNN, and we demonstrate with four backbone CNNs that its accuracy with limited training data approaches the accuracy of backbone CNNs trained with a dataset that is 50 times larger. We also introduce One-Vote Veto (OVV) self-training, a semi-supervised learning strategy that is designed specifically for MTSNs. By taking both self-predictions and contrastive predictions of the unlabeled training data into account, OVV self-training provides additional pseudo labels for fine-tuning a pre-trained MTSN. Using a large (imbalanced) dataset with 66,715 fundus photographs acquired over 15 years, extensive experimental results demonstrate the effectiveness of low-shot learning with MTSN and semi-supervised learning with OVV self-training. Three additional, smaller clinical datasets of fundus images acquired under different conditions (cameras, instruments, locations, populations) are used to demonstrate the generalizability of the proposed methods.
Collapse
|
80
|
Yue G, Yang C, Zhao Z, An Z, Yang Y. ERGPNet: lesion segmentation network for COVID-19 chest X-ray images based on embedded residual convolution and global perception. Front Physiol 2023; 14:1296185. [PMID: 38028767 PMCID: PMC10679680 DOI: 10.3389/fphys.2023.1296185] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2023] [Accepted: 11/02/2023] [Indexed: 12/01/2023] Open
Abstract
The Segmentation of infected areas from COVID-19 chest X-ray (CXR) images is of great significance for the diagnosis and treatment of patients. However, accurately and effectively segmenting infected areas of CXR images is still challenging due to the inherent ambiguity of CXR images and the cross-scale variations in infected regions. To address these issues, this article proposes a ERGPNet based on embedded residuals and global perception, to segment lesion regions in COVID-19 CXR images. First, aiming at the inherent fuzziness of CXR images, an embedded residual convolution structure is proposed to enhance the ability of internal feature extraction. Second, a global information perception module is constructed to guide the network in generating long-distance information flow, alleviating the interferences of cross-scale variations on the algorithm's discrimination ability. Finally, the network's sensitivity to target regions is improved, and the interference of noise information is suppressed through the utilization of parallel spatial and serial channel attention modules. The interactions between each module fully establish the mapping relationship between feature representation and information decision-making and improve the accuracy of lesion segmentation. Extensive experiments on three datasets of COVID-19 CXR images, and the results demonstrate that the proposed method outperforms other state-of-the-art segmentation methods of CXR images.
Collapse
Affiliation(s)
- Gongtao Yue
- School of Computer Science, Xijing University, Xi’an, China
| | - Chen Yang
- School of Computer Science, Xijing University, Xi’an, China
| | - Zhengyang Zhao
- School of Information and Navigation, Air Force Engineering University, Xi’an, China
| | - Ziheng An
- School of Integrated Circuits, Anhui University, Hefei, China
| | - Yongsheng Yang
- School of Computer Science, Xijing University, Xi’an, China
| |
Collapse
|
81
|
Sun J, Shi W, Giuste FO, Vaghani YS, Tang L, Wang MD. Improving explainable AI with patch perturbation-based evaluation pipeline: a COVID-19 X-ray image analysis case study. Sci Rep 2023; 13:19488. [PMID: 37945586 PMCID: PMC10636093 DOI: 10.1038/s41598-023-46493-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/21/2023] [Accepted: 11/01/2023] [Indexed: 11/12/2023] Open
Abstract
Recent advances in artificial intelligence (AI) have sparked interest in developing explainable AI (XAI) methods for clinical decision support systems, especially in translational research. Although using XAI methods may enhance trust in black-box models, evaluating their effectiveness has been challenging, primarily due to the absence of human (expert) intervention, additional annotations, and automated strategies. In order to conduct a thorough assessment, we propose a patch perturbation-based approach to automatically evaluate the quality of explanations in medical imaging analysis. To eliminate the need for human efforts in conventional evaluation methods, our approach executes poisoning attacks during model retraining by generating both static and dynamic triggers. We then propose a comprehensive set of evaluation metrics during the model inference stage to facilitate the evaluation from multiple perspectives, covering a wide range of correctness, completeness, consistency, and complexity. In addition, we include an extensive case study to showcase the proposed evaluation strategy by applying widely-used XAI methods on COVID-19 X-ray imaging classification tasks, as well as a thorough review of existing XAI methods in medical imaging analysis with evaluation availability. The proposed patch perturbation-based workflow offers model developers an automated and generalizable evaluation strategy to identify potential pitfalls and optimize their proposed explainable solutions, while also aiding end-users in comparing and selecting appropriate XAI methods that meet specific clinical needs in real-world clinical research and practice.
Collapse
Affiliation(s)
- Jimin Sun
- School of Computer Science and Engineering, Georgia Institute of Technology, Atlanta, 30322, USA
| | - Wenqi Shi
- School of Electrical and Computer Engineering, Georgia Institute of Technology, Atlanta, 30322, USA
| | - Felipe O Giuste
- The Wallace H. Coulter Department of Biomedical Engineering, Georgia Institute of Technology and Emory University, Atlanta, 30322, USA
| | - Yog S Vaghani
- The Wallace H. Coulter Department of Biomedical Engineering, Georgia Institute of Technology and Emory University, Atlanta, 30322, USA
| | - Lingzi Tang
- The Wallace H. Coulter Department of Biomedical Engineering, Georgia Institute of Technology and Emory University, Atlanta, 30322, USA
| | - May D Wang
- The Wallace H. Coulter Department of Biomedical Engineering, Georgia Institute of Technology and Emory University, Atlanta, 30322, USA.
| |
Collapse
|
82
|
Celniak W, Wodziński M, Jurgas A, Burti S, Zotti A, Atzori M, Müller H, Banzato T. Improving the classification of veterinary thoracic radiographs through inter-species and inter-pathology self-supervised pre-training of deep learning models. Sci Rep 2023; 13:19518. [PMID: 37945653 PMCID: PMC10636209 DOI: 10.1038/s41598-023-46345-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/27/2023] [Accepted: 10/31/2023] [Indexed: 11/12/2023] Open
Abstract
The analysis of veterinary radiographic imaging data is an essential step in the diagnosis of many thoracic lesions. Given the limited time that physicians can devote to a single patient, it would be valuable to implement an automated system to help clinicians make faster but still accurate diagnoses. Currently, most of such systems are based on supervised deep learning approaches. However, the problem with these solutions is that they need a large database of labeled data. Access to such data is often limited, as it requires a great investment of both time and money. Therefore, in this work we present a solution that allows higher classification scores to be obtained using knowledge transfer from inter-species and inter-pathology self-supervised learning methods. Before training the network for classification, pretraining of the model was performed using self-supervised learning approaches on publicly available unlabeled radiographic data of human and dog images, which allowed substantially increasing the number of images for this phase. The self-supervised learning approaches included the Beta Variational Autoencoder, the Soft-Introspective Variational Autoencoder, and a Simple Framework for Contrastive Learning of Visual Representations. After the initial pretraining, fine-tuning was performed for the collected veterinary dataset using 20% of the available data. Next, a latent space exploration was performed for each model after which the encoding part of the model was fine-tuned again, this time in a supervised manner for classification. Simple Framework for Contrastive Learning of Visual Representations proved to be the most beneficial pretraining method. Therefore, it was for this method that experiments with various fine-tuning methods were carried out. We achieved a mean ROC AUC score of 0.77 and 0.66, respectively, for the laterolateral and dorsoventral projection datasets. The results show significant improvement compared to using the model without any pretraining approach.
Collapse
Affiliation(s)
- Weronika Celniak
- University of Applied Sciences Western Switzerland (HES-SO), 3960, Sierre, Switzerland.
- Faculty of Electrical Engineering, Automatics, Computer Science and Biomedical Engineering, AGH University of Krakow, 30059, Kraków, Poland.
| | - Marek Wodziński
- University of Applied Sciences Western Switzerland (HES-SO), 3960, Sierre, Switzerland
- Faculty of Electrical Engineering, Automatics, Computer Science and Biomedical Engineering, AGH University of Krakow, 30059, Kraków, Poland
| | - Artur Jurgas
- University of Applied Sciences Western Switzerland (HES-SO), 3960, Sierre, Switzerland
- Faculty of Electrical Engineering, Automatics, Computer Science and Biomedical Engineering, AGH University of Krakow, 30059, Kraków, Poland
| | - Silvia Burti
- Department of Animal Medicine, Productions, and Health, Legnaro (PD), University of Padua, 35020, Padua, Italy
| | - Alessandro Zotti
- Department of Animal Medicine, Productions, and Health, Legnaro (PD), University of Padua, 35020, Padua, Italy
| | - Manfredo Atzori
- University of Applied Sciences Western Switzerland (HES-SO), 3960, Sierre, Switzerland
- Department of Neuroscience, University of Padua, 35121, Padua, IT, Italy
- Padova Neuroscience Center, University of Padova, Via Orus 2/B, 35129, Padova, Italy
| | - Henning Müller
- University of Applied Sciences Western Switzerland (HES-SO), 3960, Sierre, Switzerland
- Medical Faculty, University of Geneva, 1206, Geneva, Switzerland
- The Sense Research and Innovation Insitute, 1950, Sion, Switzerland
| | - Tommaso Banzato
- Department of Animal Medicine, Productions, and Health, Legnaro (PD), University of Padua, 35020, Padua, Italy
| |
Collapse
|
83
|
Saju B, Tressa N, Dhanaraj RK, Tharewal S, Mathew JC, Pelusi D. Effective multi-class lungdisease classification using the hybridfeature engineering mechanism. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2023; 20:20245-20273. [PMID: 38052644 DOI: 10.3934/mbe.2023896] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/07/2023]
Abstract
The utilization of computational models in the field of medical image classification is an ongoing and unstoppable trend, driven by the pursuit of aiding medical professionals in achieving swift and precise diagnoses. Post COVID-19, many researchers are studying better classification and diagnosis of lung diseases particularly, as it was reported that one of the very few diseases greatly affecting human beings was related to lungs. This research study, as presented in the paper, introduces an advanced computer-assisted model that is specifically tailored for the classification of 13 lung diseases using deep learning techniques, with a focus on analyzing chest radiograph images. The work flows from data collection, image quality enhancement, feature extraction to a comparative classification performance analysis. For data collection, an open-source data set consisting of 112,000 chest X-Ray images was used. Since, the quality of the pictures was significant for the work, enhanced image quality is achieved through preprocessing techniques such as Otsu-based binary conversion, contrast limited adaptive histogram equalization-driven noise reduction, and Canny edge detection. Feature extraction incorporates connected regions, histogram of oriented gradients, gray-level co-occurrence matrix and Haar wavelet transformation, complemented by feature selection via regularized neighbourhood component analysis. The paper proposes an optimized hybrid model, improved Aquila optimization convolutional neural networks (CNN), which is a combination of optimized CNN and DENSENET121 with applied batch equalization, which provides novelty for the model compared with other similar works. The comparative evaluation of classification performance among CNN, DENSENET121 and the proposed hybrid model is also done to find the results. The findings highlight the proposed hybrid model's supremacy, boasting 97.00% accuracy, 94.00% precision, 96.00% sensitivity, 96.00% specificity and 95.00% F1-score. In the future, potential avenues encompass exploring explainable machine learning for discerning model decisions and optimizing performance through strategic model restructuring.
Collapse
Affiliation(s)
- Binju Saju
- Department of Master of Computer Applications, New Horizon College of Engineering, Bengaluru, India
| | - Neethu Tressa
- Department of Master of Computer Applications, New Horizon College of Engineering, Bengaluru, India
| | - Rajesh Kumar Dhanaraj
- Symbiosis Institute of Computer Studies and Research (SICSR), Symbiosis International University, Pune, India
| | - Sumegh Tharewal
- Symbiosis Institute of Computer Studies and Research (SICSR), Symbiosis International University, Pune, India
| | | | - Danilo Pelusi
- Department of Communication Sciences, University of Teramo, Teramo, Italy
| |
Collapse
|
84
|
Ahmad I, Merla A, Ali F, Shah B, AlZubi AA, AlZubi MA. A deep transfer learning approach for COVID-19 detection and exploring a sense of belonging with Diabetes. Front Public Health 2023; 11:1308404. [PMID: 38026271 PMCID: PMC10657998 DOI: 10.3389/fpubh.2023.1308404] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/06/2023] [Accepted: 10/18/2023] [Indexed: 12/01/2023] Open
Abstract
COVID-19 is an epidemic disease that results in death and significantly affects the older adult and those afflicted with chronic medical conditions. Diabetes medication and high blood glucose levels are significant predictors of COVID-19-related death or disease severity. Diabetic individuals, particularly those with preexisting comorbidities or geriatric patients, are at a higher risk of COVID-19 infection, including hospitalization, ICU admission, and death, than those without Diabetes. Everyone's lives have been significantly changed due to the COVID-19 outbreak. Identifying patients infected with COVID-19 in a timely manner is critical to overcoming this challenge. The Real-Time Polymerase Chain Reaction (RT-PCR) diagnostic assay is currently the gold standard for COVID-19 detection. However, RT-PCR is a time-consuming and costly technique requiring a lab kit that is difficult to get in crises and epidemics. This work suggests the CIDICXR-Net50 model, a ResNet-50-based Transfer Learning (TL) method for COVID-19 detection via Chest X-ray (CXR) image classification. The presented model is developed by substituting the final ResNet-50 classifier layer with a new classification head. The model is trained on 3,923 chest X-ray images comprising a substantial dataset of 1,360 viral pneumonia, 1,363 normal, and 1,200 COVID-19 CXR images. The proposed model's performance is evaluated in contrast to the results of six other innovative pre-trained models. The proposed CIDICXR-Net50 model attained 99.11% accuracy on the provided dataset while maintaining 99.15% precision and recall. This study also explores potential relationships between COVID-19 and Diabetes.
Collapse
Affiliation(s)
- Ijaz Ahmad
- Digital Transition, Innovation and Health Service, Leonardo da Vinci Telematic University, Chieti, Italy
| | - Arcangelo Merla
- Department of Engineering and Geology (INGEO) University "G. d’Annunzio" Chieti-Pescara, Pescara, Italy
| | - Farman Ali
- Department of Computer Science and Engineering, School of Convergence, College of Computing and Informatics, Sungkyunkwan University, Seoul, Republic of Korea
| | - Babar Shah
- College of Technological Innovation, Zayed University, Dubai, United Arab Emirates
| | - Ahmad Ali AlZubi
- Department of Computer Science, Community College, King Saud University, Riyadh, Saudi Arabia
| | - Mallak Ahmad AlZubi
- Faculty of Medicine, Jordan University of Science and Technology, Irbid, Jordan
| |
Collapse
|
85
|
Nahiduzzaman M, Goni MOF, Hassan R, Islam MR, Syfullah MK, Shahriar SM, Anower MS, Ahsan M, Haider J, Kowalski M. Parallel CNN-ELM: A multiclass classification of chest X-ray images to identify seventeen lung diseases including COVID-19. EXPERT SYSTEMS WITH APPLICATIONS 2023; 229:120528. [PMID: 37274610 PMCID: PMC10223636 DOI: 10.1016/j.eswa.2023.120528] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/01/2022] [Revised: 05/19/2023] [Accepted: 05/19/2023] [Indexed: 06/06/2023]
Abstract
Numerous epidemic lung diseases such as COVID-19, tuberculosis (TB), and pneumonia have spread over the world, killing millions of people. Medical specialists have experienced challenges in correctly identifying these diseases due to their subtle differences in Chest X-ray images (CXR). To assist the medical experts, this study proposed a computer-aided lung illness identification method based on the CXR images. For the first time, 17 different forms of lung disorders were considered and the study was divided into six trials with each containing two, two, three, four, fourteen, and seventeen different forms of lung disorders. The proposed framework combined robust feature extraction capabilities of a lightweight parallel convolutional neural network (CNN) with the classification abilities of the extreme learning machine algorithm named CNN-ELM. An optimistic accuracy of 90.92% and an area under the curve (AUC) of 96.93% was achieved when 17 classes were classified side by side. It also accurately identified COVID-19 and TB with 99.37% and 99.98% accuracy, respectively, in 0.996 microseconds for a single image. Additionally, the current results also demonstrated that the framework could outperform the existing state-of-the-art (SOTA) models. On top of that, a secondary conclusion drawn from this study was that the prospective framework retained its effectiveness over a range of real-world environments, including balanced-unbalanced or large-small datasets, large multiclass or simple binary class, and high- or low-resolution images. A prototype Android App was also developed to establish the potential of the framework in real-life implementation.
Collapse
Affiliation(s)
- Md Nahiduzzaman
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Md Omaer Faruq Goni
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Rakibul Hassan
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Md Robiul Islam
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Md Khalid Syfullah
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Saleh Mohammed Shahriar
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Md Shamim Anower
- Department of Electrical & Electronic Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Mominul Ahsan
- Department of Computer Science, University of York, Deramore Lane, Heslington, York YO10 5GH, UK
| | - Julfikar Haider
- Department of Engineering, Manchester Metropolitan University, Chester St, Manchester M1 5GD, UK
| | - Marcin Kowalski
- Institute of Optoelectronics, Military University of Technology, Gen. S. Kaliskiego 2, 00-908 Warsaw, Poland
| |
Collapse
|
86
|
Schaudt D, von Schwerin R, Hafner A, Riedel P, Reichert M, von Schwerin M, Beer M, Kloth C. Augmentation strategies for an imbalanced learning problem on a novel COVID-19 severity dataset. Sci Rep 2023; 13:18299. [PMID: 37880333 PMCID: PMC10600145 DOI: 10.1038/s41598-023-45532-2] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/23/2023] [Accepted: 10/20/2023] [Indexed: 10/27/2023] Open
Abstract
Since the beginning of the COVID-19 pandemic, many different machine learning models have been developed to detect and verify COVID-19 pneumonia based on chest X-ray images. Although promising, binary models have only limited implications for medical treatment, whereas the prediction of disease severity suggests more suitable and specific treatment options. In this study, we publish severity scores for the 2358 COVID-19 positive images in the COVIDx8B dataset, creating one of the largest collections of publicly available COVID-19 severity data. Furthermore, we train and evaluate deep learning models on the newly created dataset to provide a first benchmark for the severity classification task. One of the main challenges of this dataset is the skewed class distribution, resulting in undesirable model performance for the most severe cases. We therefore propose and examine different augmentation strategies, specifically targeting majority and minority classes. Our augmentation strategies show significant improvements in precision and recall values for the rare and most severe cases. While the models might not yet fulfill medical requirements, they serve as an appropriate starting point for further research with the proposed dataset to optimize clinical resource allocation and treatment.
Collapse
Affiliation(s)
- Daniel Schaudt
- Department of Computer Science, Ulm University of Applied Science, Albert-Einstein-Allee 55, 89081, Ulm, Baden-Wurttemberg, Germany.
| | - Reinhold von Schwerin
- Department of Computer Science, Ulm University of Applied Science, Albert-Einstein-Allee 55, 89081, Ulm, Baden-Wurttemberg, Germany
| | - Alexander Hafner
- Department of Computer Science, Ulm University of Applied Science, Albert-Einstein-Allee 55, 89081, Ulm, Baden-Wurttemberg, Germany
| | - Pascal Riedel
- Department of Computer Science, Ulm University of Applied Science, Albert-Einstein-Allee 55, 89081, Ulm, Baden-Wurttemberg, Germany
| | - Manfred Reichert
- Institute of Databases and Information Systems, Ulm University, James-Franck-Ring, 89081, Ulm, Baden-Wurttemberg, Germany
| | - Marianne von Schwerin
- Department of Computer Science, Ulm University of Applied Science, Albert-Einstein-Allee 55, 89081, Ulm, Baden-Wurttemberg, Germany
| | - Meinrad Beer
- Department of Radiology, University Hospital of Ulm, Albert-Einstein-Allee 23, 89081, Ulm, Baden-Wurttemberg, Germany
| | - Christopher Kloth
- Department of Radiology, University Hospital of Ulm, Albert-Einstein-Allee 23, 89081, Ulm, Baden-Wurttemberg, Germany
| |
Collapse
|
87
|
Antony M, Kakileti ST, Shah R, Sahoo S, Bhattacharyya C, Manjunath G. Challenges of AI driven diagnosis of chest X-rays transmitted through smart phones: a case study in COVID-19. Sci Rep 2023; 13:18102. [PMID: 37872204 PMCID: PMC10593822 DOI: 10.1038/s41598-023-44653-y] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/31/2022] [Accepted: 10/11/2023] [Indexed: 10/25/2023] Open
Abstract
Healthcare delivery during the initial days of outbreak of COVID-19 pandemic was badly impacted due to large number of severely infected patients posing an unprecedented global challenge. Although the importance of Chest X-rays (CXRs) in meeting this challenge has now been widely recognized, speedy diagnosis of CXRs remains an outstanding challenge because of fewer Radiologists. The exponential increase in Smart Phone ownership globally, including LMICs, provides an opportunity for exploring AI-driven diagnostic tools when provided with large volumes of CXRs transmitted through Smart Phones. However, the challenges associated with such systems have not been studied to the best of our knowledge. In this paper, we show that the predictions of AI-driven models on CXR images transmitted through Smart Phones via applications, such as WhatsApp, suffer both in terms of Predictability and Explainability, two key aspects of any automated Medical Diagnosis system. We find that several existing Deep learning based models exhibit prediction instability-disagreement between the prediction outcome of the original image and the transmitted image. Concomitantly we find that the explainability of the models deteriorate substantially, prediction on the transmitted CXR is often driven by features present outside the lung region, clearly a manifestation of Spurious Correlations. Our study reveals that there is significant compression of high-resolution CXR images, sometimes as high as 95%, and this could be the reason behind these two problems. Apart from demonstrating these problems, our main contribution is to show that Multi-Task learning (MTL) can serve as an effective bulwark against the aforementioned problems. We show that MTL models exhibit substantially more robustness, 40% over existing baselines. Explainability of such models, when measured by a saliency score dependent on out-of-lung features, also show a 35% improvement. The study is conducted on WaCXR dataset, a curated dataset of 6562 image pairs corresponding to original uncompressed and WhatsApp compressed CXR images. Keeping in mind that there are no previous datasets to study such problems, we open-source this data along with all implementations.
Collapse
Affiliation(s)
| | | | - Rachit Shah
- Indian Institute of Science, Bangalore, India
| | | | | | | |
Collapse
|
88
|
Socha M, Prażuch W, Suwalska A, Foszner P, Tobiasz J, Jaroszewicz J, Gruszczynska K, Sliwinska M, Nowak M, Gizycka B, Zapolska G, Popiela T, Przybylski G, Fiedor P, Pawlowska M, Flisiak R, Simon K, Walecki J, Cieszanowski A, Szurowska E, Marczyk M, Polanska J. Pathological changes or technical artefacts? The problem of the heterogenous databases in COVID-19 CXR image analysis. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2023; 240:107684. [PMID: 37356354 PMCID: PMC10278898 DOI: 10.1016/j.cmpb.2023.107684] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/20/2023] [Revised: 06/11/2023] [Accepted: 06/18/2023] [Indexed: 06/27/2023]
Abstract
BACKGROUND When the COVID-19 pandemic commenced in 2020, scientists assisted medical specialists with diagnostic algorithm development. One scientific research area related to COVID-19 diagnosis was medical imaging and its potential to support molecular tests. Unfortunately, several systems reported high accuracy in development but did not fare well in clinical application. The reason was poor generalization, a long-standing issue in AI development. Researchers found many causes of this issue and decided to refer to them as confounders, meaning a set of artefacts and methodological errors associated with the method. We aim to contribute to this steed by highlighting an undiscussed confounder related to image resolution. METHODS 20 216 chest X-ray images (CXR) from worldwide centres were analyzed. The CXRs were bijectively projected into the 2D domain by performing Uniform Manifold Approximation and Projection (UMAP) embedding on the radiomic features (rUMAP) or CNN-based neural features (nUMAP) from the pre-last layer of the pre-trained classification neural network. Additional 44 339 thorax CXRs were used for validation. The comprehensive analysis of the multimodality of the density distribution in rUMAP/nUMAP domains and its relation to the original image properties was used to identify the main confounders. RESULTS nUMAP revealed a hidden bias of neural networks towards the image resolution, which the regular up-sampling procedure cannot compensate for. The issue appears regardless of the network architecture and is not observed in a high-resolution dataset. The impact of the resolution heterogeneity can be partially diminished by applying advanced deep-learning-based super-resolution networks. CONCLUSIONS rUMAP and nUMAP are great tools for image homogeneity analysis and bias discovery, as demonstrated by applying them to COVID-19 image data. Nonetheless, nUMAP could be applied to any type of data for which a deep neural network could be constructed. Advanced image super-resolution solutions are needed to reduce the impact of the resolution diversity on the classification network decision.
Collapse
Affiliation(s)
- Marek Socha
- Department of Data Science and Engineering, Silesian University of Technology, Gliwice, Poland
| | - Wojciech Prażuch
- Department of Data Science and Engineering, Silesian University of Technology, Gliwice, Poland
| | - Aleksandra Suwalska
- Department of Data Science and Engineering, Silesian University of Technology, Gliwice, Poland
| | - Paweł Foszner
- Department of Data Science and Engineering, Silesian University of Technology, Gliwice, Poland; Department of Computer Graphics, Vision and Digital Systems, Silesian University of Technology, Gliwice, Poland
| | - Joanna Tobiasz
- Department of Data Science and Engineering, Silesian University of Technology, Gliwice, Poland; Department of Computer Graphics, Vision and Digital Systems, Silesian University of Technology, Gliwice, Poland
| | - Jerzy Jaroszewicz
- Department of Infectious Diseases and Hepatology, Medical University of Silesia, Katowice, Poland
| | - Katarzyna Gruszczynska
- Department of Radiology and Nuclear Medicine, Medical University of Silesia, Katowice, Poland
| | - Magdalena Sliwinska
- Department of Diagnostic Imaging, Voivodship Specialist Hospital, Wroclaw, Poland
| | - Mateusz Nowak
- Department of Radiology, Silesian Hospital, Cieszyn, Poland
| | - Barbara Gizycka
- Department of Imaging Diagnostics, MEGREZ Hospital, Tychy, Poland
| | | | - Tadeusz Popiela
- Department of Radiology, Jagiellonian University Medical College, Krakow, Poland
| | - Grzegorz Przybylski
- Department of Lung Diseases, Cancer and Tuberculosis, Kujawsko-Pomorskie Pulmonology Center, Bydgoszcz, Poland
| | - Piotr Fiedor
- Department of General and Transplantation Surgery, Medical University of Warsaw, Warsaw, Poland
| | - Malgorzata Pawlowska
- Department of Infectious Diseases and Hepatology, Collegium Medicum in Bydgoszcz, Nicolaus Copernicus University, Torun, Poland
| | - Robert Flisiak
- Department of Infectious Diseases and Hepatology, Medical University of Bialystok, Bialystok, Poland
| | - Krzysztof Simon
- Department of Infectious Diseases and Hepatology, Wroclaw Medical University, Wroclaw, Poland
| | - Jerzy Walecki
- Department of Radiology, Centre of Postgraduate Medical Education, Central Clinical Hospital of the Ministry of Interior in Warsaw, Poland
| | - Andrzej Cieszanowski
- Department of Radiology I, The Maria Sklodowska-Curie National Research Institute of Oncology, Warsaw, Poland
| | - Edyta Szurowska
- 2nd Department of Radiology, Medical University of Gdansk, Poland
| | - Michal Marczyk
- Department of Data Science and Engineering, Silesian University of Technology, Gliwice, Poland; Yale Cancer Center, Yale School of Medicine, New Haven, CT, USA
| | - Joanna Polanska
- Department of Data Science and Engineering, Silesian University of Technology, Gliwice, Poland.
| |
Collapse
|
89
|
Chen K, Zhang X, Zhou X, Mi B, Xiao Y, Zhou L, Wu Z, Wu L, Wang X. Privacy preserving federated learning for full heterogeneity. ISA TRANSACTIONS 2023; 141:73-83. [PMID: 37105888 DOI: 10.1016/j.isatra.2023.04.020] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/17/2022] [Revised: 04/01/2023] [Accepted: 04/14/2023] [Indexed: 06/19/2023]
Abstract
Federated learning is a novel distribute machine learning paradigm to support cooperative model training among multiple participant clients, where each client keeps its private data locally to protect its data privacy. However, in practical application domains, Federated learning still meets several heterogeneous challenges such data heterogeneity, model heterogeneity, and computation heterogeneity, significantly decreasing its global model performance. To the best of our knowledge, existing solutions only focus on one or two challenges in their heterogeneous settings. In this paper, to address the above challenges simultaneously, we present a novel solution called Full Heterogeneous Federated Learning (FHFL). Firstly, we propose a synthetic data generation approach to mitigate the Non-IID data heterogeneity problem. Secondly, we use knowledge distillation to learn from heterogeneous models of participant clients for model aggregation in the central server. Finally, we produce an opportunistic computation schedule strategy to exploit the idle computation resources for fast-computing clients. Experiment results on different datasets show that our FHFL method can achieve an excellent model training performance. We believe it will serve as a pioneer work for distributed model training among heterogeneous clients in Federated learning.
Collapse
Affiliation(s)
- Kongyang Chen
- Institute of Artificial Intelligence and Blockchain, Guangzhou University, China; Pazhou Lab, Guangzhou, China; Jiangsu Key Laboratory of Media Design and Software Technology, Jiangnan University, Wuxi, China
| | - Xiaoxue Zhang
- School of Computer Science and Cyber Engineering, Guangzhou University, China
| | - Xiuhua Zhou
- School of Computer Science and Cyber Engineering, Guangzhou University, China
| | - Bing Mi
- School of Public Finance and Taxation, Guangdong University of Finance and Economics, China
| | - Yatie Xiao
- School of Computer Science and Cyber Engineering, Guangzhou University, China
| | - Lei Zhou
- Department of Otorhinolaryngology-Head and Neck Surgery, Zhongshan Hospital Affiliated to Fudan University, China
| | - Zhen Wu
- Third Affiliated Hospital, Sun Yat-sen University, China
| | - Lin Wu
- Third Affiliated Hospital, Sun Yat-sen University, China.
| | - Xiaoying Wang
- Third Affiliated Hospital, Sun Yat-sen University, China.
| |
Collapse
|
90
|
Nahiduzzaman M, Chowdhury MEH, Salam A, Nahid E, Ahmed F, Al-Emadi N, Ayari MA, Khandakar A, Haider J. Explainable deep learning model for automatic mulberry leaf disease classification. FRONTIERS IN PLANT SCIENCE 2023; 14:1175515. [PMID: 37794930 PMCID: PMC10546311 DOI: 10.3389/fpls.2023.1175515] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 02/27/2023] [Accepted: 08/28/2023] [Indexed: 10/06/2023]
Abstract
Mulberry leaves feed Bombyx mori silkworms to generate silk thread. Diseases that affect mulberry leaves have reduced crop and silk yields in sericulture, which produces 90% of the world's raw silk. Manual leaf disease identification is tedious and error-prone. Computer vision can categorize leaf diseases early and overcome the challenges of manual identification. No mulberry leaf deep learning (DL) models have been reported. Therefore, in this study, two types of leaf diseases: leaf rust and leaf spot, with disease-free leaves, were collected from two regions of Bangladesh. Sericulture experts annotated the leaf images. The images were pre-processed, and 6,000 synthetic images were generated using typical image augmentation methods from the original 764 training images. Additional 218 and 109 images were employed for testing and validation respectively. In addition, a unique lightweight parallel depth-wise separable CNN model, PDS-CNN was developed by applying depth-wise separable convolutional layers to reduce parameters, layers, and size while boosting classification performance. Finally, the explainable capability of PDS-CNN is obtained through the use of SHapley Additive exPlanations (SHAP) evaluated by a sericulture specialist. The proposed PDS-CNN outperforms well-known deep transfer learning models, achieving an optimistic accuracy of 95.05 ± 2.86% for three-class classifications and 96.06 ± 3.01% for binary classifications with only 0.53 million parameters, 8 layers, and a size of 6.3 megabytes. Furthermore, when compared with other well-known transfer models, the proposed model identified mulberry leaf diseases with higher accuracy, fewer factors, fewer layers, and lower overall size. The visually expressive SHAP explanation images validate the models' findings aligning with the predictions made the sericulture specialist. Based on these findings, it is possible to conclude that the explainable AI (XAI)-based PDS-CNN can provide sericulture specialists with an effective tool for accurately categorizing mulberry leaves.
Collapse
Affiliation(s)
- Md. Nahiduzzaman
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi, Bangladesh
- Department of Electrical Engineering, Qatar University, Doha, Qatar
| | | | - Abdus Salam
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi, Bangladesh
| | - Emama Nahid
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi, Bangladesh
| | - Faruque Ahmed
- Bangladesh Sericulture Research and Training Institute, Rajshahi, Bangladesh
| | - Nasser Al-Emadi
- Department of Electrical Engineering, Qatar University, Doha, Qatar
| | - Mohamed Arselene Ayari
- Department of Civil and Environmental Engineering, Qatar University, Doha, Qatar
- Technology Innovation and Engineering Education Unit, Qatar University, Doha, Qatar
| | - Amith Khandakar
- Department of Electrical Engineering, Qatar University, Doha, Qatar
| | - Julfikar Haider
- Department of Engineering, Manchester Metropolitan University, Manchester, United Kingdom
| |
Collapse
|
91
|
Ahmed MAO, Abbas IA, AbdelSatar Y. HDSNE a new unsupervised multiple image database fusion learning algorithm with flexible and crispy production of one database: a proof case study of lung infection diagnose In chest X-ray images. BMC Med Imaging 2023; 23:134. [PMID: 37718458 PMCID: PMC10506286 DOI: 10.1186/s12880-023-01078-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/23/2022] [Accepted: 08/16/2023] [Indexed: 09/19/2023] Open
Abstract
Continuous release of image databases with fully or partially identical inner categories dramatically deteriorates the production of autonomous Computer-Aided Diagnostics (CAD) systems for true comprehensive medical diagnostics. The first challenge is the frequent massive bulk release of medical image databases, which often suffer from two common drawbacks: image duplication and corruption. The many subsequent releases of the same data with the same classes or categories come with no clear evidence of success in the concatenation of those identical classes among image databases. This issue stands as a stumbling block in the path of hypothesis-based experiments for the production of a single learning model that can successfully classify all of them correctly. Removing redundant data, enhancing performance, and optimizing energy resources are among the most challenging aspects. In this article, we propose a global data aggregation scale model that incorporates six image databases selected from specific global resources. The proposed valid learner is based on training all the unique patterns within any given data release, thereby creating a unique dataset hypothetically. The Hash MD5 algorithm (MD5) generates a unique hash value for each image, making it suitable for duplication removal. The T-Distributed Stochastic Neighbor Embedding (t-SNE), with a tunable perplexity parameter, can represent data dimensions. Both the Hash MD5 and t-SNE algorithms are applied recursively, producing a balanced and uniform database containing equal samples per category: normal, pneumonia, and Coronavirus Disease of 2019 (COVID-19). We evaluated the performance of all proposed data and the new automated version using the Inception V3 pre-trained model with various evaluation metrics. The performance outcome of the proposed scale model showed more respectable results than traditional data aggregation, achieving a high accuracy of 98.48%, along with high precision, recall, and F1-score. The results have been proved through a statistical t-test, yielding t-values and p-values. It's important to emphasize that all t-values are undeniably significant, and the p-values provide irrefutable evidence against the null hypothesis. Furthermore, it's noteworthy that the Final dataset outperformed all other datasets across all metric values when diagnosing various lung infections with the same factors.
Collapse
Affiliation(s)
- Muhammad Atta Othman Ahmed
- Department of Computer Science, Faculty of Computers and Information, Luxor University, Luxor, 85951, Egypt.
| | - Ibrahim A Abbas
- Mathematics Department, Faculty of Science, Sohag University, Sohag, 82511, Egypt
| | - Yasser AbdelSatar
- Mathematics Department, Faculty of Science, Sohag University, Sohag, 82511, Egypt
| |
Collapse
|
92
|
Rangel G, Cuevas-Tello JC, Rivera M, Renteria O. A Deep Learning Model Based on Capsule Networks for COVID Diagnostics through X-ray Images. Diagnostics (Basel) 2023; 13:2858. [PMID: 37685396 PMCID: PMC10486517 DOI: 10.3390/diagnostics13172858] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2023] [Revised: 08/22/2023] [Accepted: 08/24/2023] [Indexed: 09/10/2023] Open
Abstract
X-ray diagnostics are widely used to detect various diseases, such as bone fracture, pneumonia, or intracranial hemorrhage. This method is simple and accessible in most hospitals, but requires an expert who is sometimes unavailable. Today, some diagnoses are made with the help of deep learning algorithms based on Convolutional Neural Networks (CNN), but these algorithms show limitations. Recently, Capsule Networks (CapsNet) have been proposed to overcome these problems. In our work, CapsNet is used to detect whether a chest X-ray image has disease (COVID or pneumonia) or is healthy. An improved model called DRCaps is proposed, which combines the advantage of CapsNet and the dilation rate (dr) parameter to manage images with 226 × 226 resolution. We performed experiments with 16,669 chest images, in which our model achieved an accuracy of 90%. Furthermore, the model size is 11M with a reconstruction stage, which helps to avoid overfitting. Experiments show how the reconstruction stage works and how we can avoid the max-pooling operation for networks with a stride and dilation rate to downsampling the convolution layers. In this paper, DRCaps is superior to other comparable models in terms of accuracy, parameters, and image size handling. The main idea is to keep the model as simple as possible without using data augmentation or a complex preprocessing stage.
Collapse
Affiliation(s)
- Gabriela Rangel
- Facultad de Ingeniería, Universidad Autonoma de San Luis Potosi, San Luis Potosi 78290, Mexico;
- Tecnologico Nacional de Mexico/ITSSLPC, San Luis Potosi 78421, Mexico
| | - Juan C. Cuevas-Tello
- Facultad de Ingeniería, Universidad Autonoma de San Luis Potosi, San Luis Potosi 78290, Mexico;
| | - Mariano Rivera
- Centro de Investigacion en Matematicas, Guanajuato 36000, Mexico; (M.R.); (O.R.)
| | - Octavio Renteria
- Centro de Investigacion en Matematicas, Guanajuato 36000, Mexico; (M.R.); (O.R.)
| |
Collapse
|
93
|
Liu W, Delalibera Rodrigues R, Yan J, Zhu YT, de Freitas Pereira EJ, Li G, Zheng Q, Zhao L. Complex network-based classification of radiographic images for COVID-19 diagnosis. PLoS One 2023; 18:e0290968. [PMID: 37656697 PMCID: PMC10473542 DOI: 10.1371/journal.pone.0290968] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/05/2022] [Accepted: 08/03/2023] [Indexed: 09/03/2023] Open
Abstract
In this work, we present a network-based technique for chest X-ray image classification to help the diagnosis and prognosis of patients with COVID-19. From visual inspection, we perceive that healthy and COVID-19 chest radiographic images present different levels of geometric complexity. Therefore, we apply fractal dimension and quadtree as feature extractors to characterize such differences. Moreover, real-world datasets often present complex patterns, which are hardly handled by only the physical features of the data (such as similarity, distance, or distribution). This issue is addressed by complex networks, which are suitable tools for characterizing data patterns and capturing spatial, topological, and functional relationships in data. Specifically, we propose a new approach combining complexity measures and complex networks to provide a modified high-level classification technique to be applied to COVID-19 chest radiographic image classification. The computational results on the Kaggle COVID-19 Radiography Database show that the proposed method can obtain high classification accuracy on X-ray images, being competitive with state-of-the-art classification techniques. Lastly, a set of network measures is evaluated according to their potential in distinguishing the network classes, which resulted in the choice of communicability measure. We expect that the present work will make significant contributions to machine learning at the semantic level and to combat COVID-19.
Collapse
Affiliation(s)
- Weiguang Liu
- School of Computer Science, Zhongyuan University of Technology, ZhengZhou, China
| | | | - Jianglong Yan
- Institute of Mathematics and Computer Science (ICMC), University of São Paulo (USP), São Carlos, Brazil
| | - Yu-tao Zhu
- China Branch of BRICS Institute of Future Networks, ShenZhen, China
| | | | - Gen Li
- Henan Key Laboratory on Public Opinion Intelligent Analysis, Zhongyuan University of Technology, ZhengZhou, China
| | - Qiusheng Zheng
- Henan Key Laboratory on Public Opinion Intelligent Analysis, Zhongyuan University of Technology, ZhengZhou, China
| | - Liang Zhao
- China Branch of BRICS Institute of Future Networks, ShenZhen, China
- Department of Computing and Mathematics, FFCLRP, University of São Paulo (USP), Ribeirão Preto, Brazil
| |
Collapse
|
94
|
Jani J, Doshi J, Kheria I, Mehta K, Bhadane C, Karani R. LayNet-A multi-layer architecture to handle imbalance in medical imaging data. Comput Biol Med 2023; 163:107179. [PMID: 37354820 DOI: 10.1016/j.compbiomed.2023.107179] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/10/2023] [Revised: 06/02/2023] [Accepted: 06/11/2023] [Indexed: 06/26/2023]
Abstract
In an imbalanced dataset, a machine learning classifier using traditional imbalance handling methods may achieve good accuracy, but in highly imbalanced datasets, it may over-predict the majority class and ignore the minority class. In the medical domain, failing to correctly estimate the minority class might lead to a false negative, which is concerning in cases of life-threatening illnesses and infectious diseases like Covid-19. Currently, classification in deep learning has a single layered architecture where a neural network is employed. This paper proposes a multilayer design entitled LayNet to address this issue. LayNet aims to lessen the class imbalance by dividing the classes among layers and achieving a balanced class distribution at each layer. To ensure that all the classes are being classified, minor classes are combined to form a single new 'hybrid' class at higher layers. The final layer has no hybrid class and only singleton(distinct) classes. Each layer of the architecture includes a separate model that determines if an input belongs to one class or a hybrid class. If it fits into the hybrid class, it advances to the following layer, which is further categorized within the hybrid class. The method to divide the classes into various architectural levels is also introduced in this paper. The Ocular Disease Intelligent Recognition Dataset, Covid-19 Radiography Dataset, and Retinal OCT Dataset are used to evaluate this methodology. The LayNet architecture performs better on these datasets when the results of the traditional single-layer architecture and the proposed multilayered architecture are compared.
Collapse
Affiliation(s)
- Jay Jani
- Computer Engineering Department, D.J. Sanghvi College of Engineering, Mumbai, India.
| | - Jay Doshi
- Computer Engineering Department, D.J. Sanghvi College of Engineering, Mumbai, India.
| | - Ishita Kheria
- Computer Engineering Department, D.J. Sanghvi College of Engineering, Mumbai, India.
| | - Karishni Mehta
- Computer Engineering Department, D.J. Sanghvi College of Engineering, Mumbai, India.
| | - Chetashri Bhadane
- Computer Engineering Department, D.J. Sanghvi College of Engineering, Mumbai, India.
| | - Ruhina Karani
- Computer Engineering Department, D.J. Sanghvi College of Engineering, Mumbai, India.
| |
Collapse
|
95
|
Hadi MU, Qureshi R, Ahmed A, Iftikhar N. A lightweight CORONA-NET for COVID-19 detection in X-ray images. EXPERT SYSTEMS WITH APPLICATIONS 2023; 225:120023. [PMID: 37063778 PMCID: PMC10088342 DOI: 10.1016/j.eswa.2023.120023] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/08/2022] [Revised: 03/28/2023] [Accepted: 03/31/2023] [Indexed: 06/19/2023]
Abstract
Since December 2019, COVID-19 has posed the most serious threat to living beings. With the advancement of vaccination programs around the globe, the need to quickly diagnose COVID-19 in general with little logistics is fore important. As a consequence, the fastest diagnostic option to stop COVID-19 from spreading, especially among senior patients, should be the development of an automated detection system. This study aims to provide a lightweight deep learning method that incorporates a convolutional neural network (CNN), discrete wavelet transform (DWT), and a long short-term memory (LSTM), called CORONA-NET for diagnosing COVID-19 from chest X-ray images. In this system, deep feature extraction is performed by CNN, the feature vector is reduced yet strengthened by DWT, and the extracted feature is detected by LSTM for prediction. The dataset included 3000 X-rays, 1000 of which were COVID-19 obtained locally. Within minutes of the test, the proposed test platform's prototype can accurately detect COVID-19 patients. The proposed method achieves state-of-the-art performance in comparison with the existing deep learning methods. We hope that the suggested method will hasten clinical diagnosis and may be used for patients in remote areas where clinical labs are not easily accessible due to a lack of resources, location, or other factors.
Collapse
Affiliation(s)
- Muhammad Usman Hadi
- Nanotechnology and Integrated Bio-Engineering Centre (NIBEC), School of Engineering, Ulster University, BT15 1AP Belfast, UK
| | - Rizwan Qureshi
- Department of Imaging Physics, MD Anderson Cancer Center, The University of Texas, Houston, TX 77030, USA
| | - Ayesha Ahmed
- Department of Radiology, Aalborg University Hospital, Aalborg 9000, Denmark
| | - Nadeem Iftikhar
- University College of Northern Denmark, Aalborg 9200, Denmark
| |
Collapse
|
96
|
Gu J, Qian X, Zhang Q, Zhang H, Wu F. Unsupervised domain adaptation for Covid-19 classification based on balanced slice Wasserstein distance. Comput Biol Med 2023; 164:107207. [PMID: 37480680 DOI: 10.1016/j.compbiomed.2023.107207] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/15/2023] [Revised: 06/06/2023] [Accepted: 06/25/2023] [Indexed: 07/24/2023]
Abstract
Covid-19 has swept the world since 2020, taking millions of lives. In order to seek a rapid diagnosis of Covid-19, deep learning-based Covid-19 classification methods have been extensively developed. However, deep learning relies on many samples with high-quality labels, which is expensive. To this end, we propose a novel unsupervised domain adaptation method to process many different but related Covid-19 X-ray images. Unlike existing unsupervised domain adaptation methods that cannot handle conditional class distributions, we adopt a balanced Slice Wasserstein distance as the metric for unsupervised domain adaptation to solve this problem. Multiple standard datasets for domain adaptation and X-ray datasets of different Covid-19 are adopted to verify the effectiveness of our proposed method. Experimented by cross-adopting multiple datasets as source and target domains, respectively, our proposed method can effectively capture discriminative and domain-invariant representations with better data distribution matching.
Collapse
Affiliation(s)
- Jiawei Gu
- Affiliated Hospital of Nantong University, Nantong, 226001, China.
| | - Xuan Qian
- Affiliated Hospital of Nantong University, Nantong, 226001, China.
| | - Qian Zhang
- School of Data Science and Artificial Intelligence, Wenzhou University of Technology, Wenzhou, 325035, China.
| | - Hongliang Zhang
- College of Computer Science and Artificial Intelligence, Wenzhou University, Wenzhou, 325035, China.
| | - Fang Wu
- Department of Gastroenterology, The First Affiliated Hospital of Wenzhou Medical University, Wenzhou, 325035, China.
| |
Collapse
|
97
|
Chen F, Wang J, Liu H, Kong W, Zhao Z, Ma L, Liao H, Zhang D. Frequency constraint-based adversarial attack on deep neural networks for medical image classification. Comput Biol Med 2023; 164:107248. [PMID: 37515875 DOI: 10.1016/j.compbiomed.2023.107248] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2023] [Revised: 06/21/2023] [Accepted: 07/07/2023] [Indexed: 07/31/2023]
Abstract
The security of AI systems has gained significant attention in recent years, particularly in the medical diagnosis field. To develop a secure medical image classification system based on deep neural networks, it is crucial to design effective adversarial attacks that can embed hidden, malicious behaviors into the system. However, designing a unified attack method that can generate imperceptible attack samples with high content similarity and be applied to diverse medical image classification systems is challenging due to the diversity of medical imaging modalities and dimensionalities. Most existing attack methods are designed to attack natural image classification models, which inevitably corrupt the semantics of pixels by applying spatial perturbations. To address this issue, we propose a novel frequency constraint-based adversarial attack method capable of delivering attacks in various medical image classification tasks. Specially, our method introduces a frequency constraint to inject perturbation into high-frequency information while preserving low-frequency information to ensure content similarity. Our experiments include four public medical image datasets, including a 3D CT dataset, a 2D chest X-Ray image dataset, a 2D breast ultrasound dataset, and a 2D thyroid ultrasound dataset, which contain different imaging modalities and dimensionalities. The results demonstrate the superior performance of our model over other state-of-the-art adversarial attack methods for attacking medical image classification tasks on different imaging modalities and dimensionalities.
Collapse
Affiliation(s)
- Fang Chen
- Key Laboratory of Brain-Machine Intelligence Technology, Ministry of Education, Nanjing University of Aeronautics and Astronautics, Nanjing China; College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, Nanjing China.
| | - Jian Wang
- Key Laboratory of Brain-Machine Intelligence Technology, Ministry of Education, Nanjing University of Aeronautics and Astronautics, Nanjing China; College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, Nanjing China
| | - Han Liu
- Department of Ultrasound, Affiliated Drum Tower Hospital, Nanjing University Medical School, Nanjing, 21008, China.
| | - Wentao Kong
- Department of Ultrasound, Affiliated Drum Tower Hospital, Nanjing University Medical School, Nanjing, 21008, China
| | - Zhe Zhao
- Department of Orthopaedics, Beijing Tsinghua Changgung Hospital, Tsinghua University, Beijing 102218, China
| | - Longfei Ma
- Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, China
| | - Hongen Liao
- Department of Biomedical Engineering, School of Medicine, Tsinghua University, Beijing, China
| | - Daoqiang Zhang
- Key Laboratory of Brain-Machine Intelligence Technology, Ministry of Education, Nanjing University of Aeronautics and Astronautics, Nanjing China; College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, Nanjing China
| |
Collapse
|
98
|
Liu Z, Lv Q, Yang Z, Li Y, Lee CH, Shen L. Recent progress in transformer-based medical image analysis. Comput Biol Med 2023; 164:107268. [PMID: 37494821 DOI: 10.1016/j.compbiomed.2023.107268] [Citation(s) in RCA: 23] [Impact Index Per Article: 11.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/21/2023] [Revised: 05/30/2023] [Accepted: 07/16/2023] [Indexed: 07/28/2023]
Abstract
The transformer is primarily used in the field of natural language processing. Recently, it has been adopted and shows promise in the computer vision (CV) field. Medical image analysis (MIA), as a critical branch of CV, also greatly benefits from this state-of-the-art technique. In this review, we first recap the core component of the transformer, the attention mechanism, and the detailed structures of the transformer. After that, we depict the recent progress of the transformer in the field of MIA. We organize the applications in a sequence of different tasks, including classification, segmentation, captioning, registration, detection, enhancement, localization, and synthesis. The mainstream classification and segmentation tasks are further divided into eleven medical image modalities. A large number of experiments studied in this review illustrate that the transformer-based method outperforms existing methods through comparisons with multiple evaluation metrics. Finally, we discuss the open challenges and future opportunities in this field. This task-modality review with the latest contents, detailed information, and comprehensive comparison may greatly benefit the broad MIA community.
Collapse
Affiliation(s)
- Zhaoshan Liu
- Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore, 117575, Singapore.
| | - Qiujie Lv
- Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore, 117575, Singapore; School of Intelligent Systems Engineering, Sun Yat-sen University, No. 66, Gongchang Road, Guangming District, 518107, China.
| | - Ziduo Yang
- Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore, 117575, Singapore; School of Intelligent Systems Engineering, Sun Yat-sen University, No. 66, Gongchang Road, Guangming District, 518107, China.
| | - Yifan Li
- Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore, 117575, Singapore.
| | - Chau Hung Lee
- Department of Radiology, Tan Tock Seng Hospital, 11 Jalan Tan Tock Seng, Singapore, 308433, Singapore.
| | - Lei Shen
- Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore, 117575, Singapore.
| |
Collapse
|
99
|
Zheng K, Wu J, Yuan Y, Liu L. From single to multiple: Generalized detection of Covid-19 under limited classes samples. Comput Biol Med 2023; 164:107298. [PMID: 37573722 DOI: 10.1016/j.compbiomed.2023.107298] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2023] [Revised: 07/13/2023] [Accepted: 07/28/2023] [Indexed: 08/15/2023]
Abstract
Amid the unfolding Covid-19 pandemic, there is a critical need for rapid and accurate diagnostic methods. In this context, the field of deep learning-based medical image diagnosis has witnessed a swift evolution. However, the prevailing methodologies often rely on large amounts of labeled data and require comprehensive medical knowledge. Both of these prerequisites pose significant challenges in real clinical settings, given the high cost of data labeling and the complexities of disease representations. Addressing this gap, we propose a novel problem setting, the Open-Set Single-Domain Generalization for Medical Image Diagnosis (OSSDG-MID). In OSSDG-MID, our aim is to train a model exclusively on a single source domain, so it can classify samples from the target domain accurately, designating them as 'unknown' if they don't belong to the source domain sample category space. Our innovative solution, the Multiple Cross-Matching method (MCM), enhances the identification of these 'unknown' categories by generating auxiliary samples that fall outside the category space of the source domain. Experimental evaluations on two diverse cross-domain image classification tasks demonstrate that our approach outperforms existing methodologies in both single-domain generalization and open-set image classification.
Collapse
Affiliation(s)
- Kaihui Zheng
- Department of Intensive Care Unit, The Second Affiliated Hospital of Shanghai University (Wenzhou Central Hospital), Wenzhou, Zhejiang, 325000, China.
| | - Jianhua Wu
- Department of Intensive Care Unit, The Second Affiliated Hospital of Shanghai University (Wenzhou Central Hospital), Wenzhou, Zhejiang, 325000, China.
| | - Youjun Yuan
- Department of Emergency, The Second Affiliated Hospital of Shanghai University (Wenzhou Central Hospital), Wenzhou, Zhejiang, 325000, China.
| | - Lei Liu
- College of Computer Science, Sichuan University, Chengdu, Sichuan, 610065, China.
| |
Collapse
|
100
|
Chen Y, Wang T, Tang H, Zhao L, Zhang X, Tan T, Gao Q, Du M, Tong T. CoTrFuse: a novel framework by fusing CNN and transformer for medical image segmentation. Phys Med Biol 2023; 68:175027. [PMID: 37605997 DOI: 10.1088/1361-6560/acede8] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/03/2023] [Accepted: 08/07/2023] [Indexed: 08/23/2023]
Abstract
Medical image segmentation is a crucial and intricate process in medical image processing and analysis. With the advancements in artificial intelligence, deep learning techniques have been widely used in recent years for medical image segmentation. One such technique is the U-Net framework based on the U-shaped convolutional neural networks (CNN) and its variants. However, these methods have limitations in simultaneously capturing both the global and the remote semantic information due to the restricted receptive domain caused by the convolution operation's intrinsic features. Transformers are attention-based models with excellent global modeling capabilities, but their ability to acquire local information is limited. To address this, we propose a network that combines the strengths of both CNN and Transformer, called CoTrFuse. The proposed CoTrFuse network uses EfficientNet and Swin Transformer as dual encoders. The Swin Transformer and CNN Fusion module are combined to fuse the features of both branches before the skip connection structure. We evaluated the proposed network on two datasets: the ISIC-2017 challenge dataset and the COVID-QU-Ex dataset. Our experimental results demonstrate that the proposed CoTrFuse outperforms several state-of-the-art segmentation methods, indicating its superiority in medical image segmentation. The codes are available athttps://github.com/BinYCn/CoTrFuse.
Collapse
Affiliation(s)
- Yuanbin Chen
- College of Physics and Information Engineering, Fuzhou University, Fuzhou 350116, People's Republic of China
- Fujian Key Lab of Medical Instrumentation & Pharmaceutical Technology, Fuzhou University, Fuzhou 350116, People's Republic of China
| | - Tao Wang
- College of Physics and Information Engineering, Fuzhou University, Fuzhou 350116, People's Republic of China
- Fujian Key Lab of Medical Instrumentation & Pharmaceutical Technology, Fuzhou University, Fuzhou 350116, People's Republic of China
| | - Hui Tang
- College of Physics and Information Engineering, Fuzhou University, Fuzhou 350116, People's Republic of China
- Fujian Key Lab of Medical Instrumentation & Pharmaceutical Technology, Fuzhou University, Fuzhou 350116, People's Republic of China
| | - Longxuan Zhao
- College of Physics and Information Engineering, Fuzhou University, Fuzhou 350116, People's Republic of China
- Fujian Key Lab of Medical Instrumentation & Pharmaceutical Technology, Fuzhou University, Fuzhou 350116, People's Republic of China
| | - Xinlin Zhang
- College of Physics and Information Engineering, Fuzhou University, Fuzhou 350116, People's Republic of China
- Fujian Key Lab of Medical Instrumentation & Pharmaceutical Technology, Fuzhou University, Fuzhou 350116, People's Republic of China
| | - Tao Tan
- Faculty of Applied Science, Macao Polytechnic University, Macao 999078, People's Republic of China
| | - Qinquan Gao
- College of Physics and Information Engineering, Fuzhou University, Fuzhou 350116, People's Republic of China
- Fujian Key Lab of Medical Instrumentation & Pharmaceutical Technology, Fuzhou University, Fuzhou 350116, People's Republic of China
| | - Min Du
- College of Physics and Information Engineering, Fuzhou University, Fuzhou 350116, People's Republic of China
- Fujian Key Lab of Medical Instrumentation & Pharmaceutical Technology, Fuzhou University, Fuzhou 350116, People's Republic of China
| | - Tong Tong
- College of Physics and Information Engineering, Fuzhou University, Fuzhou 350116, People's Republic of China
- Fujian Key Lab of Medical Instrumentation & Pharmaceutical Technology, Fuzhou University, Fuzhou 350116, People's Republic of China
| |
Collapse
|