1
|
Kang H, Kim N, Ryu J. Attentional decoder networks for chest X-ray image recognition on high-resolution features. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 251:108198. [PMID: 38718718 DOI: 10.1016/j.cmpb.2024.108198] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/17/2024] [Revised: 03/31/2024] [Accepted: 04/21/2024] [Indexed: 05/31/2024]
Abstract
BACKGROUND AND OBJECTIVE This paper introduces an encoder-decoder-based attentional decoder network to recognize small-size lesions in chest X-ray images. In the encoder-only network, small-size lesions disappear during the down-sampling steps or are indistinguishable in the low-resolution feature maps. To address these issues, the proposed network processes images in the encoder-decoder architecture similar to U-Net families and classifies lesions by globally pooling high-resolution feature maps. However, two challenging obstacles prohibit U-Net families from being extended to classification: (1) the up-sampling procedure consumes considerable resources, and (2) there needs to be an effective pooling approach for the high-resolution feature maps. METHODS Therefore, the proposed network employs a lightweight attentional decoder and harmonic magnitude transform. The attentional decoder up-samples the given features with the low-resolution features as the key and value while the high-resolution features as the query. Since multi-scaled features interact, up-sampled features embody global context at a high resolution, maintaining pathological locality. In addition, harmonic magnitude transform is devised for pooling high-resolution feature maps in the frequency domain. We borrow the shift theorem of the Fourier transform to preserve the translation invariant property and further reduce the parameters of the pooling layer by an efficient embedding strategy. RESULTS The proposed network achieves state-of-the-art classification performance on the three public chest X-ray datasets, such as NIH, CheXpert, and MIMIC-CXR. CONCLUSIONS In conclusion, the proposed efficient encoder-decoder network recognizes small-size lesions well in chest X-ray images by efficiently up-sampling feature maps through an attentional decoder and processing high-resolution feature maps with harmonic magnitude transform. We open-source our implementation at https://github.com/Lab-LVM/ADNet.
Collapse
Affiliation(s)
- Hankyul Kang
- Department of Artificial Intelligence, Ajou University, Suwon, Republic of Korea
| | - Namkug Kim
- Department of Convergence Medicine, Ulsan University, Seoul, Republic of Korea
| | - Jongbin Ryu
- Department of Artificial Intelligence, Ajou University, Suwon, Republic of Korea; Department of Software and Computer Engineering, Ajou University, Suwon, Republic of Korea.
| |
Collapse
|
2
|
Harkness R, Frangi AF, Zucker K, Ravikumar N. Multi-centre benchmarking of deep learning models for COVID-19 detection in chest x-rays. FRONTIERS IN RADIOLOGY 2024; 4:1386906. [PMID: 38836218 PMCID: PMC11148230 DOI: 10.3389/fradi.2024.1386906] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/16/2024] [Accepted: 05/07/2024] [Indexed: 06/06/2024]
Abstract
Introduction This study is a retrospective evaluation of the performance of deep learning models that were developed for the detection of COVID-19 from chest x-rays, undertaken with the goal of assessing the suitability of such systems as clinical decision support tools. Methods Models were trained on the National COVID-19 Chest Imaging Database (NCCID), a UK-wide multi-centre dataset from 26 different NHS hospitals and evaluated on independent multi-national clinical datasets. The evaluation considers clinical and technical contributors to model error and potential model bias. Model predictions are examined for spurious feature correlations using techniques for explainable prediction. Results Models performed adequately on NHS populations, with performance comparable to radiologists, but generalised poorly to international populations. Models performed better in males than females, and performance varied across age groups. Alarmingly, models routinely failed when applied to complex clinical cases with confounding pathologies and when applied to radiologist defined "mild" cases. Discussion This comprehensive benchmarking study examines the pitfalls in current practices that have led to impractical model development. Key findings highlight the need for clinician involvement at all stages of model development, from data curation and label definition, to model evaluation, to ensure that all clinical factors and disease features are appropriately considered during model design. This is imperative to ensure automated approaches developed for disease detection are fit-for-purpose in a clinical setting.
Collapse
Affiliation(s)
- Rachael Harkness
- School of Computing, University of Leeds, Leeds, United Kingdom
- Centre for Computational Imaging and Simulation Technologies in Biomedicine, Leeds, United Kingdom
| | - Alejandro F Frangi
- Division of Informatics, Imaging, and Data Sciences, School of Health Sciences, University of Manchester, Manchester, United Kingdom
- Department of Computer Science, School of Engineering, University of Manchester, Manchester, United Kingdom
| | - Kieran Zucker
- Leeds Institute of Medical Research, School of Medicine, University of Leeds, Leeds, United Kingdom
| | - Nishant Ravikumar
- School of Computing, University of Leeds, Leeds, United Kingdom
- Centre for Computational Imaging and Simulation Technologies in Biomedicine, Leeds, United Kingdom
| |
Collapse
|
3
|
Farghaly O, Deshpande P. Texture-Based Classification to Overcome Uncertainty between COVID-19 and Viral Pneumonia Using Machine Learning and Deep Learning Techniques. Diagnostics (Basel) 2024; 14:1017. [PMID: 38786315 PMCID: PMC11119936 DOI: 10.3390/diagnostics14101017] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2024] [Revised: 05/11/2024] [Accepted: 05/13/2024] [Indexed: 05/25/2024] Open
Abstract
The SARS-CoV-2 virus, responsible for COVID-19, often manifests symptoms akin to viral pneumonia, complicating early detection and potentially leading to severe COVID pneumonia and long-term effects. Particularly affecting young individuals, the elderly, and those with weakened immune systems, the accurate classification of COVID-19 poses challenges, especially with highly dimensional image data. Past studies have faced limitations due to simplistic algorithms and small, biased datasets, yielding inaccurate results. In response, our study introduces a novel classification model that integrates advanced texture feature extraction methods, including GLCM, GLDM, and wavelet transform, within a deep learning framework. This innovative approach enables the effective classification of chest X-ray images into normal, COVID-19, and viral pneumonia categories, overcoming the limitations encountered in previous studies. Leveraging the unique textures inherent to each dataset class, our model achieves superior classification performance, even amidst the complexity and diversity of the data. Moreover, we present comprehensive numerical findings demonstrating the superiority of our approach over traditional methods. The numerical results highlight the accuracy (random forest (RF): 0.85; SVM (support vector machine): 0.70; deep learning neural network (DLNN): 0.92), recall (RF: 0.85, SVM: 0.74, DLNN: 0.93), precision (RF: 0.86, SVM: 0.71, DLNN: 0.87), and F1-Score (RF: 0.86, SVM: 0.72, DLNN: 0.89) of our proposed model. Our study represents a significant advancement in AI-based diagnostic systems for COVID-19 and pneumonia, promising improved patient outcomes and healthcare management strategies.
Collapse
Affiliation(s)
- Omar Farghaly
- Data-Intensive Computing Distributed Systems Laboratory, Department of Electrical and Computer Engineering, Marquette University, Milwaukee, WI 53233, USA
| | - Priya Deshpande
- Data-Intensive Computing Distributed Systems Laboratory, Department of Electrical and Computer Engineering, Marquette University, Milwaukee, WI 53233, USA
| |
Collapse
|
4
|
Bennour A, Ben Aoun N, Khalaf OI, Ghabban F, Wong WK, Algburi S. Contribution to pulmonary diseases diagnostic from X-ray images using innovative deep learning models. Heliyon 2024; 10:e30308. [PMID: 38707425 PMCID: PMC11068804 DOI: 10.1016/j.heliyon.2024.e30308] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2024] [Revised: 04/09/2024] [Accepted: 04/23/2024] [Indexed: 05/07/2024] Open
Abstract
Pulmonary disease identification and characterization are among the most intriguing research topics of recent years since they require an accurate and prompt diagnosis. Although pulmonary radiography has helped in lung disease diagnosis, the interpretation of the radiographic image has always been a major concern for doctors and radiologists to reduce diagnosis errors. Due to their success in image classification and segmentation tasks, cutting-edge artificial intelligence techniques like machine learning (ML) and deep learning (DL) are widely encouraged to be applied in the field of diagnosing lung disorders and identifying them using medical images, particularly radiographic ones. For this end, the researchers are concurring to build systems based on these techniques in particular deep learning ones. In this paper, we proposed three deep-learning models that were trained to identify the presence of certain lung diseases using thoracic radiography. The first model, named "CovCXR-Net", identifies the COVID-19 disease (two cases: COVID-19 or normal). The second model, named "MDCXR3-Net", identifies the COVID-19 and pneumonia diseases (three cases: COVID-19, pneumonia, or normal), and the last model, named "MDCXR4-Net", is destined to identify the COVID-19, pneumonia and the pulmonary opacity diseases (4 cases: COVID-19, pneumonia, pulmonary opacity or normal). These models have proven their superiority in comparison with the state-of-the-art models and reached an accuracy of 99,09 %, 97.74 %, and 90,37 % respectively with three benchmarks.
Collapse
Affiliation(s)
- Akram Bennour
- LAMIS Laboratiry, Echahid Cheikh Larbi Tebessi University, Tebessa, Algeria
| | - Najib Ben Aoun
- College of Computer Science and Information Technology, Al-Baha University, Al Baha, Saudi Arabia
- REGIM-Lab: Research Groups in Intelligent Machines, National School of Engineers of Sfax (ENIS), University of Sfax, Tunisia
| | - Osamah Ibrahim Khalaf
- Department of Solar, Al-Nahrain Research Center for Renewable Energy, Al-Nahrain University, Jadriya, Baghdad, Iraq
| | - Fahad Ghabban
- College of Computer Science and Engineering, Taibah University, Medina, Saudi Arabia
| | | | - Sameer Algburi
- Al-Kitab University, College of Engineering Techniques, Kirkuk, Iraq
| |
Collapse
|
5
|
Jia H, Zhang J, Ma K, Qiao X, Ren L, Shi X. Application of convolutional neural networks in medical images: a bibliometric analysis. Quant Imaging Med Surg 2024; 14:3501-3518. [PMID: 38720828 PMCID: PMC11074758 DOI: 10.21037/qims-23-1600] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/10/2023] [Accepted: 03/06/2024] [Indexed: 05/12/2024]
Abstract
Background In the field of medical imaging, the rapid rise of convolutional neural networks (CNNs) has presented significant opportunities for conserving healthcare resources. However, with the wide spread application of CNNs, several challenges have emerged, such as enormous data annotation costs, difficulties in ensuring user privacy and security, weak model interpretability, and the consumption of substantial computational resources. The fundamental challenge lies in optimizing and seamlessly integrating CNN technology to enhance the precision and efficiency of medical diagnosis. Methods This study sought to provide a comprehensive bibliometric overview of current research on the application of CNNs in medical imaging. Initially, bibliometric methods were used to calculate the frequency statistics, and perform the cluster analysis and the co-citation analysis of countries, institutions, authors, keywords, and references. Subsequently, the latent Dirichlet allocation (LDA) method was employed for the topic modeling of the literature. Next, an in-depth analysis of the topics was conducted, and the topics in the medical field, technical aspects, and trends in topic evolution were summarized. Finally, by integrating the bibliometrics and LDA results, the developmental trajectory, milestones, and future directions in this field were outlined. Results A data set containing 6,310 articles in this field published from January 2013 to December 2023 was complied. With a total of 55,538 articles, the United States led in terms of the citation count, while in terms of the publication volume, China led with 2,385 articles. Harvard University emerged as the most influential institution, boasting an average of 69.92 citations per article. Within the realm of CNNs, residual neural network (ResNet) and U-Net stood out, receiving 1,602 and 1,419 citations, respectively, which highlights the significant attention these models have received. The impact of coronavirus disease 2019 (COVID-19) was unmistakable, as reflected by the publication of 597 articles, making it a focal point of research. Additionally, among various disease topics, with 290 articles, brain-related research was the most prevalent. Computed tomography (CT) imaging dominated the research landscape, representing 73% of the 30 different topics. Conclusions Over the past 11 years, CNN-related research in medical imaging has grown exponentially. The findings of the present study provide insights into the field's status and research hotspots. In addition, this article meticulously chronicled the development of CNNs and highlighted key milestones, starting with LeNet in 1989, followed by a challenging 20-year exploration period, and culminating in the breakthrough moment with AlexNet in 2012. Finally, this article explored recent advancements in CNN technology, including semi-supervised learning, efficient learning, trustworthy artificial intelligence (AI), and federated learning methods, and also addressed challenges related to data annotation costs, diagnostic efficiency, model performance, and data privacy.
Collapse
Affiliation(s)
- Huixin Jia
- School of Computer Science and Technology, Shandong Technology and Business University, Yantai, China
| | - Jiali Zhang
- School of Computer Science and Technology, Shandong Technology and Business University, Yantai, China
| | - Kejun Ma
- School of Statistics, Shandong Technology and Business University, Yantai, China
| | - Xiaoyan Qiao
- School of Mathematics and Information Science, Shandong Technology and Business University, Yantai, China
| | - Lijie Ren
- Department of Neurology, Shenzhen Institute of Translational Medicine, The First Affiliated Hospital of Shenzhen University, Shenzhen Second People’s Hospital, Shenzhen, China
| | - Xin Shi
- School of Health Management/Institute of Health Sciences, China Medical University, Shenyang, China
- Immersion Technology and Evaluation Shandong Engineering Research Center, Shandong Technology and Business University, Yantai, China
| |
Collapse
|
6
|
Kabir MM, Mridha M, Rahman A, Hamid MA, Monowar MM. Detection of COVID-19, pneumonia, and tuberculosis from radiographs using AI-driven knowledge distillation. Heliyon 2024; 10:e26801. [PMID: 38444490 PMCID: PMC10912466 DOI: 10.1016/j.heliyon.2024.e26801] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/10/2023] [Revised: 01/30/2024] [Accepted: 02/20/2024] [Indexed: 03/07/2024] Open
Abstract
Chest radiography is an essential diagnostic tool for respiratory diseases such as COVID-19, pneumonia, and tuberculosis because it accurately depicts the structures of the chest. However, accurate detection of these diseases from radiographs is a complex task that requires the availability of medical imaging equipment and trained personnel. Conventional deep learning models offer a viable automated solution for this task. However, the high complexity of these models often poses a significant obstacle to their practical deployment within automated medical applications, including mobile apps, web apps, and cloud-based platforms. This study addresses and resolves this dilemma by reducing the complexity of neural networks using knowledge distillation techniques (KDT). The proposed technique trains a neural network on an extensive collection of chest X-ray images and propagates the knowledge to a smaller network capable of real-time detection. To create a comprehensive dataset, we have integrated three popular chest radiograph datasets with chest radiographs for COVID-19, pneumonia, and tuberculosis. Our experiments show that this knowledge distillation approach outperforms conventional deep learning methods in terms of computational complexity and performance for real-time respiratory disease detection. Specifically, our system achieves an impressive average accuracy of 0.97, precision of 0.94, and recall of 0.97.
Collapse
Affiliation(s)
- Md Mohsin Kabir
- Department of Computer Science & Engineering, Bangladesh University of Business & Technology, Dhaka-1216, Bangladesh
| | - M.F. Mridha
- Department of Computer Science, American International University-Bangladesh, Dhaka-1229, Bangladesh
| | - Ashifur Rahman
- Department of Computer Science & Engineering, Bangladesh University of Business & Technology, Dhaka-1216, Bangladesh
| | - Md. Abdul Hamid
- Department of Information Technology, Faculty of Computing & Information Technology, King Abdulaziz University, Jeddah-21589, Kingdom of Saudi Arabia
| | - Muhammad Mostafa Monowar
- Department of Information Technology, Faculty of Computing & Information Technology, King Abdulaziz University, Jeddah-21589, Kingdom of Saudi Arabia
| |
Collapse
|
7
|
Du F, Zhou H, Niu Y, Han Z, Sui X. Transformaer-based model for lung adenocarcinoma subtypes. Med Phys 2024. [PMID: 38427790 DOI: 10.1002/mp.17006] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/26/2023] [Revised: 01/27/2024] [Accepted: 01/27/2024] [Indexed: 03/03/2024] Open
Abstract
BACKGROUND Lung cancer has the highest morbidity and mortality rate among all types of cancer. Histological subtypes serve as crucial markers for the development of lung cancer and possess significant clinical values for cancer diagnosis, prognosis, and prediction of treatment responses. However, existing studies only dichotomize normal and cancerous tissues, failing to capture the unique characteristics of tissue sections and cancer types. PURPOSE Therefore, we have pioneered the classification of lung adenocarcinoma (LAD) cancer tissues into five subtypes (acinar, lepidic, micropapillary, papillary, and solid) based on section data in whole-slide image sections. In addition, a novel model called HybridNet was designed to improve the classification performance. METHODS HybridNet primarily consists of two interactive streams: a Transformer and a convolutional neural network (CNN). The Transformer stream captures rich global representations using a self-attention mechanism, while the CNN stream extracts local semantic features to optimize image details. Specifically, during the dual-stream parallelism, the feature maps of the Transformer stream as weights are weighted and summed with those of the CNN stream backbone; at the end of the parallelism, the respective final features are concatenated to obtain more discriminative semantic information. RESULTS Experimental results on a private dataset of LAD showed that HybridNet achieved 95.12% classification accuracy, and the accuracy of five histological subtypes (acinar, lepidic, micropapillary, papillary, and solid) reached 94.5%, 97.1%, 94%, 91%, and 99% respectively; the experimental results on the public BreakHis dataset show that HybridNet achieves the best results in three evaluation metrics: accuracy, recall and F1-score, with 92.40%, 90.63%, and 91.43%, respectively. CONCLUSIONS The process of classifying LAD into five subtypes assists pathologists in selecting appropriate treatments and enables them to predict tumor mutation burden (TMB) and analyze the spatial distribution of immune checkpoint proteins based on this and other clinical data. In addition, the proposed HybridNet fuses CNN and Transformer information several times and is able to improve the accuracy of subtype classification, and also shows satisfactory performance on public datasets with some generalization ability.
Collapse
Affiliation(s)
- Fawen Du
- School of Information Science and Engineering, Shandong Normal University, Jinan, Shandong, China
| | - Huiyu Zhou
- School of Computing and Mathematic Sciences, University of Leicester, Leicester, UK
| | - Yi Niu
- School of Information Science and Engineering, Shandong Normal University, Jinan, Shandong, China
| | - Zeyu Han
- School of Mathematics and Statistics, Shandong University, Weihai, China
| | - Xiaodan Sui
- School of Information Science and Engineering, Shandong Normal University, Jinan, Shandong, China
| |
Collapse
|
8
|
Ju H, Cui Y, Su Q, Juan L, Manavalan B. CODENET: A deep learning model for COVID-19 detection. Comput Biol Med 2024; 171:108229. [PMID: 38447500 DOI: 10.1016/j.compbiomed.2024.108229] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2023] [Revised: 02/20/2024] [Accepted: 02/25/2024] [Indexed: 03/08/2024]
Abstract
Conventional COVID-19 testing methods have some flaws: they are expensive and time-consuming. Chest X-ray (CXR) diagnostic approaches can alleviate these flaws to some extent. However, there is no accurate and practical automatic diagnostic framework with good interpretability. The application of artificial intelligence (AI) technology to medical radiography can help to accurately detect the disease, reduce the burden on healthcare organizations, and provide good interpretability. Therefore, this study proposes a new deep neural network (CNN) based on CXR for COVID-19 diagnosis - CodeNet. This method uses contrastive learning to make full use of latent image data to enhance the model's ability to extract features and generalize across different data domains. On the evaluation dataset, the proposed method achieves an accuracy as high as 94.20%, outperforming several other existing methods used for comparison. Ablation studies validate the efficacy of the proposed method, while interpretability analysis shows that the method can effectively guide clinical professionals. This work demonstrates the superior detection performance of a CNN using contrastive learning techniques on CXR images, paving the way for computer vision and artificial intelligence technologies to leverage massive medical data for disease diagnosis.
Collapse
Affiliation(s)
- Hong Ju
- Heilongjiang Agricultural Engineering Vocational College, China
| | - Yanyan Cui
- Beidahuang Industry Group General Hospital, Harbin, China
| | - Qiaosen Su
- Department of Integrative Biotechnology, College of Biotechnology and Bioengineering, Sungkyunkwan University, Suwon, 16419, Gyeonggi-do, Republic of Korea
| | - Liran Juan
- School of Life Science and Technology, Harbin Institute of Technology, Harbin, 150001, China.
| | - Balachandran Manavalan
- Department of Integrative Biotechnology, College of Biotechnology and Bioengineering, Sungkyunkwan University, Suwon, 16419, Gyeonggi-do, Republic of Korea.
| |
Collapse
|
9
|
Xu Y, Zhang X, Huang C, Qiu X. Can using a pre-trained deep learning model as the feature extractor in the bag-of-deep-visual-words model always improve image classification accuracy? PLoS One 2024; 19:e0298228. [PMID: 38422007 PMCID: PMC10903886 DOI: 10.1371/journal.pone.0298228] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2023] [Accepted: 01/19/2024] [Indexed: 03/02/2024] Open
Abstract
This article investigates whether higher classification accuracy can always be achieved by utilizing a pre-trained deep learning model as the feature extractor in the Bag-of-Deep-Visual-Words (BoDVW) classification model, as opposed to directly using the new classification layer of the pre-trained model for classification. Considering the multiple factors related to the feature extractor -such as model architecture, fine-tuning strategy, number of training samples, feature extraction method, and feature encoding method-we investigate these factors through experiments and then provide detailed answers to the question. In our experiments, we use five feature encoding methods: hard-voting, soft-voting, locally constrained linear coding, super vector coding, and fisher vector (FV). We also employ two popular feature extraction methods: one (denoted as Ext-DFs(CP)) uses a convolutional or non-global pooling layer, and another (denoted as Ext-DFs(FC)) uses a fully-connected or global pooling layer. Three pre-trained models-VGGNet-16, ResNext-50(32×4d), and Swin-B-are utilized as feature extractors. Experimental results on six datasets (15-Scenes, TF-Flowers, MIT Indoor-67, COVID-19 CXR, NWPU-RESISC45, and Caltech-101) reveal that compared to using the pre-trained model with only the new classification layer re-trained for classification, employing it as the feature extractor in the BoDVW model improves the accuracy in 35 out of 36 experiments when using FV. With Ext-DFs(CP), the accuracy increases by 0.13% to 8.43% (averaged at 3.11%), and with Ext-DFs(FC), it increases by 1.06% to 14.63% (averaged at 5.66%). Furthermore, when all layers of the pre-trained model are fine-tuned and used as the feature extractor, the results vary depending on the methods used. If FV and Ext-DFs(FC) are used, the accuracy increases by 0.21% to 5.65% (averaged at 1.58%) in 14 out of 18 experiments. Our results suggest that while using a pre-trained deep learning model as the feature extractor does not always improve classification accuracy, it holds great potential as an accuracy improvement technique.
Collapse
Affiliation(s)
- Ye Xu
- School of IoT Technology, Wuxi Institute of Technology, Wuxi, Jiangsu, China
| | - Xin Zhang
- School of IoT Technology, Wuxi Institute of Technology, Wuxi, Jiangsu, China
| | - Chongpeng Huang
- School of IoT Technology, Wuxi Institute of Technology, Wuxi, Jiangsu, China
| | - Xiaorong Qiu
- School of IoT Technology, Wuxi Institute of Technology, Wuxi, Jiangsu, China
| |
Collapse
|
10
|
Abdulahi AT, Ogundokun RO, Adenike AR, Shah MA, Ahmed YK. PulmoNet: a novel deep learning based pulmonary diseases detection model. BMC Med Imaging 2024; 24:51. [PMID: 38418987 PMCID: PMC10903074 DOI: 10.1186/s12880-024-01227-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/02/2023] [Accepted: 02/11/2024] [Indexed: 03/02/2024] Open
Abstract
Pulmonary diseases are various pathological conditions that affect respiratory tissues and organs, making the exchange of gas challenging for animals inhaling and exhaling. It varies from gentle and self-limiting such as the common cold and catarrh, to life-threatening ones, such as viral pneumonia (VP), bacterial pneumonia (BP), and tuberculosis, as well as a severe acute respiratory syndrome, such as the coronavirus 2019 (COVID-19). The cost of diagnosis and treatment of pulmonary infections is on the high side, most especially in developing countries, and since radiography images (X-ray and computed tomography (CT) scan images) have proven beneficial in detecting various pulmonary infections, many machine learning (ML) models and image processing procedures have been utilized to identify these infections. The need for timely and accurate detection can be lifesaving, especially during a pandemic. This paper, therefore, suggested a deep convolutional neural network (DCNN) founded image detection model, optimized with image augmentation technique, to detect three (3) different pulmonary diseases (COVID-19, bacterial pneumonia, and viral pneumonia). The dataset containing four (4) different classes (healthy (10,325), COVID-19 (3,749), BP (883), and VP (1,478)) was utilized as training/testing data for the suggested model. The model's performance indicates high potential in detecting the three (3) classes of pulmonary diseases. The model recorded average detection accuracy of 94%, 95.4%, 99.4%, and 98.30%, and training/detection time of about 60/50 s. This result indicates the proficiency of the suggested approach when likened to the traditional texture descriptors technique of pulmonary disease recognition utilizing X-ray and CT scan images. This study introduces an innovative deep convolutional neural network model to enhance the detection of pulmonary diseases like COVID-19 and pneumonia using radiography. This model, notable for its accuracy and efficiency, promises significant advancements in medical diagnostics, particularly beneficial in developing countries due to its potential to surpass traditional diagnostic methods.
Collapse
Affiliation(s)
- AbdulRahman Tosho Abdulahi
- Department of Computer Science, Institute of Information and Communication Technology, Kwara State Polytechnic, Ilorin, Nigeria
| | - Roseline Oluwaseun Ogundokun
- Department of Multimedia Engineering, Kaunas University of Technology, Kaunas, Lithuania
- Department of Computer Science, Landmark University Omu Aran, Omu Aran, Nigeria
| | - Ajiboye Raimot Adenike
- Department of Statistics, Institute of Applied Sciences, Kwara State Polytechnic, Ilorin, Nigeria
| | - Mohd Asif Shah
- Department of Economics, Kebri Dehar University, Kebri Dehar, 250, Somali, Ethiopia.
- Centre of Research Impact and Outcome, Chitkara University Institute of Engineering and Technology, Chitkara University, Rajpura, Punjab, 140401, India.
- Chitkara Centre for Research and Development, Chitkara University, Baddi, Himachal Pradesh, 174103, India.
| | - Yusuf Kola Ahmed
- Department of Biomedical Engineering, University of Ilorin, Ilorin, Nigeria
- Department of Occupational Therapy, University of Alberta, Edmonton, Canada
| |
Collapse
|
11
|
Pan CT, Kumar R, Wen ZH, Wang CH, Chang CY, Shiue YL. Improving Respiratory Infection Diagnosis with Deep Learning and Combinatorial Fusion: A Two-Stage Approach Using Chest X-ray Imaging. Diagnostics (Basel) 2024; 14:500. [PMID: 38472972 DOI: 10.3390/diagnostics14050500] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/31/2023] [Revised: 02/16/2024] [Accepted: 02/18/2024] [Indexed: 03/14/2024] Open
Abstract
The challenges of respiratory infections persist as a global health crisis, placing substantial stress on healthcare infrastructures and necessitating ongoing investigation into efficacious treatment modalities. The persistent challenge of respiratory infections, including COVID-19, underscores the critical need for enhanced diagnostic methodologies to support early treatment interventions. This study introduces an innovative two-stage data analytics framework that leverages deep learning algorithms through a strategic combinatorial fusion technique, aimed at refining the accuracy of early-stage diagnosis of such infections. Utilizing a comprehensive dataset compiled from publicly available lung X-ray images, the research employs advanced pre-trained deep learning models to navigate the complexities of disease classification, addressing inherent data imbalances through methodical validation processes. The core contribution of this work lies in its novel application of combinatorial fusion, integrating select models to significantly elevate diagnostic precision. This approach not only showcases the adaptability and strength of deep learning in navigating the intricacies of medical imaging but also marks a significant step forward in the utilization of artificial intelligence to improve outcomes in healthcare diagnostics. The study's findings illuminate the path toward leveraging technological advancements in enhancing diagnostic accuracies, ultimately contributing to the timely and effective treatment of respiratory diseases.
Collapse
Affiliation(s)
- Cheng-Tang Pan
- Department of Mechanical and Electro-Mechanical Engineering, National Sun Yat-sen University, Kaohsiung 804, Taiwan
- Institute of Precision Medicine, National Sun Yat-sen University, Kaohsiung 804, Taiwan
- Taiwan Instrument Research Institute, National Applied Research Laboratories, Hsinchu 300, Taiwan
- Institute of Advanced Semiconductor Packaging and Testing, College of Semiconductor and Advanced Technology Research, National Sun Yat-sen University, Kaohsiung 804, Taiwan
| | - Rahul Kumar
- Department of Mechanical and Electro-Mechanical Engineering, National Sun Yat-sen University, Kaohsiung 804, Taiwan
| | - Zhi-Hong Wen
- Department of Marine Biotechnology and Research, National Sun Yat-sen University, Kaohsiung 804, Taiwan
| | - Chih-Hsuan Wang
- Division of Nephrology and Metabolism, Department of Internal Medicine, Kaohsiung Armed Forces General Hospital, Kaohsiung 804, Taiwan
- Institute of Medical Science and Technology, National Sun Yat-sen University, Kaohsiung 80424, Taiwan
| | - Chun-Yung Chang
- Division of Nephrology and Metabolism, Department of Internal Medicine, Kaohsiung Armed Forces General Hospital, Kaohsiung 804, Taiwan
- Institute of Medical Science and Technology, National Sun Yat-sen University, Kaohsiung 80424, Taiwan
| | - Yow-Ling Shiue
- Institute of Precision Medicine, National Sun Yat-sen University, Kaohsiung 804, Taiwan
- Institute of Biomedical Sciences, National Sun Yat-sen University, Kaohsiung 80424, Taiwan
| |
Collapse
|
12
|
Kumar S, Kumar H, Kumar G, Singh SP, Bijalwan A, Diwakar M. A methodical exploration of imaging modalities from dataset to detection through machine learning paradigms in prominent lung disease diagnosis: a review. BMC Med Imaging 2024; 24:30. [PMID: 38302883 PMCID: PMC10832080 DOI: 10.1186/s12880-024-01192-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2023] [Accepted: 01/03/2024] [Indexed: 02/03/2024] Open
Abstract
BACKGROUND Lung diseases, both infectious and non-infectious, are the most prevalent cause of mortality overall in the world. Medical research has identified pneumonia, lung cancer, and Corona Virus Disease 2019 (COVID-19) as prominent lung diseases prioritized over others. Imaging modalities, including X-rays, computer tomography (CT) scans, magnetic resonance imaging (MRIs), positron emission tomography (PET) scans, and others, are primarily employed in medical assessments because they provide computed data that can be utilized as input datasets for computer-assisted diagnostic systems. Imaging datasets are used to develop and evaluate machine learning (ML) methods to analyze and predict prominent lung diseases. OBJECTIVE This review analyzes ML paradigms, imaging modalities' utilization, and recent developments for prominent lung diseases. Furthermore, the research also explores various datasets available publically that are being used for prominent lung diseases. METHODS The well-known databases of academic studies that have been subjected to peer review, namely ScienceDirect, arXiv, IEEE Xplore, MDPI, and many more, were used for the search of relevant articles. Applied keywords and combinations used to search procedures with primary considerations for review, such as pneumonia, lung cancer, COVID-19, various imaging modalities, ML, convolutional neural networks (CNNs), transfer learning, and ensemble learning. RESULTS This research finding indicates that X-ray datasets are preferred for detecting pneumonia, while CT scan datasets are predominantly favored for detecting lung cancer. Furthermore, in COVID-19 detection, X-ray datasets are prioritized over CT scan datasets. The analysis reveals that X-rays and CT scans have surpassed all other imaging techniques. It has been observed that using CNNs yields a high degree of accuracy and practicability in identifying prominent lung diseases. Transfer learning and ensemble learning are complementary techniques to CNNs to facilitate analysis. Furthermore, accuracy is the most favored metric for assessment.
Collapse
Affiliation(s)
- Sunil Kumar
- Department of Computer Engineering, J. C. Bose University of Science and Technology, YMCA, Faridabad, India
- Department of Information Technology, School of Engineering and Technology (UIET), CSJM University, Kanpur, India
| | - Harish Kumar
- Department of Computer Engineering, J. C. Bose University of Science and Technology, YMCA, Faridabad, India
| | - Gyanendra Kumar
- Department of Computer and Communication Engineering, Manipal University Jaipur, Jaipur, India
| | | | - Anchit Bijalwan
- Faculty of Electrical and Computer Engineering, Arba Minch University, Arba Minch, Ethiopia.
| | - Manoj Diwakar
- Department of Computer Science and Engineering, Graphic Era Deemed to Be University, Dehradun, India
| |
Collapse
|
13
|
Chen S, Ren S, Wang G, Huang M, Xue C. Interpretable CNN-Multilevel Attention Transformer for Rapid Recognition of Pneumonia From Chest X-Ray Images. IEEE J Biomed Health Inform 2024; 28:753-764. [PMID: 37027681 DOI: 10.1109/jbhi.2023.3247949] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/24/2023]
Abstract
Chest imaging plays an essential role in diagnosing and predicting patients with COVID-19 with evidence of worsening respiratory status. Many deep learning-based approaches for pneumonia recognition have been developed to enable computer-aided diagnosis. However, the long training and inference time makes them inflexible, and the lack of interpretability reduces their credibility in clinical medical practice. This paper aims to develop a pneumonia recognition framework with interpretability, which can understand the complex relationship between lung features and related diseases in chest X-ray (CXR) images to provide high-speed analytics support for medical practice. To reduce the computational complexity to accelerate the recognition process, a novel multi-level self-attention mechanism within Transformer has been proposed to accelerate convergence and emphasize the task-related feature regions. Moreover, a practical CXR image data augmentation has been adopted to address the scarcity of medical image data problems to boost the model's performance. The effectiveness of the proposed method has been demonstrated on the classic COVID-19 recognition task using the widespread pneumonia CXR image dataset. In addition, abundant ablation experiments validate the effectiveness and necessity of all of the components of the proposed method.
Collapse
|
14
|
Haque SBU, Zafar A. Robust Medical Diagnosis: A Novel Two-Phase Deep Learning Framework for Adversarial Proof Disease Detection in Radiology Images. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024; 37:308-338. [PMID: 38343214 DOI: 10.1007/s10278-023-00916-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/31/2023] [Revised: 09/23/2023] [Accepted: 10/08/2023] [Indexed: 03/02/2024]
Abstract
In the realm of medical diagnostics, the utilization of deep learning techniques, notably in the context of radiology images, has emerged as a transformative force. The significance of artificial intelligence (AI), specifically machine learning (ML) and deep learning (DL), lies in their capacity to rapidly and accurately diagnose diseases from radiology images. This capability has been particularly vital during the COVID-19 pandemic, where rapid and precise diagnosis played a pivotal role in managing the spread of the virus. DL models, trained on vast datasets of radiology images, have showcased remarkable proficiency in distinguishing between normal and COVID-19-affected cases, offering a ray of hope amidst the crisis. However, as with any technological advancement, vulnerabilities emerge. Deep learning-based diagnostic models, although proficient, are not immune to adversarial attacks. These attacks, characterized by carefully crafted perturbations to input data, can potentially disrupt the models' decision-making processes. In the medical context, such vulnerabilities could have dire consequences, leading to misdiagnoses and compromised patient care. To address this, we propose a two-phase defense framework that combines advanced adversarial learning and adversarial image filtering techniques. We use a modified adversarial learning algorithm to enhance the model's resilience against adversarial examples during the training phase. During the inference phase, we apply JPEG compression to mitigate perturbations that cause misclassification. We evaluate our approach on three models based on ResNet-50, VGG-16, and Inception-V3. These models perform exceptionally in classifying radiology images (X-ray and CT) of lung regions into normal, pneumonia, and COVID-19 pneumonia categories. We then assess the vulnerability of these models to three targeted adversarial attacks: fast gradient sign method (FGSM), projected gradient descent (PGD), and basic iterative method (BIM). The results show a significant drop in model performance after the attacks. However, our defense framework greatly improves the models' resistance to adversarial attacks, maintaining high accuracy on adversarial examples. Importantly, our framework ensures the reliability of the models in diagnosing COVID-19 from clean images.
Collapse
Affiliation(s)
- Sheikh Burhan Ul Haque
- Department of Computer Science, Aligarh Muslim University, Uttar Pradesh, Aligarh, 202002, India.
| | - Aasim Zafar
- Department of Computer Science, Aligarh Muslim University, Uttar Pradesh, Aligarh, 202002, India
| |
Collapse
|
15
|
Chauhan S, Edla DR, Boddu V, Rao MJ, Cheruku R, Nayak SR, Martha S, Lavanya K, Nigat TD. Detection of COVID-19 using edge devices by a light-weight convolutional neural network from chest X-ray images. BMC Med Imaging 2024; 24:1. [PMID: 38166813 PMCID: PMC10759384 DOI: 10.1186/s12880-023-01155-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/21/2023] [Accepted: 11/14/2023] [Indexed: 01/05/2024] Open
Abstract
Deep learning is a highly significant technology in clinical treatment and diagnostics nowadays. Convolutional Neural Network (CNN) is a new idea in deep learning that is being used in the area of computer vision. The COVID-19 detection is the subject of our medical study. Researchers attempted to increase the detection accuracy but at the cost of high model complexity. In this paper, we desire to achieve better accuracy with little training space and time so that this model easily deployed in edge devices. In this paper, a new CNN design is proposed that has three stages: pre-processing, which removes the black padding on the side initially; convolution, which employs filter banks; and feature extraction, which makes use of deep convolutional layers with skip connections. In order to train the model, chest X-ray images are partitioned into three sets: learning(0.7), validation(0.1), and testing(0.2). The models are then evaluated using the test and training data. The LMNet, CoroNet, CVDNet, and Deep GRU-CNN models are the other four models used in the same experiment. The propose model achieved 99.47% & 98.91% accuracy on training and testing respectively. Additionally, it achieved 97.54%, 98.19%, 99.49%, and 97.86% scores for precision, recall, specificity, and f1-score respectively. The proposed model obtained nearly equivalent accuracy and other similar metrics when compared with other models but greatly reduced the model complexity. Moreover, it is found that proposed model is less prone to over fitting as compared to other models.
Collapse
Affiliation(s)
- Sohamkumar Chauhan
- Department of Computer Science and Engineering, National Institute of Technology Goa, Ponda, 403401, Goa, India
| | - Damoder Reddy Edla
- Department of Computer Science and Engineering, National Institute of Technology Goa, Ponda, 403401, Goa, India
| | - Vijayasree Boddu
- Department of Electronics and Communication Engineering, National Institute of Technology Warangal, Hanamkonda, 506004, Telangana, India
| | - M Jayanthi Rao
- Department of CSE, Aditya Institute of Technology and Management, Kotturu, Tekkali, Andhra Pradesh, India
| | - Ramalingaswamy Cheruku
- Department of Computer Science and Engineering, National Institute of Technology Warangal, Hanamkonda, 506004, Telangana, India
| | - Soumya Ranjan Nayak
- School of Computer Engineering, KIIT Deemed to be University, Bhubaneswar, 751024, Odisha, India
| | - Sheshikala Martha
- School of Computer Science and Artificial Intelligence, SR University, Warangal, 506004, Telangana, India
| | - Kamppa Lavanya
- University College of Sciences, Acharya Nagarjuna Univesity, Guntur, Andhra Pradesh, India
| | - Tsedenya Debebe Nigat
- Faculty of Computing and Informatics, Jimma Institute of Technology, Jimma, Oromia, Ethiopia.
| |
Collapse
|
16
|
Saha S, Nandi D. SVM-RLF-DNN: A DNN with reliefF and SVM for automatic identification of COVID from chest X-ray and CT images. Digit Health 2024; 10:20552076241257045. [PMID: 38812845 PMCID: PMC11135098 DOI: 10.1177/20552076241257045] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/17/2024] [Accepted: 05/08/2024] [Indexed: 05/31/2024] Open
Abstract
Aim To develop an advanced determination technology for detecting COVID-19 patterns from chest X-ray and CT-scan films with distinct applications of deep learning and machine learning methods. Methods and Materials The newly enhanced proposed hybrid classification network (SVM-RLF-DNN) comprises of three phases: feature extraction, selection and classification. The in-depth features are extracted from a series of 3×3 convolution, 2×2 max polling operations followed by a flattened and fully connected layer of the deep neural network (DNN). ReLU activation function and Adam optimizer are used in the model. The ReliefF is an improved feature selection algorithm of Relief that uses Manhattan distance instead of Euclidean distance. Based on the significance of the feature, the ReliefF assigns weight to each extracted feature received from a fully connected layer. The weight to each feature is the average of k closest hits and misses in each class for a neighbouring instance pair in multiclass problems. The ReliefF eliminates lower-weight features by setting the node value to zero. The higher weights of the features are kept to obtain the feature selection. At the last layer of the neural network, the multiclass Support Vector Machine (SVM) is used to classify the patterns of COVID-19, viral pneumonia and healthy cases. The three classes with three binary SVM classifiers use linear kernel function for each binary SVM following a one-versus-all approach. The hinge loss function and L2-norm regularization are selected for more stable results. The proposed method is assessed on publicly available chest X-ray and CT-scan image databases from Kaggle and GitHub. The performance of the proposed classification model has comparable training, validation, and test accuracy, as well as sensitivity, specificity, and confusion matrix for quantitative evaluation on five-fold cross-validation. Results Our proposed network has achieved test accuracy of 98.48% and 95.34% on 2-class X-rays and CT. More importantly, the proposed model's test accuracy, sensitivity, and specificity are 87.9%, 86.32%, and 90.25% for 3-class classification (COVID-19, Pneumonia, Normal) on chest X-rays. The proposed model provides the test accuracy, sensitivity, and specificity of 95.34%, 94.12%, and 96.15% for 2-class classification (COVID-19, Non-COVID) on chest CT. Conclusion Our proposed classification network experimental results indicate competitiveness with existing neural networks. The proposed neural network assists clinicians in determining and surveilling the disease.
Collapse
Affiliation(s)
- Sanjib Saha
- Department of Computer Science and Engineering, National Institute of Technology, Durgapur, India
- Department of Computer Science and Engineering, Dr. B. C. Roy Engineering College, Durgapur, India
| | - Debashis Nandi
- Department of Computer Science and Engineering, National Institute of Technology, Durgapur, India
| |
Collapse
|
17
|
Atasever ÜH, Tercan E. Deep learning-based burned forest areas mapping via Sentinel-2 imagery: a comparative study. ENVIRONMENTAL SCIENCE AND POLLUTION RESEARCH INTERNATIONAL 2024; 31:5304-5318. [PMID: 38112873 DOI: 10.1007/s11356-023-31575-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/27/2022] [Accepted: 12/11/2023] [Indexed: 12/21/2023]
Abstract
In order to evaluate the effects of forest fires on the dynamics of the function and structure of ecosystems, it is necessary to determine burned forest areas with high accuracy, effectively, economically, and practically using satellite images. Extraction of burned forest areas utilizing high-resolution satellite images and image classification algorithms and assessing the successfulness of varied classification algorithms has become a prominent research field. This study aims to indicate on the capability of the deep learning-based Stacked Autoencoders method for the burned forest areas mapping from Sentinel-2 satellite images. The Stacked Autoencoders, used in this study as an unsupervised learning method, were compared qualitatively and quantitatively with frequently used supervised learning algorithms (k-Nearest Neighbors (k-NN), Subspaced k-NN, Support Vector Machines, Random Forest, Bagged Decision Tree, Naive Bayes, Linear Discriminant Analysis) on two distinct burnt forest zones. By selecting burned forest zones with contrasting structural characteristics from one another, an objective assessment was achieved. Manually digitized burned areas from Sentinel-2 satellite images were utilized for accuracy assessment. For comparison, different classification performance and quality metrics (Overall Accuracy, Mean Squared Error, Correlation Coefficient, Structural Similarity Index Measure, Peak Signal-to-Noise Ratio, Universal Image Quality Index, and KAPPA metrics) were used. In addition, whether the Stacked Autoencoders method produces consistent results was examined through boxplots. In terms of both quantitative and qualitative analysis, the Stacked Autoencoders method showed the highest accuracy values.
Collapse
Affiliation(s)
- Ümit Haluk Atasever
- Department of Geomatics Engineering, Faculty of Engineering, Erciyes University, 38039, Kayseri, Turkey
| | - Emre Tercan
- Department of Traffic Safety, 13th Region, General Directorate of Highways, 07090, Antalya, Turkey.
| |
Collapse
|
18
|
Singh K, Kaur N, Prabhu A. Combating COVID-19 Crisis using Artificial Intelligence (AI) Based Approach: Systematic Review. Curr Top Med Chem 2024; 24:737-753. [PMID: 38318824 DOI: 10.2174/0115680266282179240124072121] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/18/2023] [Revised: 12/19/2023] [Accepted: 12/27/2023] [Indexed: 02/07/2024]
Abstract
BACKGROUND SARS-CoV-2, the unique coronavirus that causes COVID-19, has wreaked damage around the globe, with victims displaying a wide range of difficulties that have encouraged medical professionals to look for innovative technical solutions and therapeutic approaches. Artificial intelligence-based methods have contributed a significant part in tackling complicated issues, and some institutions have been quick to embrace and tailor these solutions in response to the COVID-19 pandemic's obstacles. Here, in this review article, we have covered a few DL techniques for COVID-19 detection and diagnosis, as well as ML techniques for COVID-19 identification, severity classification, vaccine and drug development, mortality rate prediction, contact tracing, risk assessment, and public distancing. This review illustrates the overall impact of AI/ML tools on tackling and managing the outbreak. PURPOSE The focus of this research was to undertake a thorough evaluation of the literature on the part of Artificial Intelligence (AI) as a complete and efficient solution in the battle against the COVID-19 epidemic in the domains of detection and diagnostics of disease, mortality prediction and vaccine as well as drug development. METHODS A comprehensive exploration of PubMed, Web of Science, and Science Direct was conducted using PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analysis) regulations to find all possibly suitable papers conducted and made publicly available between December 1, 2019, and August 2023. COVID-19, along with AI-specific words, was used to create the query syntax. RESULTS During the period covered by the search strategy, 961 articles were published and released online. Out of these, a total of 135 papers were chosen for additional investigation. Mortality rate prediction, early detection and diagnosis, vaccine as well as drug development, and lastly, incorporation of AI for supervising and controlling the COVID-19 pandemic were the four main topics focused entirely on AI applications used to tackle the COVID-19 crisis. Out of 135, 60 research papers focused on the detection and diagnosis of the COVID-19 pandemic. Next, 19 of the 135 studies applied a machine-learning approach for mortality rate prediction. Another 22 research publications emphasized the vaccine as well as drug development. Finally, the remaining studies were concentrated on controlling the COVID-19 pandemic by applying AI AI-based approach to it. CONCLUSION We compiled papers from the available COVID-19 literature that used AI-based methodologies to impart insights into various COVID-19 topics in this comprehensive study. Our results suggest crucial characteristics, data types, and COVID-19 tools that can aid in medical and translational research facilitation.
Collapse
Affiliation(s)
- Kavya Singh
- Department of Biotechnology, Banasthali University, Banasthali Vidyapith, Banasthali, 304022, Rajasthan, India
| | - Navjeet Kaur
- Department of Chemistry & Division of Research and Development, Lovely Professional University, Phagwara, 144411, Punjab, India
| | - Ashish Prabhu
- Biotechnology Department, NIT Warangal, Warangal, 506004, Telangana, India
| |
Collapse
|
19
|
Alam FB, Podder P, Mondal MRH. RVCNet: A hybrid deep neural network framework for the diagnosis of lung diseases. PLoS One 2023; 18:e0293125. [PMID: 38153925 PMCID: PMC10754458 DOI: 10.1371/journal.pone.0293125] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/03/2023] [Accepted: 10/06/2023] [Indexed: 12/30/2023] Open
Abstract
Early evaluation and diagnosis can significantly reduce the life-threatening nature of lung diseases. Computer-aided diagnostic systems (CADs) can help radiologists make more precise diagnoses and reduce misinterpretations in lung disease diagnosis. Existing literature indicates that more research is needed to correctly classify lung diseases in the presence of multiple classes for different radiographic imaging datasets. As a result, this paper proposes RVCNet, a hybrid deep neural network framework for predicting lung diseases from an X-ray dataset of multiple classes. This framework is developed based on the ideas of three deep learning techniques: ResNet101V2, VGG19, and a basic CNN model. In the feature extraction phase of this new hybrid architecture, hyperparameter fine-tuning is used. Additional layers, such as batch normalization, dropout, and a few dense layers, are applied in the classification phase. The proposed method is applied to a dataset of COVID-19, non-COVID lung infections, viral pneumonia, and normal patients' X-ray images. The experiments take into account 2262 training and 252 testing images. Results show that with the Nadam optimizer, the proposed algorithm has an overall classification accuracy, AUC, precision, recall, and F1-score of 91.27%, 92.31%, 90.48%, 98.30%, and 94.23%, respectively. Finally, these results are compared with some recent deep-learning models. For this four-class dataset, the proposed RVCNet has a classification accuracy of 91.27%, which is better than ResNet101V2, VGG19, VGG19 over CNN, and other stand-alone models. Finally, the application of the GRAD-CAM approach clearly interprets the classification of images by the RVCNet framework.
Collapse
Affiliation(s)
- Fatema Binte Alam
- Institute of Information and Communication Technology, Bangladesh University of Engineering and Technology, Dhaka, Bangladesh
| | - Prajoy Podder
- Institute of Information and Communication Technology, Bangladesh University of Engineering and Technology, Dhaka, Bangladesh
| | - M. Rubaiyat Hossain Mondal
- Institute of Information and Communication Technology, Bangladesh University of Engineering and Technology, Dhaka, Bangladesh
| |
Collapse
|
20
|
Guo K, Cheng J, Li K, Wang L, Lv Y, Cao D. Diagnosis and detection of pneumonia using weak-label based on X-ray images: a multi-center study. BMC Med Imaging 2023; 23:209. [PMID: 38087255 PMCID: PMC10717871 DOI: 10.1186/s12880-023-01174-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/13/2023] [Accepted: 12/05/2023] [Indexed: 12/18/2023] Open
Abstract
PURPOSE Development and assessment the deep learning weakly supervised algorithm for the classification and detection pneumonia via X-ray. METHODS This retrospective study analyzed two publicly available dataset that contain X-ray images of pneumonia cases and normal cases. The first dataset from Guangzhou Women and Children's Medical Center. It contains a total of 5,856 X-ray images, which are divided into training, validation, and test sets with 8:1:1 ratio for algorithm training and testing. The deep learning algorithm ResNet34 was employed to build diagnostic model. And the second public dataset were collated by researchers from Qatar University and the University of Dhaka along with collaborators from Pakistan and Malaysia and some medical doctors. A total of 1,300 images of COVID-19 positive cases, 1,300 normal images and 1,300 images of viral pneumonia for external validation. Class activation map (CAM) were used to location the pneumonia lesions. RESULTS The ResNet34 model for pneumonia detection achieved an AUC of 0.9949 [0.9910-0.9981] (with an accuracy of 98.29% a sensitivity of 99.29% and a specificity of 95.57%) in the test dataset. And for external validation dataset, the model obtained an AUC of 0.9835[0.9806-0.9864] (with an accuracy of 94.62%, a sensitivity of 92.35% and a specificity of 99.15%). Moreover, the CAM can accurately locate the pneumonia area. CONCLUSION The deep learning algorithm can accurately detect pneumonia and locate the pneumonia area based on weak supervision information, which can provide potential value for helping radiologists to improve their accuracy of detection pneumonia patients through X-ray images.
Collapse
Affiliation(s)
- Kairou Guo
- Department of Biomedical Engineering, Chinese PLA General Hospital, Beijing, 100853, P.R. China
| | - Jiangbo Cheng
- Department of Biomedical Engineering, Chinese PLA General Hospital, Beijing, 100853, P.R. China
| | - Kaiyuan Li
- Department of Biomedical Engineering, Chinese PLA General Hospital, Beijing, 100853, P.R. China
| | - Lanhui Wang
- Department of Biomedical Engineering, Chinese PLA General Hospital, Beijing, 100853, P.R. China
| | - Yadong Lv
- Department of Biomedical Engineering, Chinese PLA General Hospital, Beijing, 100853, P.R. China
| | - Desen Cao
- Department of Biomedical Engineering, Chinese PLA General Hospital, Beijing, 100853, P.R. China.
| |
Collapse
|
21
|
|
22
|
Sadeghi MH, Sina S, Alavi M, Giammarile F. The OCDA-Net: a 3D convolutional neural network-based system for classification and staging of ovarian cancer patients using [ 18F]FDG PET/CT examinations. Ann Nucl Med 2023; 37:645-654. [PMID: 37768493 DOI: 10.1007/s12149-023-01867-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/23/2023] [Accepted: 09/11/2023] [Indexed: 09/29/2023]
Abstract
OBJECTIVE To create the 3D convolutional neural network (CNN)-based system that can use whole-body [18F]FDG PET for recurrence/post-therapy surveillance in ovarian cancer (OC). METHODS In this study, 1224 image sets from OC patients who underwent whole-body [18F]FDG PET/CT at Kowsar Hospital between April 2019 and May 2022 were investigated. For recurrence/post-therapy surveillance, diagnostic classification as cancerous, and non-cancerous and staging as stage III, and stage IV were determined by pathological diagnosis and specialists' interpretation. New deep neural network algorithms, the OCDAc-Net, and the OCDAs-Net were developed for diagnostic classification and staging of OC patients using [18F]FDG PET/CT images. Examinations were divided into independent training (75%), validation (10%), and testing (15%) subsets. RESULTS This study included 37 women (mean age 56.3 years; age range 36-83 years). Data augmentation techniques were applied to the images in two phases. There were 1224 image sets for diagnostic classification and staging. For the test set, 170 image sets were considered for diagnostic classification and staging. The OCDAc-Net areas under the receiver operating characteristic curve (AUCs) and overall accuracy for diagnostic classification were 0.990 and 0.92, respectively. The OCDAs-Net achieved areas under the receiver operating characteristic curve (AUCs) of 0.995 and overall accuracy of 0.94 for staging. CONCLUSIONS The proposed 3D CNN-based models provide potential tools for recurrence/post-therapy surveillance in OC. The OCDAc-Net and the OCDAs-Net model provide a new prognostic analysis method that can utilize PET images without pathological findings for diagnostic classification and staging.
Collapse
Affiliation(s)
- Mohammad Hossein Sadeghi
- Nuclear Engineering Department, School of Mechanical Engineering, Shiraz University, Shiraz, Iran
| | - Sedigheh Sina
- Nuclear Engineering Department, School of Mechanical Engineering, Shiraz University, Shiraz, Iran.
- Radiation Research Center, School of Mechanical Engineering, Shiraz University, Shiraz, Iran.
| | - Mehrosadat Alavi
- Department of Nuclear Medicine, Shiraz University of Medical Sciences, Shiraz, Iran
| | - Francesco Giammarile
- Nuclear Medicine and Diagnostic Imaging Section, Division of Human Health, International Atomic Energy Agency, Vienna, Austria
| |
Collapse
|
23
|
Ji L, Wei Z, Hao J, Wang C. An intelligent diagnostic method of ECG signal based on Markov transition field and a ResNet. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2023; 242:107784. [PMID: 37660577 DOI: 10.1016/j.cmpb.2023.107784] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/09/2022] [Revised: 08/29/2023] [Accepted: 08/29/2023] [Indexed: 09/05/2023]
Abstract
BACKGROUND AND OBJECTIVE Heart disease seriously threatens human life and health. It has the character of abruptness and is necessary to accurately monitor and intelligently diagnose electrocardiograph signals in real-time. As part of the automation of heart monitoring, the electrocardiogram (ECG) intelligent diagnosis method based on deep learning not only meets the needs of real-time and accurate but also can abandon relevant professional knowledge, which makes it possible to be promoted in the general population. METHODS This paper presents an intelligent diagnosis method based on a ResNet. Firstly, ECG signals from MIT-BIH Database are converted into 2-dim matrices by Markov Transition Field. Secondly, the matrices are used as the input of a ResNet. Then, the ResNet is able to extract high abstract features of various diseases and realize intelligent identification of five heartbeat types, including Normal Beat, Left Bundle Branch Block Beat, Right Bundle Branch Block Beat, Premature Ventricular Contraction Beat, and Atrial Premature Contraction Beat. Eventually, the proposed model is used to identify Normal Beat and Atrial Fibrillation(AF) based on the PAF Prediction Challenge Database(the PAFPC Database) to verify its generalization ability. RESULTS The experiment result shows that the intelligent diagnosis method can reach a high F1-score of 97.7% and a high accuracy upon to 99.2% on MIT-BIH Database, which are higher than the models proposed by other researchers. Its mean sensitivity and mean specificity are 97.42% and 99.54%, respectively. Moreover, the accuracy of the generalization ability verification experiment is 94.57% on the PAFPC Database, which is also higher than the results of other studies. CONCLUSION The research results show that the method proposed in this paper still achieves higher accuracy and higher F1-score than other methods without any data preprocessing. This method has better classification performance than traditional machine learning methods and other deep learning methods. That is, the method based on Markov Transition Field and a ResNet has good application prospects. At the same time, it has been verified that the model proposed in this paper also has excellent generalization ability.
Collapse
Affiliation(s)
- Lipeng Ji
- School of Mechanical Engineering, University of Shanghai for Science and Technology, Shanghai, China.
| | - Zhonghao Wei
- School of Mechanical Engineering, University of Shanghai for Science and Technology, Shanghai, China
| | - Jian Hao
- School of Mechanical Engineering, University of Shanghai for Science and Technology, Shanghai, China
| | - Chunli Wang
- Department of Geriatrics, Xinhua Hospital Affiliated to Shanghai Jiao Tong University of Medicine, Shanghai, China
| |
Collapse
|
24
|
Tehrani SSM, Zarvani M, Amiri P, Ghods Z, Raoufi M, Safavi-Naini SAA, Soheili A, Gharib M, Abbasi H. Visual transformer and deep CNN prediction of high-risk COVID-19 infected patients using fusion of CT images and clinical data. BMC Med Inform Decis Mak 2023; 23:265. [PMID: 37978393 PMCID: PMC10656999 DOI: 10.1186/s12911-023-02344-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/08/2023] [Accepted: 10/16/2023] [Indexed: 11/19/2023] Open
Abstract
BACKGROUND Despite the globally reducing hospitalization rates and the much lower risks of Covid-19 mortality, accurate diagnosis of the infection stage and prediction of outcomes are clinically of interest. Advanced current technology can facilitate automating the process and help identifying those who are at higher risks of developing severe illness. This work explores and represents deep-learning-based schemes for predicting clinical outcomes in Covid-19 infected patients, using Visual Transformer and Convolutional Neural Networks (CNNs), fed with 3D data fusion of CT scan images and patients' clinical data. METHODS We report on the efficiency of Video Swin Transformers and several CNN models fed with fusion datasets and CT scans only vs. a set of conventional classifiers fed with patients' clinical data only. A relatively large clinical dataset from 380 Covid-19 diagnosed patients was used to train/test the models. RESULTS Results show that the 3D Video Swin Transformers fed with the fusion datasets of 64 sectional CT scans + 67 clinical labels outperformed all other approaches for predicting outcomes in Covid-19-infected patients amongst all techniques (i.e., TPR = 0.95, FPR = 0.40, F0.5 score = 0.82, AUC = 0.77, Kappa = 0.6). CONCLUSIONS We demonstrate how the utility of our proposed novel 3D data fusion approach through concatenating CT scan images with patients' clinical data can remarkably improve the performance of the models in predicting Covid-19 infection outcomes. SIGNIFICANCE Findings indicate possibilities of predicting the severity of outcome using patients' CT images and clinical data collected at the time of admission to hospital.
Collapse
Affiliation(s)
| | - Maral Zarvani
- Faculty of Engineering, Alzahra University, Tehran, Iran
| | - Paria Amiri
- University of Erlangen-Nuremberg, Bavaria, Germany
| | - Zahra Ghods
- Faculty of Engineering, Alzahra University, Tehran, Iran
| | - Masoomeh Raoufi
- Department of Radiology, School of Medicine, Imam Hossein Hospital, Shahid Beheshti, University of Medical Sciences, Tehran, Iran
| | - Seyed Amir Ahmad Safavi-Naini
- Research Institute for Gastroenterology and Liver Diseases, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - Amirali Soheili
- School of Medicine, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | | | - Hamid Abbasi
- Auckland Bioengineering Institute, University of Auckland, Auckland, 1010, New Zealand.
| |
Collapse
|
25
|
Nahiduzzaman M, Goni MOF, Hassan R, Islam MR, Syfullah MK, Shahriar SM, Anower MS, Ahsan M, Haider J, Kowalski M. Parallel CNN-ELM: A multiclass classification of chest X-ray images to identify seventeen lung diseases including COVID-19. EXPERT SYSTEMS WITH APPLICATIONS 2023; 229:120528. [PMID: 37274610 PMCID: PMC10223636 DOI: 10.1016/j.eswa.2023.120528] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/01/2022] [Revised: 05/19/2023] [Accepted: 05/19/2023] [Indexed: 06/06/2023]
Abstract
Numerous epidemic lung diseases such as COVID-19, tuberculosis (TB), and pneumonia have spread over the world, killing millions of people. Medical specialists have experienced challenges in correctly identifying these diseases due to their subtle differences in Chest X-ray images (CXR). To assist the medical experts, this study proposed a computer-aided lung illness identification method based on the CXR images. For the first time, 17 different forms of lung disorders were considered and the study was divided into six trials with each containing two, two, three, four, fourteen, and seventeen different forms of lung disorders. The proposed framework combined robust feature extraction capabilities of a lightweight parallel convolutional neural network (CNN) with the classification abilities of the extreme learning machine algorithm named CNN-ELM. An optimistic accuracy of 90.92% and an area under the curve (AUC) of 96.93% was achieved when 17 classes were classified side by side. It also accurately identified COVID-19 and TB with 99.37% and 99.98% accuracy, respectively, in 0.996 microseconds for a single image. Additionally, the current results also demonstrated that the framework could outperform the existing state-of-the-art (SOTA) models. On top of that, a secondary conclusion drawn from this study was that the prospective framework retained its effectiveness over a range of real-world environments, including balanced-unbalanced or large-small datasets, large multiclass or simple binary class, and high- or low-resolution images. A prototype Android App was also developed to establish the potential of the framework in real-life implementation.
Collapse
Affiliation(s)
- Md Nahiduzzaman
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Md Omaer Faruq Goni
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Rakibul Hassan
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Md Robiul Islam
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Md Khalid Syfullah
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Saleh Mohammed Shahriar
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Md Shamim Anower
- Department of Electrical & Electronic Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Mominul Ahsan
- Department of Computer Science, University of York, Deramore Lane, Heslington, York YO10 5GH, UK
| | - Julfikar Haider
- Department of Engineering, Manchester Metropolitan University, Chester St, Manchester M1 5GD, UK
| | - Marcin Kowalski
- Institute of Optoelectronics, Military University of Technology, Gen. S. Kaliskiego 2, 00-908 Warsaw, Poland
| |
Collapse
|
26
|
Schaudt D, von Schwerin R, Hafner A, Riedel P, Reichert M, von Schwerin M, Beer M, Kloth C. Augmentation strategies for an imbalanced learning problem on a novel COVID-19 severity dataset. Sci Rep 2023; 13:18299. [PMID: 37880333 PMCID: PMC10600145 DOI: 10.1038/s41598-023-45532-2] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/23/2023] [Accepted: 10/20/2023] [Indexed: 10/27/2023] Open
Abstract
Since the beginning of the COVID-19 pandemic, many different machine learning models have been developed to detect and verify COVID-19 pneumonia based on chest X-ray images. Although promising, binary models have only limited implications for medical treatment, whereas the prediction of disease severity suggests more suitable and specific treatment options. In this study, we publish severity scores for the 2358 COVID-19 positive images in the COVIDx8B dataset, creating one of the largest collections of publicly available COVID-19 severity data. Furthermore, we train and evaluate deep learning models on the newly created dataset to provide a first benchmark for the severity classification task. One of the main challenges of this dataset is the skewed class distribution, resulting in undesirable model performance for the most severe cases. We therefore propose and examine different augmentation strategies, specifically targeting majority and minority classes. Our augmentation strategies show significant improvements in precision and recall values for the rare and most severe cases. While the models might not yet fulfill medical requirements, they serve as an appropriate starting point for further research with the proposed dataset to optimize clinical resource allocation and treatment.
Collapse
Affiliation(s)
- Daniel Schaudt
- Department of Computer Science, Ulm University of Applied Science, Albert-Einstein-Allee 55, 89081, Ulm, Baden-Wurttemberg, Germany.
| | - Reinhold von Schwerin
- Department of Computer Science, Ulm University of Applied Science, Albert-Einstein-Allee 55, 89081, Ulm, Baden-Wurttemberg, Germany
| | - Alexander Hafner
- Department of Computer Science, Ulm University of Applied Science, Albert-Einstein-Allee 55, 89081, Ulm, Baden-Wurttemberg, Germany
| | - Pascal Riedel
- Department of Computer Science, Ulm University of Applied Science, Albert-Einstein-Allee 55, 89081, Ulm, Baden-Wurttemberg, Germany
| | - Manfred Reichert
- Institute of Databases and Information Systems, Ulm University, James-Franck-Ring, 89081, Ulm, Baden-Wurttemberg, Germany
| | - Marianne von Schwerin
- Department of Computer Science, Ulm University of Applied Science, Albert-Einstein-Allee 55, 89081, Ulm, Baden-Wurttemberg, Germany
| | - Meinrad Beer
- Department of Radiology, University Hospital of Ulm, Albert-Einstein-Allee 23, 89081, Ulm, Baden-Wurttemberg, Germany
| | - Christopher Kloth
- Department of Radiology, University Hospital of Ulm, Albert-Einstein-Allee 23, 89081, Ulm, Baden-Wurttemberg, Germany
| |
Collapse
|
27
|
Bhandari M, Shahi TB, Neupane A. Evaluating Retinal Disease Diagnosis with an Interpretable Lightweight CNN Model Resistant to Adversarial Attacks. J Imaging 2023; 9:219. [PMID: 37888326 PMCID: PMC10607865 DOI: 10.3390/jimaging9100219] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/01/2023] [Revised: 09/29/2023] [Accepted: 10/09/2023] [Indexed: 10/28/2023] Open
Abstract
Optical Coherence Tomography (OCT) is an imperative symptomatic tool empowering the diagnosis of retinal diseases and anomalies. The manual decision towards those anomalies by specialists is the norm, but its labor-intensive nature calls for more proficient strategies. Consequently, the study recommends employing a Convolutional Neural Network (CNN) for the classification of OCT images derived from the OCT dataset into distinct categories, including Choroidal NeoVascularization (CNV), Diabetic Macular Edema (DME), Drusen, and Normal. The average k-fold (k = 10) training accuracy, test accuracy, validation accuracy, training loss, test loss, and validation loss values of the proposed model are 96.33%, 94.29%, 94.12%, 0.1073, 0.2002, and 0.1927, respectively. Fast Gradient Sign Method (FGSM) is employed to introduce non-random noise aligned with the cost function's data gradient, with varying epsilon values scaling the noise, and the model correctly handles all noise levels below 0.1 epsilon. Explainable AI algorithms: Local Interpretable Model-Agnostic Explanations (LIME) and SHapley Additive exPlanations (SHAP) are utilized to provide human interpretable explanations approximating the behaviour of the model within the region of a particular retinal image. Additionally, two supplementary datasets, namely, COVID-19 and Kidney Stone, are assimilated to enhance the model's robustness and versatility, resulting in a level of precision comparable to state-of-the-art methodologies. Incorporating a lightweight CNN model with 983,716 parameters, 2.37×108 floating point operations per second (FLOPs) and leveraging explainable AI strategies, this study contributes to efficient OCT-based diagnosis, underscores its potential in advancing medical diagnostics, and offers assistance in the Internet-of-Medical-Things.
Collapse
Affiliation(s)
- Mohan Bhandari
- Department of Science and Technology, Samriddhi College, Bhaktapur 44800, Nepal;
| | - Tej Bahadur Shahi
- School of Engineering and Technology, Central Queensland University, Norman Gardens, Rockhampton, QLD 4701, Australia;
- Central Department of Computer Science and IT, Tribhuvan University, Kathmandu 44600, Nepal
| | - Arjun Neupane
- School of Engineering and Technology, Central Queensland University, Norman Gardens, Rockhampton, QLD 4701, Australia;
| |
Collapse
|
28
|
Socha M, Prażuch W, Suwalska A, Foszner P, Tobiasz J, Jaroszewicz J, Gruszczynska K, Sliwinska M, Nowak M, Gizycka B, Zapolska G, Popiela T, Przybylski G, Fiedor P, Pawlowska M, Flisiak R, Simon K, Walecki J, Cieszanowski A, Szurowska E, Marczyk M, Polanska J. Pathological changes or technical artefacts? The problem of the heterogenous databases in COVID-19 CXR image analysis. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2023; 240:107684. [PMID: 37356354 PMCID: PMC10278898 DOI: 10.1016/j.cmpb.2023.107684] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/20/2023] [Revised: 06/11/2023] [Accepted: 06/18/2023] [Indexed: 06/27/2023]
Abstract
BACKGROUND When the COVID-19 pandemic commenced in 2020, scientists assisted medical specialists with diagnostic algorithm development. One scientific research area related to COVID-19 diagnosis was medical imaging and its potential to support molecular tests. Unfortunately, several systems reported high accuracy in development but did not fare well in clinical application. The reason was poor generalization, a long-standing issue in AI development. Researchers found many causes of this issue and decided to refer to them as confounders, meaning a set of artefacts and methodological errors associated with the method. We aim to contribute to this steed by highlighting an undiscussed confounder related to image resolution. METHODS 20 216 chest X-ray images (CXR) from worldwide centres were analyzed. The CXRs were bijectively projected into the 2D domain by performing Uniform Manifold Approximation and Projection (UMAP) embedding on the radiomic features (rUMAP) or CNN-based neural features (nUMAP) from the pre-last layer of the pre-trained classification neural network. Additional 44 339 thorax CXRs were used for validation. The comprehensive analysis of the multimodality of the density distribution in rUMAP/nUMAP domains and its relation to the original image properties was used to identify the main confounders. RESULTS nUMAP revealed a hidden bias of neural networks towards the image resolution, which the regular up-sampling procedure cannot compensate for. The issue appears regardless of the network architecture and is not observed in a high-resolution dataset. The impact of the resolution heterogeneity can be partially diminished by applying advanced deep-learning-based super-resolution networks. CONCLUSIONS rUMAP and nUMAP are great tools for image homogeneity analysis and bias discovery, as demonstrated by applying them to COVID-19 image data. Nonetheless, nUMAP could be applied to any type of data for which a deep neural network could be constructed. Advanced image super-resolution solutions are needed to reduce the impact of the resolution diversity on the classification network decision.
Collapse
Affiliation(s)
- Marek Socha
- Department of Data Science and Engineering, Silesian University of Technology, Gliwice, Poland
| | - Wojciech Prażuch
- Department of Data Science and Engineering, Silesian University of Technology, Gliwice, Poland
| | - Aleksandra Suwalska
- Department of Data Science and Engineering, Silesian University of Technology, Gliwice, Poland
| | - Paweł Foszner
- Department of Data Science and Engineering, Silesian University of Technology, Gliwice, Poland; Department of Computer Graphics, Vision and Digital Systems, Silesian University of Technology, Gliwice, Poland
| | - Joanna Tobiasz
- Department of Data Science and Engineering, Silesian University of Technology, Gliwice, Poland; Department of Computer Graphics, Vision and Digital Systems, Silesian University of Technology, Gliwice, Poland
| | - Jerzy Jaroszewicz
- Department of Infectious Diseases and Hepatology, Medical University of Silesia, Katowice, Poland
| | - Katarzyna Gruszczynska
- Department of Radiology and Nuclear Medicine, Medical University of Silesia, Katowice, Poland
| | - Magdalena Sliwinska
- Department of Diagnostic Imaging, Voivodship Specialist Hospital, Wroclaw, Poland
| | - Mateusz Nowak
- Department of Radiology, Silesian Hospital, Cieszyn, Poland
| | - Barbara Gizycka
- Department of Imaging Diagnostics, MEGREZ Hospital, Tychy, Poland
| | | | - Tadeusz Popiela
- Department of Radiology, Jagiellonian University Medical College, Krakow, Poland
| | - Grzegorz Przybylski
- Department of Lung Diseases, Cancer and Tuberculosis, Kujawsko-Pomorskie Pulmonology Center, Bydgoszcz, Poland
| | - Piotr Fiedor
- Department of General and Transplantation Surgery, Medical University of Warsaw, Warsaw, Poland
| | - Malgorzata Pawlowska
- Department of Infectious Diseases and Hepatology, Collegium Medicum in Bydgoszcz, Nicolaus Copernicus University, Torun, Poland
| | - Robert Flisiak
- Department of Infectious Diseases and Hepatology, Medical University of Bialystok, Bialystok, Poland
| | - Krzysztof Simon
- Department of Infectious Diseases and Hepatology, Wroclaw Medical University, Wroclaw, Poland
| | - Jerzy Walecki
- Department of Radiology, Centre of Postgraduate Medical Education, Central Clinical Hospital of the Ministry of Interior in Warsaw, Poland
| | - Andrzej Cieszanowski
- Department of Radiology I, The Maria Sklodowska-Curie National Research Institute of Oncology, Warsaw, Poland
| | - Edyta Szurowska
- 2nd Department of Radiology, Medical University of Gdansk, Poland
| | - Michal Marczyk
- Department of Data Science and Engineering, Silesian University of Technology, Gliwice, Poland; Yale Cancer Center, Yale School of Medicine, New Haven, CT, USA
| | - Joanna Polanska
- Department of Data Science and Engineering, Silesian University of Technology, Gliwice, Poland.
| |
Collapse
|
29
|
Ahmed MAO, Abbas IA, AbdelSatar Y. HDSNE a new unsupervised multiple image database fusion learning algorithm with flexible and crispy production of one database: a proof case study of lung infection diagnose In chest X-ray images. BMC Med Imaging 2023; 23:134. [PMID: 37718458 PMCID: PMC10506286 DOI: 10.1186/s12880-023-01078-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/23/2022] [Accepted: 08/16/2023] [Indexed: 09/19/2023] Open
Abstract
Continuous release of image databases with fully or partially identical inner categories dramatically deteriorates the production of autonomous Computer-Aided Diagnostics (CAD) systems for true comprehensive medical diagnostics. The first challenge is the frequent massive bulk release of medical image databases, which often suffer from two common drawbacks: image duplication and corruption. The many subsequent releases of the same data with the same classes or categories come with no clear evidence of success in the concatenation of those identical classes among image databases. This issue stands as a stumbling block in the path of hypothesis-based experiments for the production of a single learning model that can successfully classify all of them correctly. Removing redundant data, enhancing performance, and optimizing energy resources are among the most challenging aspects. In this article, we propose a global data aggregation scale model that incorporates six image databases selected from specific global resources. The proposed valid learner is based on training all the unique patterns within any given data release, thereby creating a unique dataset hypothetically. The Hash MD5 algorithm (MD5) generates a unique hash value for each image, making it suitable for duplication removal. The T-Distributed Stochastic Neighbor Embedding (t-SNE), with a tunable perplexity parameter, can represent data dimensions. Both the Hash MD5 and t-SNE algorithms are applied recursively, producing a balanced and uniform database containing equal samples per category: normal, pneumonia, and Coronavirus Disease of 2019 (COVID-19). We evaluated the performance of all proposed data and the new automated version using the Inception V3 pre-trained model with various evaluation metrics. The performance outcome of the proposed scale model showed more respectable results than traditional data aggregation, achieving a high accuracy of 98.48%, along with high precision, recall, and F1-score. The results have been proved through a statistical t-test, yielding t-values and p-values. It's important to emphasize that all t-values are undeniably significant, and the p-values provide irrefutable evidence against the null hypothesis. Furthermore, it's noteworthy that the Final dataset outperformed all other datasets across all metric values when diagnosing various lung infections with the same factors.
Collapse
Affiliation(s)
- Muhammad Atta Othman Ahmed
- Department of Computer Science, Faculty of Computers and Information, Luxor University, Luxor, 85951, Egypt.
| | - Ibrahim A Abbas
- Mathematics Department, Faculty of Science, Sohag University, Sohag, 82511, Egypt
| | - Yasser AbdelSatar
- Mathematics Department, Faculty of Science, Sohag University, Sohag, 82511, Egypt
| |
Collapse
|
30
|
Celik G. CovidCoughNet: A new method based on convolutional neural networks and deep feature extraction using pitch-shifting data augmentation for covid-19 detection from cough, breath, and voice signals. Comput Biol Med 2023; 163:107153. [PMID: 37321101 PMCID: PMC10249348 DOI: 10.1016/j.compbiomed.2023.107153] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/09/2023] [Revised: 05/25/2023] [Accepted: 06/07/2023] [Indexed: 06/17/2023]
Abstract
This study proposes a new deep learning-based method that demonstrates high performance in detecting Covid-19 disease from cough, breath, and voice signals. This impressive method, named CovidCoughNet, consists of a deep feature extraction network (InceptionFireNet) and a prediction network (DeepConvNet). The InceptionFireNet architecture, based on Inception and Fire modules, was designed to extract important feature maps. The DeepConvNet architecture, which is made up of convolutional neural network blocks, was developed to predict the feature vectors obtained from the InceptionFireNet architecture. The COUGHVID dataset containing cough data and the Coswara dataset containing cough, breath, and voice signals were used as the data sets. The pitch-shifting technique was used to data augmentation the signal data, which significantly contributed to improving performance. Additionally, Chroma features (CF), Root mean square energy (RMSE), Spectral centroid (SC), Spectral bandwidth (SB), Spectral rolloff (SR), Zero crossing rate (ZCR), and Mel frequency cepstral coefficients (MFCC) feature extraction techniques were used to extract important features from voice signals. Experimental studies have shown that using the pitch-shifting technique improved performance by around 3% compared to raw signals. When the proposed model was used with the COUGHVID dataset (Healthy, Covid-19, and Symptomatic), a high performance of 99.19% accuracy, 0.99 precision, 0.98 recall, 0.98 F1-Score, 97.77% specificity, and 98.44% AUC was achieved. Similarly, when the voice data in the Coswara dataset was used, higher performance was achieved compared to the cough and breath studies, with 99.63% accuracy, 100% precision, 0.99 recall, 0.99 F1-Score, 99.24% specificity, and 99.24% AUC. Moreover, when compared with current studies in the literature, the proposed model was observed to exhibit highly successful performance. The codes and details of the experimental studies can be accessed from the relevant Github page: (https://github.com/GaffariCelik/CovidCoughNet).
Collapse
Affiliation(s)
- Gaffari Celik
- Agri Ibrahim Cecen University, Department of Computer Technology, Agri, Turkey.
| |
Collapse
|
31
|
Choudhry IA, Qureshi AN, Aurangzeb K, Iqbal S, Alhussein M. Hybrid Diagnostic Model for Improved COVID-19 Detection in Lung Radiographs Using Deep and Traditional Features. Biomimetics (Basel) 2023; 8:406. [PMID: 37754157 PMCID: PMC10526442 DOI: 10.3390/biomimetics8050406] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/22/2023] [Revised: 08/25/2023] [Accepted: 08/29/2023] [Indexed: 09/28/2023] Open
Abstract
A recently discovered coronavirus (COVID-19) poses a major danger to human life and health across the planet. The most important step in managing and combating COVID-19 is to accurately screen and diagnose affected people. The imaging technology of lung X-ray is a useful imaging identification/detection approach among them. The help of such computer-aided machines and diagnoses to examine lung X-ray images of COVID-19 instances can give supplemental assessment ideas to specialists, easing their workload to some level. The novel concept of this study is a hybridized approach merging pertinent manual features with deep spatial features for the classification of COVID-19. Further, we employed traditional transfer learning techniques in this investigation, utilizing four different pre-trained CNN-based deep learning models, with the Inception model showing a reasonably accurate result and a diagnosis accuracy of 82.17%. We provide a successful diagnostic approach that blends deep characteristics with machine learning classification to further increase clinical performance. It employs a complete diagnostic model. Two datasets were used to test the suggested approach, and it did quite well on several of them. On 1102 lung X-ray scans, the model was originally evaluated. The results of the experiments indicate that the suggested SVM model has a diagnostic accuracy of 95.57%. When compared to the Xception model's baseline, the diagnostic accuracy had risen by 17.58 percent. The sensitivity, specificity, and AUC of the proposed models were 95.37 percent, 95.39%, and 95.77%, respectively. To show the adaptability of our approach, we also verified our proposed model on other datasets. Finally, we arrived at results that were conclusive. When compared to research of a comparable kind, our suggested CNN model has a greater accuracy of classification and diagnostic effectiveness.
Collapse
Affiliation(s)
- Imran Arshad Choudhry
- Department of Computer Science, Faculty of Information Technology & Computer Science, University of Central Punjab, Lahore 54000, Pakistan
| | - Adnan N. Qureshi
- Faculty of Arts, Society and Professional Studies, Newman University, Birmingham B32 3NT, UK;
| | - Khursheed Aurangzeb
- Department of Computer Engineering, College of Computer and Information Sciences, King Saud University, P.O. Box 51178, Riyadh 11543, Saudi Arabia; (K.A.); (M.A.)
| | - Saeed Iqbal
- Department of Computer Science, Faculty of Information Technology & Computer Science, University of Central Punjab, Lahore 54000, Pakistan
| | - Musaed Alhussein
- Department of Computer Engineering, College of Computer and Information Sciences, King Saud University, P.O. Box 51178, Riyadh 11543, Saudi Arabia; (K.A.); (M.A.)
| |
Collapse
|
32
|
Joloudari JH, Azizi F, Nodehi I, Nematollahi MA, Kamrannejhad F, Hassannatajjeloudari E, Alizadehsani R, Islam SMS. Developing a Deep Neural Network model for COVID-19 diagnosis based on CT scan images. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2023; 20:16236-16258. [PMID: 37920011 DOI: 10.3934/mbe.2023725] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/04/2023]
Abstract
COVID-19 is most commonly diagnosed using a testing kit but chest X-rays and computed tomography (CT) scan images have a potential role in COVID-19 diagnosis. Currently, CT diagnosis systems based on Artificial intelligence (AI) models have been used in some countries. Previous research studies used complex neural networks, which led to difficulty in network training and high computation rates. Hence, in this study, we developed the 6-layer Deep Neural Network (DNN) model for COVID-19 diagnosis based on CT scan images. The proposed DNN model is generated to improve accurate diagnostics for classifying sick and healthy persons. Also, other classification models, such as decision trees, random forests and standard neural networks, have been investigated. One of the main contributions of this study is the use of the global feature extractor operator for feature extraction from the images. Furthermore, the 10-fold cross-validation technique is utilized for partitioning the data into training, testing and validation. During the DNN training, the model is generated without dropping out of neurons in the layers. The experimental results of the lightweight DNN model demonstrated that this model has the best accuracy of 96.71% compared to the previous classification models for COVID-19 diagnosis.
Collapse
Affiliation(s)
| | - Faezeh Azizi
- Department of Computer Engineering, Faculty of Engineering, University of Birjand, Birjand, Iran
| | - Issa Nodehi
- Department of Computer Engineering, University of Qom, Qom, Iran
| | | | - Fateme Kamrannejhad
- Department of Computer Engineering, Faculty of Engineering, University of Birjand, Birjand, Iran
| | - Edris Hassannatajjeloudari
- Department of Nursing, School of Nursing and Allied Medical Sciences, Maragheh Faculty of Medical Sciences, Maragheh, Iran
| | - Roohallah Alizadehsani
- Institute for Intelligent Systems Research and Innovation, Deakin University, Geelong, VIC 3216, Australia
| | - Sheikh Mohammed Shariful Islam
- Institute for Physical Activity and Nutrition, School of Exercise and Nutrition Sciences, Deakin University, Geelong, VIC, Australia
| |
Collapse
|
33
|
Arora M, Davis CM, Gowda NR, Foster DG, Mondal A, Coopersmith CM, Kamaleswaran R. Uncertainty-Aware Convolutional Neural Network for Identifying Bilateral Opacities on Chest X-rays: A Tool to Aid Diagnosis of Acute Respiratory Distress Syndrome. Bioengineering (Basel) 2023; 10:946. [PMID: 37627831 PMCID: PMC10451804 DOI: 10.3390/bioengineering10080946] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/30/2023] [Revised: 07/26/2023] [Accepted: 08/03/2023] [Indexed: 08/27/2023] Open
Abstract
Acute Respiratory Distress Syndrome (ARDS) is a severe lung injury with high mortality, primarily characterized by bilateral pulmonary opacities on chest radiographs and hypoxemia. In this work, we trained a convolutional neural network (CNN) model that can reliably identify bilateral opacities on routine chest X-ray images of critically ill patients. We propose this model as a tool to generate predictive alerts for possible ARDS cases, enabling early diagnosis. Our team created a unique dataset of 7800 single-view chest-X-ray images labeled for the presence of bilateral or unilateral pulmonary opacities, or 'equivocal' images, by three blinded clinicians. We used a novel training technique that enables the CNN to explicitly predict the 'equivocal' class using an uncertainty-aware label smoothing loss. We achieved an Area under the Receiver Operating Characteristic Curve (AUROC) of 0.82 (95% CI: 0.80, 0.85), a precision of 0.75 (95% CI: 0.73, 0.78), and a sensitivity of 0.76 (95% CI: 0.73, 0.78) on the internal test set while achieving an (AUROC) of 0.84 (95% CI: 0.81, 0.86), a precision of 0.73 (95% CI: 0.63, 0.69), and a sensitivity of 0.73 (95% CI: 0.70, 0.75) on an external validation set. Further, our results show that this approach improves the model calibration and diagnostic odds ratio of the hypothesized alert tool, making it ideal for clinical decision support systems.
Collapse
Affiliation(s)
- Mehak Arora
- Department of Electrical and Computer Engineering, Georgia Institute of Technology, Atlanta, GA 30332, USA
- Department of Biomedical Informatics, Emory University School of Medicine, Atlanta, GA 30332, USA;
| | - Carolyn M. Davis
- Department of Surgery, Emory University School of Medicine, Atlanta, GA 30332, USA; (C.M.D.); (D.G.F.); (C.M.C.)
- Emory Critical Care Center, Emory University School of Medicine, Atlanta, GA 30332, USA
| | - Niraj R. Gowda
- Division of Pulmonary, Critical Care, Allergy and Sleep Medicine, Emory University School of Medicine, Atlanta, GA 30332, USA;
| | - Dennis G. Foster
- Department of Surgery, Emory University School of Medicine, Atlanta, GA 30332, USA; (C.M.D.); (D.G.F.); (C.M.C.)
| | - Angana Mondal
- Department of Biomedical Informatics, Emory University School of Medicine, Atlanta, GA 30332, USA;
| | - Craig M. Coopersmith
- Department of Surgery, Emory University School of Medicine, Atlanta, GA 30332, USA; (C.M.D.); (D.G.F.); (C.M.C.)
- Emory Critical Care Center, Emory University School of Medicine, Atlanta, GA 30332, USA
| | - Rishikesan Kamaleswaran
- Department of Biomedical Informatics, Emory University School of Medicine, Atlanta, GA 30332, USA;
- Emory Critical Care Center, Emory University School of Medicine, Atlanta, GA 30332, USA
| |
Collapse
|
34
|
Zaeri N. Artificial intelligence and machine learning responses to COVID-19 related inquiries. J Med Eng Technol 2023; 47:301-320. [PMID: 38625639 DOI: 10.1080/03091902.2024.2321846] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/14/2021] [Accepted: 02/18/2024] [Indexed: 04/17/2024]
Abstract
Researchers and scientists can use computational-based models to turn linked data into useful information, aiding in disease diagnosis, examination, and viral containment due to recent artificial intelligence and machine learning breakthroughs. In this paper, we extensively study the role of artificial intelligence and machine learning in delivering efficient responses to the COVID-19 pandemic almost four years after its start. In this regard, we examine a large number of critical studies conducted by various academic and research communities from multiple disciplines, as well as practical implementations of artificial intelligence algorithms that suggest potential solutions in investigating different COVID-19 decision-making scenarios. We identify numerous areas where artificial intelligence and machine learning can impact this context, including diagnosis (using chest X-ray imaging and CT imaging), severity, tracking, treatment, and the drug industry. Furthermore, we analyse the dilemma's limits, restrictions, and hazards.
Collapse
Affiliation(s)
- Naser Zaeri
- Faculty of Computer Studies, Arab Open University, Kuwait
| |
Collapse
|
35
|
Gopatoti A, Vijayalakshmi P. MTMC-AUR2CNet: Multi-textural multi-class attention recurrent residual convolutional neural network for COVID-19 classification using chest X-ray images. Biomed Signal Process Control 2023; 85:104857. [PMID: 36968651 PMCID: PMC10027978 DOI: 10.1016/j.bspc.2023.104857] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2022] [Revised: 02/13/2023] [Accepted: 03/11/2023] [Indexed: 03/24/2023]
Abstract
Coronavirus disease (COVID-19) has infected over 603 million confirmed cases as of September 2022, and its rapid spread has raised concerns worldwide. More than 6.4 million fatalities in confirmed patients have been reported. According to reports, the COVID-19 virus causes lung damage and rapidly mutates before the patient receives any diagnosis-specific medicine. Daily increasing COVID-19 cases and the limited number of diagnosis tool kits encourage the use of deep learning (DL) models to assist health care practitioners using chest X-ray (CXR) images. The CXR is a low radiation radiography tool available in hospitals to diagnose COVID-19 and combat this spread. We propose a Multi-Textural Multi-Class (MTMC) UNet-based Recurrent Residual Convolutional Neural Network (MTMC-UR2CNet) and MTMC-UR2CNet with attention mechanism (MTMC-AUR2CNet) for multi-class lung lobe segmentation of CXR images. The lung lobe segmentation output of MTMC-UR2CNet and MTMC-AUR2CNet are mapped individually with their input CXRs to generate the region of interest (ROI). The multi-textural features are extracted from the ROI of each proposed MTMC network. The extracted multi-textural features from ROI are fused and are trained to the Whale optimization algorithm (WOA) based DeepCNN classifier on classifying the CXR images into normal (healthy), COVID-19, viral pneumonia, and lung opacity. The experimental result shows that the MTMC-AUR2CNet has superior performance in multi-class lung lobe segmentation of CXR images with an accuracy of 99.47%, followed by MTMC-UR2CNet with an accuracy of 98.39%. Also, MTMC-AUR2CNet improves the multi-textural multi-class classification accuracy of the WOA-based DeepCNN classifier to 97.60% compared to MTMC-UR2CNet.
Collapse
Affiliation(s)
- Anandbabu Gopatoti
- Department of Electronics and Communication Engineering, Hindusthan College of Engineering and Technology, Coimbatore, Tamil Nadu, India
- Centre for Research, Anna University, Chennai, Tamil Nadu, India
| | - P Vijayalakshmi
- Department of Electronics and Communication Engineering, Hindusthan College of Engineering and Technology, Coimbatore, Tamil Nadu, India
| |
Collapse
|
36
|
Hussein HI, Mohammed AO, Hassan MM, Mstafa RJ. Lightweight deep CNN-based models for early detection of COVID-19 patients from chest X-ray images. EXPERT SYSTEMS WITH APPLICATIONS 2023; 223:119900. [PMID: 36969370 PMCID: PMC10023206 DOI: 10.1016/j.eswa.2023.119900] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/27/2022] [Revised: 03/05/2023] [Accepted: 03/15/2023] [Indexed: 06/18/2023]
Abstract
Hundreds of millions of people worldwide have recently been infected by the novel Coronavirus disease (COVID-19), causing significant damage to the health, economy, and welfare of the world's population. Moreover, the unprecedented number of patients with COVID-19 has placed a massive burden on healthcare centers, making timely and rapid diagnosis challenging. A crucial step in minimizing the impact of such problems is to automatically detect infected patients and place them under special care as quickly as possible. Deep learning algorithms, such as Convolutional Neural Networks (CNN), can be used to meet this need. Despite the desired results, most of the existing deep learning-based models were built on millions of parameters (weights), which are not applicable to devices with limited resources. Inspired by such fact, in this research, we developed two new lightweight CNN-based diagnostic models for the automatic and early detection of COVID-19 subjects from chest X-ray images. The first model was built for binary classification (COVID-19 and Normal), whereas the second one was built for multiclass classification (COVID-19, viral pneumonia, or normal). The proposed models were tested on a relatively large dataset of chest X-ray images, and the results showed that the accuracy rates of the 2- and 3-class-based classification models are 98.55% and 96.83%, respectively. The results also revealed that our models achieved competitive performance compared with the existing heavyweight models while significantly reducing cost and memory requirements for computing resources. With these findings, we can indicate that our models are helpful to clinicians in making insightful diagnoses of COVID-19 and are potentially easily deployable on devices with limited computational power and resources.
Collapse
Affiliation(s)
- Haval I Hussein
- Department of Computer Science, Faculty of Science, University of Zakho. Zakho, Kurdistan Region, Iraq
| | - Abdulhakeem O Mohammed
- Department of Information Technology Management, Technical College of Administration, Duhok Polytechnic University, Duhok, Iraq
| | - Masoud M Hassan
- Department of Computer Science, Faculty of Science, University of Zakho. Zakho, Kurdistan Region, Iraq
| | - Ramadhan J Mstafa
- Department of Computer Science, Faculty of Science, University of Zakho. Zakho, Kurdistan Region, Iraq
- Department of Computer Science, College of Science, Nawroz University, Duhok, Kurdistan Region, Iraq
| |
Collapse
|
37
|
Azeem M, Javaid S, Khalil RA, Fahim H, Althobaiti T, Alsharif N, Saeed N. Neural Networks for the Detection of COVID-19 and Other Diseases: Prospects and Challenges. Bioengineering (Basel) 2023; 10:850. [PMID: 37508877 PMCID: PMC10416184 DOI: 10.3390/bioengineering10070850] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/29/2023] [Revised: 07/09/2023] [Accepted: 07/12/2023] [Indexed: 07/30/2023] Open
Abstract
Artificial neural networks (ANNs) ability to learn, correct errors, and transform a large amount of raw data into beneficial medical decisions for treatment and care has increased in popularity for enhanced patient safety and quality of care. Therefore, this paper reviews the critical role of ANNs in providing valuable insights for patients' healthcare decisions and efficient disease diagnosis. We study different types of ANNs in the existing literature that advance ANNs' adaptation for complex applications. Specifically, we investigate ANNs' advances for predicting viral, cancer, skin, and COVID-19 diseases. Furthermore, we propose a deep convolutional neural network (CNN) model called ConXNet, based on chest radiography images, to improve the detection accuracy of COVID-19 disease. ConXNet is trained and tested using a chest radiography image dataset obtained from Kaggle, achieving more than 97% accuracy and 98% precision, which is better than other existing state-of-the-art models, such as DeTraC, U-Net, COVID MTNet, and COVID-Net, having 93.1%, 94.10%, 84.76%, and 90% accuracy and 94%, 95%, 85%, and 92% precision, respectively. The results show that the ConXNet model performed significantly well for a relatively large dataset compared with the aforementioned models. Moreover, the ConXNet model reduces the time complexity by using dropout layers and batch normalization techniques. Finally, we highlight future research directions and challenges, such as the complexity of the algorithms, insufficient available data, privacy and security, and integration of biosensing with ANNs. These research directions require considerable attention for improving the scope of ANNs for medical diagnostic and treatment applications.
Collapse
Affiliation(s)
- Muhammad Azeem
- School of Science, Engineering & Environment, University of Salford, Manchester M5 4WT, UK;
| | - Shumaila Javaid
- Department of Control Science and Engineering, College of Electronics and Information Engineering, Tongji University, Shanghai 201804, China; (S.J.); (H.F.)
| | - Ruhul Amin Khalil
- Department of Electrical Engineering, University of Engineering and Technology, Peshawar 25120, Pakistan;
- Department of Electrical and Communication Engineering, United Arab Emirates University (UAEU), Al-Ain 15551, United Arab Emirates
| | - Hamza Fahim
- Department of Control Science and Engineering, College of Electronics and Information Engineering, Tongji University, Shanghai 201804, China; (S.J.); (H.F.)
| | - Turke Althobaiti
- Department of Computer Science, Faculty of Science, Northern Border University, Arar 73222, Saudi Arabia;
| | - Nasser Alsharif
- Department of Administrative and Financial Sciences, Ranyah University Collage, Taif University, P.O. Box 11099, Taif 21944, Saudi Arabia;
| | - Nasir Saeed
- Department of Electrical and Communication Engineering, United Arab Emirates University (UAEU), Al-Ain 15551, United Arab Emirates
| |
Collapse
|
38
|
Pérez E, Ventura S. Progressive growing of Generative Adversarial Networks for improving data augmentation and skin cancer diagnosis. Artif Intell Med 2023; 141:102556. [PMID: 37295899 DOI: 10.1016/j.artmed.2023.102556] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/12/2022] [Revised: 04/06/2023] [Accepted: 04/14/2023] [Indexed: 06/12/2023]
Abstract
Early melanoma diagnosis is the most important factor in the treatment of skin cancer and can effectively reduce mortality rates. Recently, Generative Adversarial Networks have been used to augment data, prevent overfitting and improve the diagnostic capacity of models. However, its application remains a challenging task due to the high levels of inter and intra-class variance seen in skin images, limited amounts of data, and model instability. We present a more robust Progressive Growing of Adversarial Networks based on residual learning, which is highly recommended to ease the training of deep networks. The stability of the training process was increased by receiving additional inputs from preceding blocks. The architecture is able to produce plausible photorealistic synthetic 512 × 512 skin images, even with small dermoscopic and non-dermoscopic skin image datasets as problem domains. In this manner, we tackle the lack of data and the imbalance problems. Additionally, the proposed approach leverages a skin lesion boundary segmentation algorithm and transfer learning to enhance the diagnosis of melanoma. Inception score and Matthews Correlation Coefficient were used to measure the performance of the models. The architecture was evaluated qualitatively and quantitatively through the use of an extensive experimental study on sixteen datasets, illustrating its effectiveness in the diagnosis of melanoma. Finally, four state-of-the-art data augmentation techniques applied in five convolutional neural network models were significantly outperformed. The results indicated that a bigger number of trainable parameters will not necessarily obtain a better performance in melanoma diagnosis.
Collapse
Affiliation(s)
- Eduardo Pérez
- Andalusian Research Institute in Data Science and Computational Intelligence (DaSCI). University of Córdoba, Córdoba, Spain; Maimónides Biomedical Research Institute of Córdoba (IMIBIC). University of Córdoba, Córdoba, Spain
| | - Sebastián Ventura
- Andalusian Research Institute in Data Science and Computational Intelligence (DaSCI). University of Córdoba, Córdoba, Spain; Maimónides Biomedical Research Institute of Córdoba (IMIBIC). University of Córdoba, Córdoba, Spain.
| |
Collapse
|
39
|
Ukwuoma CC, Cai D, Heyat MBB, Bamisile O, Adun H, Al-Huda Z, Al-Antari MA. Deep learning framework for rapid and accurate respiratory COVID-19 prediction using chest X-ray images. JOURNAL OF KING SAUD UNIVERSITY. COMPUTER AND INFORMATION SCIENCES 2023; 35:101596. [PMID: 37275558 PMCID: PMC10211254 DOI: 10.1016/j.jksuci.2023.101596] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/19/2022] [Revised: 05/17/2023] [Accepted: 05/19/2023] [Indexed: 06/07/2023]
Abstract
COVID-19 is a contagious disease that affects the human respiratory system. Infected individuals may develop serious illnesses, and complications may result in death. Using medical images to detect COVID-19 from essentially identical thoracic anomalies is challenging because it is time-consuming, laborious, and prone to human error. This study proposes an end-to-end deep-learning framework based on deep feature concatenation and a Multi-head Self-attention network. Feature concatenation involves fine-tuning the pre-trained backbone models of DenseNet, VGG-16, and InceptionV3, which are trained on a large-scale ImageNet, whereas a Multi-head Self-attention network is adopted for performance gain. End-to-end training and evaluation procedures are conducted using the COVID-19_Radiography_Dataset for binary and multi-classification scenarios. The proposed model achieved overall accuracies (96.33% and 98.67%) and F1_scores (92.68% and 98.67%) for multi and binary classification scenarios, respectively. In addition, this study highlights the difference in accuracy (98.0% vs. 96.33%) and F_1 score (97.34% vs. 95.10%) when compared with feature concatenation against the highest individual model performance. Furthermore, a virtual representation of the saliency maps of the employed attention mechanism focusing on the abnormal regions is presented using explainable artificial intelligence (XAI) technology. The proposed framework provided better COVID-19 prediction results outperforming other recent deep learning models using the same dataset.
Collapse
Affiliation(s)
- Chiagoziem C Ukwuoma
- The College of Nuclear Technology and Automation Engineering, Chengdu University of Technology, Sichuan, 610059, China
| | - Dongsheng Cai
- The College of Nuclear Technology and Automation Engineering, Chengdu University of Technology, Sichuan, 610059, China
| | - Md Belal Bin Heyat
- IoT Research Center, College of Computer Science and Software Engineering, Shenzhen University, Shenzhen, Guangdong 518060, China
| | - Olusola Bamisile
- Sichuan Industrial Internet Intelligent Monitoring and Application Engineering Technology Research Center, Chengdu University of Technology, China
| | - Humphrey Adun
- Department of Mechanical and Energy Systems Engineering, Cyprus International University, Nicosia, North Nicosia, Cyprus
| | - Zaid Al-Huda
- School of Computing and Artificial Intelligence, Southwest Jiaotong University, Chengdu, Sichuan, China
| | - Mugahed A Al-Antari
- Department of Artificial Intelligence, College of Software & Convergence Technology, Daeyang AI Center, Sejong University, Seoul 05006, Republic of Korea
| |
Collapse
|
40
|
Iqbal A, Usman M, Ahmed Z. Tuberculosis chest X-ray detection using CNN-based hybrid segmentation and classification approach. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2023.104667] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/06/2023]
|
41
|
Li H, Drukker K, Hu Q, Whitney HM, Fuhrman JD, Giger ML. Predicting intensive care need for COVID-19 patients using deep learning on chest radiography. J Med Imaging (Bellingham) 2023; 10:044504. [PMID: 37608852 PMCID: PMC10440543 DOI: 10.1117/1.jmi.10.4.044504] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/01/2023] [Revised: 07/12/2023] [Accepted: 08/01/2023] [Indexed: 08/24/2023] Open
Abstract
Purpose Image-based prediction of coronavirus disease 2019 (COVID-19) severity and resource needs can be an important means to address the COVID-19 pandemic. In this study, we propose an artificial intelligence/machine learning (AI/ML) COVID-19 prognosis method to predict patients' needs for intensive care by analyzing chest X-ray radiography (CXR) images using deep learning. Approach The dataset consisted of 8357 CXR exams from 5046 COVID-19-positive patients as confirmed by reverse transcription polymerase chain reaction (RT-PCR) tests for the SARS-CoV-2 virus with a training/validation/test split of 64%/16%/20% on a by patient level. Our model involved a DenseNet121 network with a sequential transfer learning technique employed to train on a sequence of gradually more specific and complex tasks: (1) fine-tuning a model pretrained on ImageNet using a previously established CXR dataset with a broad spectrum of pathologies; (2) refining on another established dataset to detect pneumonia; and (3) fine-tuning using our in-house training/validation datasets to predict patients' needs for intensive care within 24, 48, 72, and 96 h following the CXR exams. The classification performances were evaluated on our independent test set (CXR exams of 1048 patients) using the area under the receiver operating characteristic curve (AUC) as the figure of merit in the task of distinguishing between those COVID-19-positive patients who required intensive care following the imaging exam and those who did not. Results Our proposed AI/ML model achieved an AUC (95% confidence interval) of 0.78 (0.74, 0.81) when predicting the need for intensive care 24 h in advance, and at least 0.76 (0.73, 0.80) for 48 h or more in advance using predictions based on the AI prognostic marker derived from CXR images. Conclusions This AI/ML prediction model for patients' needs for intensive care has the potential to support both clinical decision-making and resource management.
Collapse
Affiliation(s)
- Hui Li
- The University of Chicago, Department of Radiology, Chicago, Illinois, United States
| | - Karen Drukker
- The University of Chicago, Department of Radiology, Chicago, Illinois, United States
| | - Qiyuan Hu
- The University of Chicago, Department of Radiology, Chicago, Illinois, United States
| | - Heather M. Whitney
- The University of Chicago, Department of Radiology, Chicago, Illinois, United States
| | - Jordan D. Fuhrman
- The University of Chicago, Department of Radiology, Chicago, Illinois, United States
| | - Maryellen L. Giger
- The University of Chicago, Department of Radiology, Chicago, Illinois, United States
| |
Collapse
|
42
|
Zhu H, Zhu Z, Wang S, Zhang Y. CovC-ReDRNet: A Deep Learning Model for COVID-19 Classification. MACHINE LEARNING AND KNOWLEDGE EXTRACTION 2023; 5:684-712. [PMID: 38560420 PMCID: PMC7615781 DOI: 10.3390/make5030037] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Indexed: 04/04/2024]
Abstract
Since the COVID-19 pandemic outbreak, over 760 million confirmed cases and over 6.8 million deaths have been reported globally, according to the World Health Organization. While the SARS-CoV-2 virus carried by COVID-19 patients can be identified though the reverse transcription-polymerase chain reaction (RT-PCR) test with high accuracy, clinical misdiagnosis between COVID-19 and pneumonia patients remains a challenge. Therefore, we developed a novel CovC-ReDRNet model to distinguish COVID-19 patients from pneumonia patients as well as normal cases. ResNet-18 was introduced as the backbone model and tailored for the feature representation afterward. In our feature-based randomized neural network (RNN) framework, the feature representation automatically pairs with the deep random vector function link network (dRVFL) as the optimal classifier, producing a CovC-ReDRNet model for the classification task. Results based on five-fold cross-validation reveal that our method achieved 94.94%, 97.01%, 97.56%, 96.81%, and 95.84% MA sensitivity, MA specificity, MA accuracy, MA precision, and MA F1-score, respectively. Ablation studies evidence the superiority of ResNet-18 over different backbone networks, RNNs over traditional classifiers, and deep RNNs over shallow RNNs. Moreover, our proposed model achieved a better MA accuracy than the state-of-the-art (SOTA) methods, the highest score of which was 95.57%. To conclude, our CovC-ReDRNet model could be perceived as an advanced computer-aided diagnostic model with high speed and high accuracy for classifying and predicting COVID-19 diseases.
Collapse
Affiliation(s)
- Hanruo Zhu
- School of Computing and Mathematical Sciences, University of Leicester, Leicester LE1 7RH, UK
| | - Ziquan Zhu
- School of Computing and Mathematical Sciences, University of Leicester, Leicester LE1 7RH, UK
| | - Shuihua Wang
- School of Computing and Mathematical Sciences, University of Leicester, Leicester LE1 7RH, UK
| | - Yudong Zhang
- School of Computing and Mathematical Sciences, University of Leicester, Leicester LE1 7RH, UK
- School of Computer Science and Technology, Henan Polytechnic University, Jiaozuo 454000, China
- Department of Information Systems, Faculty of Computing and Information Technology, King Abdulaziz University, Jeddah 21589, Saudi Arabia
| |
Collapse
|
43
|
Nahiduzzaman M, Faruq Goni MO, Robiul Islam M, Sayeed A, Shamim Anower M, Ahsan M, Haider J, Kowalski M. Detection of various lung diseases including COVID-19 using extreme learning machine algorithm based on the features extracted from a lightweight CNN architecture. Biocybern Biomed Eng 2023; 43:S0208-5216(23)00037-2. [PMID: 38620111 PMCID: PMC10292668 DOI: 10.1016/j.bbe.2023.06.003] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/18/2022] [Revised: 04/04/2023] [Accepted: 06/16/2023] [Indexed: 11/09/2023]
Abstract
Around the world, several lung diseases such as pneumonia, cardiomegaly, and tuberculosis (TB) contribute to severe illness, hospitalization or even death, particularly for elderly and medically vulnerable patients. In the last few decades, several new types of lung-related diseases have taken the lives of millions of people, and COVID-19 has taken almost 6.27 million lives. To fight against lung diseases, timely and correct diagnosis with appropriate treatment is crucial in the current COVID-19 pandemic. In this study, an intelligent recognition system for seven lung diseases has been proposed based on machine learning (ML) techniques to aid the medical experts. Chest X-ray (CXR) images of lung diseases were collected from several publicly available databases. A lightweight convolutional neural network (CNN) has been used to extract characteristic features from the raw pixel values of the CXR images. The best feature subset has been identified using the Pearson Correlation Coefficient (PCC). Finally, the extreme learning machine (ELM) has been used to perform the classification task to assist faster learning and reduced computational complexity. The proposed CNN-PCC-ELM model achieved an accuracy of 96.22% with an Area Under Curve (AUC) of 99.48% for eight class classification. The outcomes from the proposed model demonstrated better performance than the existing state-of-the-art (SOTA) models in the case of COVID-19, pneumonia, and tuberculosis detection in both binary and multiclass classifications. For eight class classification, the proposed model achieved precision, recall and fi-score and ROC are 100%, 99%, 100% and 99.99% respectively for COVID-19 detection demonstrating its robustness. Therefore, the proposed model has overshadowed the existing pioneering models to accurately differentiate COVID-19 from the other lung diseases that can assist the medical physicians in treating the patient effectively.
Collapse
Affiliation(s)
- Md Nahiduzzaman
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Md Omaer Faruq Goni
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Md Robiul Islam
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Abu Sayeed
- Department of Computer Science & Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Md Shamim Anower
- Department of Electrical & Electronic Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Mominul Ahsan
- Department of Computer Science, University of York, Deramore Lane, Heslington, York YO10 5GH, UK
| | - Julfikar Haider
- Department of Engineering, Manchester Metropolitan University, Chester St, Manchester M1 5GD, UK
| | - Marcin Kowalski
- Institute of Optoelectronics, Military University of Technology, Gen. S. Kaliskiego 2, Warsaw, Poland
| |
Collapse
|
44
|
Merrouchi M, Benyoussef Y, Skittou M, Atifi K, Gadi T. ConvCoroNet: a deep convolutional neural network optimized with iterative thresholding algorithm for Covid-19 detection using chest X-ray images. J Biomol Struct Dyn 2023:1-14. [PMID: 37354142 DOI: 10.1080/07391102.2023.2227726] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/30/2022] [Accepted: 06/15/2023] [Indexed: 06/26/2023]
Abstract
Covid-19 is a global pandemic. Early and accurate detection of positive cases prevent the further spread of this epidemic and help to treat rapidly the infected patients. During the peak of this epidemic, there was an insufficiency of Covid-19 test kits. In addition, this technique takes a considerable time in the diagnosis. Hence the need to find fast, accurate and low-cost method to replace or supplement RT PCR-based methods. Covid-19 is a respiratory disease, chest X-ray images are often used to diagnose pneumonia. From this perspective, these images can play an important role in the Covid-19 detection. In this article, we propose ConvCoroNet, a deep convolutional neural network model optimized with new method based on iterative thresholding algorithm to detect coronavirus automatically from chest X-ray images. ConvCoroNet is trained on a dataset prepared by collecting chest X-ray images of Covid-19, pneumonia and normal cases from publically datasets. The experimental results of our proposed model show a high accuracy of 99.50%, sensitivity of 98.80% and specificity of 99.85% when detecting Covid-19 from chest X-ray images. ConvCoroNet achieves promising results in the automatic detection of Covid-19 from chest X-ray images. It may be able to help radiologists in the Covid-19 detection by reducing the examination time of X-ray images.Communicated by Ramaswamy H. Sarma.
Collapse
Affiliation(s)
- M Merrouchi
- Faculty of Science and Technology, Hassan First, Settat, Morocco
| | - Y Benyoussef
- National School of Applied Sciences, Hassan First, Berrechid, Morocco
| | - M Skittou
- Faculty of Science and Technology, Hassan First, Settat, Morocco
| | - K Atifi
- Faculty of Science and Technology, Hassan First, Settat, Morocco
| | - T Gadi
- Faculty of Science and Technology, Hassan First, Settat, Morocco
| |
Collapse
|
45
|
Mehrdad S, Shamout FE, Wang Y, Atashzar SF. Deep learning for deterioration prediction of COVID-19 patients based on time-series of three vital signs. Sci Rep 2023; 13:9968. [PMID: 37339986 DOI: 10.1038/s41598-023-37013-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/24/2022] [Accepted: 06/14/2023] [Indexed: 06/22/2023] Open
Abstract
Unrecognized deterioration of COVID-19 patients can lead to high morbidity and mortality. Most existing deterioration prediction models require a large number of clinical information, typically collected in hospital settings, such as medical images or comprehensive laboratory tests. This is infeasible for telehealth solutions and highlights a gap in deterioration prediction models based on minimal data, which can be recorded at a large scale in any clinic, nursing home, or even at the patient's home. In this study, we develop and compare two prognostic models that predict if a patient will experience deterioration in the forthcoming 3 to 24 h. The models sequentially process routine triadic vital signs: (a) oxygen saturation, (b) heart rate, and (c) temperature. These models are also provided with basic patient information, including sex, age, vaccination status, vaccination date, and status of obesity, hypertension, or diabetes. The difference between the two models is the way that the temporal dynamics of the vital signs are processed. Model #1 utilizes a temporally-dilated version of the Long-Short Term Memory model (LSTM) for temporal processes, and Model #2 utilizes a residual temporal convolutional network (TCN) for this purpose. We train and evaluate the models using data collected from 37,006 COVID-19 patients at NYU Langone Health in New York, USA. The convolution-based model outperforms the LSTM based model, achieving a high AUROC of 0.8844-0.9336 for 3 to 24 h deterioration prediction on a held-out test set. We also conduct occlusion experiments to evaluate the importance of each input feature, which reveals the significance of continuously monitoring the variation of the vital signs. Our results show the prospect for accurate deterioration forecast using a minimum feature set that can be relatively easily obtained using wearable devices and self-reported patient information.
Collapse
Affiliation(s)
- Sarmad Mehrdad
- Department of Electrical and Computer Engineering, New York University (NYU), New York, USA
| | - Farah E Shamout
- Department of Biomedical Engineering, New York University (NYU), New York, USA
- Division of Engineering, New York University Abu Dhabi (NYUAD), Abu Dhabi, UAE
- Computer Science and Engineering, New York University (NYU), New York, USA
| | - Yao Wang
- Department of Electrical and Computer Engineering, New York University (NYU), New York, USA
- Department of Biomedical Engineering, New York University (NYU), New York, USA
| | - S Farokh Atashzar
- Department of Electrical and Computer Engineering, New York University (NYU), New York, USA.
- Department of Biomedical Engineering, New York University (NYU), New York, USA.
- Department of Mechanical and Aerospace Engineering, New York University (NYU), New York, USA.
| |
Collapse
|
46
|
JavadiMoghaddam S. A novel framework based on deep learning for COVID-19 diagnosis from X-ray images. PeerJ Comput Sci 2023; 9:e1375. [PMID: 37346600 PMCID: PMC10280393 DOI: 10.7717/peerj-cs.1375] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/24/2023] [Accepted: 04/10/2023] [Indexed: 06/23/2023]
Abstract
Background The coronavirus infection has endangered human health because of the high speed of the outbreak. A rapid and accurate diagnosis of the infection is essential to avoid further spread. Due to the cost of diagnostic kits and the availability of radiology equipment in most parts of the world, the COVID-19 detection method using X-ray images is still used in underprivileged countries. However, they are challenging due to being prone to human error, time-consuming, and demanding. The success of deep learning (DL) in automatic COVID-19 diagnosis systems has necessitated a detection system using these techniques. The most critical challenge in using deep learning techniques in diagnosing COVID-19 is accuracy because it plays an essential role in controlling the spread of the disease. Methods This article presents a new framework for detecting COVID-19 using X-ray images. The model uses a modified version of DenseNet-121 for the network layer, an image data loader to separate images in batches, a loss function to reduce the prediction error, and a weighted random sampler to balance the training phase. Finally, an optimizer changes the attributes of the neural networks. Results Extensive experiments using different types of pneumonia expresses satisfactory diagnosis performance with an accuracy of 99.81%. Conclusion This work aims to design a new deep neural network for highly accurate online recognition of medical images. The evaluation results show that the proposed framework can be considered an auxiliary device to help radiologists accurately confirm initial screening.
Collapse
|
47
|
Schaudt D, von Schwerin R, Hafner A, Riedel P, Späte C, Reichert M, Hinteregger A, Beer M, Kloth C. Leveraging human expert image annotations to improve pneumonia differentiation through human knowledge distillation. Sci Rep 2023; 13:9203. [PMID: 37280219 DOI: 10.1038/s41598-023-36148-7] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/20/2022] [Accepted: 05/30/2023] [Indexed: 06/08/2023] Open
Abstract
In medical imaging, deep learning models can be a critical tool to shorten time-to-diagnosis and support specialized medical staff in clinical decision making. The successful training of deep learning models usually requires large amounts of quality data, which are often not available in many medical imaging tasks. In this work we train a deep learning model on university hospital chest X-ray data, containing 1082 images. The data was reviewed, differentiated into 4 causes for pneumonia, and annotated by an expert radiologist. To successfully train a model on this small amount of complex image data, we propose a special knowledge distillation process, which we call Human Knowledge Distillation. This process enables deep learning models to utilize annotated regions in the images during the training process. This form of guidance by a human expert improves model convergence and performance. We evaluate the proposed process on our study data for multiple types of models, all of which show improved results. The best model of this study, called PneuKnowNet, shows an improvement of + 2.3% points in overall accuracy compared to a baseline model and also leads to more meaningful decision regions. Utilizing this implicit data quality-quantity trade-off can be a promising approach for many scarce data domains beyond medical imaging.
Collapse
Affiliation(s)
- Daniel Schaudt
- Department of Computer Science, Ulm University of Applied Science, Albert-Einstein-Allee 55, 89081, Ulm, Baden-Wurttemberg, Germany.
| | - Reinhold von Schwerin
- Department of Computer Science, Ulm University of Applied Science, Albert-Einstein-Allee 55, 89081, Ulm, Baden-Wurttemberg, Germany
| | - Alexander Hafner
- Department of Computer Science, Ulm University of Applied Science, Albert-Einstein-Allee 55, 89081, Ulm, Baden-Wurttemberg, Germany
| | - Pascal Riedel
- Department of Computer Science, Ulm University of Applied Science, Albert-Einstein-Allee 55, 89081, Ulm, Baden-Wurttemberg, Germany
| | - Christian Späte
- Department of Computer Science, Ulm University of Applied Science, Albert-Einstein-Allee 55, 89081, Ulm, Baden-Wurttemberg, Germany
| | - Manfred Reichert
- Institute of Databases and Information Systems, Ulm University, James-Franck-Ring, 89081, Ulm, Baden-Wurttemberg, Germany
| | - Andreas Hinteregger
- Department of Radiology, University Hospital of Ulm, Albert-Einstein-Allee 23, 89081, Ulm, Baden-Wurttemberg, Germany
| | - Meinrad Beer
- Department of Radiology, University Hospital of Ulm, Albert-Einstein-Allee 23, 89081, Ulm, Baden-Wurttemberg, Germany
| | - Christopher Kloth
- Department of Radiology, University Hospital of Ulm, Albert-Einstein-Allee 23, 89081, Ulm, Baden-Wurttemberg, Germany
| |
Collapse
|
48
|
Zhang XY, Wei Q, Wu GG, Tang Q, Pan XF, Chen GQ, Zhang D, Dietrich CF, Cui XW. Artificial intelligence - based ultrasound elastography for disease evaluation - a narrative review. Front Oncol 2023; 13:1197447. [PMID: 37333814 PMCID: PMC10272784 DOI: 10.3389/fonc.2023.1197447] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/31/2023] [Accepted: 05/22/2023] [Indexed: 06/20/2023] Open
Abstract
Ultrasound elastography (USE) provides complementary information of tissue stiffness and elasticity to conventional ultrasound imaging. It is noninvasive and free of radiation, and has become a valuable tool to improve diagnostic performance with conventional ultrasound imaging. However, the diagnostic accuracy will be reduced due to high operator-dependence and intra- and inter-observer variability in visual observations of radiologists. Artificial intelligence (AI) has great potential to perform automatic medical image analysis tasks to provide a more objective, accurate and intelligent diagnosis. More recently, the enhanced diagnostic performance of AI applied to USE have been demonstrated for various disease evaluations. This review provides an overview of the basic concepts of USE and AI techniques for clinical radiologists and then introduces the applications of AI in USE imaging that focus on the following anatomical sites: liver, breast, thyroid and other organs for lesion detection and segmentation, machine learning (ML) - assisted classification and prognosis prediction. In addition, the existing challenges and future trends of AI in USE are also discussed.
Collapse
Affiliation(s)
- Xian-Ya Zhang
- Department of Medical Ultrasound, Tongji Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China
| | - Qi Wei
- Department of Medical Ultrasound, Tongji Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China
| | - Ge-Ge Wu
- Department of Medical Ultrasound, Tongji Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China
| | - Qi Tang
- Department of Ultrasonography, The First Hospital of Changsha, Changsha, China
| | - Xiao-Fang Pan
- Health Medical Department, Dalian Municipal Central Hospital, Dalian, China
| | - Gong-Quan Chen
- Department of Medical Ultrasound, Minda Hospital of Hubei Minzu University, Enshi, China
| | - Di Zhang
- Department of Medical Ultrasound, The First Affiliated Hospital of Anhui Medical University, Hefei, China
| | | | - Xin-Wu Cui
- Department of Medical Ultrasound, Tongji Hospital, Tongji Medical College, Huazhong University of Science and Technology, Wuhan, China
| |
Collapse
|
49
|
Wang T, Nie Z, Wang R, Xu Q, Huang H, Xu H, Xie F, Liu XJ. PneuNet: deep learning for COVID-19 pneumonia diagnosis on chest X-ray image analysis using Vision Transformer. Med Biol Eng Comput 2023; 61:1395-1408. [PMID: 36719562 PMCID: PMC9887581 DOI: 10.1007/s11517-022-02746-2] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/23/2022] [Accepted: 12/22/2022] [Indexed: 02/01/2023]
Abstract
A long-standing challenge in pneumonia diagnosis is recognizing the pathological lung texture, especially the ground-glass appearance pathological texture. One main difficulty lies in precisely extracting and recognizing the pathological features. The patients, especially those with mild symptoms, show very little difference in lung texture, neither conventional computer vision methods nor convolutional neural networks perform well on pneumonia diagnosis based on chest X-ray (CXR) images. In the meanwhile, the Coronavirus Disease 2019 (COVID-19) pandemic continues wreaking havoc around the world, where quick and accurate diagnosis backed by CXR images is in high demand. Rather than simply recognizing the patterns, extracting feature maps from the original CXR image is what we need in the classification process. Thus, we propose a Vision Transformer (VIT)-based model called PneuNet to make an accurate diagnosis backed by channel-based attention through X-ray images of the lung, where multi-head attention is applied on channel patches rather than feature patches. The techniques presented in this paper are oriented toward the medical application of deep neural networks and VIT. Extensive experiment results show that our method can reach 94.96% accuracy in the three-categories classification problem on the test set, which outperforms previous deep learning models.
Collapse
Affiliation(s)
- Tianmu Wang
- Department of Mechanical Engineering, Tsinghua University, Beijing, 100084 China
- State Key Laboratory of Tribology in Advanced Equipment, Tsinghua University, Beijing, 100084 China
- Beijing Key Lab of Precision/Ultra-precision Manufacturing Equipments and Control, Tsinghua University, Beijing, 100084 China
| | - Zhenguo Nie
- Department of Mechanical Engineering, Tsinghua University, Beijing, 100084 China
- State Key Laboratory of Tribology in Advanced Equipment, Tsinghua University, Beijing, 100084 China
- Beijing Key Lab of Precision/Ultra-precision Manufacturing Equipments and Control, Tsinghua University, Beijing, 100084 China
| | - Ruijing Wang
- School of System & Enterprises, Stevens Institute of Technology, Hoboken, NJ 07030 USA
| | - Qingfeng Xu
- Department of Mechanical Engineering, Tsinghua University, Beijing, 100084 China
- National Cancer Center, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, 100060 China
| | - Hongshi Huang
- Institute of Sports Medicine, Peking University Third Hospital, Beijing, 100091 China
| | - Handing Xu
- Department of Mechanical Engineering, Tsinghua University, Beijing, 100084 China
- State Key Laboratory of Tribology in Advanced Equipment, Tsinghua University, Beijing, 100084 China
- Beijing Key Lab of Precision/Ultra-precision Manufacturing Equipments and Control, Tsinghua University, Beijing, 100084 China
| | - Fugui Xie
- Department of Mechanical Engineering, Tsinghua University, Beijing, 100084 China
- State Key Laboratory of Tribology in Advanced Equipment, Tsinghua University, Beijing, 100084 China
- Beijing Key Lab of Precision/Ultra-precision Manufacturing Equipments and Control, Tsinghua University, Beijing, 100084 China
| | - Xin-Jun Liu
- Department of Mechanical Engineering, Tsinghua University, Beijing, 100084 China
- State Key Laboratory of Tribology in Advanced Equipment, Tsinghua University, Beijing, 100084 China
- Beijing Key Lab of Precision/Ultra-precision Manufacturing Equipments and Control, Tsinghua University, Beijing, 100084 China
| |
Collapse
|
50
|
Yuan J, Wu F, Li Y, Li J, Huang G, Huang Q. DPDH-CapNet: A Novel Lightweight Capsule Network with Non-routing for COVID-19 Diagnosis Using X-ray Images. J Digit Imaging 2023; 36:988-1000. [PMID: 36813978 PMCID: PMC9946284 DOI: 10.1007/s10278-023-00791-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/31/2022] [Revised: 01/26/2023] [Accepted: 01/29/2023] [Indexed: 02/24/2023] Open
Abstract
COVID-19 has claimed millions of lives since its outbreak in December 2019, and the damage continues, so it is urgent to develop new technologies to aid its diagnosis. However, the state-of-the-art deep learning methods often rely on large-scale labeled data, limiting their clinical application in COVID-19 identification. Recently, capsule networks have achieved highly competitive performance for COVID-19 detection, but they require expensive routing computation or traditional matrix multiplication to deal with the capsule dimensional entanglement. A more lightweight capsule network is developed to effectively address these problems, namely DPDH-CapNet, which aims to enhance the technology of automated diagnosis for COVID-19 chest X-ray images. It adopts depthwise convolution (D), point convolution (P), and dilated convolution (D) to construct a new feature extractor, thus successfully capturing the local and global dependencies of COVID-19 pathological features. Simultaneously, it constructs the classification layer by homogeneous (H) vector capsules with an adaptive, non-iterative, and non-routing mechanism. We conduct experiments on two publicly available combined datasets, including normal, pneumonia, and COVID-19 images. With a limited number of samples, the parameters of the proposed model are reduced by 9x compared to the state-of-the-art capsule network. Moreover, our model has faster convergence speed and better generalization, and its accuracy, precision, recall, and F-measure are improved to 97.99%, 98.05%, 98.02%, and 98.03%, respectively. In addition, experimental results demonstrate that, contrary to the transfer learning method, the proposed model does not require pre-training and a large number of training samples.
Collapse
Affiliation(s)
- Jianjun Yuan
- College of Artificial Intelligence, Southwest University, Chongqing, 40075, China.
| | - Fujun Wu
- College of Artificial Intelligence, Southwest University, Chongqing, 40075, China
| | - Yuxi Li
- College of Artificial Intelligence, Southwest University, Chongqing, 40075, China
| | - Jinyi Li
- College of Artificial Intelligence, Southwest University, Chongqing, 40075, China
| | - Guojun Huang
- College of Artificial Intelligence, Southwest University, Chongqing, 40075, China
| | - Quanyong Huang
- College of Machinery and Automation, Wuhan University of Science and Technology, Heping Avenue No. 947, Wuhan, Hubei Province, 430091, China.
| |
Collapse
|