1
|
Bombiński P, Szatkowski P, Sobieski B, Kwieciński T, Płotka S, Adamek M, Banasiuk M, Furmanek MI, Biecek P. Underestimation of lung regions on chest X-ray segmentation masks assessed by comparison with total lung volume evaluated on computed tomography. Radiography (Lond) 2025; 31:102930. [PMID: 40174327 DOI: 10.1016/j.radi.2025.102930] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/14/2024] [Revised: 03/03/2025] [Accepted: 03/14/2025] [Indexed: 04/04/2025]
Abstract
INTRODUCTION The lung regions on chest X-ray segmentation masks created according to the current gold standard method for AI-driven applications are underestimated. This can be evaluated by comparison with computed tomography. METHODS This retrospective study included data from non-contrast chest low-dose CT examinations of 55 individuals without pulmonary pathology. Synthetic X-ray images were generated by projecting a 3D CT examination onto a 2D image plane. Two experienced radiologists manually created two types of lung masks: 3D lung masks from CT examinations (ground truth for further calculations) and 2D lung masks from synthetic X-ray images (according to the current gold standard method: following the contours of other anatomical structures). Overlapping and non-overlapping lung regions covered by both types of masks were analyzed. Volume of the overlapping regions was compared with total lung volume, and volume fractions of non-overlapping lung regions in relation to the total lung volume were calculated. The performance results between the two radiologists were compared. RESULTS Significant differences were observed between lung regions covered by CT and synthetic X-ray masks. The mean volume fractions of the lung regions not covered by synthetic X-ray masks for the right lung, the left lung, and both lungs were 22.8 %, 32.9 %, and 27.3 %, respectively, for Radiologist 1 and 22.7 %, 32.9 %, and 27.3 %, respectively, for Radiologist 2. There was excellent spatial agreement between the masks created by the two radiologists. CONCLUSIONS Lung X-ray masks created according to the current gold standard method significantly underestimate lung regions and do not cover substantial portions of the lungs. IMPLICATIONS FOR PRACTICE Standard lung masks fail to encompass the whole range of the lungs and significantly restrict the field of analysis in AI-driven applications, which may lead to false conclusions and diagnoses.
Collapse
Affiliation(s)
- P Bombiński
- Department of Pediatric Radiology, Medical University of Warsaw, Pediatric Clinical Hospital, 63A Żwirki i Wigury St, Warsaw 02-091, Poland; upmedic, 2/11 Sądowa St, 20-027 Lublin, Poland.
| | - P Szatkowski
- 2nd Department of Clinical Radiology, Medical University of Warsaw, Central Clinical Hospital, 1A Banacha St, Warsaw 02-097, Poland.
| | - B Sobieski
- Faculty of Mathematics and Information Science, Warsaw University of Technology, 75 Koszykowa St, Warsaw 00-661, Poland; MI2.ai, Warsaw University of Technology, 75 Koszykowa St, Warsaw 00-661, Poland.
| | - T Kwieciński
- Faculty of Mathematics and Information Science, Warsaw University of Technology, 75 Koszykowa St, Warsaw 00-661, Poland; MI2.ai, Warsaw University of Technology, 75 Koszykowa St, Warsaw 00-661, Poland.
| | - S Płotka
- Faculty of Mathematics and Information Science, Warsaw University of Technology, 75 Koszykowa St, Warsaw 00-661, Poland; MI2.ai, Warsaw University of Technology, 75 Koszykowa St, Warsaw 00-661, Poland; Informatics Institute, University of Amsterdam, Science Park 900, 1098 XH Amsterdam, the Netherlands; Department of Biomedical Engineering and Physics, Amsterdam University Medical Center, Meibergdreef 9, 1105 AZ Amsterdam, the Netherlands.
| | - M Adamek
- Department of Thoracic Surgery, Medical University of Silesia, 35 Ceglana St, 40-514 Katowice, Poland; Department of Thoracic Surgery, Medical University of Gdańsk, 17 Smoluchowskiego St, 80-214 Gdańsk, Poland.
| | - M Banasiuk
- Department of Pediatric Gastroenterology and Nutrition, Medical University of Warsaw, Pediatric Clinical Hospital, 63A Żwirki i Wigury, Warsaw 02-091, Poland.
| | - M I Furmanek
- Department of Pediatric Radiology, Medical University of Warsaw, Pediatric Clinical Hospital, 63A Żwirki i Wigury St, Warsaw 02-091, Poland.
| | - P Biecek
- Faculty of Mathematics and Information Science, Warsaw University of Technology, 75 Koszykowa St, Warsaw 00-661, Poland; MI2.ai, Warsaw University of Technology, 75 Koszykowa St, Warsaw 00-661, Poland; Faculty of Mathematics, Informatics, and Mechanics, University of Warsaw, 1A Banacha St, Warsaw 02-097, Poland.
| |
Collapse
|
2
|
Alaoui Abdalaoui Slimani F, Bentourkia M. Improving deep learning U-Net++ by discrete wavelet and attention gate mechanisms for effective pathological lung segmentation in chest X-ray imaging. Phys Eng Sci Med 2025; 48:59-73. [PMID: 39495449 DOI: 10.1007/s13246-024-01489-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/16/2024] [Accepted: 10/09/2024] [Indexed: 11/05/2024]
Abstract
Since its introduction in 2015, the U-Net architecture used in Deep Learning has played a crucial role in medical imaging. Recognized for its ability to accurately discriminate small structures, the U-Net has received more than 2600 citations in academic literature, which motivated continuous enhancements to its architecture. In hospitals, chest radiography is the primary diagnostic method for pulmonary disorders, however, accurate lung segmentation in chest X-ray images remains a challenging task, primarily due to the significant variations in lung shapes and the presence of intense opacities caused by various diseases. This article introduces a new approach for the segmentation of lung X-ray images. Traditional max-pooling operations, commonly employed in conventional U-Net++ models, were replaced with the discrete wavelet transform (DWT), offering a more accurate down-sampling technique that potentially captures detailed features of lung structures. Additionally, we used attention gate (AG) mechanisms that enable the model to focus on specific regions in the input image, which improves the accuracy of the segmentation process. When compared with current techniques like Atrous Convolutions, Improved FCN, Improved SegNet, U-Net, and U-Net++, our method (U-Net++-DWT) showed remarkable efficacy, particularly on the Japanese Society of Radiological Technology dataset, achieving an accuracy of 99.1%, specificity of 98.9%, sensitivity of 97.8%, Dice Coefficient of 97.2%, and Jaccard Index of 96.3%. Its performance on the Montgomery County dataset further demonstrated its consistent effectiveness. Moreover, when applied to additional datasets of Chest X-ray Masks and Labels and COVID-19, our method maintained high performance levels, achieving up to 99.3% accuracy, thereby underscoring its adaptability and potential for broad applications in medical imaging diagnostics.
Collapse
Affiliation(s)
| | - M'hamed Bentourkia
- Department of Nuclear Medicine and Radiobiology, 12th Avenue North, 3001, Sherbrooke, QC, J1H5N4, Canada.
| |
Collapse
|
3
|
Okumu A, Orwa J, Sitati R, Omondi I, Odhiambo B, Ogoro J, Oballa G, Ochieng B, Wandiga S, Ouma C. Factors associated with tuberculosis drug resistance among presumptive multidrug resistance tuberculosis patients identified in a DRTB surveillance study in western Kenya. J Clin Tuberc Other Mycobact Dis 2024; 37:100466. [PMID: 39188352 PMCID: PMC11345928 DOI: 10.1016/j.jctube.2024.100466] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/28/2024] Open
Abstract
Multidrug-resistant tuberculosis (MDR-TB) is caused by M. tuberculosis (Mtb) with resistance to the first-line anti-TB medicines isoniazid (INH) and rifampicin (RIF). In Western Kenya, there is reported low prevalence of drug resistant strains among HIV tuberculosis patients, creating a need to determine factors associated with drug resistance patterns among presumptive MDR-TB patients. To determine factors associated with drug resistance patterns among presumptive MDR-TB patients in western Kenya. Three hundred and ninety (3 9 0) sputum sample isolates from among presumptive multidrug TB patients, were analyzed for TB drug resistance as per Ministry of Health (MoH) TB program diagnostic algorithm. Frequency and percentages were used to summarize categorical data while median and interquartile range (IQR) were used for continuous data. Multivariable logistic regression was carried out to identify factors associated with TB drug resistance. Out of 390 participants enrolled, 302/390 (77.4 %) were males, with a median age of 34 years. The HIV-infected were 118/390 (30.3 %). Samples included 322 (82.6 %) from presumptive patients, while 68/390 (17.4 %) were either lost to follow-up patients, failures to first-line treatment or newly diagnosed cases. A total of 64/390 (16.4 %) of the isolates had at least some form of drug resistance. Out of 390, 14/390 (3.6 %) had MDR, 12 (3.1 %) were RIF mono-resistance, 34 (8.7 %) had INH, while 4 (1 %) had ethambutol resistance. The category of previously treated patients (those who received or are currently on TB treatment) had a 70 % reduced likelihood of resistance (aOR: 0.30; 95 % CI: 0.13-0.70). In contrast, older age was associated with an increased likelihood of resistance to INH and RIF, with an adjusted odds ratio of 1.04 per year (95 % CI: 1.00-1.08). Prompt MDR-TB diagnosis is essential for appropriate patient care, management, and disease prevention and control. We recommend active surveillance on drug resistant TB in these regions to detect drug resistance patterns for rapid disease management.
Collapse
Affiliation(s)
- Albert Okumu
- Kenya Medical Research Institute, Centre for Global Health Research, PO Box 1578 -40100, Kisumu, Kenya
- Department of Biomedical Sciences and Technology, Maseno University, PO Box 333-40105, Maseno, Kenya
| | - James Orwa
- The Aga Khan University, Department of Population Health Science, University Center, PO BOX 30270- 00100, Nairobi, Kenya
| | - Ruth Sitati
- Kenya Medical Research Institute, Centre for Global Health Research, PO Box 1578 -40100, Kisumu, Kenya
| | - Isaiah Omondi
- Kenya Medical Research Institute, Centre for Global Health Research, PO Box 1578 -40100, Kisumu, Kenya
| | - Ben Odhiambo
- Kenya Medical Research Institute, Centre for Global Health Research, PO Box 1578 -40100, Kisumu, Kenya
| | - Jeremiah Ogoro
- Ministry of Health, National Leprosy and Tuberculosis Program, NLTP, Afya House Annex, Kenyatta National Hospital, Hospital Road, Nairobi P.O. Box, 30016-00100, Kenya
| | - George Oballa
- Ministry of Health, National Leprosy and Tuberculosis Program, NLTP, Afya House Annex, Kenyatta National Hospital, Hospital Road, Nairobi P.O. Box, 30016-00100, Kenya
| | - Benjamin Ochieng
- Kenya Medical Research Institute, Centre for Global Health Research, PO Box 1578 -40100, Kisumu, Kenya
| | - Steve Wandiga
- Kenya Medical Research Institute, Centre for Global Health Research, PO Box 1578 -40100, Kisumu, Kenya
| | - Collins Ouma
- Department of Biomedical Sciences and Technology, Maseno University, PO Box 333-40105, Maseno, Kenya
| |
Collapse
|
4
|
Xu Y, Quan R, Xu W, Huang Y, Chen X, Liu F. Advances in Medical Image Segmentation: A Comprehensive Review of Traditional, Deep Learning and Hybrid Approaches. Bioengineering (Basel) 2024; 11:1034. [PMID: 39451409 PMCID: PMC11505408 DOI: 10.3390/bioengineering11101034] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/23/2024] [Revised: 10/08/2024] [Accepted: 10/11/2024] [Indexed: 10/26/2024] Open
Abstract
Medical image segmentation plays a critical role in accurate diagnosis and treatment planning, enabling precise analysis across a wide range of clinical tasks. This review begins by offering a comprehensive overview of traditional segmentation techniques, including thresholding, edge-based methods, region-based approaches, clustering, and graph-based segmentation. While these methods are computationally efficient and interpretable, they often face significant challenges when applied to complex, noisy, or variable medical images. The central focus of this review is the transformative impact of deep learning on medical image segmentation. We delve into prominent deep learning architectures such as Convolutional Neural Networks (CNNs), Fully Convolutional Networks (FCNs), U-Net, Recurrent Neural Networks (RNNs), Adversarial Networks (GANs), and Autoencoders (AEs). Each architecture is analyzed in terms of its structural foundation and specific application to medical image segmentation, illustrating how these models have enhanced segmentation accuracy across various clinical contexts. Finally, the review examines the integration of deep learning with traditional segmentation methods, addressing the limitations of both approaches. These hybrid strategies offer improved segmentation performance, particularly in challenging scenarios involving weak edges, noise, or inconsistent intensities. By synthesizing recent advancements, this review provides a detailed resource for researchers and practitioners, offering valuable insights into the current landscape and future directions of medical image segmentation.
Collapse
Affiliation(s)
- Yan Xu
- School of Electrical, Electronic and Mechanical Engineering, University of Bristol, Bristol BS8 1QU, UK; (Y.X.); (R.Q.); (W.X.)
| | - Rixiang Quan
- School of Electrical, Electronic and Mechanical Engineering, University of Bristol, Bristol BS8 1QU, UK; (Y.X.); (R.Q.); (W.X.)
| | - Weiting Xu
- School of Electrical, Electronic and Mechanical Engineering, University of Bristol, Bristol BS8 1QU, UK; (Y.X.); (R.Q.); (W.X.)
| | - Yi Huang
- Bristol Medical School, University of Bristol, Bristol BS8 1UD, UK;
| | - Xiaolong Chen
- Department of Mechanical, Materials and Manufacturing Engineering, University of Nottingham, Nottingham NG7 2RD, UK;
| | - Fengyuan Liu
- School of Electrical, Electronic and Mechanical Engineering, University of Bristol, Bristol BS8 1QU, UK; (Y.X.); (R.Q.); (W.X.)
| |
Collapse
|
5
|
Zhen A, Wang X. The deep learning-based physical education course recommendation system under the internet of things. Heliyon 2024; 10:e38907. [PMID: 39435083 PMCID: PMC11492338 DOI: 10.1016/j.heliyon.2024.e38907] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/17/2024] [Revised: 08/30/2024] [Accepted: 10/02/2024] [Indexed: 10/23/2024] Open
Abstract
This study aims to propose a deep learning (DL)-based physical education course recommendation system by combining the Internet of Things (IoT) technology and DL, to improve the accuracy and personalization of recommendation. Firstly, IoT devices such as smart bracelets and smart clothing are used to monitor students' physiological data in real-time, and IoT sensors are utilized to sense the environment around students. Secondly, IoT devices capture students' social interactions with their peers, recommending socially oriented courses. Meanwhile, by integrating IoT data with students' academic data, course recommendations are optimized to match students' learning progress and schedule. Finally, Generative Adversarial Network (GAN) models, especially the improved Regularization Penalty Conditional Feature Generative Adversarial Network (RP-CFGAN) model, deal with data sparsity and cold start problems. The experimental results show that this model performs well in TopN evaluation and is markedly enhanced compared with traditional models. This study denotes that integrating IoT technology and GAN models can more accurately understand student needs and provide personalized recommendations. Although the model performs well, there is still room for improvement, such as exploring more regularization techniques, protecting user privacy, and extending the system to diverse platforms and scenarios.
Collapse
Affiliation(s)
- Aiyuan Zhen
- School of Physical Education, Shanghai Normal University, Shanghai, 200234, China
| | - Xin Wang
- Faculty of Physical Education, China West Normal University, Nanchong, 637000, China
| |
Collapse
|
6
|
Cai J, Zhu H, Liu S, Qi Y, Chen R. Lung image segmentation via generative adversarial networks. Front Physiol 2024; 15:1408832. [PMID: 39219839 PMCID: PMC11365075 DOI: 10.3389/fphys.2024.1408832] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/28/2024] [Accepted: 08/01/2024] [Indexed: 09/04/2024] Open
Abstract
Introduction Lung image segmentation plays an important role in computer-aid pulmonary disease diagnosis and treatment. Methods This paper explores the lung CT image segmentation method by generative adversarial networks. We employ a variety of generative adversarial networks and used their capability of image translation to perform image segmentation. The generative adversarial network is employed to translate the original lung image into the segmented image. Results The generative adversarial networks-based segmentation method is tested on real lung image data set. Experimental results show that the proposed method outperforms the state-of-the-art method. Discussion The generative adversarial networks-based method is effective for lung image segmentation.
Collapse
Affiliation(s)
- Jiaxin Cai
- School of Mathematics and Statistics, Xiamen University of Technology, Xiamen, China
| | - Hongfeng Zhu
- School of Mathematics and Statistics, Xiamen University of Technology, Xiamen, China
| | - Siyu Liu
- School of Computer and Information Engineering, Xiamen University of Technology, Xiamen, China
| | - Yang Qi
- School of Computer and Information Engineering, Xiamen University of Technology, Xiamen, China
| | - Rongshang Chen
- School of Computer and Information Engineering, Xiamen University of Technology, Xiamen, China
| |
Collapse
|
7
|
Rendon-Atehortua JC, Cardenas-Pena D, Daza-Santacoloma G, Orozco-Gutierrez AA, Jaramillo-Robledo O. Efficient Lung Segmentation from Chest Radiographs using Transfer Learning and Lightweight Deep Architecture. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2024; 2024:1-5. [PMID: 40039676 DOI: 10.1109/embc53108.2024.10782198] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/06/2025]
Abstract
Lung delineation constitutes a critical preprocessing stage for X-ray-based diagnosis and follow-up. However, automatic lung segmentation from chest radiographs (CXR) poses a challenging problem due to anatomical structures' varying shapes and sizes, the differences between radio-opacity, contrast, and image quality, and the requirement of complex models for automatic detection of regions of interest. This work proposes the automated lung segmentation methodology DenseCX, based on U-Net architectures and transfer learning techniques. Unlike other U-Net networks, DenseCX includes an encoder built from Dense blocks, promoting a meaningful feature extraction with lightweight layers. Then, a homogeneous domain adaptation transfers the knowledge from classifying a large cohort of CXR to the DenseCX, reducing the overfitting risk due to the lack of manually labeled images. The experimental setup evaluates the proposed methodology on three public datasets, namely Shenzhen Hospital Chest X-ray, the Japan Society of Radiological Technology, and Montgomery County Chest X-ray, in a leave-one-group-out validation strategy for warranting the generalization. The attained Dice, Sensitivity, and Specificity metrics evidence that DenseCX outperforms other conventional ImageNet initialization while providing the best trade-off between performance and model complexity than state-of-the-art approaches, with a much lighter architecture and an improved convergence.
Collapse
|
8
|
Gopatoti A, Jayakumar R, Billa P, Patteeswaran V. DDA-SSNets: Dual decoder attention-based semantic segmentation networks for COVID-19 infection segmentation and classification using chest X-Ray images. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2024; 32:623-649. [PMID: 38607728 DOI: 10.3233/xst-230421] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/14/2024]
Abstract
BACKGROUND COVID-19 needs to be diagnosed and staged to be treated accurately. However, prior studies' diagnostic and staging abilities for COVID-19 infection needed to be improved. Therefore, new deep learning-based approaches are required to aid radiologists in detecting and quantifying COVID-19-related lung infections. OBJECTIVE To develop deep learning-based models to classify and quantify COVID-19-related lung infections. METHODS Initially, Dual Decoder Attention-based Semantic Segmentation Networks (DDA-SSNets) such as Dual Decoder Attention-UNet (DDA-UNet) and Dual Decoder Attention-SegNet (DDA-SegNet) are proposed to facilitate the dual segmentation tasks such as lung lobes and infection segmentation in chest X-ray (CXR) images. The lung lobe and infection segmentations are mapped to grade the severity of COVID-19 infection in both the lungs of CXRs. Later, a Genetic algorithm-based Deep Convolutional Neural Network classifier with the optimum number of layers, namely GADCNet, is proposed to classify the extracted regions of interest (ROI) from the CXR lung lobes into COVID-19 and non-COVID-19. RESULTS The DDA-SegNet shows better segmentation with an average BCSSDC of 99.53% and 99.97% for lung lobes and infection segmentations, respectively, compared with DDA-UNet with an average BCSSDC of 99.14% and 99.92%. The proposed DDA-SegNet with GADCNet classifier offered excellent classification results with an average BCCAC of 99.98%, followed by the GADCNet with DDA-UNet with an average BCCAC of 99.92% after extensive testing and analysis. CONCLUSIONS The results show that the proposed DDA-SegNet has superior performance in the segmentation of lung lobes and COVID-19-infected regions in CXRs, along with improved severity grading compared to the DDA-UNet and improved accuracy of the GADCNet classifier in classifying the CXRs into COVID-19, and non-COVID-19.
Collapse
Affiliation(s)
- Anandbabu Gopatoti
- Department of Electronics and Communication Engineering, Hindusthan College of Engineering and Technology, Coimbatore, Tamil Nadu, India
| | - Ramya Jayakumar
- Department of Electronics and Communication Engineering, Hindusthan College of Engineering and Technology, Coimbatore, Tamil Nadu, India
| | - Poornaiah Billa
- Department of Electronics and Communication Engineering, Lakireddy Bali Reddy College of Engineering, Mylavaram, Andhra Pradesh, India
| | - Vijayalakshmi Patteeswaran
- Department of Electronics and Communication Engineering, Hindusthan College of Engineering and Technology, Coimbatore, Tamil Nadu, India
| |
Collapse
|
9
|
Horry MJ, Chakraborty S, Pradhan B, Paul M, Zhu J, Loh HW, Barua PD, Acharya UR. Development of Debiasing Technique for Lung Nodule Chest X-ray Datasets to Generalize Deep Learning Models. SENSORS (BASEL, SWITZERLAND) 2023; 23:6585. [PMID: 37514877 PMCID: PMC10385599 DOI: 10.3390/s23146585] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/14/2023] [Revised: 07/16/2023] [Accepted: 07/20/2023] [Indexed: 07/30/2023]
Abstract
Screening programs for early lung cancer diagnosis are uncommon, primarily due to the challenge of reaching at-risk patients located in rural areas far from medical facilities. To overcome this obstacle, a comprehensive approach is needed that combines mobility, low cost, speed, accuracy, and privacy. One potential solution lies in combining the chest X-ray imaging mode with federated deep learning, ensuring that no single data source can bias the model adversely. This study presents a pre-processing pipeline designed to debias chest X-ray images, thereby enhancing internal classification and external generalization. The pipeline employs a pruning mechanism to train a deep learning model for nodule detection, utilizing the most informative images from a publicly available lung nodule X-ray dataset. Histogram equalization is used to remove systematic differences in image brightness and contrast. Model training is then performed using combinations of lung field segmentation, close cropping, and rib/bone suppression. The resulting deep learning models, generated through this pre-processing pipeline, demonstrate successful generalization on an independent lung nodule dataset. By eliminating confounding variables in chest X-ray images and suppressing signal noise from the bone structures, the proposed deep learning lung nodule detection algorithm achieves an external generalization accuracy of 89%. This approach paves the way for the development of a low-cost and accessible deep learning-based clinical system for lung cancer screening.
Collapse
Affiliation(s)
- Michael J Horry
- Centre for Advanced Modelling and Geospatial Information Systems (CAMGIS), Faculty of Engineering and Information Technology, University of Technology Sydney, Ultimo, NSW 2007, Australia
- IBM Australia Limited, Sydney, NSW 2000, Australia
| | - Subrata Chakraborty
- Centre for Advanced Modelling and Geospatial Information Systems (CAMGIS), Faculty of Engineering and Information Technology, University of Technology Sydney, Ultimo, NSW 2007, Australia
- Faculty of Science, Agriculture, Business and Law, University of New England, Armidale, NSW 2351, Australia
| | - Biswajeet Pradhan
- Centre for Advanced Modelling and Geospatial Information Systems (CAMGIS), Faculty of Engineering and Information Technology, University of Technology Sydney, Ultimo, NSW 2007, Australia
- Earth Observation Center, Institute of Climate Change, Universiti Kebangsaan Malaysia, Bangi 43600, Malaysia
| | - Manoranjan Paul
- Machine Vision and Digital Health (MaViDH), School of Computing and Mathematics, Charles Sturt University, Bathurst, NSW 2795, Australia
| | - Jing Zhu
- Department of Radiology, Westmead Hospital, Westmead, NSW 2145, Australia
| | - Hui Wen Loh
- School of Science and Technology, Singapore University of Social Sciences, Singapore 599494, Singapore
| | - Prabal Datta Barua
- Centre for Advanced Modelling and Geospatial Information Systems (CAMGIS), Faculty of Engineering and Information Technology, University of Technology Sydney, Ultimo, NSW 2007, Australia
- Faculty of Science, Agriculture, Business and Law, University of New England, Armidale, NSW 2351, Australia
- Cogninet Brain Team, Cogninet Australia, Sydney, NSW 2010, Australia
- School of Business (Information Systems), Faculty of Business, Education, Law & Arts, University of Southern Queensland, Toowoomba, QLD 4350, Australia
| | - U Rajendra Acharya
- School of Mathematics, Physics and Computing, University of Southern Queensland, Springfield, QLD 4300, Australia
| |
Collapse
|
10
|
Thirumagal E, Saruladha K. Lung cancer diagnosis using Hessian adaptive learning optimization in generative adversarial networks. Soft comput 2023. [DOI: 10.1007/s00500-023-07877-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/08/2023]
|
11
|
Osuala R, Kushibar K, Garrucho L, Linardos A, Szafranowska Z, Klein S, Glocker B, Diaz O, Lekadir K. Data synthesis and adversarial networks: A review and meta-analysis in cancer imaging. Med Image Anal 2023; 84:102704. [PMID: 36473414 DOI: 10.1016/j.media.2022.102704] [Citation(s) in RCA: 17] [Impact Index Per Article: 8.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2021] [Revised: 11/02/2022] [Accepted: 11/21/2022] [Indexed: 11/26/2022]
Abstract
Despite technological and medical advances, the detection, interpretation, and treatment of cancer based on imaging data continue to pose significant challenges. These include inter-observer variability, class imbalance, dataset shifts, inter- and intra-tumour heterogeneity, malignancy determination, and treatment effect uncertainty. Given the recent advancements in image synthesis, Generative Adversarial Networks (GANs), and adversarial training, we assess the potential of these technologies to address a number of key challenges of cancer imaging. We categorise these challenges into (a) data scarcity and imbalance, (b) data access and privacy, (c) data annotation and segmentation, (d) cancer detection and diagnosis, and (e) tumour profiling, treatment planning and monitoring. Based on our analysis of 164 publications that apply adversarial training techniques in the context of cancer imaging, we highlight multiple underexplored solutions with research potential. We further contribute the Synthesis Study Trustworthiness Test (SynTRUST), a meta-analysis framework for assessing the validation rigour of medical image synthesis studies. SynTRUST is based on 26 concrete measures of thoroughness, reproducibility, usefulness, scalability, and tenability. Based on SynTRUST, we analyse 16 of the most promising cancer imaging challenge solutions and observe a high validation rigour in general, but also several desirable improvements. With this work, we strive to bridge the gap between the needs of the clinical cancer imaging community and the current and prospective research on data synthesis and adversarial networks in the artificial intelligence community.
Collapse
Affiliation(s)
- Richard Osuala
- Artificial Intelligence in Medicine Lab (BCN-AIM), Facultat de Matemàtiques i Informàtica, Universitat de Barcelona, Spain.
| | - Kaisar Kushibar
- Artificial Intelligence in Medicine Lab (BCN-AIM), Facultat de Matemàtiques i Informàtica, Universitat de Barcelona, Spain
| | - Lidia Garrucho
- Artificial Intelligence in Medicine Lab (BCN-AIM), Facultat de Matemàtiques i Informàtica, Universitat de Barcelona, Spain
| | - Akis Linardos
- Artificial Intelligence in Medicine Lab (BCN-AIM), Facultat de Matemàtiques i Informàtica, Universitat de Barcelona, Spain
| | - Zuzanna Szafranowska
- Artificial Intelligence in Medicine Lab (BCN-AIM), Facultat de Matemàtiques i Informàtica, Universitat de Barcelona, Spain
| | - Stefan Klein
- Biomedical Imaging Group Rotterdam, Department of Radiology & Nuclear Medicine, Erasmus MC, Rotterdam, The Netherlands
| | - Ben Glocker
- Biomedical Image Analysis Group, Department of Computing, Imperial College London, UK
| | - Oliver Diaz
- Artificial Intelligence in Medicine Lab (BCN-AIM), Facultat de Matemàtiques i Informàtica, Universitat de Barcelona, Spain
| | - Karim Lekadir
- Artificial Intelligence in Medicine Lab (BCN-AIM), Facultat de Matemàtiques i Informàtica, Universitat de Barcelona, Spain
| |
Collapse
|
12
|
Segmentation-Based Classification Deep Learning Model Embedded with Explainable AI for COVID-19 Detection in Chest X-ray Scans. Diagnostics (Basel) 2022; 12:diagnostics12092132. [PMID: 36140533 PMCID: PMC9497601 DOI: 10.3390/diagnostics12092132] [Citation(s) in RCA: 21] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/20/2022] [Revised: 08/26/2022] [Accepted: 08/30/2022] [Indexed: 12/16/2022] Open
Abstract
Background and Motivation: COVID-19 has resulted in a massive loss of life during the last two years. The current imaging-based diagnostic methods for COVID-19 detection in multiclass pneumonia-type chest X-rays are not so successful in clinical practice due to high error rates. Our hypothesis states that if we can have a segmentation-based classification error rate <5%, typically adopted for 510 (K) regulatory purposes, the diagnostic system can be adapted in clinical settings. Method: This study proposes 16 types of segmentation-based classification deep learning-based systems for automatic, rapid, and precise detection of COVID-19. The two deep learning-based segmentation networks, namely UNet and UNet+, along with eight classification models, namely VGG16, VGG19, Xception, InceptionV3, Densenet201, NASNetMobile, Resnet50, and MobileNet, were applied to select the best-suited combination of networks. Using the cross-entropy loss function, the system performance was evaluated by Dice, Jaccard, area-under-the-curve (AUC), and receiver operating characteristics (ROC) and validated using Grad-CAM in explainable AI framework. Results: The best performing segmentation model was UNet, which exhibited the accuracy, loss, Dice, Jaccard, and AUC of 96.35%, 0.15%, 94.88%, 90.38%, and 0.99 (p-value <0.0001), respectively. The best performing segmentation-based classification model was UNet+Xception, which exhibited the accuracy, precision, recall, F1-score, and AUC of 97.45%, 97.46%, 97.45%, 97.43%, and 0.998 (p-value <0.0001), respectively. Our system outperformed existing methods for segmentation-based classification models. The mean improvement of the UNet+Xception system over all the remaining studies was 8.27%. Conclusion: The segmentation-based classification is a viable option as the hypothesis (error rate <5%) holds true and is thus adaptable in clinical practice.
Collapse
|
13
|
|
14
|
Zhang Z, Li Y, Shin BS. C 2 -GAN: Content-consistent generative adversarial networks for unsupervised domain adaptation in medical image segmentation. Med Phys 2022; 49:6491-6504. [PMID: 35981348 DOI: 10.1002/mp.15944] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2022] [Revised: 07/19/2022] [Accepted: 07/31/2022] [Indexed: 11/05/2022] Open
Abstract
PURPOSE In clinical practice, medical image analysis has played a key role in disease diagnosis. One of the important steps is to perform an accurate organ or tissue segmentation for assisting medical professionals in making correct diagnoses. Despite the tremendous progress in the deep learning-based medical image segmentation approaches, they often fail to generalize to test datasets due to distribution discrepancies across domains. Recent advances aligning the domain gaps by using bi-directional GANs (e.g., CycleGAN) have shown promising results, but the strict constraints of the cycle consistency hamper these methods from yielding better performance. The purpose of this study is to propose a novel bi-directional GAN-based segmentation model with fewer constraints on the cycle consistency to improve the generalized segmentation results. METHODS We propose a novel unsupervised domain adaptation approach by designing content-consistent generative adversarial networks (C2 -GAN) for medical image segmentation. Firstly, we introduce content consistency instead of cycle consistency to relax the constraint of the invertibility map to encourage the synthetic domain generated with a large domain transportation distance. The synthetic domain is thus pulled close to the target domain for the reduction of domain discrepancy. Secondly, we suggest a novel style transfer loss based on the difference in low-frequency magnitude to further mitigate the appearance shifts across domains. RESULTS We validate our proposed approach on three public X-ray datasets including the Montgomery, JSRT, and Shenzhen datasets. For an accurate evaluation, we randomly divided the images of each dataset into 70% for training, 10% for evaluation, and 20% for testing. The mean Dice was 95.73 ± 0.22%, 95.16 ± 1.42% for JSRT and Shenzhen datasets, respectively. For the recall and precision metrics, our model also achieved better or comparable performance than the state-of-the-art CycleGAN-based UDA approaches. CONCLUSIONS The experimental results validate the effectiveness of our method in mitigating the domain gaps and improving generalized segmentation results for X-ray image segmentation. This article is protected by copyright. All rights reserved.
Collapse
Affiliation(s)
- Zuyu Zhang
- Department of Electrical and Computer Engineering, Inha University, Incheon, 22212, Korea
| | - Yan Li
- Department of Electrical and Computer Engineering, Inha University, Incheon, 22212, Korea
| | - Byeong-Seok Shin
- Department of Electrical and Computer Engineering, Inha University, Incheon, 22212, Korea
| |
Collapse
|
15
|
Ali H, Shah Z. Combating COVID-19 Using Generative Adversarial Networks and Artificial Intelligence for Medical Images: Scoping Review. JMIR Med Inform 2022; 10:e37365. [PMID: 35709336 PMCID: PMC9246088 DOI: 10.2196/37365] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/17/2022] [Revised: 03/06/2022] [Accepted: 03/11/2022] [Indexed: 12/02/2022] Open
Abstract
BACKGROUND Research on the diagnosis of COVID-19 using lung images is limited by the scarcity of imaging data. Generative adversarial networks (GANs) are popular for synthesis and data augmentation. GANs have been explored for data augmentation to enhance the performance of artificial intelligence (AI) methods for the diagnosis of COVID-19 within lung computed tomography (CT) and X-ray images. However, the role of GANs in overcoming data scarcity for COVID-19 is not well understood. OBJECTIVE This review presents a comprehensive study on the role of GANs in addressing the challenges related to COVID-19 data scarcity and diagnosis. It is the first review that summarizes different GAN methods and lung imaging data sets for COVID-19. It attempts to answer the questions related to applications of GANs, popular GAN architectures, frequently used image modalities, and the availability of source code. METHODS A search was conducted on 5 databases, namely PubMed, IEEEXplore, Association for Computing Machinery (ACM) Digital Library, Scopus, and Google Scholar. The search was conducted from October 11-13, 2021. The search was conducted using intervention keywords, such as "generative adversarial networks" and "GANs," and application keywords, such as "COVID-19" and "coronavirus." The review was performed following the Preferred Reporting Items for Systematic Reviews and Meta-Analyses Extension for Scoping Reviews (PRISMA-ScR) guidelines for systematic and scoping reviews. Only those studies were included that reported GAN-based methods for analyzing chest X-ray images, chest CT images, and chest ultrasound images. Any studies that used deep learning methods but did not use GANs were excluded. No restrictions were imposed on the country of publication, study design, or outcomes. Only those studies that were in English and were published from 2020 to 2022 were included. No studies before 2020 were included. RESULTS This review included 57 full-text studies that reported the use of GANs for different applications in COVID-19 lung imaging data. Most of the studies (n=42, 74%) used GANs for data augmentation to enhance the performance of AI techniques for COVID-19 diagnosis. Other popular applications of GANs were segmentation of lungs and superresolution of lung images. The cycleGAN and the conditional GAN were the most commonly used architectures, used in 9 studies each. In addition, 29 (51%) studies used chest X-ray images, while 21 (37%) studies used CT images for the training of GANs. For the majority of the studies (n=47, 82%), the experiments were conducted and results were reported using publicly available data. A secondary evaluation of the results by radiologists/clinicians was reported by only 2 (4%) studies. CONCLUSIONS Studies have shown that GANs have great potential to address the data scarcity challenge for lung images in COVID-19. Data synthesized with GANs have been helpful to improve the training of the convolutional neural network (CNN) models trained for the diagnosis of COVID-19. In addition, GANs have also contributed to enhancing the CNNs' performance through the superresolution of the images and segmentation. This review also identified key limitations of the potential transformation of GAN-based methods in clinical applications.
Collapse
Affiliation(s)
- Hazrat Ali
- College of Science and Engineering, Hamad Bin Khalifa University, Doha, Qatar
| | - Zubair Shah
- College of Science and Engineering, Hamad Bin Khalifa University, Doha, Qatar
| |
Collapse
|
16
|
You A, Kim JK, Ryu IH, Yoo TK. Application of generative adversarial networks (GAN) for ophthalmology image domains: a survey. EYE AND VISION (LONDON, ENGLAND) 2022; 9:6. [PMID: 35109930 PMCID: PMC8808986 DOI: 10.1186/s40662-022-00277-3] [Citation(s) in RCA: 57] [Impact Index Per Article: 19.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/15/2021] [Accepted: 01/11/2022] [Indexed: 12/12/2022]
Abstract
BACKGROUND Recent advances in deep learning techniques have led to improved diagnostic abilities in ophthalmology. A generative adversarial network (GAN), which consists of two competing types of deep neural networks, including a generator and a discriminator, has demonstrated remarkable performance in image synthesis and image-to-image translation. The adoption of GAN for medical imaging is increasing for image generation and translation, but it is not familiar to researchers in the field of ophthalmology. In this work, we present a literature review on the application of GAN in ophthalmology image domains to discuss important contributions and to identify potential future research directions. METHODS We performed a survey on studies using GAN published before June 2021 only, and we introduced various applications of GAN in ophthalmology image domains. The search identified 48 peer-reviewed papers in the final review. The type of GAN used in the analysis, task, imaging domain, and the outcome were collected to verify the usefulness of the GAN. RESULTS In ophthalmology image domains, GAN can perform segmentation, data augmentation, denoising, domain transfer, super-resolution, post-intervention prediction, and feature extraction. GAN techniques have established an extension of datasets and modalities in ophthalmology. GAN has several limitations, such as mode collapse, spatial deformities, unintended changes, and the generation of high-frequency noises and artifacts of checkerboard patterns. CONCLUSIONS The use of GAN has benefited the various tasks in ophthalmology image domains. Based on our observations, the adoption of GAN in ophthalmology is still in a very early stage of clinical validation compared with deep learning classification techniques because several problems need to be overcome for practical use. However, the proper selection of the GAN technique and statistical modeling of ocular imaging will greatly improve the performance of each image analysis. Finally, this survey would enable researchers to access the appropriate GAN technique to maximize the potential of ophthalmology datasets for deep learning research.
Collapse
Affiliation(s)
- Aram You
- School of Architecture, Kumoh National Institute of Technology, Gumi, Gyeongbuk, South Korea
| | - Jin Kuk Kim
- B&VIIT Eye Center, Seoul, South Korea
- VISUWORKS, Seoul, South Korea
| | - Ik Hee Ryu
- B&VIIT Eye Center, Seoul, South Korea
- VISUWORKS, Seoul, South Korea
| | - Tae Keun Yoo
- B&VIIT Eye Center, Seoul, South Korea.
- Department of Ophthalmology, Aerospace Medical Center, Republic of Korea Air Force, 635 Danjae-ro, Namil-myeon, Cheongwon-gun, Cheongju, Chungcheongbuk-do, 363-849, South Korea.
| |
Collapse
|
17
|
Agrawal T, Choudhary P. Segmentation and classification on chest radiography: a systematic survey. THE VISUAL COMPUTER 2022; 39:875-913. [PMID: 35035008 PMCID: PMC8741572 DOI: 10.1007/s00371-021-02352-7] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Accepted: 11/01/2021] [Indexed: 06/14/2023]
Abstract
Chest radiography (X-ray) is the most common diagnostic method for pulmonary disorders. A trained radiologist is required for interpreting the radiographs. But sometimes, even experienced radiologists can misinterpret the findings. This leads to the need for computer-aided detection diagnosis. For decades, researchers were automatically detecting pulmonary disorders using the traditional computer vision (CV) methods. Now the availability of large annotated datasets and computing hardware has made it possible for deep learning to dominate the area. It is now the modus operandi for feature extraction, segmentation, detection, and classification tasks in medical imaging analysis. This paper focuses on the research conducted using chest X-rays for the lung segmentation and detection/classification of pulmonary disorders on publicly available datasets. The studies performed using the Generative Adversarial Network (GAN) models for segmentation and classification on chest X-rays are also included in this study. GAN has gained the interest of the CV community as it can help with medical data scarcity. In this study, we have also included the research conducted before the popularity of deep learning models to have a clear picture of the field. Many surveys have been published, but none of them is dedicated to chest X-rays. This study will help the readers to know about the existing techniques, approaches, and their significance.
Collapse
Affiliation(s)
- Tarun Agrawal
- Department of Computer Science and Engineering, National Institute of Technology Hamirpur, Hamirpur, Himachal Pradesh 177005 India
| | - Prakash Choudhary
- Department of Computer Science and Engineering, National Institute of Technology Hamirpur, Hamirpur, Himachal Pradesh 177005 India
| |
Collapse
|
18
|
Singh A, Lall B, Panigrahi B, Agrawal A, Agrawal A, Thangakunam B, Christopher D. Deep LF-Net: Semantic lung segmentation from Indian chest radiographs including severely unhealthy images. Biomed Signal Process Control 2021. [DOI: 10.1016/j.bspc.2021.102666] [Citation(s) in RCA: 7] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|