1
|
Rajaraman S, Liang Z, Xue Z, Antani S. Addressing Class Imbalance with Latent Diffusion-based Data Augmentation for Improving Disease Classification in Pediatric Chest X-rays. PROCEEDINGS. IEEE INTERNATIONAL CONFERENCE ON BIOINFORMATICS AND BIOMEDICINE 2024; 2024:5059-5066. [PMID: 40134830 PMCID: PMC11936509 DOI: 10.1109/bibm62325.2024.10822172] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/27/2025]
Abstract
Deep learning (DL) has transformed medical image classification; however, its efficacy is often limited by significant data imbalance due to far fewer cases (minority class) compared to controls (majority class). It has been shown that synthetic image augmentation techniques can simulate clinical variability, leading to enhanced model performance. We hypothesize that they could also mitigate the challenge of data imbalance, thereby addressing overfitting to the majority class and enhancing generalization. Recently, latent diffusion models (LDMs) have shown promise in synthesizing high-quality medical images. This study evaluates the effectiveness of a text-guided image-to-image LDM in synthesizing disease-positive chest X-rays (CXRs) and augmenting a pediatric CXR dataset to improve classification performance. We first establish baseline performance by fine-tuning an ImageNet-pretrained Inception-V3 model on class-imbalanced data for two tasks-normal vs. pneumonia and normal vs. bronchopneumonia. Next, we fine-tune individual text-guided image-to-image LDMs to generate CXRs showing signs of pneumonia and bronchopneumonia. The Inception-V3 model is retrained on an updated data set that includes these synthesized images as part of augmented training and validation sets. Classification performance is compared using balanced accuracy, sensitivity, specificity, F-score, Matthews correlation coefficient (MCC), Kappa, and Youden's index against the baseline performance. Results show that the augmentation significantly improves Youden's index (p<0.05) and markedly enhances other metrics, indicating that data augmentation using LDM-synthesized images is an effective strategy for addressing class imbalance in medical image classification.
Collapse
Affiliation(s)
- Sivaramakrishnan Rajaraman
- Division of Intramural Research, National Library of Medicine, National Institutes of Health, Bethesda, MD, USA
| | - Zhaohui Liang
- Division of Intramural Research, National Library of Medicine, National Institutes of Health, Bethesda, MD, USA
| | - Zhiyun Xue
- Division of Intramural Research, National Library of Medicine, National Institutes of Health, Bethesda, MD, USA
| | - Sameer Antani
- Division of Intramural Research, National Library of Medicine, National Institutes of Health, Bethesda, MD, USA
| |
Collapse
|
2
|
Ganesan P, Feng R, Deb B, Tjong FVY, Rogers AJ, Ruipérez-Campillo S, Somani S, Clopton P, Baykaner T, Rodrigo M, Zou J, Haddad F, Zaharia M, Narayan SM. Novel Domain Knowledge-Encoding Algorithm Enables Label-Efficient Deep Learning for Cardiac CT Segmentation to Guide Atrial Fibrillation Treatment in a Pilot Dataset. Diagnostics (Basel) 2024; 14:1538. [PMID: 39061675 PMCID: PMC11276420 DOI: 10.3390/diagnostics14141538] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/15/2024] [Revised: 07/07/2024] [Accepted: 07/10/2024] [Indexed: 07/28/2024] Open
Abstract
Background: Segmenting computed tomography (CT) is crucial in various clinical applications, such as tailoring personalized cardiac ablation for managing cardiac arrhythmias. Automating segmentation through machine learning (ML) is hindered by the necessity for large, labeled training data, which can be challenging to obtain. This article proposes a novel approach for automated, robust labeling using domain knowledge to achieve high-performance segmentation by ML from a small training set. The approach, the domain knowledge-encoding (DOKEN) algorithm, reduces the reliance on large training datasets by encoding cardiac geometry while automatically labeling the training set. The method was validated in a hold-out dataset of CT results from an atrial fibrillation (AF) ablation study. Methods: The DOKEN algorithm parses left atrial (LA) structures, extracts "anatomical knowledge" by leveraging digital LA models (available publicly), and then applies this knowledge to achieve high ML segmentation performance with a small number of training samples. The DOKEN-labeled training set was used to train a nnU-Net deep neural network (DNN) model for segmenting cardiac CT in N = 20 patients. Subsequently, the method was tested in a hold-out set with N = 100 patients (five times larger than training set) who underwent AF ablation. Results: The DOKEN algorithm integrated with the nn-Unet model achieved high segmentation performance with few training samples, with a training to test ratio of 1:5. The Dice score of the DOKEN-enhanced model was 96.7% (IQR: 95.3% to 97.7%), with a median error in surface distance of boundaries of 1.51 mm (IQR: 0.72 to 3.12) and a mean centroid-boundary distance of 1.16 mm (95% CI: -4.57 to 6.89), similar to expert results (r = 0.99; p < 0.001). In digital hearts, the novel DOKEN approach segmented the LA structures with a mean difference for the centroid-boundary distances of -0.27 mm (95% CI: -3.87 to 3.33; r = 0.99; p < 0.0001). Conclusions: The proposed novel domain knowledge-encoding algorithm was able to perform the segmentation of six substructures of the LA, reducing the need for large training data sets. The combination of domain knowledge encoding and a machine learning approach could reduce the dependence of ML on large training datasets and could potentially be applied to AF ablation procedures and extended in the future to other imaging, 3D printing, and data science applications.
Collapse
Affiliation(s)
- Prasanth Ganesan
- Department of Medicine and Stanford Cardiovascular Institute (CVI), Stanford University, Stanford, CA 94305, USA; (P.G.); (R.F.)
| | - Ruibin Feng
- Department of Medicine and Stanford Cardiovascular Institute (CVI), Stanford University, Stanford, CA 94305, USA; (P.G.); (R.F.)
| | - Brototo Deb
- Department of Medicine and Stanford Cardiovascular Institute (CVI), Stanford University, Stanford, CA 94305, USA; (P.G.); (R.F.)
| | - Fleur V. Y. Tjong
- Department of Medicine and Stanford Cardiovascular Institute (CVI), Stanford University, Stanford, CA 94305, USA; (P.G.); (R.F.)
- Heart Center, Department of Clinical and Experimental Cardiology, Amsterdam UMC, University of Amsterdam, 1105 AZ Amsterdam, The Netherlands
| | - Albert J. Rogers
- Department of Medicine and Stanford Cardiovascular Institute (CVI), Stanford University, Stanford, CA 94305, USA; (P.G.); (R.F.)
| | - Samuel Ruipérez-Campillo
- Department of Medicine and Stanford Cardiovascular Institute (CVI), Stanford University, Stanford, CA 94305, USA; (P.G.); (R.F.)
- Department of Computer Science, ETH Zurich, 8092 Zurich, Switzerland
| | - Sulaiman Somani
- Department of Medicine and Stanford Cardiovascular Institute (CVI), Stanford University, Stanford, CA 94305, USA; (P.G.); (R.F.)
| | - Paul Clopton
- Department of Medicine and Stanford Cardiovascular Institute (CVI), Stanford University, Stanford, CA 94305, USA; (P.G.); (R.F.)
| | - Tina Baykaner
- Department of Medicine and Stanford Cardiovascular Institute (CVI), Stanford University, Stanford, CA 94305, USA; (P.G.); (R.F.)
| | - Miguel Rodrigo
- Department of Medicine and Stanford Cardiovascular Institute (CVI), Stanford University, Stanford, CA 94305, USA; (P.G.); (R.F.)
- CoMMLab, Universitat de València, 46100 Valencia, Spain
| | - James Zou
- Department of Biomedical Data Science, Stanford University, Stanford, CA 94305, USA
| | - Francois Haddad
- Department of Medicine and Stanford Cardiovascular Institute (CVI), Stanford University, Stanford, CA 94305, USA; (P.G.); (R.F.)
| | - Matei Zaharia
- Department of Computer Science, University of California Berkeley, Berkeley, CA 94720, USA
| | - Sanjiv M. Narayan
- Department of Medicine and Stanford Cardiovascular Institute (CVI), Stanford University, Stanford, CA 94305, USA; (P.G.); (R.F.)
| |
Collapse
|
3
|
Rehman A, Khan A, Fatima G, Naz S, Razzak I. Review on chest pathogies detection systems using deep learning techniques. Artif Intell Rev 2023; 56:1-47. [PMID: 37362896 PMCID: PMC10027283 DOI: 10.1007/s10462-023-10457-9] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/29/2023]
Abstract
Chest radiography is the standard and most affordable way to diagnose, analyze, and examine different thoracic and chest diseases. Typically, the radiograph is examined by an expert radiologist or physician to decide about a particular anomaly, if exists. Moreover, computer-aided methods are used to assist radiologists and make the analysis process accurate, fast, and more automated. A tremendous improvement in automatic chest pathologies detection and analysis can be observed with the emergence of deep learning. The survey aims to review, technically evaluate, and synthesize the different computer-aided chest pathologies detection systems. The state-of-the-art of single and multi-pathologies detection systems, which are published in the last five years, are thoroughly discussed. The taxonomy of image acquisition, dataset preprocessing, feature extraction, and deep learning models are presented. The mathematical concepts related to feature extraction model architectures are discussed. Moreover, the different articles are compared based on their contributions, datasets, methods used, and the results achieved. The article ends with the main findings, current trends, challenges, and future recommendations.
Collapse
Affiliation(s)
- Arshia Rehman
- COMSATS University Islamabad, Abbottabad-Campus, Abbottabad, Pakistan
| | - Ahmad Khan
- COMSATS University Islamabad, Abbottabad-Campus, Abbottabad, Pakistan
| | - Gohar Fatima
- The Islamia University of Bahawalpur, Bahawal Nagar Campus, Bahawal Nagar, Pakistan
| | - Saeeda Naz
- Govt Girls Post Graduate College No.1, Abbottabad, Pakistan
| | - Imran Razzak
- School of Computer Science and Engineering, University of New South Wales, Sydney, Australia
| |
Collapse
|
4
|
Rajaraman S, Zamzmi G, Yang F, Xue Z, Antani SK. Data Characterization for Reliable AI in Medicine. RECENT TRENDS IN IMAGE PROCESSING AND PATTERN RECOGNITION : 5TH INTERNATIONAL CONFERENCE, RTIP2R 2022, KINGSVILLE, TX, USA, DECEMBER 01-02, 2022, REVISED SELECTED PAPERS. INTERNATIONAL CONFERENCE ON RECENT TRENDS IN IMAGE PROCESSING AND... 2023; 1704:3-11. [PMID: 36780238 PMCID: PMC9912175 DOI: 10.1007/978-3-031-23599-3_1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/12/2023]
Abstract
Research in Artificial Intelligence (AI)-based medical computer vision algorithms bear promises to improve disease screening, diagnosis, and subsequently patient care. However, these algorithms are highly impacted by the characteristics of the underlying data. In this work, we discuss various data characteristics, namely Volume, Veracity, Validity, Variety, and Velocity, that impact the design, reliability, and evolution of machine learning in medical computer vision. Further, we discuss each characteristic and the recent works conducted in our research lab that informed our understanding of the impact of these characteristics on the design of medical decision-making algorithms and outcome reliability.
Collapse
Affiliation(s)
- Sivaramakrishnan Rajaraman
- Computational Health Research Branch, National Library of Medicine, National Institutes of Health, Bethesda MD 20894, USA
| | - Ghada Zamzmi
- Computational Health Research Branch, National Library of Medicine, National Institutes of Health, Bethesda MD 20894, USA
| | - Feng Yang
- Computational Health Research Branch, National Library of Medicine, National Institutes of Health, Bethesda MD 20894, USA
| | - Zhiyun Xue
- Computational Health Research Branch, National Library of Medicine, National Institutes of Health, Bethesda MD 20894, USA
| | - Sameer K Antani
- Computational Health Research Branch, National Library of Medicine, National Institutes of Health, Bethesda MD 20894, USA
| |
Collapse
|
5
|
Sun C, Wang R, Zhao L, Han L, Ma S, Liang D, Wang L, Tuo X, Zhang Y, Zhong D, Li Q. A Computer-Aided Diagnosis System of Fetal Nucleated Red Blood Cells With Convolutional Neural Network. Arch Pathol Lab Med 2022; 146:1395-1401. [PMID: 35293972 DOI: 10.5858/arpa.2021-0142-oa] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 10/18/2021] [Indexed: 11/06/2022]
Abstract
CONTEXT.— The rapid recognition of fetal nucleated red blood cells (fNRBCs) presents considerable challenges. OBJECTIVE.— To establish a computer-aided diagnosis system for rapid recognition of fNRBCs by convolutional neural network. DESIGN.— We adopted density gradient centrifugation and magnetic-activated cell sorting to extract fNRBCs from umbilical cord blood samples. The cell-block method was used to embed fNRBCs for routine formalin-fixed paraffin sectioning and hematoxylin-eosin staining. Then, we proposed a convolutional neural network-based, computer-aided diagnosis system to automatically discriminate features and recognize fNRBCs. Extracting methods of interested region were used to automatically segment individual cells in cell slices. The discriminant information from cellular-level regions of interest was encoded into a feature vector. Pathologic diagnoses were also provided by the network. RESULTS.— In total, 4760 pictures of fNRBCs from 260 cell-slides of 4 umbilical cord blood samples were collected. On the premise of 100% accuracy in the training set (3720 pictures), the sensitivity, specificity, and accuracy of cellular intelligent recognition were 96.5%, 100%, and 98.5%, respectively, in the test set (1040 pictures). CONCLUSIONS.— We established a computer-aided diagnosis system for effective and accurate fNRBC recognition based on a convolutional neural network.
Collapse
Affiliation(s)
- Chao Sun
- From the Department of Obstetrics and Gynecology, The First Affiliated Hospital of Xi'an Jiaotong University, Xi'an, Shannxi, China (Sun, Zhao, Han, Ma, Liang, L. Wang, Tuo, Zhang, Li)
| | - Ruijie Wang
- School of Automation Science and Engineering, Xi'an Jiaotong University Xi'an, Shannxi, China (R. Wang, Zhong)
| | - Lanbo Zhao
- From the Department of Obstetrics and Gynecology, The First Affiliated Hospital of Xi'an Jiaotong University, Xi'an, Shannxi, China (Sun, Zhao, Han, Ma, Liang, L. Wang, Tuo, Zhang, Li)
| | - Lu Han
- From the Department of Obstetrics and Gynecology, The First Affiliated Hospital of Xi'an Jiaotong University, Xi'an, Shannxi, China (Sun, Zhao, Han, Ma, Liang, L. Wang, Tuo, Zhang, Li)
| | - Sijia Ma
- From the Department of Obstetrics and Gynecology, The First Affiliated Hospital of Xi'an Jiaotong University, Xi'an, Shannxi, China (Sun, Zhao, Han, Ma, Liang, L. Wang, Tuo, Zhang, Li)
| | - Dongxin Liang
- From the Department of Obstetrics and Gynecology, The First Affiliated Hospital of Xi'an Jiaotong University, Xi'an, Shannxi, China (Sun, Zhao, Han, Ma, Liang, L. Wang, Tuo, Zhang, Li)
| | - Lei Wang
- From the Department of Obstetrics and Gynecology, The First Affiliated Hospital of Xi'an Jiaotong University, Xi'an, Shannxi, China (Sun, Zhao, Han, Ma, Liang, L. Wang, Tuo, Zhang, Li)
| | - Xiaoqian Tuo
- From the Department of Obstetrics and Gynecology, The First Affiliated Hospital of Xi'an Jiaotong University, Xi'an, Shannxi, China (Sun, Zhao, Han, Ma, Liang, L. Wang, Tuo, Zhang, Li)
| | - Yu Zhang
- From the Department of Obstetrics and Gynecology, The First Affiliated Hospital of Xi'an Jiaotong University, Xi'an, Shannxi, China (Sun, Zhao, Han, Ma, Liang, L. Wang, Tuo, Zhang, Li)
| | - Dexing Zhong
- School of Automation Science and Engineering, Xi'an Jiaotong University Xi'an, Shannxi, China (R. Wang, Zhong)
- Pazhou Laboratory, Guangzhou, China (Zhong)
- The State Key Laboratory of Novel Software Technology, Nanjing University, Nanjing, China (Zhong)
| | | |
Collapse
|
6
|
Deep learning model calibration for improving performance in class-imbalanced medical image classification tasks. PLoS One 2022; 17:e0262838. [PMID: 35085334 PMCID: PMC8794113 DOI: 10.1371/journal.pone.0262838] [Citation(s) in RCA: 14] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2021] [Accepted: 01/05/2022] [Indexed: 11/19/2022] Open
Abstract
In medical image classification tasks, it is common to find that the number of normal samples far exceeds the number of abnormal samples. In such class-imbalanced situations, reliable training of deep neural networks continues to be a major challenge, therefore biasing the predicted class probabilities toward the majority class. Calibration has been proposed to alleviate some of these effects. However, there is insufficient analysis explaining whether and when calibrating a model would be beneficial. In this study, we perform a systematic analysis of the effect of model calibration on its performance on two medical image modalities, namely, chest X-rays and fundus images, using various deep learning classifier backbones. For this, we study the following variations: (i) the degree of imbalances in the dataset used for training; (ii) calibration methods; and (iii) two classification thresholds, namely, default threshold of 0.5, and optimal threshold from precision-recall (PR) curves. Our results indicate that at the default classification threshold of 0.5, the performance achieved through calibration is significantly superior (p < 0.05) to using uncalibrated probabilities. However, at the PR-guided threshold, these gains are not significantly different (p > 0.05). This observation holds for both image modalities and at varying degrees of imbalance. The code is available at https://github.com/sivaramakrishnan-rajaraman/Model_calibration.
Collapse
|
7
|
Yoon D, Kong HJ, Kim BS, Cho WS, Lee JC, Cho M, Lim MH, Yang SY, Lim SH, Lee J, Song JH, Chung GE, Choi JM, Kang HY, Bae JH, Kim S. Colonoscopic image synthesis with generative adversarial network for enhanced detection of sessile serrated lesions using convolutional neural network. Sci Rep 2022; 12:261. [PMID: 34997124 PMCID: PMC8741803 DOI: 10.1038/s41598-021-04247-y] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/29/2021] [Accepted: 12/20/2021] [Indexed: 12/28/2022] Open
Abstract
Computer-aided detection (CADe) systems have been actively researched for polyp detection in colonoscopy. To be an effective system, it is important to detect additional polyps that may be easily missed by endoscopists. Sessile serrated lesions (SSLs) are a precursor to colorectal cancer with a relatively higher miss rate, owing to their flat and subtle morphology. Colonoscopy CADe systems could help endoscopists; however, the current systems exhibit a very low performance for detecting SSLs. We propose a polyp detection system that reflects the morphological characteristics of SSLs to detect unrecognized or easily missed polyps. To develop a well-trained system with imbalanced polyp data, a generative adversarial network (GAN) was used to synthesize high-resolution whole endoscopic images, including SSL. Quantitative and qualitative evaluations on GAN-synthesized images ensure that synthetic images are realistic and include SSL endoscopic features. Moreover, traditional augmentation methods were used to compare the efficacy of the GAN augmentation method. The CADe system augmented with GAN synthesized images showed a 17.5% improvement in sensitivity on SSLs. Consequently, we verified the potential of the GAN to synthesize high-resolution images with endoscopic features and the proposed system was found to be effective in detecting easily missed polyps during a colonoscopy.
Collapse
Affiliation(s)
- Dan Yoon
- Interdisciplinary Program in Bioengineering, Graduate School, Seoul National University, Seoul, 08826, South Korea
| | - Hyoun-Joong Kong
- Transdisciplinary Department of Medicine and Advanced Technology, Seoul National University Hospital, Seoul, 03080, South Korea.,Department of Biomedical Engineering, Seoul National University College of Medicine, Seoul, 03080, South Korea.,Medical Big Data Research Center, Seoul National University College of Medicine, Seoul, 03080, South Korea.,Artificial Intelligence Institute, Seoul National University, Seoul, 08826, South Korea
| | - Byeong Soo Kim
- Interdisciplinary Program in Bioengineering, Graduate School, Seoul National University, Seoul, 08826, South Korea
| | - Woo Sang Cho
- Interdisciplinary Program in Bioengineering, Graduate School, Seoul National University, Seoul, 08826, South Korea
| | - Jung Chan Lee
- Department of Biomedical Engineering, Seoul National University College of Medicine, Seoul, 03080, South Korea.,Institute of Medical and Biological Engineering, Medical Research Center, Seoul National University, Seoul, 03080, South Korea.,Institute of Bioengineering, Seoul National University, Seoul, 08826, South Korea
| | - Minwoo Cho
- Transdisciplinary Department of Medicine and Advanced Technology, Seoul National University Hospital, Seoul, 03080, South Korea.,Biomedical Research Institute, Seoul National University Hospital, Seoul, 03080, South Korea
| | - Min Hyuk Lim
- Department of Biomedical Engineering, Seoul National University College of Medicine, Seoul, 03080, South Korea
| | - Sun Young Yang
- Department of Internal Medicine and Healthcare Research Institute, Healthcare System Gangnam Center, Seoul National University Hospital, Seoul, 06236, South Korea
| | - Seon Hee Lim
- Department of Internal Medicine and Healthcare Research Institute, Healthcare System Gangnam Center, Seoul National University Hospital, Seoul, 06236, South Korea
| | - Jooyoung Lee
- Department of Internal Medicine and Healthcare Research Institute, Healthcare System Gangnam Center, Seoul National University Hospital, Seoul, 06236, South Korea
| | - Ji Hyun Song
- Department of Internal Medicine and Healthcare Research Institute, Healthcare System Gangnam Center, Seoul National University Hospital, Seoul, 06236, South Korea
| | - Goh Eun Chung
- Department of Internal Medicine and Healthcare Research Institute, Healthcare System Gangnam Center, Seoul National University Hospital, Seoul, 06236, South Korea
| | - Ji Min Choi
- Department of Internal Medicine and Healthcare Research Institute, Healthcare System Gangnam Center, Seoul National University Hospital, Seoul, 06236, South Korea
| | - Hae Yeon Kang
- Department of Internal Medicine and Healthcare Research Institute, Healthcare System Gangnam Center, Seoul National University Hospital, Seoul, 06236, South Korea
| | - Jung Ho Bae
- Department of Internal Medicine and Healthcare Research Institute, Healthcare System Gangnam Center, Seoul National University Hospital, Seoul, 06236, South Korea.
| | - Sungwan Kim
- Department of Biomedical Engineering, Seoul National University College of Medicine, Seoul, 03080, South Korea. .,Artificial Intelligence Institute, Seoul National University, Seoul, 08826, South Korea. .,Institute of Bioengineering, Seoul National University, Seoul, 08826, South Korea.
| |
Collapse
|
8
|
Zhang R, Lu W, Wei X, Zhu J, Jiang H, Liu Z, Gao J, Li X, Yu J, Yu M, Yu R. A Progressive Generative Adversarial Method for Structurally Inadequate Medical Image Data Augmentation. IEEE J Biomed Health Inform 2021; 26:7-16. [PMID: 34347609 DOI: 10.1109/jbhi.2021.3101551] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
The generation-based data augmentation method can overcome the challenge caused by the imbalance of medical image data to a certain extent. However, most of the current research focus on images with unified structure which are easy to learn. What is different is that ultrasound images are structurally inadequate, making it difficult for the structure to be captured by the generative network, resulting in the generated image lacks structural legitimacy. Therefore, a Progressive Generative Adversarial Method for Structurally Inadequate Medical Image Data Augmentation is proposed in this paper, including a network and a strategy. Our Progressive Texture Generative Adversarial Network alleviates the adverse effect of completely truncating the reconstruction of structure and texture during the generation process and enhances the implicit association between structure and texture. The Image Data Augmentation Strategy based on Mask-Reconstruction overcomes data imbalance from a novel perspective, maintains the legitimacy of the structure in the generated data, as well as increases the diversity of disease data interpretably. The experiments prove the effectiveness of our method on data augmentation and image reconstruction on Structurally Inadequate Medical Image both qualitatively and quantitatively. Finally, the weakly supervised segmentation of the lesion is the additional contribution of our method.
Collapse
|
9
|
Serena Low WC, Chuah JH, Tee CATH, Anis S, Shoaib MA, Faisal A, Khalil A, Lai KW. An Overview of Deep Learning Techniques on Chest X-Ray and CT Scan Identification of COVID-19. COMPUTATIONAL AND MATHEMATICAL METHODS IN MEDICINE 2021; 2021:5528144. [PMID: 34194535 PMCID: PMC8184329 DOI: 10.1155/2021/5528144] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/05/2021] [Revised: 04/19/2021] [Accepted: 05/19/2021] [Indexed: 12/15/2022]
Abstract
Pneumonia is an infamous life-threatening lung bacterial or viral infection. The latest viral infection endangering the lives of many people worldwide is the severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), which causes COVID-19. This paper is aimed at detecting and differentiating viral pneumonia and COVID-19 disease using digital X-ray images. The current practices include tedious conventional processes that solely rely on the radiologist or medical consultant's technical expertise that are limited, time-consuming, inefficient, and outdated. The implementation is easily prone to human errors of being misdiagnosed. The development of deep learning and technology improvement allows medical scientists and researchers to venture into various neural networks and algorithms to develop applications, tools, and instruments that can further support medical radiologists. This paper presents an overview of deep learning techniques made in the chest radiography on COVID-19 and pneumonia cases.
Collapse
Affiliation(s)
- Woan Ching Serena Low
- Department of Electrical Engineering, Faculty of Engineering, University of Malaya, 40603 Kuala Lumpur, Malaysia
| | - Joon Huang Chuah
- Department of Electrical Engineering, Faculty of Engineering, University of Malaya, 40603 Kuala Lumpur, Malaysia
| | - Clarence Augustine T. H. Tee
- Department of Electrical Engineering, Faculty of Engineering, University of Malaya, 40603 Kuala Lumpur, Malaysia
| | - Shazia Anis
- Department of Biomedical Engineering, Faculty of Engineering, University of Malaya, 40603 Kuala Lumpur, Malaysia
| | - Muhammad Ali Shoaib
- Department of Electrical Engineering, Faculty of Engineering, University of Malaya, 40603 Kuala Lumpur, Malaysia
| | - Amir Faisal
- Department of Biomedical Engineering, Faculty of Production and Industrial Technology, Institut Teknologi Sumatera, Lampung 35365, Indonesia
| | - Azira Khalil
- Faculty of Science and Technology, Universiti Sains Islam Malaysia, 71800 Nilai, Negeri Sembilan, Malaysia
| | - Khin Wee Lai
- Department of Biomedical Engineering, Faculty of Engineering, University of Malaya, 40603 Kuala Lumpur, Malaysia
| |
Collapse
|
10
|
Rajaraman S, Antani S. Weakly Labeled Data Augmentation for Deep Learning: A Study on COVID-19 Detection in Chest X-Rays. Diagnostics (Basel) 2020; 10:E358. [PMID: 32486140 PMCID: PMC7345787 DOI: 10.3390/diagnostics10060358] [Citation(s) in RCA: 51] [Impact Index Per Article: 10.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/24/2020] [Revised: 05/26/2020] [Accepted: 05/29/2020] [Indexed: 01/05/2023] Open
Abstract
The novel severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2) has caused a pandemic resulting in over 2.7 million infected individuals and over 190,000 deaths and growing. Assertions in the literature suggest that respiratory disorders due to COVID-19 commonly present with pneumonia-like symptoms which are radiologically confirmed as opacities. Radiology serves as an adjunct to the reverse transcription-polymerase chain reaction test for confirmation and evaluating disease progression. While computed tomography (CT) imaging is more specific than chest X-rays (CXR), its use is limited due to cross-contamination concerns. CXR imaging is commonly used in high-demand situations, placing a significant burden on radiology services. The use of artificial intelligence (AI) has been suggested to alleviate this burden. However, there is a dearth of sufficient training data for developing image-based AI tools. We propose increasing training data for recognizing COVID-19 pneumonia opacities using weakly labeled data augmentation. This follows from a hypothesis that the COVID-19 manifestation would be similar to that caused by other viral pathogens affecting the lungs. We expand the training data distribution for supervised learning through the use of weakly labeled CXR images, automatically pooled from publicly available pneumonia datasets, to classify them into those with bacterial or viral pneumonia opacities. Next, we use these selected images in a stage-wise, strategic approach to train convolutional neural network-based algorithms and compare against those trained with non-augmented data. Weakly labeled data augmentation expands the learned feature space in an attempt to encompass variability in unseen test distributions, enhance inter-class discrimination, and reduce the generalization error. Empirical evaluations demonstrate that simple weakly labeled data augmentation (Acc: 0.5555 and Acc: 0.6536) is better than baseline non-augmented training (Acc: 0.2885 and Acc: 0.5028) in identifying COVID-19 manifestations as viral pneumonia. Interestingly, adding COVID-19 CXRs to simple weakly labeled augmented training data significantly improves the performance (Acc: 0.7095 and Acc: 0.8889), suggesting that COVID-19, though viral in origin, creates a uniquely different presentation in CXRs compared with other viral pneumonia manifestations.
Collapse
Affiliation(s)
- Sivaramakrishnan Rajaraman
- Lister Hill National Center for Biomedical Communications, National Library of Medicine, 8600 Rockville Pike, Bethesda, MD 20894, USA;
| | | |
Collapse
|