1
|
Lamouadene H, El Kassaoui M, El Yadari M, El Kenz A, Benyoussef A, El Moutaouakil A, Mounkachi O. Detection of COVID-19, lung opacity, and viral pneumonia via X-ray using machine learning and deep learning. Comput Biol Med 2025; 191:110131. [PMID: 40198984 DOI: 10.1016/j.compbiomed.2025.110131] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/10/2024] [Revised: 03/26/2025] [Accepted: 04/01/2025] [Indexed: 04/10/2025]
Abstract
The COVID-19 pandemic has significantly strained healthcare systems, highlighting the need for early diagnosis to isolate positive cases and prevent the spread. This study combines machine learning, deep learning, and transfer learning techniques to automatically diagnose COVID-19 and other pulmonary conditions from radiographic images. First, we used Convolutional Neural Networks (CNNs) and a Support Vector Machine (SVM) classifier on a dataset of 21,165 chest X-ray images. Our model achieved an accuracy of 86.18 %. This approach aids medical experts in rapidly and accurateky detecting lung diseases. Next, we applied transfer learning using ResNet18 combined with SVM on a dataset comprising normal, COVID-19, lung opacity, and viral pneumonia images. This model outperformed traditional methods, with classification rates of 98 % with Stochastic Gradient Descent (SGD), 97 % with Adam, 96 % with RMSProp, and 94 % with Adagrad optimizers. Additionally, we incorporated two additional transfer learning models, EfficientNet-CNN and Xception-CNN, which achieved classification accuracies of 99.20 % and 98.80 %, respectively. However, we observed limitations in dataset diversity and representativeness, which may affect model generalization. Future work will focus on implementing advanced data augmentation techniques and collaborations with medical experts to enhance model performance.This research demonstrates the potential of cutting-edge deep learning techniques to improve diagnostic accuracy and efficiency in medical imaging applications.
Collapse
Affiliation(s)
- Hajar Lamouadene
- Laboratory of Condensed Matter and Interdisciplinary Sciences, Physics Department, Faculty of Sciences, Mohammed V University in Rabat, Morocco
| | - Majid El Kassaoui
- Laboratory of Condensed Matter and Interdisciplinary Sciences, Physics Department, Faculty of Sciences, Mohammed V University in Rabat, Morocco; College of computing, Mohammed VI Polytechnic University, Lot 660, Hay Moulay Rachid Ben Guerir, 43150, Morocco
| | | | - Abdallah El Kenz
- Laboratory of Condensed Matter and Interdisciplinary Sciences, Physics Department, Faculty of Sciences, Mohammed V University in Rabat, Morocco
| | - Abdelilah Benyoussef
- Laboratory of Condensed Matter and Interdisciplinary Sciences, Physics Department, Faculty of Sciences, Mohammed V University in Rabat, Morocco; Hassan II Academy of Science and Technology, Rabat, Morocco
| | - Amine El Moutaouakil
- Department of Electrical and Communication Engineering, College of Engineering, UAE University, P.O. Box: 15551, Al Ain, United Arab Emirates.
| | - Omar Mounkachi
- Laboratory of Condensed Matter and Interdisciplinary Sciences, Physics Department, Faculty of Sciences, Mohammed V University in Rabat, Morocco; College of computing, Mohammed VI Polytechnic University, Lot 660, Hay Moulay Rachid Ben Guerir, 43150, Morocco
| |
Collapse
|
2
|
Kang HYJ, Ko M, Ryu KS. Tabular transformer generative adversarial network for heterogeneous distribution in healthcare. Sci Rep 2025; 15:10254. [PMID: 40133347 PMCID: PMC11937286 DOI: 10.1038/s41598-025-93077-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/30/2024] [Accepted: 03/04/2025] [Indexed: 03/27/2025] Open
Abstract
In healthcare, the most common type of data is tabular data, which holds high significance and potential in the field of medical AI. However, privacy concerns have hindered their widespread use. Despite the emergence of synthetic data as a viable solution, the generation of healthcare tabular data (HTD) is complex owing to the extensive interdependencies between the variables within each record that incorporate diverse clinical characteristics, including sensitive information. To overcome these issues, this study proposed a tabular transformer generative adversarial network (TT-GAN) to generate synthetic data that can effectively consider the relationships between variables potentially present in the HTD dataset. Transformers can consider the relationships between the columns in each record using a multi-attention mechanism. In addition, to address the potential risk of restoring sensitive data in patient information, a Transformer was employed in a generative adversarial network (GAN) architecture, to ensure an implicit-based algorithm. To consider the heterogeneous characteristics of the continuous variables in the HTD dataset, the discretization and converter methodology were applied. The experimental results confirmed the superior performance of the TT-GAN than the Conditional Tabular GAN (CTGAN) and copula GAN. Discretization and converters were proven to be effective using our proposed Transformer algorithm. However, the application of the same methodology to Transformer-based models without discretization and converters exhibited a significantly inferior performance. The CTGAN and copula GAN indicated minimal effectiveness with discretization and converter methodologies. Thus, the TT-GAN exhibited considerable potential in healthcare, demonstrating its ability to generate artificial data that closely resembled real healthcare datasets. The ability of the algorithm to handle different types of mixed variables efficiently, including polynomial, discrete, and continuous variables, demonstrated its versatility and practicality in health care research and data synthesis.
Collapse
Affiliation(s)
- Ha Ye Jin Kang
- Department of Applied Artificial Intelligence, Hanyang University, Seoul, Republic of Korea
- Department of Public Health & AI, Graduate School of Cancer Science and Policy, National Cancer Center, Goyang, Republic of Korea
| | - Minsam Ko
- Department of Applied Artificial Intelligence, Hanyang University, Seoul, Republic of Korea
| | - Kwang Sun Ryu
- Department of Public Health & AI, Graduate School of Cancer Science and Policy, National Cancer Center, Goyang, Republic of Korea.
- National Cancer Data Center, National Cancer Center, Goyang, Republic of Korea.
| |
Collapse
|
3
|
Krishna A, Wang G, Mueller K. Guided synthesis of annotated lung CT images with pathologies using a multi-conditioned denoising diffusion probabilistic model (mDDPM). Phys Med Biol 2025; 70:10.1088/1361-6560/adb9b3. [PMID: 39993375 PMCID: PMC12020743 DOI: 10.1088/1361-6560/adb9b3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2024] [Accepted: 02/24/2025] [Indexed: 02/26/2025]
Abstract
Objective. The training of AI models for medical image diagnostics requires highly accurate, diverse, and large training datasets with annotations and pathologies. Unfortunately, due to privacy and other constraints the amount of medical image data available for AI training remains limited, and this scarcity is exacerbated by the high overhead required for annotation. We address this challenge by introducing a new controlled framework for the generation of synthetic images complete with annotations, incorporating multiple conditional specifications as inputs.Approach. Using lung CT as a case study, we employ a denoising diffusion probabilistic model to train an unconditional large-scale generative model. We extend this with a classifier-free sampling strategy to develop a robust generation framework. This approach enables the generation of constrained and annotated lung CT images that accurately depict anatomy, successfully deceiving experts into perceiving them as real. Most notably, we demonstrate the generalizability of our multi-conditioned sampling approach by producing images with specific pathologies, such as lung nodules at designated locations, within the constrained anatomy.Main results. Our experiments reveal that our proposed approach can effectively produce constrained, annotated and diverse lung CT images that maintain anatomical consistency and fidelity, even for annotations not present in the training datasets. Moreover, our results highlight the superior performance of controlled generative frameworks of this nature compared to nearly every state-of-the-art image generative model when trained on comparable large medical datasets. Finally, we highlight how our approach can be extended to other medical imaging domains, further underscoring the versatility of our method.Significance. The significance of our work lies in its robust approach for generating synthetic images with annotations, facilitating the creation of highly accurate and diverse training datasets for AI applications and its wider applicability to other imaging modalities in medical domains. Our demonstrated capability to faithfully represent anatomy and pathology in generated medical images holds significant potential for various medical imaging applications, with high promise to lead to improved diagnostic accuracy and patient care.
Collapse
Affiliation(s)
- Arjun Krishna
- Arjun Krishna is in the Computer Science Department, Stony Brook University, Stony Brook, NY 11794, USA
| | - Ge Wang
- Ge Wang is with the Biomedical Imaging Center, Rensselaer Polytechnic Institute, Troy, NY 12180, USA
| | - Klaus Mueller
- Klaus Mueller is with the Computer Science Department, Stony Brook University, Stony Brook, NY 11794, USA
| |
Collapse
|
4
|
Patel AN, Srinivasan K. Deep learning paradigms in lung cancer diagnosis: A methodological review, open challenges, and future directions. Phys Med 2025; 131:104914. [PMID: 39938402 DOI: 10.1016/j.ejmp.2025.104914] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/13/2024] [Revised: 12/19/2024] [Accepted: 01/30/2025] [Indexed: 02/14/2025] Open
Abstract
Lung cancer is the leading cause of global cancer-related deaths, which emphasizes the critical importance of early diagnosis in enhancing patient outcomes. Deep learning has demonstrated significant promise in lung cancer diagnosis, excelling in nodule detection, classification, and prognosis prediction. This methodological review comprehensively explores deep learning models' application in lung cancer diagnosis, uncovering their integration across various imaging modalities. Deep learning consistently achieves state-of-the-art performance, occasionally surpassing human expert accuracy. Notably, deep neural networks excel in detecting lung nodules, distinguishing between benign and malignant nodules, and predicting patient prognosis. They have also led to the development of computer-aided diagnosis systems, enhancing diagnostic accuracy for radiologists. This review follows the specified criteria for article selection outlined by PRISMA framework. Despite challenges such as data quality and interpretability limitations, this review emphasizes the potential of deep learning to significantly improve the precision and efficiency of lung cancer diagnosis, facilitating continued research efforts to overcome these obstacles and fully harness neural network's transformative impact in this field.
Collapse
Affiliation(s)
- Aryan Nikul Patel
- School of Computer Science and Engineering, Vellore Institute of Technology, Vellore, India.
| | - Kathiravan Srinivasan
- School of Computer Science and Engineering, Vellore Institute of Technology, Vellore, India.
| |
Collapse
|
5
|
Ou Z, Pan Y, Xie F, Guo Q, Shen D. Image-and-Label Conditioning Latent Diffusion Model: Synthesizing A$\beta$-PET From MRI for Detecting Amyloid Status. IEEE J Biomed Health Inform 2025; 29:1221-1231. [PMID: 40030191 DOI: 10.1109/jbhi.2024.3492020] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/06/2025]
Abstract
Deposition of $\beta$-amyloid (A$\beta$), which is generally observed by A$\beta$-PET, is an important biomarker to evaluate subjects with early-onset dementia. However, acquisition of A$\beta$-PET usually suffers from high expense and radiation hazards, making A$\beta$-PET not commonly used as MRI. As A$\beta$-PET scans are only used to determine whether A$\beta$ deposition is positive or not, it is highly valuable to capture the underlying relationship between A$\beta$ deposition and other neuroimages (i.e., MRI) and detect amyloid status based on other neuroimages to reduce necessity of acquiring A$\beta$-PET. To this end, we propose an image-and-label conditioning latent diffusion model (IL-CLDM) to synthesize A$\beta$-PET scans from MRI scans by enhancing critical shared information to finally achieve MRI-based A$\beta$ classification. Specifically, two conditioning modules are introduced to enable IL-CLDM to implicitly learn joint image synthesis and diagnosis: 1) an image conditioning module, to extract meaningful features from source MRI scans to provide structural information, and 2) a label conditioning module, to guide the alignment of generated scans to the diagnosed label. Experiments on a clinical dataset of 510 subjects demonstrate that our proposed IL-CLDM achieves image quality superior to five widely used models, and our synthesized A$\beta$-PET scans (by IL-CLDM) can significantly help classification of A$\beta$ as positive or negative.
Collapse
|
6
|
Chaari A, Kallel IF, Kammoun S, Frikha M. Hybrid data augmentation strategies for robust deep learning classification of corneal topographic maptopographic map. Biomed Phys Eng Express 2025; 11:025017. [PMID: 39832385 DOI: 10.1088/2057-1976/adabea] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/06/2024] [Accepted: 01/20/2025] [Indexed: 01/22/2025]
Abstract
Deep learning has emerged as a powerful tool in medical imaging, particularly for corneal topographic map classification. However, the scarcity of labeled data poses a significant challenge to achieving robust performance. This study investigates the impact of various data augmentation strategies on enhancing the performance of a customized convolutional neural network model for corneal topographic map classification. We propose a hybrid data augmentation approach that combines traditional transformations, generative adversarial networks, and specific generative models. Experimental results demonstrate that the hybrid data augmentation method, achieves the highest accuracy of 99.54%, significantly outperforming individual data augmentation techniques. This hybrid approach not only improves model accuracy but also mitigates overfitting issues, making it a promising solution for medical image classification tasks with limited data availability.
Collapse
Affiliation(s)
- Abir Chaari
- ATISP laboratory, ENET'com, University of Sfax, Tunisia
| | | | - Sonda Kammoun
- Department of ophthalmology, Habib Bourguiba Hospital, University of Sfax, Tunisia
| | | |
Collapse
|
7
|
Janumpally R, Nanua S, Ngo A, Youens K. Generative artificial intelligence in graduate medical education. Front Med (Lausanne) 2025; 11:1525604. [PMID: 39867924 PMCID: PMC11758457 DOI: 10.3389/fmed.2024.1525604] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/10/2024] [Accepted: 12/23/2024] [Indexed: 01/28/2025] Open
Abstract
Generative artificial intelligence (GenAI) is rapidly transforming various sectors, including healthcare and education. This paper explores the potential opportunities and risks of GenAI in graduate medical education (GME). We review the existing literature and provide commentary on how GenAI could impact GME, including five key areas of opportunity: electronic health record (EHR) workload reduction, clinical simulation, individualized education, research and analytics support, and clinical decision support. We then discuss significant risks, including inaccuracy and overreliance on AI-generated content, challenges to authenticity and academic integrity, potential biases in AI outputs, and privacy concerns. As GenAI technology matures, it will likely come to have an important role in the future of GME, but its integration should be guided by a thorough understanding of both its benefits and limitations.
Collapse
Affiliation(s)
| | | | | | - Kenneth Youens
- Clinical Informatics Fellowship Program, Baylor Scott & White Health, Round Rock, TX, United States
| |
Collapse
|
8
|
Acien A, Morales A, Giancardo L, Vera-Rodriguez R, Holmes AA, Fierrez J, Arroyo-Gallego T. KeyGAN: Synthetic keystroke data generation in the context of digital phenotyping. Comput Biol Med 2025; 184:109460. [PMID: 39615234 DOI: 10.1016/j.compbiomed.2024.109460] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/20/2024] [Revised: 11/06/2024] [Accepted: 11/19/2024] [Indexed: 12/22/2024]
Abstract
OBJECTIVE This paper aims to introduce and assess KeyGAN, a generative modeling-based keystroke data synthesizer. The synthesizer is designed to generate realistic synthetic keystroke data capturing the nuances of fine motor control and cognitive processes that govern finger-keyboard kinematics, thereby paving the way to support biomarker development for psychomotor impairment due to neurodegeneration. METHODS KeyGAN is designed with two primary objectives: (i) to ensure high realism in the synthetic distributions of the keystroke features and (ii) to analyze its ability to replicate the subtleties of natural typing for enhancing biomarker development. The quality of synthetic keystroke data produced by KeyGAN is evaluated against two keystroke-based applications, TypeNet and nQiMechPD, employed as'referee' controls. The performance of KeyGAN is compared with a reference random Gaussian generator, testing its ability to fool the biometric authentication method TypeNet, and its ability to characterize fine motor impairment in Parkinson's Disease using nQiMechPD. RESULTS KeyGAN outperformed the reference comparator in fooling the biometric authentication method TypeNet. It also exhibited a superior approximation to real data than the reference comparator when using nQiMechPD, showcasing its adaptability and versatility in mimicking early signs of Parkinson's Disease in natural typing. KeyGAN's synthetic data demonstrated that almost 20% of real PD samples could be replaced in the training set without a decline in classification performance on the real test set. Low Fréchet Distance (<0.03) and Kullback-Leibler Divergence (<700) between KeyGAN outputs and real data distributions underline the high performance of KeyGAN. CONCLUSION KeyGAN presents strong potential as a realistic keystroke data synthesizer, displaying impressive capability to reproduce complex typing patterns relevant to biomarkers for neurological disorders, like Parkinson's Disease. The ability of its synthetic data to effectively supplement real data for training algorithms without affecting performance implies significant promise for advancing research in digital biomarkers for neurodegenerative and psychomotor disorders.
Collapse
Affiliation(s)
- Alejandro Acien
- Area 2 AI Corporation, 245 Main Street, Cambridge, 02142, MA, United States.
| | - Aythami Morales
- Universidad Autonoma de Madrid, School of Engineering, Madrid, 28049, Spain
| | - Luca Giancardo
- McWilliams School of Biomedical Informatics, University of Texas Health Science Center, Houston, 77030, TX, United States
| | | | - Ashley A Holmes
- ProKidney Corporation, 3929 W Pt Blvd, Winston-Salem, 27103, NC, United States
| | - Julian Fierrez
- Universidad Autonoma de Madrid, School of Engineering, Madrid, 28049, Spain
| | | |
Collapse
|
9
|
Jaradat AS, Wedyan M, Alomari S, Barhoush MM. Using Machine Learning to Diagnose Autism Based on Eye Tracking Technology. Diagnostics (Basel) 2024; 15:66. [PMID: 39795594 PMCID: PMC11719697 DOI: 10.3390/diagnostics15010066] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2024] [Revised: 11/22/2024] [Accepted: 11/26/2024] [Indexed: 01/13/2025] Open
Abstract
Background/Objectives: One of the key challenges in autism is early diagnosis. Early diagnosis leads to early interventions that improve the condition and not worsen autism in the future. Currently, autism diagnoses are based on monitoring by a doctor or specialist after the child reaches a certain age exceeding three years after the parents observe the child's abnormal behavior. Methods: The paper aims to find another way to diagnose autism that is effective and earlier than traditional methods of diagnosis. Therefore, we used the Eye Gaze fixes map dataset and Eye Tracking Scanpath dataset (ETSDS) to diagnose Autistic Spectrum Disorder (ASDs), while a subset of the ETSDS was used to recognize autism scores. Results: The experimental results showed that the higher accuracy rate reached 96.1% and 98.0% for the hybrid model on Eye Gaze fixes map datasets and ETSDS, respectively. A higher accuracy rate was reached (98.1%) on the ETSDS used to recognize autism scores. Furthermore, the results showed the outperformer for the proposed method results compared to previous works. Conclusions: This confirms the effectiveness of using artificial intelligence techniques in diagnosing diseases in general and diagnosing autism, in addition to the need to increase research in the field of diagnosing diseases using advanced techniques.
Collapse
Affiliation(s)
| | - Mohammad Wedyan
- Computer Science Department, Yarmouk University, Irbid 21163, Jordan; (A.S.J.); (S.A.); (M.M.B.)
| | | | | |
Collapse
|
10
|
Tur K. Multi-Modal Machine Learning Approach for COVID-19 Detection Using Biomarkers and X-Ray Imaging. Diagnostics (Basel) 2024; 14:2800. [PMID: 39767161 PMCID: PMC11674685 DOI: 10.3390/diagnostics14242800] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/05/2024] [Revised: 12/11/2024] [Accepted: 12/12/2024] [Indexed: 01/11/2025] Open
Abstract
Background: Accurate and rapid detection of COVID-19 remains critical for clinical management, especially in resource-limited settings. Current diagnostic methods face challenges in terms of speed and reliability, creating a need for complementary AI-based models that integrate diverse data sources. Objectives: This study aimed to develop and evaluate a multi-modal machine learning model that combines clinical biomarkers and chest X-ray images to enhance diagnostic accuracy and provide interpretable insights. Methods: We used a dataset of 250 patients (180 COVID-19 positive and 70 negative cases) collected from clinical settings. Biomarkers such as CRP, ferritin, NLR, and albumin were included alongside chest X-ray images. Random Forest and Gradient Boosting models were used for biomarkers, and ResNet and VGG CNN architectures were applied to imaging data. A late-fusion strategy integrated predictions from these modalities. Stratified k-fold cross-validation ensured robust evaluation while preventing data leakage. Model performance was assessed using AUC-ROC, F1-score, Specificity, Negative Predictive Value (NPV), and Matthews Correlation Coefficient (MCC), with confidence intervals calculated via bootstrap resampling. Results: The Gradient Boosting + VGG fusion model achieved the highest performance, with an AUC-ROC of 0.94, F1-score of 0.93, Specificity of 93%, NPV of 96%, and MCC of 0.91. SHAP and LIME interpretability analyses identified CRP, ferritin, and specific lung regions as key contributors to predictions. Discussion: The proposed multi-modal approach significantly enhances diagnostic accuracy compared to single-modality models. Its interpretability aligns with clinical understanding, supporting its potential for real-world application.
Collapse
Affiliation(s)
- Kagan Tur
- Internal Medicine Department, Faculty of Medicine, Ahi Evran University, Kirsehir 40200, Turkey
| |
Collapse
|
11
|
Smolyak D, Bjarnadóttir MV, Crowley K, Agarwal R. Large language models and synthetic health data: progress and prospects. JAMIA Open 2024; 7:ooae114. [PMID: 39464796 PMCID: PMC11512648 DOI: 10.1093/jamiaopen/ooae114] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2024] [Revised: 06/27/2024] [Accepted: 10/11/2024] [Indexed: 10/29/2024] Open
Abstract
Objectives Given substantial obstacles surrounding health data acquisition, high-quality synthetic health data are needed to meet a growing demand for the application of advanced analytics for clinical discovery, prediction, and operational excellence. We highlight how recent advances in large language models (LLMs) present new opportunities for progress, as well as new risks, in synthetic health data generation (SHDG). Materials and Methods We synthesized systematic scoping reviews in the SHDG domain, recent LLM methods for SHDG, and papers investigating the capabilities and limits of LLMs. Results We summarize the current landscape of generative machine learning models (eg, Generative Adversarial Networks) for SHDG, describe remaining challenges and limitations, and identify how recent LLM approaches can potentially help mitigate them. Discussion Six research directions are outlined for further investigation of LLMs for SHDG: evaluation metrics, LLM adoption, data efficiency, generalization, health equity, and regulatory challenges. Conclusion LLMs have already demonstrated both high potential and risks in the health domain, and it is important to study their advantages and disadvantages for SHDG.
Collapse
Affiliation(s)
- Daniel Smolyak
- Department of Computer Science, University of Maryland, College Park, College Park, MD 20742, United States
| | - Margrét V Bjarnadóttir
- Robert H. Smith School of Business, University of Maryland, College Park, College Park, MD 20740, United States
| | - Kenyon Crowley
- Accenture Federal Services, Arlington, VA 22203, United States
| | - Ritu Agarwal
- Center for Digital Health and Artificial Intelligence, Carey Business School, Johns Hopkins University, Baltimore, MD 21202, United States
| |
Collapse
|
12
|
Gupta KA, Ikonomidou VN, Glancey M, Faiman R, Talafha S, Ford T, Jenkins T, Goodwin A. Mosquito species identification accuracy of early deployed algorithms in IDX, A vector identification tool. Acta Trop 2024; 260:107392. [PMID: 39255861 DOI: 10.1016/j.actatropica.2024.107392] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2024] [Revised: 09/06/2024] [Accepted: 09/07/2024] [Indexed: 09/12/2024]
Abstract
Mosquito-borne diseases continue to pose a great threat to global public health systems due to increased insecticide resistance and climate change. Accurate vector identification is crucial for effective control, yet it presents significant challenges. IDX - an automated computer vision-based device capable of capturing mosquito images and outputting mosquito species ID has been deployed globally resulting in algorithms currently capable of identifying 53 mosquito species. In this study, we evaluate deployed performance of the IDX mosquito species identification algorithms using data from partners in the Southeastern United States (SE US) and Papua New Guinea (PNG) in 2023 and 2024. This preliminary assessment indicates continued improvement of the IDX mosquito species identification algorithms over the study period for individual species as well as average regional accuracy with macro average recall improving from 55.3 % [Confidence Interval (CI) 48.9, 61.7] to 80.2 % [CI 77.3, 84.9] for SE US, and 84.1 % [CI 75.1, 93.1] to 93.6 % [CI 91.6, 95.6] for PNG using a CI of 90 %. This study underscores the importance of algorithm refinement and dataset expansion covering more species and regions to enhance identification systems thereby reducing the workload for human experts, addressing taxonomic expertise gaps, and improving vector control efforts.
Collapse
|
13
|
Akpinar MH, Sengur A, Salvi M, Seoni S, Faust O, Mir H, Molinari F, Acharya UR. Synthetic Data Generation via Generative Adversarial Networks in Healthcare: A Systematic Review of Image- and Signal-Based Studies. IEEE OPEN JOURNAL OF ENGINEERING IN MEDICINE AND BIOLOGY 2024; 6:183-192. [PMID: 39698120 PMCID: PMC11655107 DOI: 10.1109/ojemb.2024.3508472] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/01/2024] [Revised: 11/13/2024] [Accepted: 11/26/2024] [Indexed: 12/20/2024] Open
Abstract
Generative Adversarial Networks (GANs) have emerged as a powerful tool in artificial intelligence, particularly for unsupervised learning. This systematic review analyzes GAN applications in healthcare, focusing on image and signal-based studies across various clinical domains. Following Preferred Reporting Items for Systematic reviews and Meta-Analyses (PRISMA) guidelines, we reviewed 72 relevant journal articles. Our findings reveal that magnetic resonance imaging (MRI) and electrocardiogram (ECG) signal acquisition techniques were most utilized, with brain studies (22%), cardiology (18%), cancer (15%), ophthalmology (12%), and lung studies (10%) being the most researched areas. We discuss key GAN architectures, including cGAN (31%) and CycleGAN (18%), along with datasets, evaluation metrics, and performance outcomes. The review highlights promising data augmentation, anonymization, and multi-task learning results. We identify current limitations, such as the lack of standardized metrics and direct comparisons, and propose future directions, including the development of no-reference metrics, immersive simulation scenarios, and enhanced interpretability.
Collapse
Affiliation(s)
- Muhammed Halil Akpinar
- Vocational School of Technical SciencesIstanbul University-Cerrahpasa34320IstanbulTürkiye
| | | | - Massimo Salvi
- Department of Electronics and TelecommunicationsPolitecnico di Torino10129TurinItaly
| | - Silvia Seoni
- Department of Electronics and TelecommunicationsPolitecnico di Torino10129TurinItaly
| | - Oliver Faust
- Anglia Ruskin University Cambridge CampusCB1 1PTCambridgeU.K.
| | - Hasan Mir
- American University of SharjahSharjah26666UAE
| | - Filippo Molinari
- Department of Electronics and TelecommunicationsPolitecnico di Torino10129TurinItaly
| | | |
Collapse
|
14
|
Al-Haddad AA, Al-Haddad LA, Al-Haddad SA, Jaber AA, Khan ZH, Rehman HZU. Towards dental diagnostic systems: Synergizing wavelet transform with generative adversarial networks for enhanced image data fusion. Comput Biol Med 2024; 182:109241. [PMID: 39362006 DOI: 10.1016/j.compbiomed.2024.109241] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/23/2024] [Revised: 09/05/2024] [Accepted: 09/30/2024] [Indexed: 10/05/2024]
Abstract
The advent of precision diagnostics in pediatric dentistry is shifting towards ensuring early detection of dental diseases, a critical factor in safeguarding the oral health of the younger population. In this study, an innovative approach is introduced, wherein Discrete Wavelet Transform (DWT) and Generative Adversarial Networks (GANs) are synergized within an Image Data Fusion (IDF) framework to enhance the accuracy of dental disease diagnosis through dental diagnostic systems. Dental panoramic radiographs from pediatric patients were utilized to demonstrate how the integration of DWT and GANs can significantly improve the informativeness of dental images. In the IDF process, the original images, GAN-augmented images, and wavelet-transformed images are combined to create a comprehensive dataset. DWT was employed for the decomposition of images into frequency components to enhance the visibility of subtle pathological features. Simultaneously, GANs were used to augment the dataset with high-quality, synthetic radiographic images indistinguishable from real ones, to provide robust data training. These integrated images are then fed into an Artificial Neural Network (ANN) for the classification of dental diseases. The utilization of the ANN in this context demonstrates the system's robustness and culminates in achieving an unprecedented accuracy rate of 0.897, 0.905 precision, recall of 0.897, and specificity of 0.968. Additionally, this study explores the feasibility of embedding the diagnostic system into dental X-ray scanners by leveraging lightweight models and cloud-based solutions to minimize resource constraints. Such integration is posited to revolutionize dental care by providing real-time, accurate disease detection capabilities, which significantly reduces diagnostical delays and enhances treatment outcomes.
Collapse
Affiliation(s)
| | - Luttfi A Al-Haddad
- Training and Workshops Center, University of Technology- Iraq, Baghdad, Iraq.
| | - Sinan A Al-Haddad
- Civil Engineering Department, University of Technology- Iraq, Baghdad, Iraq
| | - Alaa Abdulhady Jaber
- Mechanical Engineering Department, University of Technology- Iraq, Baghdad, Iraq
| | - Zeashan Hameed Khan
- Interdisciplinary Research Center for Intelligent Manufacturing & Robotics (IRC-IMR), King Fahd University of Petroleum & Minerals (KFUPM), Dhahran, Saudi Arabia
| | - Hafiz Zia Ur Rehman
- Department of Mechatronics and Biomedical Engineering, Air University (AU), Islamabad, Pakistan
| |
Collapse
|
15
|
Gopinath K, Hoopes A, Alexander DC, Arnold SE, Balbastre Y, Billot B, Casamitjana A, Cheng Y, Chua RYZ, Edlow BL, Fischl B, Gazula H, Hoffmann M, Keene CD, Kim S, Kimberly WT, Laguna S, Larson KE, Van Leemput K, Puonti O, Rodrigues LM, Rosen MS, Tregidgo HFJ, Varadarajan D, Young SI, Dalca AV, Iglesias JE. Synthetic data in generalizable, learning-based neuroimaging. IMAGING NEUROSCIENCE (CAMBRIDGE, MASS.) 2024; 2:1-22. [PMID: 39850547 PMCID: PMC11752692 DOI: 10.1162/imag_a_00337] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 01/19/2024] [Revised: 09/20/2024] [Accepted: 09/20/2024] [Indexed: 01/25/2025]
Abstract
Synthetic data have emerged as an attractive option for developing machine-learning methods in human neuroimaging, particularly in magnetic resonance imaging (MRI)-a modality where image contrast depends enormously on acquisition hardware and parameters. This retrospective paper reviews a family of recently proposed methods, based on synthetic data, for generalizable machine learning in brain MRI analysis. Central to this framework is the concept of domain randomization, which involves training neural networks on a vastly diverse array of synthetically generated images with random contrast properties. This technique has enabled robust, adaptable models that are capable of handling diverse MRI contrasts, resolutions, and pathologies, while working out-of-the-box, without retraining. We have successfully applied this method to tasks such as whole-brain segmentation (SynthSeg), skull-stripping (SynthStrip), registration (SynthMorph, EasyReg), super-resolution, and MR contrast transfer (SynthSR). Beyond these applications, the paper discusses other possible use cases and future work in our methodology. Neural networks trained with synthetic data enable the analysis of clinical MRI, including large retrospective datasets, while greatly alleviating (and sometimes eliminating) the need for substantial labeled datasets, and offer enormous potential as robust tools to address various research goals.
Collapse
Affiliation(s)
- Karthik Gopinath
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | - Andrew Hoopes
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
- Massachusetts Institute of Technology, Cambridge, MA, United States
| | | | - Steven E. Arnold
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | - Yael Balbastre
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | - Benjamin Billot
- Computer Science & Artificial Intelligence Laboratory, Massachusetts Institute of Technology, Cambridge, MA, United States
| | | | - You Cheng
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | - Russ Yue Zhi Chua
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
- Massachusetts Institute of Technology, Cambridge, MA, United States
| | - Brian L. Edlow
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | - Bruce Fischl
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | | | - Malte Hoffmann
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | - C. Dirk Keene
- University of Washington, Seattle, WA, United States
| | | | - W. Taylor Kimberly
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | | | - Kathleen E. Larson
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | - Koen Van Leemput
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | - Oula Puonti
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
- Copenhagen University Hospital, København, Denmark
| | - Livia M. Rodrigues
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
- Universidade Estadual de Campinas, São Paulo, Brazil
| | - Matthew S. Rosen
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | | | - Divya Varadarajan
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | - Sean I. Young
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
- Massachusetts Institute of Technology, Cambridge, MA, United States
| | - Adrian V. Dalca
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
- Massachusetts Institute of Technology, Cambridge, MA, United States
| | - Juan Eugenio Iglesias
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
- Massachusetts Institute of Technology, Cambridge, MA, United States
- University College London, London, England
| |
Collapse
|
16
|
Feng S, Huang Y, Zhang N. An Improved YOLOv8 OBB Model for Ship Detection through Stable Diffusion Data Augmentation. SENSORS (BASEL, SWITZERLAND) 2024; 24:5850. [PMID: 39275764 PMCID: PMC11397737 DOI: 10.3390/s24175850] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/30/2024] [Revised: 09/03/2024] [Accepted: 09/04/2024] [Indexed: 09/16/2024]
Abstract
Unmanned aerial vehicles (UAVs) with cameras offer extensive monitoring capabilities and exceptional maneuverability, making them ideal for real-time ship detection and effective ship management. However, ship detection by camera-equipped UAVs faces challenges when it comes to multi-viewpoints, multi-scales, environmental variability, and dataset scarcity. To overcome these challenges, we proposed a data augmentation method based on stable diffusion to generate new images for expanding the dataset. Additionally, we improve the YOLOv8n OBB model by incorporating the BiFPN structure and EMA module, enhancing its ability to detect multi-viewpoint and multi-scale ship instances. Through multiple comparative experiments, we evaluated the effectiveness of our proposed data augmentation method and the improved model. The results indicated that our proposed data augmentation method is effective for low-volume datasets with complex object features. The YOLOv8n-BiFPN-EMA OBB model we proposed performed well in detecting multi-viewpoint and multi-scale ship instances, achieving the mAP (@0.5) of 92.3%, the mAP (@0.5:0.95) of 77.5%, a reduction of 0.8 million in model parameters, and a detection speed that satisfies real-time ship detection requirements.
Collapse
Affiliation(s)
- Sang Feng
- School of Electromechanical Engineering, Guangdong University of Technology, Guangzhou 510006, China
| | - Yi Huang
- School of Electromechanical Engineering, Guangdong University of Technology, Guangzhou 510006, China
| | - Ning Zhang
- School of Electromechanical Engineering, Guangdong University of Technology, Guangzhou 510006, China
| |
Collapse
|
17
|
Merrouchi M, Benyoussef Y, Skittou M, Atifi K, Gadi T. ConvCoroNet: a deep convolutional neural network optimized with iterative thresholding algorithm for Covid-19 detection using chest X-ray images. J Biomol Struct Dyn 2024; 42:5699-5712. [PMID: 37354142 DOI: 10.1080/07391102.2023.2227726] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/30/2022] [Accepted: 06/15/2023] [Indexed: 06/26/2023]
Abstract
Covid-19 is a global pandemic. Early and accurate detection of positive cases prevent the further spread of this epidemic and help to treat rapidly the infected patients. During the peak of this epidemic, there was an insufficiency of Covid-19 test kits. In addition, this technique takes a considerable time in the diagnosis. Hence the need to find fast, accurate and low-cost method to replace or supplement RT PCR-based methods. Covid-19 is a respiratory disease, chest X-ray images are often used to diagnose pneumonia. From this perspective, these images can play an important role in the Covid-19 detection. In this article, we propose ConvCoroNet, a deep convolutional neural network model optimized with new method based on iterative thresholding algorithm to detect coronavirus automatically from chest X-ray images. ConvCoroNet is trained on a dataset prepared by collecting chest X-ray images of Covid-19, pneumonia and normal cases from publically datasets. The experimental results of our proposed model show a high accuracy of 99.50%, sensitivity of 98.80% and specificity of 99.85% when detecting Covid-19 from chest X-ray images. ConvCoroNet achieves promising results in the automatic detection of Covid-19 from chest X-ray images. It may be able to help radiologists in the Covid-19 detection by reducing the examination time of X-ray images.Communicated by Ramaswamy H. Sarma.
Collapse
Affiliation(s)
- M Merrouchi
- Faculty of Science and Technology, Hassan First, Settat, Morocco
| | - Y Benyoussef
- National School of Applied Sciences, Hassan First, Berrechid, Morocco
| | - M Skittou
- Faculty of Science and Technology, Hassan First, Settat, Morocco
| | - K Atifi
- Faculty of Science and Technology, Hassan First, Settat, Morocco
| | - T Gadi
- Faculty of Science and Technology, Hassan First, Settat, Morocco
| |
Collapse
|
18
|
Chung J, Zhang J, Saimon AI, Liu Y, Johnson BN, Kong Z. Imbalanced spectral data analysis using data augmentation based on the generative adversarial network. Sci Rep 2024; 14:13230. [PMID: 38853181 PMCID: PMC11163007 DOI: 10.1038/s41598-024-63285-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/18/2024] [Accepted: 05/27/2024] [Indexed: 06/11/2024] Open
Abstract
Spectroscopic techniques generate one-dimensional spectra with distinct peaks and specific widths in the frequency domain. These features act as unique identities for material characteristics. Deep neural networks (DNNs) has recently been considered a powerful tool for automatically categorizing experimental spectra data by supervised classification to evaluate material characteristics. However, most existing work assumes balanced spectral data among various classes in the training data, contrary to actual experiments, where the spectral data is usually imbalanced. The imbalanced training data deteriorates the supervised classification performance, hindering understanding of the phase behavior, specifically, sol-gel transition (gelation) of soft materials and glycomaterials. To address this issue, this paper applies a novel data augmentation method based on a generative adversarial network (GAN) proposed by the authors in their prior work. To demonstrate the effectiveness of the proposed method, the actual imbalanced spectral data from Pluronic F-127 hydrogel and Alpha-Cyclodextrin hydrogel are used to classify the phases of data. Specifically, our approach improves 8.8%, 6.4%, and 6.2% of the performance of the existing data augmentation methods regarding the classifier's F-score, Precision, and Recall on average, respectively. Specifically, our method consists of three DNNs: the generator, discriminator, and classifier. The method generates samples that are not only authentic but emphasize the differentiation between material characteristics to provide balanced training data, improving the classification results. Based on these validated results, we expect the method's broader applications in addressing imbalanced measurement data across diverse domains in materials science and chemical engineering.
Collapse
Affiliation(s)
- Jihoon Chung
- Department of Industrial Engineering, Pusan National University, Busan, South Korea
| | - Junru Zhang
- Grado Department of Industrial and Systems Engineering, Virginia Tech, Blacksburg, VA, USA
| | - Amirul Islam Saimon
- Grado Department of Industrial and Systems Engineering, Virginia Tech, Blacksburg, VA, USA
| | - Yang Liu
- Grado Department of Industrial and Systems Engineering, Virginia Tech, Blacksburg, VA, USA
| | - Blake N Johnson
- Grado Department of Industrial and Systems Engineering, Virginia Tech, Blacksburg, VA, USA.
| | - Zhenyu Kong
- Grado Department of Industrial and Systems Engineering, Virginia Tech, Blacksburg, VA, USA.
| |
Collapse
|
19
|
Fu Z, Liu Z, Ping S, Li W, Liu J. TRA-ACGAN: A motor bearing fault diagnosis model based on an auxiliary classifier generative adversarial network and transformer network. ISA TRANSACTIONS 2024; 149:381-393. [PMID: 38604873 DOI: 10.1016/j.isatra.2024.03.033] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/10/2023] [Revised: 03/26/2024] [Accepted: 03/27/2024] [Indexed: 04/13/2024]
Abstract
Motor bearing fault diagnosis is essential to guarantee production efficiency and avoid catastrophic accidents. Deep learning-based methods have been developed and widely used for fault diagnosis, and these methods have proven to be very effective in accurately diagnosing bearing faults. In this paper, study the application of generative adversarial networks (GANs) in motor bearing fault diagnosis to address the practical issue of insufficient fault data in industrial testing. Focus on the auxiliary classifier generative adversarial network (ACGAN), and the data expansion is carried out for small datasets. This paper present a novel transformer network and auxiliary classifier generative adversarial network (TRA-ACGAN) for motor bearing fault diagnosis, where the TRA-ACGAN combines an ACGAN with a transformer network to avoid the traditional iterative and convolutional structures. The attention mechanism is fully utilized to extract more effective features, and the dual-task coupling problem encountered in classical ACGANs is avoided. Experimental results with the CWRU dataset and the PU dataset in the field of motor bearing fault diagnosis demonstrate the suitability and superiority of the TRA-ACGAN.
Collapse
Affiliation(s)
- Zhaoyang Fu
- School of Automation, Northwestern Polytechnical University, Xi'an 710129, China.
| | - Zheng Liu
- Luoyang Power Supply Company, State Grid Corporation of China, Luoyang 471023, China
| | - Shuangrui Ping
- School of Automation, Northwestern Polytechnical University, Xi'an 710129, China
| | - Weilin Li
- School of Automation, Northwestern Polytechnical University, Xi'an 710129, China
| | - Jinglin Liu
- School of Automation, Northwestern Polytechnical University, Xi'an 710129, China
| |
Collapse
|
20
|
Kosar A, Asif M, Ahmad MB, Akram W, Mahmood K, Kumari S. Towards classification and comprehensive analysis of AI-based COVID-19 diagnostic techniques: A survey. Artif Intell Med 2024; 151:102858. [PMID: 38583369 DOI: 10.1016/j.artmed.2024.102858] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/22/2023] [Revised: 01/02/2024] [Accepted: 03/25/2024] [Indexed: 04/09/2024]
Abstract
The unpredictable pandemic came to light at the end of December 2019, known as the novel coronavirus, also termed COVID-19, identified by the World Health Organization (WHO). The virus first originated in Wuhan (China) and rapidly affected most of the world's population. This outbreak's impact is experienced worldwide because it causes high mortality risk, many cases, and economic falls. Around the globe, the total number of cases and deaths reported till November 12, 2022, were >600 million and 6.6 million, respectively. During the period of COVID-19, several diverse diagnostic techniques have been proposed. This work presents a systematic review of COVID-19 diagnostic techniques in response to such acts. Initially, these techniques are classified into different categories based on their working principle and detection modalities, i.e. chest X-ray imaging, cough sound or respiratory patterns, RT-PCR, antigen testing, and antibody testing. After that, a comparative analysis is performed to evaluate these techniques' efficacy which may help to determine an optimum solution for a particular scenario. The findings of the proposed work show that Artificial Intelligence plays a vital role in developing COVID-19 diagnostic techniques which support the healthcare system. The related work can be a footprint for all the researchers, available under a single umbrella. Additionally, all the techniques are long-lasting and can be used for future pandemics.
Collapse
Affiliation(s)
- Amna Kosar
- Department of Computer Science, Lahore Garrison University, Lahore, Pakistan
| | - Muhammad Asif
- Department of Computer Science, Lahore Garrison University, Lahore, Pakistan
| | - Maaz Bin Ahmad
- College of Computing and Information Sciences, Karachi Institute of Economics and Technology (KIET), Karachi, Pakistan
| | - Waseem Akram
- Graduate School of Engineering Science and Technology, National Yunlin University of Science and Technology, Douliu, Taiwan, ROC
| | - Khalid Mahmood
- Graduate School of Intelligent Data Science, National Yunlin University of Science and Technology, Douliu, Taiwan, ROC.
| | - Saru Kumari
- Departement of Mathematics, Chaudhary Charan Singh University, Meerut, India
| |
Collapse
|
21
|
Assaf JF, Yazbeck H, Reinstein DZ, Archer TJ, Arbelaez J, Bteich Y, Arbelaez MC, Abou Mrad A, Awwad ST. Enhancing the Automated Detection of Implantable Collamer Lens Vault Using Generative Adversarial Networks and Synthetic Data on Optical Coherence Tomography. J Refract Surg 2024; 40:e199-e207. [PMID: 38593258 DOI: 10.3928/1081597x-20240214-01] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/11/2024]
Abstract
PURPOSE To investigate the efficacy of incorporating Generative Adversarial Network (GAN) and synthetic images in enhancing the performance of a convolutional neural network (CNN) for automated estimation of Implantable Collamer Lens (ICL) vault using anterior segment optical coherence tomography (AS-OCT). METHODS This study was a retrospective evaluation using synthetic data and real patient images in a deep learning framework. Synthetic ICL AS-OCT scans were generated using GANs and a secondary image editing algorithm, creating approximately 100,000 synthetic images. These were used alongside real patient scans to train a CNN for estimating ICL vault distance. The model's performance was evaluated using statistical metrics such as mean absolute percentage error (MAPE), mean absolute error (MAE), root mean squared error (RMSE), and coefficient of determination (R2) for the estimation of ICL vault distance. RESULTS The study analyzed 4,557 AS-OCT B-scans from 138 eyes of 103 patients for training. An independent, retrospectively collected dataset of 2,454 AS-OCT images from 88 eyes of 56 patients, used prospectively for evaluation, served as the test set. When trained solely on real images, the CNN achieved a MAPE of 15.31%, MAE of 44.68 µm, and RMSE of 63.3 µm. However, with the inclusion of GAN-generated and algorithmically edited synthetic images, the performance significantly improved, achieving a MAPE of 8.09%, MAE of 24.83 µm, and RMSE of 32.26 µm. The R2 value was +0.98, indicating a strong positive correlation between actual and predicted ICL vault distances (P < .01). No statistically significant difference was observed between measured and predicted vault values (P = .58). CONCLUSIONS The integration of GAN-generated and edited synthetic images substantially enhanced ICL vault estimation, demonstrating the efficacy of GANs and synthetic data in enhancing OCT image analysis accuracy. This model not only shows potential for assisting postoperative ICL evaluations, but also for improving OCT automation when data paucity is an issue. [J Refract Surg. 2024;40(4):e199-e207.].
Collapse
|
22
|
Arefin MS, Rahman MM, Hasan MT, Mahmud M. A Topical Review on Enabling Technologies for the Internet of Medical Things: Sensors, Devices, Platforms, and Applications. MICROMACHINES 2024; 15:479. [PMID: 38675290 PMCID: PMC11051832 DOI: 10.3390/mi15040479] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/24/2023] [Revised: 03/17/2024] [Accepted: 03/22/2024] [Indexed: 04/28/2024]
Abstract
The Internet of Things (IoT) is still a relatively new field of research, and its potential to be used in the healthcare and medical sectors is enormous. In the last five years, IoT has been a go-to option for various applications such as using sensors for different features, machine-to-machine communication, etc., but precisely in the medical sector, it is still lagging far behind compared to other sectors. Hence, this study emphasises IoT applications in medical fields, Medical IoT sensors and devices, IoT platforms for data visualisation, and artificial intelligence in medical applications. A systematic review considering PRISMA guidelines on research articles as well as the websites on IoMT sensors and devices has been carried out. After the year 2001, an integrated outcome of 986 articles was initially selected, and by applying the inclusion-exclusion criterion, a total of 597 articles were identified. 23 new studies have been finally found, including records from websites and citations. This review then analyses different sensor monitoring circuits in detail, considering an Intensive Care Unit (ICU) scenario, device applications, and the data management system, including IoT platforms for the patients. Lastly, detailed discussion and challenges have been outlined, and possible prospects have been presented.
Collapse
Affiliation(s)
- Md. Shamsul Arefin
- Department of Electrical and Electronic Engineering (EEE), Bangladesh University of Business & Technology, Dhaka 1216, Bangladesh;
| | | | - Md. Tanvir Hasan
- Department of Electrical and Electronic Engineering (EEE), Jashore University of Science & Technology, Jashore 7408, Bangladesh;
- Department of Electrical Engineering, University of South Carolina, Columbia, SC 29208, USA
| | - Mufti Mahmud
- Department of Computer Science, Nottingham Trent University, Nottingham NG11 8NS, UK
- Computing and Informatics Research Centre, Nottingham Trent University, Nottingham NG11 8NS, UK
- Medical Technologies Innovation Facility, Nottingham Trent University, Nottingham NG11 8NS, UK
| |
Collapse
|
23
|
Kabir MM, Mridha M, Rahman A, Hamid MA, Monowar MM. Detection of COVID-19, pneumonia, and tuberculosis from radiographs using AI-driven knowledge distillation. Heliyon 2024; 10:e26801. [PMID: 38444490 PMCID: PMC10912466 DOI: 10.1016/j.heliyon.2024.e26801] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/10/2023] [Revised: 01/30/2024] [Accepted: 02/20/2024] [Indexed: 03/07/2024] Open
Abstract
Chest radiography is an essential diagnostic tool for respiratory diseases such as COVID-19, pneumonia, and tuberculosis because it accurately depicts the structures of the chest. However, accurate detection of these diseases from radiographs is a complex task that requires the availability of medical imaging equipment and trained personnel. Conventional deep learning models offer a viable automated solution for this task. However, the high complexity of these models often poses a significant obstacle to their practical deployment within automated medical applications, including mobile apps, web apps, and cloud-based platforms. This study addresses and resolves this dilemma by reducing the complexity of neural networks using knowledge distillation techniques (KDT). The proposed technique trains a neural network on an extensive collection of chest X-ray images and propagates the knowledge to a smaller network capable of real-time detection. To create a comprehensive dataset, we have integrated three popular chest radiograph datasets with chest radiographs for COVID-19, pneumonia, and tuberculosis. Our experiments show that this knowledge distillation approach outperforms conventional deep learning methods in terms of computational complexity and performance for real-time respiratory disease detection. Specifically, our system achieves an impressive average accuracy of 0.97, precision of 0.94, and recall of 0.97.
Collapse
Affiliation(s)
- Md Mohsin Kabir
- Department of Computer Science & Engineering, Bangladesh University of Business & Technology, Dhaka-1216, Bangladesh
| | - M.F. Mridha
- Department of Computer Science, American International University-Bangladesh, Dhaka-1229, Bangladesh
| | - Ashifur Rahman
- Department of Computer Science & Engineering, Bangladesh University of Business & Technology, Dhaka-1216, Bangladesh
| | - Md. Abdul Hamid
- Department of Information Technology, Faculty of Computing & Information Technology, King Abdulaziz University, Jeddah-21589, Kingdom of Saudi Arabia
| | - Muhammad Mostafa Monowar
- Department of Information Technology, Faculty of Computing & Information Technology, King Abdulaziz University, Jeddah-21589, Kingdom of Saudi Arabia
| |
Collapse
|
24
|
Singh K, Kaur N, Prabhu A. Combating COVID-19 Crisis using Artificial Intelligence (AI) Based Approach: Systematic Review. Curr Top Med Chem 2024; 24:737-753. [PMID: 38318824 DOI: 10.2174/0115680266282179240124072121] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/18/2023] [Revised: 12/19/2023] [Accepted: 12/27/2023] [Indexed: 02/07/2024]
Abstract
BACKGROUND SARS-CoV-2, the unique coronavirus that causes COVID-19, has wreaked damage around the globe, with victims displaying a wide range of difficulties that have encouraged medical professionals to look for innovative technical solutions and therapeutic approaches. Artificial intelligence-based methods have contributed a significant part in tackling complicated issues, and some institutions have been quick to embrace and tailor these solutions in response to the COVID-19 pandemic's obstacles. Here, in this review article, we have covered a few DL techniques for COVID-19 detection and diagnosis, as well as ML techniques for COVID-19 identification, severity classification, vaccine and drug development, mortality rate prediction, contact tracing, risk assessment, and public distancing. This review illustrates the overall impact of AI/ML tools on tackling and managing the outbreak. PURPOSE The focus of this research was to undertake a thorough evaluation of the literature on the part of Artificial Intelligence (AI) as a complete and efficient solution in the battle against the COVID-19 epidemic in the domains of detection and diagnostics of disease, mortality prediction and vaccine as well as drug development. METHODS A comprehensive exploration of PubMed, Web of Science, and Science Direct was conducted using PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analysis) regulations to find all possibly suitable papers conducted and made publicly available between December 1, 2019, and August 2023. COVID-19, along with AI-specific words, was used to create the query syntax. RESULTS During the period covered by the search strategy, 961 articles were published and released online. Out of these, a total of 135 papers were chosen for additional investigation. Mortality rate prediction, early detection and diagnosis, vaccine as well as drug development, and lastly, incorporation of AI for supervising and controlling the COVID-19 pandemic were the four main topics focused entirely on AI applications used to tackle the COVID-19 crisis. Out of 135, 60 research papers focused on the detection and diagnosis of the COVID-19 pandemic. Next, 19 of the 135 studies applied a machine-learning approach for mortality rate prediction. Another 22 research publications emphasized the vaccine as well as drug development. Finally, the remaining studies were concentrated on controlling the COVID-19 pandemic by applying AI AI-based approach to it. CONCLUSION We compiled papers from the available COVID-19 literature that used AI-based methodologies to impart insights into various COVID-19 topics in this comprehensive study. Our results suggest crucial characteristics, data types, and COVID-19 tools that can aid in medical and translational research facilitation.
Collapse
Affiliation(s)
- Kavya Singh
- Department of Biotechnology, Banasthali University, Banasthali Vidyapith, Banasthali, 304022, Rajasthan, India
| | - Navjeet Kaur
- Department of Chemistry & Division of Research and Development, Lovely Professional University, Phagwara, 144411, Punjab, India
| | - Ashish Prabhu
- Biotechnology Department, NIT Warangal, Warangal, 506004, Telangana, India
| |
Collapse
|
25
|
Gopatoti A, Jayakumar R, Billa P, Patteeswaran V. DDA-SSNets: Dual decoder attention-based semantic segmentation networks for COVID-19 infection segmentation and classification using chest X-Ray images. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2024; 32:623-649. [PMID: 38607728 DOI: 10.3233/xst-230421] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/14/2024]
Abstract
BACKGROUND COVID-19 needs to be diagnosed and staged to be treated accurately. However, prior studies' diagnostic and staging abilities for COVID-19 infection needed to be improved. Therefore, new deep learning-based approaches are required to aid radiologists in detecting and quantifying COVID-19-related lung infections. OBJECTIVE To develop deep learning-based models to classify and quantify COVID-19-related lung infections. METHODS Initially, Dual Decoder Attention-based Semantic Segmentation Networks (DDA-SSNets) such as Dual Decoder Attention-UNet (DDA-UNet) and Dual Decoder Attention-SegNet (DDA-SegNet) are proposed to facilitate the dual segmentation tasks such as lung lobes and infection segmentation in chest X-ray (CXR) images. The lung lobe and infection segmentations are mapped to grade the severity of COVID-19 infection in both the lungs of CXRs. Later, a Genetic algorithm-based Deep Convolutional Neural Network classifier with the optimum number of layers, namely GADCNet, is proposed to classify the extracted regions of interest (ROI) from the CXR lung lobes into COVID-19 and non-COVID-19. RESULTS The DDA-SegNet shows better segmentation with an average BCSSDC of 99.53% and 99.97% for lung lobes and infection segmentations, respectively, compared with DDA-UNet with an average BCSSDC of 99.14% and 99.92%. The proposed DDA-SegNet with GADCNet classifier offered excellent classification results with an average BCCAC of 99.98%, followed by the GADCNet with DDA-UNet with an average BCCAC of 99.92% after extensive testing and analysis. CONCLUSIONS The results show that the proposed DDA-SegNet has superior performance in the segmentation of lung lobes and COVID-19-infected regions in CXRs, along with improved severity grading compared to the DDA-UNet and improved accuracy of the GADCNet classifier in classifying the CXRs into COVID-19, and non-COVID-19.
Collapse
Affiliation(s)
- Anandbabu Gopatoti
- Department of Electronics and Communication Engineering, Hindusthan College of Engineering and Technology, Coimbatore, Tamil Nadu, India
| | - Ramya Jayakumar
- Department of Electronics and Communication Engineering, Hindusthan College of Engineering and Technology, Coimbatore, Tamil Nadu, India
| | - Poornaiah Billa
- Department of Electronics and Communication Engineering, Lakireddy Bali Reddy College of Engineering, Mylavaram, Andhra Pradesh, India
| | - Vijayalakshmi Patteeswaran
- Department of Electronics and Communication Engineering, Hindusthan College of Engineering and Technology, Coimbatore, Tamil Nadu, India
| |
Collapse
|
26
|
Zahedi Nasab R, Mohseni H, Montazeri M, Ghasemian F, Amin S. AFEX-Net: Adaptive feature extraction convolutional neural network for classifying computerized tomography images. Digit Health 2024; 10:20552076241232882. [PMID: 38406769 PMCID: PMC10894540 DOI: 10.1177/20552076241232882] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/20/2023] [Accepted: 01/23/2024] [Indexed: 02/27/2024] Open
Abstract
Purpose Deep convolutional neural networks are favored methods that are widely used in medical image processing due to their demonstrated performance in this area. Recently, the emergence of new lung diseases, such as COVID-19, and the possibility of early detection of their symptoms from chest computerized tomography images has attracted many researchers to classify diseases by training deep convolutional neural networks on lung computerized tomography images. The trained networks are expected to distinguish between different lung indications in various diseases, especially at the early stages. The purpose of this study is to introduce and assess an efficient deep convolutional neural network, called AFEX-Net, that can classify different lung diseases from chest computerized tomography images. Methods We designed a lightweight convolutional neural network called AFEX-Net with adaptive feature extraction layers, adaptive pooling layers, and adaptive activation functions. We trained and tested AFEX-Net on a dataset of more than 10,000 chest computerized tomography slices from different lung diseases (CC dataset), using an effective pre-processing method to remove bias. We also applied AFEX-Net to the public COVID-CTset dataset to assess its generalizability. The study was mainly conducted based on data collected over approximately six months during the pandemic outbreak in Afzalipour Hospital, Iran, which is the largest hospital in Southeast Iran. Results AFEX-Net achieved high accuracy and fast training on both datasets, outperforming several state-of-the-art convolutional neural networks. It has an accuracy of 99.7 % and 98.8 % on the CC and COVID-CTset datasets, respectively, with a learning speed that is 3 times faster compared to similar methods due to its lightweight structure. AFEX-Net was able to extract distinguishing features and classify chest computerized tomography images, especially at the early stages of lung diseases. Conclusion The AFEX-Net is a high-performing convolutional neural network for classifying lung diseases from chest CT images. It is efficient, adaptable, and compatible with input data, making it a reliable tool for early detection and diagnosis of lung diseases.
Collapse
Affiliation(s)
- Roxana Zahedi Nasab
- Department of Computer Engineering, Faculty of Engineering, Shahid Bahonar University of Kerman, Kerman, Iran
| | - Hadis Mohseni
- Department of Computer Engineering, Faculty of Engineering, Shahid Bahonar University of Kerman, Kerman, Iran
| | - Mahdieh Montazeri
- Health Information Sciences Department, Kerman University of Medical Sciences, Kerman, Iran
| | - Fahimeh Ghasemian
- Department of Computer Engineering, Faculty of Engineering, Shahid Bahonar University of Kerman, Kerman, Iran
| | - Sobhan Amin
- Kazerun Branch, Islamic Azad University, Kazerun, Iran
| |
Collapse
|
27
|
Amgothu S, Koppu S. COVID-19 prediction using Caviar Squirrel Jellyfish Search Optimization technique in fog-cloud based architecture. PLoS One 2023; 18:e0295599. [PMID: 38127990 PMCID: PMC10735048 DOI: 10.1371/journal.pone.0295599] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/15/2023] [Accepted: 11/24/2023] [Indexed: 12/23/2023] Open
Abstract
In the pandemic of COVID-19 patients approach to the hospital for prescription, yet due to extreme line up the patient gets treatment after waiting for more than one hour. Generally, wearable devices directly measure the preliminary data of the patient stored in capturing mode. In order to store the data, the hospitals require large storage devices that make the progression of data more complex. To bridge this gap, a potent scheme is established for COVID-19 prediction based fog-cloud named Caviar Squirrel Jellyfish Search Optimization (CSJSO). Here, CSJSO is the amalgamation of CAViar Squirrel Search Algorithm (CSSA) and Jellyfish Search Optimization (JSO), where CSSA is blended by the Conditional Autoregressive Value-at-Risk (CAViar) and Squirrel Search Algorithm (SSA). This architecture comprises the healthcare IoT sensor layer, fog layer and cloud layer. In the healthcare IoT sensor layer, the routing process with the collection of patient health condition data is carried out. On the other hand, in the fog layer COVID-19 detection is performed by employing a Deep Neuro Fuzzy Network (DNFN) trained by the proposed Remora Namib Beetle JSO (RNBJSO). Here, RNBJSO is the combination of Namib Beetle Optimization (NBO), Remora Optimization Algorithm (ROA) and Jellyfish Search optimization (JSO). Finally, in the cloud layer, the detection of COVID-19 employing Deep Long Short Term Memory (Deep LSTM) trained utilizing proposed CSJSO is performed. The evaluation measures utilized for CSJSO_Deep LSTM in database-1, such as Mean Squared Error (MSE) and Root Mean Squared Error (RMSE) observed 0.062 and 0.252 in confirmed cases. The measures employed in database-2 are accuracy, sensitivity and specificity achieved 0.925, 0.928 and 0.925 in K-set.
Collapse
Affiliation(s)
- Shanthi Amgothu
- School of Computer Science Engineering and Information Systems, Vellore, India
| | - Srinivas Koppu
- School of Computer Science Engineering and Information Systems, Vellore, India
| |
Collapse
|
28
|
Schaudt D, Späte C, von Schwerin R, Reichert M, von Schwerin M, Beer M, Kloth C. A Critical Assessment of Generative Models for Synthetic Data Augmentation on Limited Pneumonia X-ray Data. Bioengineering (Basel) 2023; 10:1421. [PMID: 38136012 PMCID: PMC10741143 DOI: 10.3390/bioengineering10121421] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/08/2023] [Revised: 11/28/2023] [Accepted: 12/12/2023] [Indexed: 12/24/2023] Open
Abstract
In medical imaging, deep learning models serve as invaluable tools for expediting diagnoses and aiding specialized medical professionals in making clinical decisions. However, effectively training deep learning models typically necessitates substantial quantities of high-quality data, a resource often lacking in numerous medical imaging scenarios. One way to overcome this deficiency is to artificially generate such images. Therefore, in this comparative study we train five generative models to artificially increase the amount of available data in such a scenario. This synthetic data approach is evaluated on a a downstream classification task, predicting four causes for pneumonia as well as healthy cases on 1082 chest X-ray images. Quantitative and medical assessments show that a Generative Adversarial Network (GAN)-based approach significantly outperforms more recent diffusion-based approaches on this limited dataset with better image quality and pathological plausibility. We show that better image quality surprisingly does not translate to improved classification performance by evaluating five different classification models and varying the amount of additional training data. Class-specific metrics like precision, recall, and F1-score show a substantial improvement by using synthetic images, emphasizing the data rebalancing effect of less frequent classes. However, overall performance does not improve for most models and configurations, except for a DreamBooth approach which shows a +0.52 improvement in overall accuracy. The large variance of performance impact in this study suggests a careful consideration of utilizing generative models for limited data scenarios, especially with an unexpected negative correlation between image quality and downstream classification improvement.
Collapse
Affiliation(s)
- Daniel Schaudt
- Institute of Databases and Information Systems, Ulm University, James-Franck-Ring, 89081 Ulm, Germany
| | - Christian Späte
- DASU Transferzentrum für Digitalisierung, Analytics und Data Science Ulm, Olgastraße 94, 89073 Ulm, Germany
| | - Reinhold von Schwerin
- Department of Computer Science, Ulm University of Applied Science, Albert–Einstein–Allee 55, 89081 Ulm, Germany
| | - Manfred Reichert
- Institute of Databases and Information Systems, Ulm University, James-Franck-Ring, 89081 Ulm, Germany
| | - Marianne von Schwerin
- Department of Computer Science, Ulm University of Applied Science, Albert–Einstein–Allee 55, 89081 Ulm, Germany
| | - Meinrad Beer
- Department of Radiology, University Hospital of Ulm, Albert–Einstein–Allee 23, 89081 Ulm, Germany
| | - Christopher Kloth
- Department of Radiology, University Hospital of Ulm, Albert–Einstein–Allee 23, 89081 Ulm, Germany
| |
Collapse
|
29
|
Wang Y, Liu L, Wang C. Trends in using deep learning algorithms in biomedical prediction systems. Front Neurosci 2023; 17:1256351. [PMID: 38027475 PMCID: PMC10665494 DOI: 10.3389/fnins.2023.1256351] [Citation(s) in RCA: 12] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/10/2023] [Accepted: 09/25/2023] [Indexed: 12/01/2023] Open
Abstract
In the domain of using DL-based methods in medical and healthcare prediction systems, the utilization of state-of-the-art deep learning (DL) methodologies assumes paramount significance. DL has attained remarkable achievements across diverse domains, rendering its efficacy particularly noteworthy in this context. The integration of DL with health and medical prediction systems enables real-time analysis of vast and intricate datasets, yielding insights that significantly enhance healthcare outcomes and operational efficiency in the industry. This comprehensive literature review systematically investigates the latest DL solutions for the challenges encountered in medical healthcare, with a specific emphasis on DL applications in the medical domain. By categorizing cutting-edge DL approaches into distinct categories, including convolutional neural networks (CNNs), recurrent neural networks (RNNs), generative adversarial networks (GANs), long short-term memory (LSTM) models, support vector machine (SVM), and hybrid models, this study delves into their underlying principles, merits, limitations, methodologies, simulation environments, and datasets. Notably, the majority of the scrutinized articles were published in 2022, underscoring the contemporaneous nature of the research. Moreover, this review accentuates the forefront advancements in DL techniques and their practical applications within the realm of medical prediction systems, while simultaneously addressing the challenges that hinder the widespread implementation of DL in image segmentation within the medical healthcare domains. These discerned insights serve as compelling impetuses for future studies aimed at the progressive advancement of using DL-based methods in medical and health prediction systems. The evaluation metrics employed across the reviewed articles encompass a broad spectrum of features, encompassing accuracy, precision, specificity, F-score, adoptability, adaptability, and scalability.
Collapse
Affiliation(s)
- Yanbu Wang
- School of Strength and Conditioning, Beijing Sport University, Beijing, China
| | - Linqing Liu
- Department of Physical Education, Peking University, Beijing, China
| | - Chao Wang
- Institute of Competitive Sports, Beijing Sport University, Beijing, China
| |
Collapse
|
30
|
Chu WT, Reza SMS, Anibal JT, Landa A, Crozier I, Bağci U, Wood BJ, Solomon J. Artificial Intelligence and Infectious Disease Imaging. J Infect Dis 2023; 228:S322-S336. [PMID: 37788501 PMCID: PMC10547369 DOI: 10.1093/infdis/jiad158] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/22/2022] [Accepted: 05/06/2023] [Indexed: 10/05/2023] Open
Abstract
The mass production of the graphics processing unit and the coronavirus disease 2019 (COVID-19) pandemic have provided the means and the motivation, respectively, for rapid developments in artificial intelligence (AI) and medical imaging techniques. This has led to new opportunities to improve patient care but also new challenges that must be overcome before these techniques are put into practice. In particular, early AI models reported high performances but failed to perform as well on new data. However, these mistakes motivated further innovation focused on developing models that were not only accurate but also stable and generalizable to new data. The recent developments in AI in response to the COVID-19 pandemic will reap future dividends by facilitating, expediting, and informing other medical AI applications and educating the broad academic audience on the topic. Furthermore, AI research on imaging animal models of infectious diseases offers a unique problem space that can fill in evidence gaps that exist in clinical infectious disease research. Here, we aim to provide a focused assessment of the AI techniques leveraged in the infectious disease imaging research space, highlight the unique challenges, and discuss burgeoning solutions.
Collapse
Affiliation(s)
- Winston T Chu
- Center for Infectious Disease Imaging, Radiology and Imaging Sciences, Clinical Center, National Institutes of Health, Bethesda, Maryland, USA
- Integrated Research Facility at Fort Detrick, Division of Clinical Research, National Institute of Allergy and Infectious Diseases, National Institutes of Health, Frederick, Maryland, USA
| | - Syed M S Reza
- Center for Infectious Disease Imaging, Radiology and Imaging Sciences, Clinical Center, National Institutes of Health, Bethesda, Maryland, USA
| | - James T Anibal
- Center for Interventional Oncology, Clinical Center, National Institutes of Health, Bethesda, Maryland, USA
| | - Adam Landa
- Center for Interventional Oncology, Clinical Center, National Institutes of Health, Bethesda, Maryland, USA
| | - Ian Crozier
- Clinical Monitoring Research Program Directorate, Frederick National Laboratory for Cancer Research sponsored by the National Cancer Institute, Frederick, Maryland, USA
| | - Ulaş Bağci
- Department of Radiology, Feinberg School of Medicine, Northwestern University, Chicago, Illinois, USA
| | - Bradford J Wood
- Center for Interventional Oncology, Clinical Center, National Institutes of Health, Bethesda, Maryland, USA
- Center for Interventional Oncology, National Cancer Institute, National Institutes of Health, Bethesda, Maryland, USA
| | - Jeffrey Solomon
- Clinical Monitoring Research Program Directorate, Frederick National Laboratory for Cancer Research sponsored by the National Cancer Institute, Frederick, Maryland, USA
| |
Collapse
|
31
|
Wang J, Dvornek NC, Staib LH, Duncan JS. Learning Sequential Information in Task-based fMRI for Synthetic Data Augmentation. MACHINE LEARNING IN CLINICAL NEUROIMAGING : 6TH INTERNATIONAL WORKSHOP, MLCN 2023, HELD IN CONJUNCTION WITH MICCAI 2023, VANCOUVER, BC, CANADA, OCTOBER 8, 2023, PROCEEDINGS. MLCN (WORKSHOP) (6TH : 2023 : VANCOUVER, B.C.) 2023; 14312:79-88. [PMID: 39281201 PMCID: PMC11395879 DOI: 10.1007/978-3-031-44858-4_8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/18/2024]
Abstract
Insufficiency of training data is a persistent issue in medical image analysis, especially for task-based functional magnetic resonance images (fMRI) with spatio-temporal imaging data acquired using specific cognitive tasks. In this paper, we propose an approach for generating synthetic fMRI sequences that can then be used to create augmented training datasets in downstream learning tasks. To synthesize high-resolution task-specific fMRI, we adapt the α-GAN structure, leveraging advantages of both GAN and variational autoencoder models, and propose different alternatives in aggregating temporal information. The synthetic images are evaluated from multiple perspectives including visualizations and an autism spectrum disorder (ASD) classification task. The results show that the synthetic task-based fMRI can provide effective data augmentation in learning the ASD classification task.
Collapse
Affiliation(s)
- Jiyao Wang
- Biomedical Engineering, Yale University, New Haven, CT 06511, USA
| | - Nicha C Dvornek
- Biomedical Engineering, Yale University, New Haven, CT 06511, USA
- Radiology & Biomedical Imaging, Yale School of Medicine, New Haven, CT 06511, USA
| | - Lawrence H Staib
- Biomedical Engineering, Yale University, New Haven, CT 06511, USA
- Radiology & Biomedical Imaging, Yale School of Medicine, New Haven, CT 06511, USA
| | - James S Duncan
- Biomedical Engineering, Yale University, New Haven, CT 06511, USA
- Radiology & Biomedical Imaging, Yale School of Medicine, New Haven, CT 06511, USA
- Electrical Engineering, Yale University, New Haven, CT 06511, USA
- Statistics & Data Science, Yale University New Haven, CT, 06511, USA
| |
Collapse
|
32
|
Ukwuoma CC, Cai D, Heyat MBB, Bamisile O, Adun H, Al-Huda Z, Al-Antari MA. Deep learning framework for rapid and accurate respiratory COVID-19 prediction using chest X-ray images. JOURNAL OF KING SAUD UNIVERSITY. COMPUTER AND INFORMATION SCIENCES 2023; 35:101596. [PMID: 37275558 PMCID: PMC10211254 DOI: 10.1016/j.jksuci.2023.101596] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/19/2022] [Revised: 05/17/2023] [Accepted: 05/19/2023] [Indexed: 06/07/2023]
Abstract
COVID-19 is a contagious disease that affects the human respiratory system. Infected individuals may develop serious illnesses, and complications may result in death. Using medical images to detect COVID-19 from essentially identical thoracic anomalies is challenging because it is time-consuming, laborious, and prone to human error. This study proposes an end-to-end deep-learning framework based on deep feature concatenation and a Multi-head Self-attention network. Feature concatenation involves fine-tuning the pre-trained backbone models of DenseNet, VGG-16, and InceptionV3, which are trained on a large-scale ImageNet, whereas a Multi-head Self-attention network is adopted for performance gain. End-to-end training and evaluation procedures are conducted using the COVID-19_Radiography_Dataset for binary and multi-classification scenarios. The proposed model achieved overall accuracies (96.33% and 98.67%) and F1_scores (92.68% and 98.67%) for multi and binary classification scenarios, respectively. In addition, this study highlights the difference in accuracy (98.0% vs. 96.33%) and F_1 score (97.34% vs. 95.10%) when compared with feature concatenation against the highest individual model performance. Furthermore, a virtual representation of the saliency maps of the employed attention mechanism focusing on the abnormal regions is presented using explainable artificial intelligence (XAI) technology. The proposed framework provided better COVID-19 prediction results outperforming other recent deep learning models using the same dataset.
Collapse
Affiliation(s)
- Chiagoziem C Ukwuoma
- The College of Nuclear Technology and Automation Engineering, Chengdu University of Technology, Sichuan, 610059, China
| | - Dongsheng Cai
- The College of Nuclear Technology and Automation Engineering, Chengdu University of Technology, Sichuan, 610059, China
| | - Md Belal Bin Heyat
- IoT Research Center, College of Computer Science and Software Engineering, Shenzhen University, Shenzhen, Guangdong 518060, China
| | - Olusola Bamisile
- Sichuan Industrial Internet Intelligent Monitoring and Application Engineering Technology Research Center, Chengdu University of Technology, China
| | - Humphrey Adun
- Department of Mechanical and Energy Systems Engineering, Cyprus International University, Nicosia, North Nicosia, Cyprus
| | - Zaid Al-Huda
- School of Computing and Artificial Intelligence, Southwest Jiaotong University, Chengdu, Sichuan, China
| | - Mugahed A Al-Antari
- Department of Artificial Intelligence, College of Software & Convergence Technology, Daeyang AI Center, Sejong University, Seoul 05006, Republic of Korea
| |
Collapse
|
33
|
Kumar S, Bhowmik B. COVID-19 Waves and Their Impacts to Society. 2023 IEEE GUWAHATI SUBSECTION CONFERENCE (GCON) 2023:1-5. [DOI: 10.1109/gcon58516.2023.10183544] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/02/2025]
Affiliation(s)
- Sunil Kumar
- National Institute of Technology Karnataka,Maharshi Patanjali CPS Lab, BRICS Laboratory,Department of Computer Science and Engineering,Mangalore,Bharat,575025
| | - Biswajit Bhowmik
- National Institute of Technology Karnataka,Maharshi Patanjali CPS Lab, BRICS Laboratory,Department of Computer Science and Engineering,Mangalore,Bharat,575025
| |
Collapse
|
34
|
Mehrdad S, Shamout FE, Wang Y, Atashzar SF. Deep learning for deterioration prediction of COVID-19 patients based on time-series of three vital signs. Sci Rep 2023; 13:9968. [PMID: 37339986 PMCID: PMC10282033 DOI: 10.1038/s41598-023-37013-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/24/2022] [Accepted: 06/14/2023] [Indexed: 06/22/2023] Open
Abstract
Unrecognized deterioration of COVID-19 patients can lead to high morbidity and mortality. Most existing deterioration prediction models require a large number of clinical information, typically collected in hospital settings, such as medical images or comprehensive laboratory tests. This is infeasible for telehealth solutions and highlights a gap in deterioration prediction models based on minimal data, which can be recorded at a large scale in any clinic, nursing home, or even at the patient's home. In this study, we develop and compare two prognostic models that predict if a patient will experience deterioration in the forthcoming 3 to 24 h. The models sequentially process routine triadic vital signs: (a) oxygen saturation, (b) heart rate, and (c) temperature. These models are also provided with basic patient information, including sex, age, vaccination status, vaccination date, and status of obesity, hypertension, or diabetes. The difference between the two models is the way that the temporal dynamics of the vital signs are processed. Model #1 utilizes a temporally-dilated version of the Long-Short Term Memory model (LSTM) for temporal processes, and Model #2 utilizes a residual temporal convolutional network (TCN) for this purpose. We train and evaluate the models using data collected from 37,006 COVID-19 patients at NYU Langone Health in New York, USA. The convolution-based model outperforms the LSTM based model, achieving a high AUROC of 0.8844-0.9336 for 3 to 24 h deterioration prediction on a held-out test set. We also conduct occlusion experiments to evaluate the importance of each input feature, which reveals the significance of continuously monitoring the variation of the vital signs. Our results show the prospect for accurate deterioration forecast using a minimum feature set that can be relatively easily obtained using wearable devices and self-reported patient information.
Collapse
Affiliation(s)
- Sarmad Mehrdad
- Department of Electrical and Computer Engineering, New York University (NYU), New York, USA
| | - Farah E Shamout
- Department of Biomedical Engineering, New York University (NYU), New York, USA
- Division of Engineering, New York University Abu Dhabi (NYUAD), Abu Dhabi, UAE
- Computer Science and Engineering, New York University (NYU), New York, USA
| | - Yao Wang
- Department of Electrical and Computer Engineering, New York University (NYU), New York, USA
- Department of Biomedical Engineering, New York University (NYU), New York, USA
| | - S Farokh Atashzar
- Department of Electrical and Computer Engineering, New York University (NYU), New York, USA.
- Department of Biomedical Engineering, New York University (NYU), New York, USA.
- Department of Mechanical and Aerospace Engineering, New York University (NYU), New York, USA.
| |
Collapse
|
35
|
Eshraghi MA, Ayatollahi A, Shokouhi SB. COV-MobNets: a mobile networks ensemble model for diagnosis of COVID-19 based on chest X-ray images. BMC Med Imaging 2023; 23:83. [PMID: 37322450 PMCID: PMC10273540 DOI: 10.1186/s12880-023-01039-w] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2023] [Accepted: 06/01/2023] [Indexed: 06/17/2023] Open
Abstract
BACKGROUND The medical profession is facing an excessive workload, which has led to the development of various Computer-Aided Diagnosis (CAD) systems as well as Mobile-Aid Diagnosis (MAD) systems. These technologies enhance the speed and accuracy of diagnoses, particularly in areas with limited resources or remote regions during the pandemic. The primary purpose of this research is to predict and diagnose COVID-19 infection from chest X-ray images by developing a mobile-friendly deep learning framework, which has the potential for deployment in portable devices such as mobile or tablet, especially in situations where the workload of radiology specialists may be high. Moreover, this could improve the accuracy and transparency of population screening to assist radiologists during the pandemic. METHODS In this study, the Mobile Networks ensemble model called COV-MobNets is proposed to classify positive COVID-19 X-ray images from negative ones and can have an assistant role in diagnosing COVID-19. The proposed model is an ensemble model, combining two lightweight and mobile-friendly models: MobileViT based on transformer structure and MobileNetV3 based on Convolutional Neural Network. Hence, COV-MobNets can extract the features of chest X-ray images in two different methods to achieve better and more accurate results. In addition, data augmentation techniques were applied to the dataset to avoid overfitting during the training process. The COVIDx-CXR-3 benchmark dataset was used for training and evaluation. RESULTS The classification accuracy of the improved MobileViT and MobileNetV3 models on the test set has reached 92.5% and 97%, respectively, while the accuracy of the proposed model (COV-MobNets) has reached 97.75%. The sensitivity and specificity of the proposed model have also reached 98.5% and 97%, respectively. Experimental comparison proves the result is more accurate and balanced than other methods. CONCLUSION The proposed method can distinguish between positive and negative COVID-19 cases more accurately and quickly. The proposed method proves that utilizing two automatic feature extractors with different structures as an overall framework of COVID-19 diagnosis can lead to improved performance, enhanced accuracy, and better generalization to new or unseen data. As a result, the proposed framework in this study can be used as an effective method for computer-aided diagnosis and mobile-aided diagnosis of COVID-19. The code is available publicly for open access at https://github.com/MAmirEshraghi/COV-MobNets .
Collapse
Affiliation(s)
- Mohammad Amir Eshraghi
- School of Electrical Engineering, Iran University of Science and Technology, Tehran, Iran
| | - Ahmad Ayatollahi
- School of Electrical Engineering, Iran University of Science and Technology, Tehran, Iran
| | | |
Collapse
|
36
|
Butt MJ, Malik AK, Qamar N, Yar S, Malik AJ, Rauf U. A Survey on COVID-19 Data Analysis Using AI, IoT, and Social Media. SENSORS (BASEL, SWITZERLAND) 2023; 23:5543. [PMID: 37420714 DOI: 10.3390/s23125543] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/30/2023] [Revised: 06/04/2023] [Accepted: 06/07/2023] [Indexed: 07/09/2023]
Abstract
Coronaviruses are a well-established and deadly group of viruses that cause illness in both humans and animals. The novel type of this virus group, named COVID-19, was firstly reported in December 2019, and, with the passage of time, coronavirus has spread to almost all parts of the world. Coronavirus has been the cause of millions of deaths around the world. Furthermore, many countries are struggling with COVID-19 and have experimented with various kinds of vaccines to eliminate the deadly virus and its variants. This survey deals with COVID-19 data analysis and its impact on human social life. Data analysis and information related to coronavirus can greatly help scientists and governments in controlling the spread and symptoms of the deadly coronavirus. In this survey, we cover many areas of discussion related to COVID-19 data analysis, such as how artificial intelligence, along with machine learning, deep learning, and IoT, have worked together to fight against COVID-19. We also discuss artificial intelligence and IoT techniques used to forecast, detect, and diagnose patients of the novel coronavirus. Moreover, this survey also describes how fake news, doctored results, and conspiracy theories were spread over social media sites, such as Twitter, by applying various social network analysis and sentimental analysis techniques. A comprehensive comparative analysis of existing techniques has also been conducted. In the end, the Discussion section presents different data analysis techniques, provides future directions for research, and suggests general guidelines for handling coronavirus, as well as changing work and life conditions.
Collapse
Affiliation(s)
- Muhammad Junaid Butt
- Department of Computer Science, COMSATS University Islamabad (CUI), Islamabad 45550, Pakistan
| | - Ahmad Kamran Malik
- Department of Computer Science, COMSATS University Islamabad (CUI), Islamabad 45550, Pakistan
| | - Nafees Qamar
- School of Health and Behavioral Sciences, Bryant University, Smithfield, RI 02917, USA
| | - Samad Yar
- Department of Computer Science, COMSATS University Islamabad (CUI), Islamabad 45550, Pakistan
| | - Arif Jamal Malik
- Department of Software Engineering, Foundation University, Islamabad 44000, Pakistan
| | - Usman Rauf
- Department of Mathematics and Computer Science, Mercy College, Dobbs Ferry, NY 10522, USA
| |
Collapse
|
37
|
Shanthi A, Koppu S. Remora Namib Beetle Optimization Enabled Deep Learning for Severity of COVID-19 Lung Infection Identification and Classification Using CT Images. SENSORS (BASEL, SWITZERLAND) 2023; 23:s23115316. [PMID: 37300043 DOI: 10.3390/s23115316] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/16/2022] [Revised: 02/26/2023] [Accepted: 04/14/2023] [Indexed: 06/12/2023]
Abstract
Coronavirus disease 2019 (COVID-19) has seen a crucial outburst for both females and males worldwide. Automatic lung infection detection from medical imaging modalities provides high potential for increasing the treatment for patients to tackle COVID-19 disease. COVID-19 detection from lung CT images is a rapid way of diagnosing patients. However, identifying the occurrence of infectious tissues and segmenting this from CT images implies several challenges. Therefore, efficient techniques termed as Remora Namib Beetle Optimization_ Deep Quantum Neural Network (RNBO_DQNN) and RNBO_Deep Neuro Fuzzy Network (RNBO_DNFN) are introduced for the identification as well as classification of COVID-19 lung infection. Here, the pre-processing of lung CT images is performed utilizing an adaptive Wiener filter, whereas lung lobe segmentation is performed employing the Pyramid Scene Parsing Network (PSP-Net). Afterwards, feature extraction is carried out wherein features are extracted for the classification phase. In the first level of classification, DQNN is utilized, tuned by RNBO. Furthermore, RNBO is designed by merging the Remora Optimization Algorithm (ROA) and Namib Beetle Optimization (NBO). If a classified output is COVID-19, then the second-level classification is executed using DNFN for further classification. Additionally, DNFN is also trained by employing the newly proposed RNBO. Furthermore, the devised RNBO_DNFN achieved maximum testing accuracy, with TNR and TPR obtaining values of 89.4%, 89.5% and 87.5%.
Collapse
Affiliation(s)
- Amgothu Shanthi
- School of Information Technology and Engineering, Vellore Institute of Technology, Vellore 632014, India
| | - Srinivas Koppu
- School of Information Technology and Engineering, Vellore Institute of Technology, Vellore 632014, India
| |
Collapse
|
38
|
Sistaninejhad B, Rasi H, Nayeri P. A Review Paper about Deep Learning for Medical Image Analysis. COMPUTATIONAL AND MATHEMATICAL METHODS IN MEDICINE 2023; 2023:7091301. [PMID: 37284172 PMCID: PMC10241570 DOI: 10.1155/2023/7091301] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/03/2023] [Revised: 02/12/2023] [Accepted: 04/21/2023] [Indexed: 06/08/2023]
Abstract
Medical imaging refers to the process of obtaining images of internal organs for therapeutic purposes such as discovering or studying diseases. The primary objective of medical image analysis is to improve the efficacy of clinical research and treatment options. Deep learning has revamped medical image analysis, yielding excellent results in image processing tasks such as registration, segmentation, feature extraction, and classification. The prime motivations for this are the availability of computational resources and the resurgence of deep convolutional neural networks. Deep learning techniques are good at observing hidden patterns in images and supporting clinicians in achieving diagnostic perfection. It has proven to be the most effective method for organ segmentation, cancer detection, disease categorization, and computer-assisted diagnosis. Many deep learning approaches have been published to analyze medical images for various diagnostic purposes. In this paper, we review the work exploiting current state-of-the-art deep learning approaches in medical image processing. We begin the survey by providing a synopsis of research works in medical imaging based on convolutional neural networks. Second, we discuss popular pretrained models and general adversarial networks that aid in improving convolutional networks' performance. Finally, to ease direct evaluation, we compile the performance metrics of deep learning models focusing on COVID-19 detection and child bone age prediction.
Collapse
Affiliation(s)
| | - Habib Rasi
- Sahand University of Technology, East Azerbaijan, New City of Sahand, Iran
| | - Parisa Nayeri
- Khoy University of Medical Sciences, West Azerbaijan, Khoy, Iran
| |
Collapse
|
39
|
Pati SK, Gupta MK, Banerjee A, Shai R, Shivakumara P. Drug discovery through Covid-19 genome sequencing with siamese graph convolutional neural network. MULTIMEDIA TOOLS AND APPLICATIONS 2023; 83:1-35. [PMID: 37362739 PMCID: PMC10170456 DOI: 10.1007/s11042-023-15270-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 04/16/2022] [Revised: 09/23/2022] [Accepted: 04/06/2023] [Indexed: 06/28/2023]
Abstract
After several waves of COVID-19 led to a massive loss of human life worldwide due to the changes in its variants and the vast explosion. Several researchers proposed neural network-based drug discovery techniques to fight against the pandemic; utilizing neural networks has limitations (Exponential time complexity, Non-Convergence, Mode Collapse, and Diminished Gradient). To overcome those difficulties, this paper proposed a hybrid architecture that will help to repurpose the most appropriate medicines for the treatment of COVID-19. A brief investigation of the sequences has been made to discover the gene density and noncoding proportion through the next gene sequencing. The paper tracks the exceptional locales in the virus DNA sequence as a Drug Target Region (DTR). Then the variable DNA neighborhood search is applied to this DTR to obtain the DNA interaction network to show how the genes are correlated. A drug database has been obtained based on the ontological property of the genomes with advanced D3Similarity so that all the chemical components of the drug database have been identified. Other methods obtained hydroxychloroquine as an effective drug which was rejected by WHO. However, The experimental results show that Remdesivir and Dexamethasone are the most effective drugs, with 97.41 and 97.93%, respectively.
Collapse
Affiliation(s)
- Soumen Kumar Pati
- Department of Bioinformatics, Maulana Abul Kalam Azad University of Technology, Haringhata, West Bengal 741249 India
| | - Manan Kumar Gupta
- Department of Bioinformatics, Maulana Abul Kalam Azad University of Technology, Haringhata, West Bengal 741249 India
| | - Ayan Banerjee
- Department of Computer Science & Engineering, Jalpaiguri Governmemt Engineering College, Jalpaiguri, West Bengal 735102 India
| | - Rinita Shai
- Department of Mathematics, Behala College, Calcutta University, Kolkata, West Bengal 700060 India
| | | |
Collapse
|
40
|
Guo W, Wang Y, Chen X, Jiang P. Federated transfer learning for auxiliary classifier generative adversarial networks: framework and industrial application. JOURNAL OF INTELLIGENT MANUFACTURING 2023:1-16. [PMID: 37361337 PMCID: PMC10162656 DOI: 10.1007/s10845-023-02126-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 07/12/2022] [Accepted: 04/01/2023] [Indexed: 06/28/2023]
Abstract
Machine learning with considering data privacy-preservation and personalized models has received attentions, especially in the manufacturing field. The data often exist in the form of isolated islands and cannot be shared because of data privacy in real industrial scenarios. It is difficult to gather the data to train a personalized model without compromising data privacy. To address this issue, we proposed a Federated Transfer Learning framework based on Auxiliary Classifier Generative Adversarial Networks named ACGAN-FTL. In the framework, Federated Learning (FL) trains a global model on decentralized datasets of the clients with data privacy-preservation and Transfer Learning (TL) transfers the knowledge from the global model to a personalized model with a relatively small data volume. ACGAN acts as a data bridge to connect FL and TL by generating similar probability distribution data of clients since the client datasets in FL cannot be directly used in TL for data privacy-preservation. A real industrial scenario of pre-baked carbon anode quality prediction is applied to verify the performance of the proposed framework. The results show that ACGAN-FTL can not only obtain acceptable performance on 0.81 accuracy, 0.86 precision, 0.74 recall, and 0.79 F1 but also ensure data privacy-preservation in the whole learning process. Compared to the baseline method without FL and TL, the former metrics have increased by 13%, 11%, 16%, and 15% respectively. The experiments verify that the performance of the proposed ACGAN-FTL framework fulfills the requirements of industrial scenarios.
Collapse
Affiliation(s)
- Wei Guo
- State Key Laboratory for Manufacturing Systems Engineering, Xi’an Jiaotong University, Xi’an, China
| | - Yijin Wang
- State Key Laboratory for Manufacturing Systems Engineering, Xi’an Jiaotong University, Xi’an, China
| | - Xin Chen
- State Key Laboratory for Manufacturing Systems Engineering, Xi’an Jiaotong University, Xi’an, China
| | - Pingyu Jiang
- State Key Laboratory for Manufacturing Systems Engineering, Xi’an Jiaotong University, Xi’an, China
| |
Collapse
|
41
|
Novel prediction model on OSCC histopathological images via deep transfer learning combined with Grad-CAM interpretation. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2023.104704] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/22/2023]
|
42
|
Li G, Togo R, Ogawa T, Haseyama M. Boosting automatic COVID-19 detection performance with self-supervised learning and batch knowledge ensembling. Comput Biol Med 2023; 158:106877. [PMID: 37019015 PMCID: PMC10063457 DOI: 10.1016/j.compbiomed.2023.106877] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2022] [Revised: 03/15/2023] [Accepted: 03/30/2023] [Indexed: 04/03/2023]
Abstract
PROBLEM Detecting COVID-19 from chest X-ray (CXR) images has become one of the fastest and easiest methods for detecting COVID-19. However, the existing methods usually use supervised transfer learning from natural images as a pretraining process. These methods do not consider the unique features of COVID-19 and the similar features between COVID-19 and other pneumonia. AIM In this paper, we want to design a novel high-accuracy COVID-19 detection method that uses CXR images, which can consider the unique features of COVID-19 and the similar features between COVID-19 and other pneumonia. METHODS Our method consists of two phases. One is self-supervised learning-based pertaining; the other is batch knowledge ensembling-based fine-tuning. Self-supervised learning-based pretraining can learn distinguished representations from CXR images without manually annotated labels. On the other hand, batch knowledge ensembling-based fine-tuning can utilize category knowledge of images in a batch according to their visual feature similarities to improve detection performance. Unlike our previous implementation, we introduce batch knowledge ensembling into the fine-tuning phase, reducing the memory used in self-supervised learning and improving COVID-19 detection accuracy. RESULTS On two public COVID-19 CXR datasets, namely, a large dataset and an unbalanced dataset, our method exhibited promising COVID-19 detection performance. Our method maintains high detection accuracy even when annotated CXR training images are reduced significantly (e.g., using only 10% of the original dataset). In addition, our method is insensitive to changes in hyperparameters. CONCLUSION The proposed method outperforms other state-of-the-art COVID-19 detection methods in different settings. Our method can reduce the workloads of healthcare providers and radiologists.
Collapse
Affiliation(s)
- Guang Li
- Graduate School of Information Science and Technology, Hokkaido University, N-14, W-9, Kita-Ku, Sapporo, 060-0814, Japan.
| | - Ren Togo
- Faculty of Information Science and Technology, Hokkaido University, N-14, W-9, Kita-Ku, Sapporo, 060-0814, Japan.
| | - Takahiro Ogawa
- Faculty of Information Science and Technology, Hokkaido University, N-14, W-9, Kita-Ku, Sapporo, 060-0814, Japan.
| | - Miki Haseyama
- Faculty of Information Science and Technology, Hokkaido University, N-14, W-9, Kita-Ku, Sapporo, 060-0814, Japan.
| |
Collapse
|
43
|
Ullah Z, Usman M, Gwak J. MTSS-AAE: Multi-task semi-supervised adversarial autoencoding for COVID-19 detection based on chest X-ray images. EXPERT SYSTEMS WITH APPLICATIONS 2023; 216:119475. [PMID: 36619348 PMCID: PMC9810379 DOI: 10.1016/j.eswa.2022.119475] [Citation(s) in RCA: 9] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/15/2021] [Revised: 07/28/2022] [Accepted: 12/22/2022] [Indexed: 06/12/2023]
Abstract
Efficient diagnosis of COVID-19 plays an important role in preventing the spread of the disease. There are three major modalities to diagnose COVID-19 which include polymerase chain reaction tests, computed tomography scans, and chest X-rays (CXRs). Among these, diagnosis using CXRs is the most economical approach; however, it requires extensive human expertise to diagnose COVID-19 in CXRs, which may deprive it of cost-effectiveness. The computer-aided diagnosis with deep learning has the potential to perform accurate detection of COVID-19 in CXRs without human intervention while preserving its cost-effectiveness. Many efforts have been made to develop a highly accurate and robust solution. However, due to the limited amount of labeled data, existing solutions are evaluated on a small set of test dataset. In this work, we proposed a solution to this problem by using a multi-task semi-supervised learning (MTSSL) framework that utilized auxiliary tasks for which adequate data is publicly available. Specifically, we utilized Pneumonia, Lung Opacity, and Pleural Effusion as additional tasks using the ChesXpert dataset. We illustrated that the primary task of COVID-19 detection, for which only limited labeled data is available, can be improved by using this additional data. We further employed an adversarial autoencoder (AAE), which has a strong capability to learn powerful and discriminative features, within our MTSSL framework to maximize the benefit of multi-task learning. In addition, the supervised classification networks in combination with the unsupervised AAE empower semi-supervised learning, which includes a discriminative part in the unsupervised AAE training pipeline. The generalization of our framework is improved due to this semi-supervised learning and thus it leads to enhancement in COVID-19 detection performance. The proposed model is rigorously evaluated on the largest publicly available COVID-19 dataset and experimental results show that the proposed model attained state-of-the-art performance.
Collapse
Affiliation(s)
- Zahid Ullah
- Department of Software, Korea National University of Transportation, Chungju 27469, South Korea
| | - Muhammad Usman
- Department of Computer Science and Engineering, Seoul National University, Seoul 08826, South Korea
| | - Jeonghwan Gwak
- Department of Software, Korea National University of Transportation, Chungju 27469, South Korea
- Department of Biomedical Engineering, Korea National University of Transportation, Chungju 27469, South Korea
- Department of AI Robotics Engineering, Korea National University of Transportation, Chungju 27469, South Korea
- Department of IT Energy Convergence (BK21 FOUR), Korea National University of Transportation, Chungju 27469, South Korea
| |
Collapse
|
44
|
Kebaili A, Lapuyade-Lahorgue J, Ruan S. Deep Learning Approaches for Data Augmentation in Medical Imaging: A Review. J Imaging 2023; 9:81. [PMID: 37103232 PMCID: PMC10144738 DOI: 10.3390/jimaging9040081] [Citation(s) in RCA: 36] [Impact Index Per Article: 18.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/13/2023] [Revised: 03/31/2023] [Accepted: 04/07/2023] [Indexed: 04/28/2023] Open
Abstract
Deep learning has become a popular tool for medical image analysis, but the limited availability of training data remains a major challenge, particularly in the medical field where data acquisition can be costly and subject to privacy regulations. Data augmentation techniques offer a solution by artificially increasing the number of training samples, but these techniques often produce limited and unconvincing results. To address this issue, a growing number of studies have proposed the use of deep generative models to generate more realistic and diverse data that conform to the true distribution of the data. In this review, we focus on three types of deep generative models for medical image augmentation: variational autoencoders, generative adversarial networks, and diffusion models. We provide an overview of the current state of the art in each of these models and discuss their potential for use in different downstream tasks in medical imaging, including classification, segmentation, and cross-modal translation. We also evaluate the strengths and limitations of each model and suggest directions for future research in this field. Our goal is to provide a comprehensive review about the use of deep generative models for medical image augmentation and to highlight the potential of these models for improving the performance of deep learning algorithms in medical image analysis.
Collapse
Affiliation(s)
| | | | - Su Ruan
- Université Rouen Normandie, INSA Rouen Normandie, Université Le Havre Normandie, Normandie Univ, LITIS UR 4108, F-76000 Rouen, France
| |
Collapse
|
45
|
Wali A, Ahmad M, Naseer A, Tamoor M, Gilani S. StynMedGAN: Medical images augmentation using a new GAN model for improved diagnosis of diseases. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS 2023. [DOI: 10.3233/jifs-223996] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/07/2023]
Abstract
Deep networks require a considerable amount of training data otherwise these networks generalize poorly. Data Augmentation techniques help the network generalize better by providing more variety in the training data. Standard data augmentation techniques such as flipping, and scaling, produce new data that is a modified version of the original data. Generative Adversarial networks (GANs) have been designed to generate new data that can be exploited. In this paper, we propose a new GAN model, named StynMedGAN for synthetically generating medical images to improve the performance of classification models. StynMedGAN builds upon the state-of-the-art styleGANv2 that has produced remarkable results generating all kinds of natural images. We introduce a regularization term that is a normalized loss factor in the existing discriminator loss of styleGANv2. It is used to force the generator to produce normalized images and penalize it if it fails. Medical imaging modalities, such as X-Rays, CT-Scans, and MRIs are different in nature, we show that the proposed GAN extends the capacity of styleGANv2 to handle medical images in a better way. This new GAN model (StynMedGAN) is applied to three types of medical imaging: X-Rays, CT scans, and MRI to produce more data for the classification tasks. To validate the effectiveness of the proposed model for the classification, 3 classifiers (CNN, DenseNet121, and VGG-16) are used. Results show that the classifiers trained with StynMedGAN-augmented data outperform other methods that only used the original data. The proposed model achieved 100%, 99.6%, and 100% for chest X-Ray, Chest CT-Scans, and Brain MRI respectively. The results are promising and favor a potentially important resource that can be used by practitioners and radiologists to diagnose different diseases.
Collapse
Affiliation(s)
- Aamir Wali
- Department of Computer Science, National University of Computer and Emerging Science, Faisal Town, Lahore, Pakistan
| | - Muzammil Ahmad
- Department of Computer Science, National University of Computer and Emerging Science, Faisal Town, Lahore, Pakistan
| | - Asma Naseer
- Department of Computer Science, National University of Computer and Emerging Science, Faisal Town, Lahore, Pakistan
| | - Maria Tamoor
- Department of Computer Science, Forman Christian College University, Zahoor Ilahi Road, Lahore, Pakistan
| | - S.A.M. Gilani
- Department of Computer Science, National University of Computer and Emerging Science, Faisal Town, Lahore, Pakistan
| |
Collapse
|
46
|
Shukla AK, Seth T, Muhuri PK. Artificial intelligence centric scientific research on COVID-19: an analysis based on scientometrics data. MULTIMEDIA TOOLS AND APPLICATIONS 2023; 82:1-33. [PMID: 37362722 PMCID: PMC9978294 DOI: 10.1007/s11042-023-14642-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 02/07/2022] [Revised: 07/01/2022] [Accepted: 02/03/2023] [Indexed: 06/28/2023]
Abstract
With the spread of the deadly coronavirus disease throughout the geographies of the globe, expertise from every field has been sought to fight the impact of the virus. The use of Artificial Intelligence (AI), especially, has been the center of attention due to its capability to produce trustworthy results in a reasonable time. As a result, AI centric based research on coronavirus (or COVID-19) has been receiving growing attention from different domains ranging from medicine, virology, and psychiatry etc. We present this comprehensive study that closely monitors the impact of the pandemic on global research activities related exclusively to AI. In this article, we produce highly informative insights pertaining to publications, such as the best articles, research areas, most productive and influential journals, authors, and institutions. Studies are made on top 50 most cited articles to identify the most influential AI subcategories. We also study the outcome of research from different geographic areas while identifying the research collaborations that have had an impact. This study also compares the outcome of research from the different countries around the globe and produces insights on the same.
Collapse
Affiliation(s)
- Amit K. Shukla
- Faculty of Information Technology, University of Jyväskylä, Box 35 (Agora), Jyväskylä, 40014 Finland
| | - Taniya Seth
- Department of Computer Science, South Asian University, Akbar Bhawan, Chanakyapuri, New Delhi 110021 India
| | - Pranab K. Muhuri
- Department of Computer Science, South Asian University, Akbar Bhawan, Chanakyapuri, New Delhi 110021 India
| |
Collapse
|
47
|
Wu Y, Qi Q, Qi S, Yang L, Wang H, Yu H, Li J, Wang G, Zhang P, Liang Z, Chen R. Classification of COVID-19 from community-acquired pneumonia: Boosting the performance with capsule network and maximum intensity projection image of CT scans. Comput Biol Med 2023; 154:106567. [PMID: 36738705 PMCID: PMC9869624 DOI: 10.1016/j.compbiomed.2023.106567] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/30/2022] [Revised: 12/30/2022] [Accepted: 01/22/2023] [Indexed: 01/24/2023]
Abstract
BACKGROUND The coronavirus disease 2019 (COVID-19) and community-acquired pneumonia (CAP) present a high degree of similarity in chest computed tomography (CT) images. Therefore, a procedure for accurately and automatically distinguishing between them is crucial. METHODS A deep learning method for distinguishing COVID-19 from CAP is developed using maximum intensity projection (MIP) images from CT scans. LinkNet is employed for lung segmentation of chest CT images. MIP images are produced by superposing the maximum gray of intrapulmonary CT values. The MIP images are input into a capsule network for patient-level pred iction and diagnosis of COVID-19. The network is trained using 333 CT scans (168 COVID-19/165 CAP) and validated on three external datasets containing 3581 CT scans (2110 COVID-19/1471 CAP). RESULTS LinkNet achieves the highest Dice coefficient of 0.983 for lung segmentation. For the classification of COVID-19 and CAP, the capsule network with the DenseNet-121 feature extractor outperforms ResNet-50 and Inception-V3, achieving an accuracy of 0.970 on the training dataset. Without MIP or the capsule network, the accuracy decreases to 0.857 and 0.818, respectively. Accuracy scores of 0.961, 0.997, and 0.949 are achieved on the external validation datasets. The proposed method has higher or comparable sensitivity compared with ten state-of-the-art methods. CONCLUSIONS The proposed method illustrates the feasibility of applying MIP images from CT scans to distinguish COVID-19 from CAP using capsule networks. MIP images provide conspicuous benefits when exploiting deep learning to detect COVID-19 lesions from CT scans and the capsule network improves COVID-19 diagnosis.
Collapse
Affiliation(s)
- Yanan Wu
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China; Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, China.
| | - Qianqian Qi
- Research Center for Healthcare Data Science, Zhejiang Lab, Hangzhou, China.
| | - Shouliang Qi
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China; Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, China.
| | - Liming Yang
- Department of Radiology, The Affiliated Hospital of Guizhou Medical University, Guiyang, China.
| | - Hanlin Wang
- Department of Radiology, General Hospital of the Yangtze River Shipping, Wuhan, China.
| | - Hui Yu
- General Practice Center, The Seventh Affiliated Hospital, Southern Medical University, Guangzhou, China.
| | - Jianpeng Li
- Department of Radiology, Affiliated Dongguan Hospital, Southern Medical University, Dongguan, China.
| | - Gang Wang
- Department of Radiology, Affiliated Dongguan Hospital, Southern Medical University, Dongguan, China.
| | - Ping Zhang
- Department of Pulmonary and Critical Care Medicine, Affiliated Dongguan Hospital, Southern Medical University, Dongguan, China.
| | - Zhenyu Liang
- State Key Laboratory of Respiratory Disease, National Clinical Research Center for Respiratory Disease, Guangzhou Institute of Respiratory Health, The First Affiliated Hospital of Guangzhou Medical University, Guangzhou, China.
| | - Rongchang Chen
- Key Laboratory of Respiratory Disease of Shenzhen, Shenzhen Institute of Respiratory Disease, Shenzhen People's Hospital (Second Affiliated Hospital of Jinan University, First Affiliated Hospital of South University of Science and Technology of China), Shenzhen, China.
| |
Collapse
|
48
|
Chadebec C, Thibeau-Sutre E, Burgos N, Allassonniere S. Data Augmentation in High Dimensional Low Sample Size Setting Using a Geometry-Based Variational Autoencoder. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2023; 45:2879-2896. [PMID: 35749321 DOI: 10.1109/tpami.2022.3185773] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
In this paper, we propose a new method to perform data augmentation in a reliable way in the High Dimensional Low Sample Size (HDLSS) setting using a geometry-based variational autoencoder (VAE). Our approach combines the proposal of 1) a new VAE model, the latent space of which is modeled as a Riemannian manifold and which combines both Riemannian metric learning and normalizing flows and 2) a new generation scheme which produces more meaningful samples especially in the context of small data sets. The method is tested through a wide experimental study where its robustness to data sets, classifiers and training samples size is stressed. It is also validated on a medical imaging classification task on the challenging ADNI database where a small number of 3D brain magnetic resonance images (MRIs) are considered and augmented using the proposed VAE framework. In each case, the proposed method allows for a significant and reliable gain in the classification metrics. For instance, balanced accuracy jumps from 66.3% to 74.3% for a state-of-the-art convolutional neural network classifier trained with 50 MRIs of cognitively normal (CN) and 50 Alzheimer disease (AD) patients and from 77.7% to 86.3% when trained with 243 CN and 210 AD while improving greatly sensitivity and specificity metrics.
Collapse
|
49
|
Gao C, Killeen BD, Hu Y, Grupp RB, Taylor RH, Armand M, Unberath M. Synthetic data accelerates the development of generalizable learning-based algorithms for X-ray image analysis. NAT MACH INTELL 2023; 5:294-308. [PMID: 38523605 PMCID: PMC10959504 DOI: 10.1038/s42256-023-00629-1] [Citation(s) in RCA: 23] [Impact Index Per Article: 11.5] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/14/2022] [Accepted: 02/06/2023] [Indexed: 03/26/2024]
Abstract
Artificial intelligence (AI) now enables automated interpretation of medical images. However, AI's potential use for interventional image analysis remains largely untapped. This is because the post hoc analysis of data collected during live procedures has fundamental and practical limitations, including ethical considerations, expense, scalability, data integrity and a lack of ground truth. Here we demonstrate that creating realistic simulated images from human models is a viable alternative and complement to large-scale in situ data collection. We show that training AI image analysis models on realistically synthesized data, combined with contemporary domain generalization techniques, results in machine learning models that on real data perform comparably to models trained on a precisely matched real data training set. We find that our model transfer paradigm for X-ray image analysis, which we refer to as SyntheX, can even outperform real-data-trained models due to the effectiveness of training on a larger dataset. SyntheX provides an opportunity to markedly accelerate the conception, design and evaluation of X-ray-based intelligent systems. In addition, SyntheX provides the opportunity to test novel instrumentation, design complementary surgical approaches, and envision novel techniques that improve outcomes, save time or mitigate human error, free from the ethical and practical considerations of live human data collection.
Collapse
Affiliation(s)
- Cong Gao
- Department of Computer Science, Johns Hopkins University, Baltimore, MD, USA
| | - Benjamin D. Killeen
- Department of Computer Science, Johns Hopkins University, Baltimore, MD, USA
| | - Yicheng Hu
- Department of Computer Science, Johns Hopkins University, Baltimore, MD, USA
| | - Robert B. Grupp
- Department of Computer Science, Johns Hopkins University, Baltimore, MD, USA
| | - Russell H. Taylor
- Department of Computer Science, Johns Hopkins University, Baltimore, MD, USA
| | - Mehran Armand
- Department of Computer Science, Johns Hopkins University, Baltimore, MD, USA
- Department of Orthopaedic Surgery, Johns Hopkins Applied Physics Laboratory, Baltimore, MD, USA
| | - Mathias Unberath
- Department of Computer Science, Johns Hopkins University, Baltimore, MD, USA
| |
Collapse
|
50
|
Suresh K, Cohen MS, Hartnick CJ, Bartholomew RA, Lee DJ, Crowson MG. Generation of synthetic tympanic membrane images: Development, human validation, and clinical implications of synthetic data. PLOS DIGITAL HEALTH 2023; 2:e0000202. [PMID: 36827244 PMCID: PMC9956018 DOI: 10.1371/journal.pdig.0000202] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 07/27/2022] [Accepted: 01/24/2023] [Indexed: 02/25/2023]
Abstract
Synthetic clinical images could augment real medical image datasets, a novel approach in otolaryngology-head and neck surgery (OHNS). Our objective was to develop a generative adversarial network (GAN) for tympanic membrane images and to validate the quality of synthetic images with human reviewers. Our model was developed using a state-of-the-art GAN architecture, StyleGAN2-ADA. The network was trained on intraoperative high-definition (HD) endoscopic images of tympanic membranes collected from pediatric patients undergoing myringotomy with possible tympanostomy tube placement. A human validation survey was administered to a cohort of OHNS and pediatrics trainees at our institution. The primary measure of model quality was the Frechet Inception Distance (FID), a metric comparing the distribution of generated images with the distribution of real images. The measures used for human reviewer validation were the sensitivity, specificity, and area under the curve (AUC) for humans' ability to discern synthetic from real images. Our dataset comprised 202 images. The best GAN was trained at 512x512 image resolution with a FID of 47.0. The progression of images through training showed stepwise "learning" of the anatomic features of a tympanic membrane. The validation survey was taken by 65 persons who reviewed 925 images. Human reviewers demonstrated a sensitivity of 66%, specificity of 73%, and AUC of 0.69 for the detection of synthetic images. In summary, we successfully developed a GAN to produce synthetic tympanic membrane images and validated this with human reviewers. These images could be used to bolster real datasets with various pathologies and develop more robust deep learning models such as those used for diagnostic predictions from otoscopic images. However, caution should be exercised with the use of synthetic data given issues regarding data diversity and performance validation. Any model trained using synthetic data will require robust external validation to ensure validity and generalizability.
Collapse
Affiliation(s)
- Krish Suresh
- Department of Otolaryngology-Head & Neck Surgery, Massachusetts Eye & Ear, Boston, Massachusetts, United States of America
- Department of Otolaryngology-Head & Neck Surgery, Harvard Medical School, Boston, Massachusetts, United States of America
- * E-mail:
| | - Michael S. Cohen
- Department of Otolaryngology-Head & Neck Surgery, Massachusetts Eye & Ear, Boston, Massachusetts, United States of America
- Department of Otolaryngology-Head & Neck Surgery, Harvard Medical School, Boston, Massachusetts, United States of America
| | - Christopher J. Hartnick
- Department of Otolaryngology-Head & Neck Surgery, Massachusetts Eye & Ear, Boston, Massachusetts, United States of America
- Department of Otolaryngology-Head & Neck Surgery, Harvard Medical School, Boston, Massachusetts, United States of America
| | - Ryan A. Bartholomew
- Department of Otolaryngology-Head & Neck Surgery, Massachusetts Eye & Ear, Boston, Massachusetts, United States of America
- Department of Otolaryngology-Head & Neck Surgery, Harvard Medical School, Boston, Massachusetts, United States of America
| | - Daniel J. Lee
- Department of Otolaryngology-Head & Neck Surgery, Massachusetts Eye & Ear, Boston, Massachusetts, United States of America
- Department of Otolaryngology-Head & Neck Surgery, Harvard Medical School, Boston, Massachusetts, United States of America
| | - Matthew G. Crowson
- Department of Otolaryngology-Head & Neck Surgery, Massachusetts Eye & Ear, Boston, Massachusetts, United States of America
- Department of Otolaryngology-Head & Neck Surgery, Harvard Medical School, Boston, Massachusetts, United States of America
| |
Collapse
|