1
|
Yuan H, Zhu M, Yang R, Liu H, Li I, Hong C. Rethinking Domain-Specific Pretraining by Supervised or Self-Supervised Learning for Chest Radiograph Classification: A Comparative Study Against ImageNet Counterparts in Cold-Start Active Learning. HEALTH CARE SCIENCE 2025; 4:110-143. [PMID: 40241982 PMCID: PMC11997468 DOI: 10.1002/hcs2.70009] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 10/10/2024] [Revised: 01/05/2025] [Accepted: 01/26/2025] [Indexed: 04/18/2025]
Abstract
Objective Deep learning (DL) has become the prevailing method in chest radiograph analysis, yet its performance heavily depends on large quantities of annotated images. To mitigate the cost, cold-start active learning (AL), comprising an initialization followed by subsequent learning, selects a small subset of informative data points for labeling. Recent advancements in pretrained models by supervised or self-supervised learning tailored to chest radiograph have shown broad applicability to diverse downstream tasks. However, their potential in cold-start AL remains unexplored. Methods To validate the efficacy of domain-specific pretraining, we compared two foundation models: supervised TXRV and self-supervised REMEDIS with their general domain counterparts pretrained on ImageNet. Model performance was evaluated at both initialization and subsequent learning stages on two diagnostic tasks: psychiatric pneumonia and COVID-19. For initialization, we assessed their integration with three strategies: diversity, uncertainty, and hybrid sampling. For subsequent learning, we focused on uncertainty sampling powered by different pretrained models. We also conducted statistical tests to compare the foundation models with ImageNet counterparts, investigate the relationship between initialization and subsequent learning, examine the performance of one-shot initialization against the full AL process, and investigate the influence of class balance in initialization samples on initialization and subsequent learning. Results First, domain-specific foundation models failed to outperform ImageNet counterparts in six out of eight experiments on informative sample selection. Both domain-specific and general pretrained models were unable to generate representations that could substitute for the original images as model inputs in seven of the eight scenarios. However, pretrained model-based initialization surpassed random sampling, the default approach in cold-start AL. Second, initialization performance was positively correlated with subsequent learning performance, highlighting the importance of initialization strategies. Third, one-shot initialization performed comparably to the full AL process, demonstrating the potential of reducing experts' repeated waiting during AL iterations. Last, a U-shaped correlation was observed between the class balance of initialization samples and model performance, suggesting that the class balance is more strongly associated with performance at middle budget levels than at low or high budgets. Conclusions In this study, we highlighted the limitations of medical pretraining compared to general pretraining in the context of cold-start AL. We also identified promising outcomes related to cold-start AL, including initialization based on pretrained models, the positive influence of initialization on subsequent learning, the potential for one-shot initialization, and the influence of class balance on middle-budget AL. Researchers are encouraged to improve medical pretraining for versatile DL foundations and explore novel AL methods.
Collapse
Affiliation(s)
- Han Yuan
- Duke‐NUS Medical School, Centre for Quantitative MedicineSingaporeSingapore
| | - Mingcheng Zhu
- Duke‐NUS Medical School, Centre for Quantitative MedicineSingaporeSingapore
- Department of Engineering ScienceUniversity of OxfordOxfordUK
| | - Rui Yang
- Duke‐NUS Medical School, Centre for Quantitative MedicineSingaporeSingapore
| | - Han Liu
- Department of Computer ScienceVanderbilt UniversityNashvilleTennesseeUSA
| | - Irene Li
- Information Technology CenterUniversity of TokyoBunkyo‐kuJapan
| | - Chuan Hong
- Department of Biostatistics and BioinformaticsDuke UniversityDurhamNorth CarolinaUSA
| |
Collapse
|
2
|
Fu X, Lin R, Du W, Tavares A, Liang Y. Explainable hybrid transformer for multi-classification of lung disease using chest X-rays. Sci Rep 2025; 15:6650. [PMID: 39994381 PMCID: PMC11850790 DOI: 10.1038/s41598-025-90607-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/23/2024] [Accepted: 02/14/2025] [Indexed: 02/26/2025] Open
Abstract
Lung disease is an infection that causes chronic inflammation of the human lung cells, which is one of the major causes of death around the world. Thoracic X-ray medical image is a well-known cheap screening approach used for lung disease detection. Deep learning networks, which are used to identify disease features in X-rays medical images, diagnosing a variety of lung diseases, are playing an increasingly important role in assisting clinical diagnosis. This paper proposes an explainable transformer with a hybrid network structure (LungMaxViT) combining CNN initial stage block with SE block to improve feature recognition for predicting Chest X-ray images for multiple lung disease classification. We contrast four classical pre-training models (ResNet50, MobileNetV2, ViT and MaxViT) through transfer learning based on two public datasets. The LungMaxVit, based on maxvit pre-trained with ImageNet 1K datasets, is a hybrid transformer with fine-tuning hyperparameters on the both X-ray datasets. The LungMaxVit outperforms all the four mentioned models, achieving a classification accuracy of 96.8%, AUC scores of 98.3%, and F1 scores of 96.7% on the COVID-19 dataset, while AUC scores of 93.2% and F1 scores of 70.7% on the Chest X-ray 14 dataset. The LungMaxVit distinguishes by its superior performance in terms of Accuracy, AUC and F1-score compared with other hybrids Networks. Several enhancement techniques, such as CLAHE, flipping and denoising, are employed to improve the classification performance of our study. The Grad-CAM visual technique is leveraged to represent the heat map of disease detection, explaining the consistency among clinical doctors and neural network models in the treatment of lung disease from Chest X-ray. The LungMaxVit shows the robust results and generalization in detecting multiple lung lesions and COVID-19 on Chest X-ray images.
Collapse
Affiliation(s)
- Xiaoyang Fu
- School of Computer Science, Zhuhai College of Science and Technology, Zhuhai, 519040, China
| | - Rongbin Lin
- School of Computer Science, Zhuhai College of Science and Technology, Zhuhai, 519040, China
| | - Wei Du
- School of Computer Science and Technology, Jilin University, Changchun, 130012, China
| | - Adriano Tavares
- Department of Industrial Electronics, University of Minho, 4800-058, Guimarães, Portugal
| | - Yanchun Liang
- School of Computer Science, Zhuhai College of Science and Technology, Zhuhai, 519040, China.
- School of Computer Science and Technology, Jilin University, Changchun, 130012, China.
| |
Collapse
|
3
|
Milne MR, Ahmad HK, Buchlak QD, Esmaili N, Tang C, Seah J, Ektas N, Brotchie P, Marwick TH, Jones CM. Applications and potential of machine, learning augmented chest X-ray interpretation in cardiology. Minerva Cardiol Angiol 2025; 73:8-22. [PMID: 39535525 DOI: 10.23736/s2724-5683.24.06288-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2024]
Abstract
The chest X-ray (CXR) has a wide range of clinical indications in the field of cardiology, from the assessment of acute pathology to disease surveillance and screening. Despite many technological advancements, CXR interpretation error rates have remained constant for decades. The application of machine learning has the potential to substantially improve clinical workflow efficiency, pathology detection accuracy, error rates and clinical decision making in cardiology. To date, machine learning has been developed to improve image processing, facilitate pathology detection, optimize the clinical workflow, and facilitate risk stratification. This review explores the current and potential future applications of machine learning for chest radiography to facilitate clinical decision making in cardiology. It maps the current state of the science and considers additional potential use cases from the perspective of clinicians and technologists actively engaged in the development and deployment of deep learning driven clinical decision support systems.
Collapse
Affiliation(s)
| | | | - Quinlan D Buchlak
- Annalise.ai, Sydney, Australia
- School of Medicine, University of Notre Dame Australia, Sydney, Australia
- Department of Neurosurgery, Monash Health, Melbourne, Australia
| | - Nazanin Esmaili
- School of Medicine, University of Notre Dame Australia, Sydney, Australia
- Faculty of Engineering and Information Technology, University of Technology Sydney, Sydney, Australia
| | | | - Jarrel Seah
- Annalise.ai, Sydney, Australia
- Department of Radiology, Alfred Health, Melbourne, Australia
| | | | | | | | - Catherine M Jones
- Annalise.ai, Sydney, Australia
- I-MED Radiology Network, Brisbane, Australia
- Faculty of Medicine and Health, University of Sydney, Sydney, Australia
| |
Collapse
|
4
|
Singh A, Gorade V, Mishra D. MLVICX: Multi-Level Variance-Covariance Exploration for Chest X-Ray Self-Supervised Representation Learning. IEEE J Biomed Health Inform 2024; 28:7480-7490. [PMID: 39240749 DOI: 10.1109/jbhi.2024.3455337] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/08/2024]
Abstract
Self-supervised learning (SSL) reduces the need for manual annotation in deep learning models for medical image analysis. By learning the representations from unablelled data, self-supervised models perform well on tasks that require little to no fine-tuning. However, for medical images, like chest X-rays, characterised by complex anatomical structures and diverse clinical conditions, a need arises for representation learning techniques that encode fine-grained details while preserving the broader contextual information. In this context, we introduce MLVICX (Multi-Level Variance-Covariance Exploration for Chest X-ray Self-Supervised Representation Learning), an approach to capture rich representations in the form of embeddings from chest X-ray images. Central to our approach is a novel multi-level variance and covariance exploration strategy that effectively enables the model to detect diagnostically meaningful patterns while reducing redundancy. MLVICX promotes the retention of critical medical insights by adapting global and local contextual details and enhancing the variance and covariance of the learned embeddings. We demonstrate the performance of MLVICX in advancing self-supervised chest X-ray representation learning through comprehensive experiments. The performance enhancements we observe across various downstream tasks highlight the significance of the proposed approach in enhancing the utility of chest X-ray embeddings for precision medical diagnosis and comprehensive image analysis. For pertaining, we used the NIH-Chest X-ray dataset. Downstream tasks utilized NIH-Chest X-ray, Vinbig-CXR, RSNA pneumonia, and SIIM-ACR Pneumothorax datasets. Overall, we observe up to 3% performance gain over SOTA SSL approaches in various downstream tasks. Additionally, to demonstrate generalizability of our method, we conducted additional experiments on fundus images and observed superior performance on multiple datasets. Codes are available at GitHub.
Collapse
|
5
|
Anderson PG, Tarder-Stoll H, Alpaslan M, Keathley N, Levin DL, Venkatesh S, Bartel E, Sicular S, Howell S, Lindsey RV, Jones RM. Deep learning improves physician accuracy in the comprehensive detection of abnormalities on chest X-rays. Sci Rep 2024; 14:25151. [PMID: 39448764 PMCID: PMC11502915 DOI: 10.1038/s41598-024-76608-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2024] [Accepted: 10/15/2024] [Indexed: 10/26/2024] Open
Abstract
Chest X-rays are the most commonly performed medical imaging exam, yet they are often misinterpreted by physicians. Here, we present an FDA-cleared, artificial intelligence (AI) system which uses a deep learning algorithm to assist physicians in the comprehensive detection and localization of abnormalities on chest X-rays. We trained and tested the AI system on a large dataset, assessed generalizability on publicly available data, and evaluated radiologist and non-radiologist physician accuracy when unaided and aided by the AI system. The AI system accurately detected chest X-ray abnormalities (AUC: 0.976, 95% bootstrap CI: 0.975, 0.976) and generalized to a publicly available dataset (AUC: 0.975, 95% bootstrap CI: 0.971, 0.978). Physicians showed significant improvements in detecting abnormalities on chest X-rays when aided by the AI system compared to when unaided (difference in AUC: 0.101, p < 0.001). Non-radiologist physicians detected abnormalities on chest X-ray exams as accurately as radiologists when aided by the AI system and were faster at evaluating chest X-rays when aided compared to unaided. Together, these results show that the AI system is accurate and reduces physician errors in chest X-ray evaluation, which highlights the potential of AI systems to improve access to fast, high-quality radiograph interpretation.
Collapse
Affiliation(s)
- Pamela G Anderson
- Imagen Technologies, 224 W 35th St Ste 500, New York, NY, 10001, USA.
| | | | - Mehmet Alpaslan
- Imagen Technologies, 224 W 35th St Ste 500, New York, NY, 10001, USA
| | - Nora Keathley
- Imagen Technologies, 224 W 35th St Ste 500, New York, NY, 10001, USA
| | - David L Levin
- Department of Radiology, Stanford University School of Medicine, 453 Quarry Rd, Palo Alto, CA, 94305, USA
| | - Srivas Venkatesh
- Imagen Technologies, 224 W 35th St Ste 500, New York, NY, 10001, USA
| | - Elliot Bartel
- Imagen Technologies, 224 W 35th St Ste 500, New York, NY, 10001, USA
| | - Serge Sicular
- Imagen Technologies, 224 W 35th St Ste 500, New York, NY, 10001, USA
- The Mount Sinai Hospital, 1 Gustave L. Levy Place, New York, NY, 10029, USA
| | - Scott Howell
- Imagen Technologies, 224 W 35th St Ste 500, New York, NY, 10001, USA
| | - Robert V Lindsey
- Imagen Technologies, 224 W 35th St Ste 500, New York, NY, 10001, USA
| | - Rebecca M Jones
- Imagen Technologies, 224 W 35th St Ste 500, New York, NY, 10001, USA
| |
Collapse
|
6
|
Cheng CT, Kuo LW, Ouyang CH, Hsu CP, Lin WC, Fu CY, Kang SC, Liao CH. Development and evaluation of a deep learning-based model for simultaneous detection and localization of rib and clavicle fractures in trauma patients' chest radiographs. Trauma Surg Acute Care Open 2024; 9:e001300. [PMID: 38646620 PMCID: PMC11029226 DOI: 10.1136/tsaco-2023-001300] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/23/2024] Open
Abstract
Purpose To develop a rib and clavicle fracture detection model for chest radiographs in trauma patients using a deep learning (DL) algorithm. Materials and methods We retrospectively collected 56 145 chest X-rays (CXRs) from trauma patients in a trauma center between August 2008 and December 2016. A rib/clavicle fracture detection DL algorithm was trained using this data set with 991 (1.8%) images labeled by experts with fracture site locations. The algorithm was tested on independently collected 300 CXRs in 2017. An external test set was also collected from hospitalized trauma patients in a regional hospital for evaluation. The receiver operating characteristic curve with area under the curve (AUC), accuracy, sensitivity, specificity, precision, and negative predictive value of the model on each test set was evaluated. The prediction probability on the images was visualized as heatmaps. Results The trained DL model achieved an AUC of 0.912 (95% CI 87.8 to 94.7) on the independent test set. The accuracy, sensitivity, and specificity on the given cut-off value are 83.7, 86.8, and 80.4, respectively. On the external test set, the model had a sensitivity of 88.0 and an accuracy of 72.5. While the model exhibited a slight decrease in accuracy on the external test set, it maintained its sensitivity in detecting fractures. Conclusion The algorithm detects rib and clavicle fractures concomitantly in the CXR of trauma patients with high accuracy in locating lesions through heatmap visualization.
Collapse
Affiliation(s)
- Chi-Tung Cheng
- Department of Trauma and Emergency Surgery, Chang Gung Memorial Hospital Linkou, Taoyuan, Taiwan
- Department of medicine, Chang Gung university, Taoyuan, Taiwan
| | - Ling-Wei Kuo
- Department of Trauma and Emergency Surgery, Chang Gung Memorial Hospital Linkou, Taoyuan, Taiwan
- Department of medicine, Chang Gung university, Taoyuan, Taiwan
| | - Chun-Hsiang Ouyang
- Department of Trauma and Emergency Surgery, Chang Gung Memorial Hospital Linkou, Taoyuan, Taiwan
- Department of medicine, Chang Gung university, Taoyuan, Taiwan
| | - Chi-Po Hsu
- Department of Trauma and Emergency Surgery, Chang Gung Memorial Hospital Linkou, Taoyuan, Taiwan
- Department of medicine, Chang Gung university, Taoyuan, Taiwan
| | - Wei-Cheng Lin
- Department of Electrical Engineering, Chang Gung University, Taoyuan, Taiwan
| | - Chih-Yuan Fu
- Department of Trauma and Emergency Surgery, Chang Gung Memorial Hospital Linkou, Taoyuan, Taiwan
- Department of medicine, Chang Gung university, Taoyuan, Taiwan
| | - Shih-Ching Kang
- Department of Trauma and Emergency Surgery, Chang Gung Memorial Hospital Linkou, Taoyuan, Taiwan
- Department of medicine, Chang Gung university, Taoyuan, Taiwan
| | - Chien-Hung Liao
- Department of Trauma and Emergency Surgery, Chang Gung Memorial Hospital Linkou, Taoyuan, Taiwan
- Department of medicine, Chang Gung university, Taoyuan, Taiwan
| |
Collapse
|
7
|
Fan W, Yang Y, Qi J, Zhang Q, Liao C, Wen L, Wang S, Wang G, Xia Y, Wu Q, Fan X, Chen X, He M, Xiao J, Yang L, Liu Y, Chen J, Wang B, Zhang L, Yang L, Gan H, Zhang S, Liu G, Ge X, Cai Y, Zhao G, Zhang X, Xie M, Xu H, Zhang Y, Chen J, Li J, Han S, Mu K, Xiao S, Xiong T, Nian Y, Zhang D. A deep-learning-based framework for identifying and localizing multiple abnormalities and assessing cardiomegaly in chest X-ray. Nat Commun 2024; 15:1347. [PMID: 38355644 PMCID: PMC10867134 DOI: 10.1038/s41467-024-45599-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/19/2022] [Accepted: 01/30/2024] [Indexed: 02/16/2024] Open
Abstract
Accurate identification and localization of multiple abnormalities are crucial steps in the interpretation of chest X-rays (CXRs); however, the lack of a large CXR dataset with bounding boxes severely constrains accurate localization research based on deep learning. We created a large CXR dataset named CXR-AL14, containing 165,988 CXRs and 253,844 bounding boxes. On the basis of this dataset, a deep-learning-based framework was developed to identify and localize 14 common abnormalities and calculate the cardiothoracic ratio (CTR) simultaneously. The mean average precision values obtained by the model for 14 abnormalities reached 0.572-0.631 with an intersection-over-union threshold of 0.5, and the intraclass correlation coefficient of the CTR algorithm exceeded 0.95 on the held-out, multicentre and prospective test datasets. This framework shows an excellent performance, good generalization ability and strong clinical applicability, which is superior to senior radiologists and suitable for routine clinical settings.
Collapse
Affiliation(s)
- Weijie Fan
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Yi Yang
- Department of Digital Medicine, School of Biomedical Engineering and Imaging Medicine, Army Medical University, Chongqing, 400038, P. R. China
| | - Jing Qi
- Department of Digital Medicine, School of Biomedical Engineering and Imaging Medicine, Army Medical University, Chongqing, 400038, P. R. China
| | - Qichuan Zhang
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Cuiwei Liao
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Li Wen
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Shuang Wang
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Guangxian Wang
- Department of Radiology, People's Hospital of Banan, Chongqing Medical University, Chongqing, 401320, P. R. China
| | - Yu Xia
- Department of Radiology, Xishui hospital of Traditional Chinese Medicine, Zunyi of Guizhou province, 564600, P. R. China
| | - Qihua Wu
- Department of Radiology, People's Hospital of Nanchuan, Chongqing, 408400, P. R. China
| | - Xiaotao Fan
- Department of Radiology, Fengdu People's Hospital, Chongqing, 408200, P. R. China
| | - Xingcai Chen
- Department of Digital Medicine, School of Biomedical Engineering and Imaging Medicine, Army Medical University, Chongqing, 400038, P. R. China
| | - Mi He
- Department of Digital Medicine, School of Biomedical Engineering and Imaging Medicine, Army Medical University, Chongqing, 400038, P. R. China
| | - JingJing Xiao
- Department of Medical Engineering, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Liu Yang
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Yun Liu
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Jia Chen
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Bing Wang
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Lei Zhang
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Liuqing Yang
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Hui Gan
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Shushu Zhang
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Guofang Liu
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Xiaodong Ge
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Yuanqing Cai
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Gang Zhao
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Xi Zhang
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Mingxun Xie
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Huilin Xu
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Yi Zhang
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Jiao Chen
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Jun Li
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Shuang Han
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Ke Mu
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Shilin Xiao
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Tingwei Xiong
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China
| | - Yongjian Nian
- Department of Digital Medicine, School of Biomedical Engineering and Imaging Medicine, Army Medical University, Chongqing, 400038, P. R. China.
| | - Dong Zhang
- Department of Radiology, Second Affiliated Hospital, Army Medical University, Chongqing, 400037, P. R. China.
| |
Collapse
|
8
|
Cid YD, Macpherson M, Gervais-Andre L, Zhu Y, Franco G, Santeramo R, Lim C, Selby I, Muthuswamy K, Amlani A, Hopewell H, Indrajeet D, Liakata M, Hutchinson CE, Goh V, Montana G. Development and validation of open-source deep neural networks for comprehensive chest x-ray reading: a retrospective, multicentre study. Lancet Digit Health 2024; 6:e44-e57. [PMID: 38071118 DOI: 10.1016/s2589-7500(23)00218-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/29/2022] [Revised: 10/03/2023] [Accepted: 10/16/2023] [Indexed: 12/22/2023]
Abstract
BACKGROUND Artificial intelligence (AI) systems for automated chest x-ray interpretation hold promise for standardising reporting and reducing delays in health systems with shortages of trained radiologists. Yet, there are few freely accessible AI systems trained on large datasets for practitioners to use with their own data with a view to accelerating clinical deployment of AI systems in radiology. We aimed to contribute an AI system for comprehensive chest x-ray abnormality detection. METHODS In this retrospective cohort study, we developed open-source neural networks, X-Raydar and X-Raydar-NLP, for classifying common chest x-ray findings from images and their free-text reports. Our networks were developed using data from six UK hospitals from three National Health Service (NHS) Trusts (University Hospitals Coventry and Warwickshire NHS Trust, University Hospitals Birmingham NHS Foundation Trust, and University Hospitals Leicester NHS Trust) collectively contributing 2 513 546 chest x-ray studies taken from a 13-year period (2006-19), which yielded 1 940 508 usable free-text radiological reports written by the contemporary assessing radiologist (collectively referred to as the "historic reporters") and 1 896 034 frontal images. Chest x-rays were labelled using a taxonomy of 37 findings by a custom-trained natural language processing (NLP) algorithm, X-Raydar-NLP, from the original free-text reports. X-Raydar-NLP was trained on 23 230 manually annotated reports and tested on 4551 reports from all hospitals. 1 694 921 labelled images from the training set and 89 238 from the validation set were then used to train a multi-label image classifier. Our algorithms were evaluated on three retrospective datasets: a set of exams sampled randomly from the full NHS dataset reported during clinical practice and annotated using NLP (n=103 328); a consensus set sampled from all six hospitals annotated by three expert radiologists (two independent annotators for each image and a third consultant to facilitate disagreement resolution) under research conditions (n=1427); and an independent dataset, MIMIC-CXR, consisting of NLP-annotated exams (n=252 374). FINDINGS X-Raydar achieved a mean AUC of 0·919 (SD 0·039) on the auto-labelled set, 0·864 (0·102) on the consensus set, and 0·842 (0·074) on the MIMIC-CXR test, demonstrating similar performance to the historic clinical radiologist reporters, as assessed on the consensus set, for multiple clinically important findings, including pneumothorax, parenchymal opacification, and parenchymal mass or nodules. On the consensus set, X-Raydar outperformed historical reporter balanced accuracy with significance on 27 of 37 findings, was non-inferior on nine, and inferior on one finding, resulting in an average improvement of 13·3% (SD 13·1) to 0·763 (0·110), including a mean 5·6% (13·2) improvement in critical findings to 0·826 (0·119). INTERPRETATION Our study shows that automated classification of chest x-rays under a comprehensive taxonomy can achieve performance levels similar to those of historical reporters and exhibit robust generalisation to external data. The open-sourced neural networks can serve as foundation models for further research and are freely available to the research community. FUNDING Wellcome Trust.
Collapse
Affiliation(s)
| | - Matthew Macpherson
- WMG, University of Warwick, Coventry, UK; Mathematics Institute, University of Warwick, Coventry, UK
| | - Louise Gervais-Andre
- School of Biomedical Engineering and Imaging Sciences, King's College London, London, UK
| | - Yuanyi Zhu
- WMG, University of Warwick, Coventry, UK; Mathematics Institute, University of Warwick, Coventry, UK
| | | | | | - Chee Lim
- Department of Radiology, University Hospitals Birmingham NHS Foundation Trust, Birmingham, UK
| | - Ian Selby
- Department of Radiology, University of Cambridge, Cambridge, UK
| | | | - Ashik Amlani
- Department of Radiology, Guy's and St Thomas' NHS Foundation Trust, London, UK
| | - Heath Hopewell
- Department of Radiology, University Hospitals of Leicester NHS Trust, Leicester, UK
| | - Das Indrajeet
- Department of Radiology, University Hospitals of Leicester NHS Trust, Leicester, UK
| | - Maria Liakata
- The Alan Turing Institute, London, UK; Institute of Applied Data Science, Queen Mary University of London, London, UK
| | - Charles E Hutchinson
- Warwick Medical School, University of Warwick, Coventry, UK; Department of Radiology, University Hospitals Coventry and Warwickshire NHS Trust, Coventry, UK
| | - Vicky Goh
- School of Biomedical Engineering and Imaging Sciences, King's College London, London, UK; Department of Radiology, Guy's and St Thomas' NHS Foundation Trust, London, UK
| | - Giovanni Montana
- WMG, University of Warwick, Coventry, UK; Department of Statistics, University of Warwick, Coventry, UK; The Alan Turing Institute, London, UK.
| |
Collapse
|
9
|
Ahmad Z, Malik AK, Qamar N, Islam SU. Efficient Thorax Disease Classification and Localization Using DCNN and Chest X-ray Images. Diagnostics (Basel) 2023; 13:3462. [PMID: 37998598 PMCID: PMC10669971 DOI: 10.3390/diagnostics13223462] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/31/2023] [Revised: 11/05/2023] [Accepted: 11/09/2023] [Indexed: 11/25/2023] Open
Abstract
Thorax disease is a life-threatening disease caused by bacterial infections that occur in the lungs. It could be deadly if not treated at the right time, so early diagnosis of thoracic diseases is vital. The suggested study can assist radiologists in more swiftly diagnosing thorax disorders and in the rapid airport screening of patients with a thorax disease, such as pneumonia. This paper focuses on automatically detecting and localizing thorax disease using chest X-ray images. It provides accurate detection and localization using DenseNet-121 which is foundation of our proposed framework, called Z-Net. The proposed framework utilizes the weighted cross-entropy loss function (W-CEL) that manages class imbalance issue in the ChestX-ray14 dataset, which helped in achieving the highest performance as compared to the previous models. The 112,120 images contained in the ChestX-ray14 dataset (60,412 images are normal, and the rest contain thorax diseases) were preprocessed and then trained for classification and localization. This work uses computer-aided diagnosis (CAD) system that supports development of highly accurate and precise computer-aided systems. We aim to develop a CAD system using a deep learning approach. Our quantitative results show high AUC scores in comparison with the latest research works. The proposed approach achieved the highest mean AUC score of 85.8%. This is the highest accuracy documented in the literature for any related model.
Collapse
Affiliation(s)
- Zeeshan Ahmad
- Department of Computer Science, COMSATS University Islamabad, Islamabad 45550, Pakistan
| | - Ahmad Kamran Malik
- Department of Computer Science, COMSATS University Islamabad, Islamabad 45550, Pakistan
| | - Nafees Qamar
- School of Health and Behavioral Sciences, Bryant University, Smithfield, RI 02917, USA
| | - Saif Ul Islam
- Department of Computer Science, Institute of Space Technology, Islamabad 44000, Pakistan
| |
Collapse
|
10
|
Liu Z, Huang B, Wen H, Lu Z, Huang Q, Jiang M, Dong C, Liu Y, Chen X, Lin H. Automatic Diagnosis of Significant Liver Fibrosis From Ultrasound B-Mode Images Using a Handcrafted-Feature-Assisted Deep Convolutional Neural Network. IEEE J Biomed Health Inform 2023; 27:4938-4949. [PMID: 37471184 DOI: 10.1109/jbhi.2023.3295078] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/22/2023]
Abstract
The accurate diagnosis of significant liver fibrosis ( ≥ F2) in patients with chronic liver disease (CLD) is critical, as ≥ F2 is a crucial factor that should be considered in selecting an antiviral therapy for these patients. This article proposes a handcrafted-feature-assisted deep convolutional neural network (HFA-DCNN) that helps radiologists automatically and accurately diagnose significant liver fibrosis from ultrasound (US) brightness (B)-mode images. The HFA-DCNN model has three main branches: one for automatic region of interest (ROI) segmentation in the US images, another for attention deep feature learning from the segmented ROI, and the third for handcrafted feature extraction. The attention deep learning features and handcrafted features are fused in the back end of the model to enable more accurate diagnosis of significant liver fibrosis. The usefulness and effectiveness of the proposed model were validated on a dataset built upon 321 CLD patients with liver fibrosis stages confirmed by pathological evaluations. In a fivefold cross validation (FFCV), the proposed model achieves accuracy, sensitivity, specificity, and area under the receiver-operating-characteristic (ROC) curve (AUC) values of 0.863 (95% confidence interval (CI) 0.820-0.899), 0.879 (95% CI 0.823-0.920), 0.872 (95% CI 0.800-0.925), and 0.925 (95% CI 0.891-0.952), which are significantly better than those obtained by the comparative methods. Given its excellent performance, the proposed HFA-DCNN model can serve as a promising tool for the noninvasive and accurate diagnosis of significant liver fibrosis in CLD patients.
Collapse
|
11
|
Devasia J, Goswami H, Lakshminarayanan S, Rajaram M, Adithan S. Observer Performance Evaluation of a Deep Learning Model for Multilabel Classification of Active Tuberculosis Lung Zone-Wise Manifestations. Cureus 2023; 15:e44954. [PMID: 37818499 PMCID: PMC10561790 DOI: 10.7759/cureus.44954] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 09/09/2023] [Indexed: 10/12/2023] Open
Abstract
Background Chest X-rays (CXRs) are widely used for cost-effective screening of active pulmonary tuberculosis despite their limitations in sensitivity and specificity when interpreted by clinicians or radiologists. To address this issue, computer-aided detection (CAD) algorithms, particularly deep learning architectures based on convolution, have been developed to automate the analysis of radiography imaging. Deep learning algorithms have shown promise in accurately classifying lung abnormalities using chest X-ray images. In this study, we utilized the EfficientNet B4 model, which was pre-trained on ImageNet with 380x380 input dimensions, using its weights for transfer learning, and was modified with a series of components including global average pooling, batch normalization, dropout, and a classifier with 12 image-wise and 44 segment-wise lung zone evaluation classes using sigmoid activation. Objectives Assess the clinical usefulness of our previously created EfficientNet B4 model in identifying lung zone-specific abnormalities related to active tuberculosis through an observer performance test involving a skilled clinician operating in tuberculosis-specific environments. Methods The ground truth was established by a radiologist who examined all sample CXRs to identify lung zone-wise abnormalities. An expert clinician working in tuberculosis-specific settings independently reviewed the same CXR with blinded access to the ground truth. Simultaneously, the CXRs were classified using the EfficientNet B4 model. The clinician's assessments were then compared with the model's predictions, and the agreement between the two was measured using the kappa coefficient, evaluating the model's performance in classifying active tuberculosis manifestations across lung zones. Results The results show a strong agreement (Kappa ≥0.81) seen for lung zone-wise abnormalities of pneumothorax, mediastinal shift, emphysema, fibrosis, calcifications, pleural effusion, and cavity. Substantial agreement (Kappa = 0.61-0.80) for cavity, mediastinal shift, volume loss, and collapsed lungs. The Kappa score for lung zone-wise abnormalities is moderate (0.41-0.60) for 39% of cases. In image-wise agreement, the EfficientNet B4 model's performance ranges from moderate to almost perfect across categories, while in lung zone-wise agreement, it varies from fair to almost perfect. The results show strong agreement between the EfficientNet B4 model and the human reader in detecting lung zone-wise and image-wise manifestations. Conclusion The clinical utility of the EfficientNet B4 models to detect the abnormalities can aid clinicians in primary care settings for screening and triaging tuberculosis where resources are constrained or overburdened.
Collapse
Affiliation(s)
- James Devasia
- Preventive Medicine, Jawaharlal Institute of Postgraduate Medical Education and Research, Puducherry, IND
| | | | - Subitha Lakshminarayanan
- Preventive Medicine, Jawaharlal Institute of Postgraduate Medical Education and Research, Puducherry, IND
| | - Manju Rajaram
- Pulmonary Medicine, Jawaharlal Institute of Postgraduate Medical Education and Research, Puducherry, IND
| | - Subathra Adithan
- Radiodiagnosis, Jawaharlal Institute of Postgraduate Medical Education and Research, Puducherry, IND
| |
Collapse
|
12
|
Liu Z, Cheng Y, Tamura S. Multi-Label Local to Global Learning: A Novel Learning Paradigm for Chest X-Ray Abnormality Classification. IEEE J Biomed Health Inform 2023; 27:4409-4420. [PMID: 37252867 DOI: 10.1109/jbhi.2023.3281466] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/01/2023]
Abstract
Deep neural network (DNN) approaches have shown remarkable progress in automatic Chest X-rays classification. However, existing methods use a training scheme that simultaneously trains all abnormalities without considering their learning priority. Inspired by the clinical practice of radiologists progressively recognizing more abnormalities and the observation that existing curriculum learning (CL) methods based on image difficulty may not be suitable for disease diagnosis, we propose a novel CL paradigm, named multi-label local to global (ML-LGL). This approach iteratively trains DNN models on gradually increasing abnormalities within the dataset, i,e, from fewer abnormalities (local) to more ones (global). At each iteration, we first build the local category by adding high-priority abnormalities for training, and the abnormality's priority is determined by our three proposed clinical knowledge-leveraged selection functions. Then, images containing abnormalities in the local category are gathered to form a new training set. The model is lastly trained on this set using a dynamic loss. Additionally, we demonstrate the superiority of ML-LGL from the perspective of the model's initial stability during training. Experimental results on three open-source datasets, PLCO, ChestX-ray14 and CheXpert show that our proposed learning paradigm outperforms baselines and achieves comparable results to state-of-the-art methods. The improved performance promises potential applications in multi-label Chest X-ray classification.
Collapse
|
13
|
Sugibayashi T, Walston SL, Matsumoto T, Mitsuyama Y, Miki Y, Ueda D. Deep learning for pneumothorax diagnosis: a systematic review and meta-analysis. Eur Respir Rev 2023; 32:32/168/220259. [PMID: 37286217 DOI: 10.1183/16000617.0259-2022] [Citation(s) in RCA: 18] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/23/2022] [Accepted: 03/16/2023] [Indexed: 06/09/2023] Open
Abstract
BACKGROUND Deep learning (DL), a subset of artificial intelligence (AI), has been applied to pneumothorax diagnosis to aid physician diagnosis, but no meta-analysis has been performed. METHODS A search of multiple electronic databases through September 2022 was performed to identify studies that applied DL for pneumothorax diagnosis using imaging. Meta-analysis via a hierarchical model to calculate the summary area under the curve (AUC) and pooled sensitivity and specificity for both DL and physicians was performed. Risk of bias was assessed using a modified Prediction Model Study Risk of Bias Assessment Tool. RESULTS In 56 of the 63 primary studies, pneumothorax was identified from chest radiography. The total AUC was 0.97 (95% CI 0.96-0.98) for both DL and physicians. The total pooled sensitivity was 84% (95% CI 79-89%) for DL and 85% (95% CI 73-92%) for physicians and the pooled specificity was 96% (95% CI 94-98%) for DL and 98% (95% CI 95-99%) for physicians. More than half of the original studies (57%) had a high risk of bias. CONCLUSIONS Our review found the diagnostic performance of DL models was similar to that of physicians, although the majority of studies had a high risk of bias. Further pneumothorax AI research is needed.
Collapse
Affiliation(s)
- Takahiro Sugibayashi
- Department of Diagnostic and Interventional Radiology, Graduate School of Medicine, Osaka Metropolitan University, Osaka, Japan
| | - Shannon L Walston
- Department of Diagnostic and Interventional Radiology, Graduate School of Medicine, Osaka Metropolitan University, Osaka, Japan
| | - Toshimasa Matsumoto
- Department of Diagnostic and Interventional Radiology, Graduate School of Medicine, Osaka Metropolitan University, Osaka, Japan
- Smart Life Science Lab, Center for Health Science Innovation, Osaka Metropolitan University, Osaka, Japan
| | - Yasuhito Mitsuyama
- Department of Diagnostic and Interventional Radiology, Graduate School of Medicine, Osaka Metropolitan University, Osaka, Japan
| | - Yukio Miki
- Department of Diagnostic and Interventional Radiology, Graduate School of Medicine, Osaka Metropolitan University, Osaka, Japan
| | - Daiju Ueda
- Department of Diagnostic and Interventional Radiology, Graduate School of Medicine, Osaka Metropolitan University, Osaka, Japan
- Smart Life Science Lab, Center for Health Science Innovation, Osaka Metropolitan University, Osaka, Japan
| |
Collapse
|
14
|
Li Q, Chen M, Geng J, Adamu MJ, Guan X. High-Resolution Network with Dynamic Convolution and Coordinate Attention for Classification of Chest X-ray Images. Diagnostics (Basel) 2023; 13:2165. [PMID: 37443559 DOI: 10.3390/diagnostics13132165] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/09/2023] [Revised: 06/12/2023] [Accepted: 06/21/2023] [Indexed: 07/15/2023] Open
Abstract
The development of automatic chest X-ray (CXR) disease classification algorithms is significant for diagnosing thoracic diseases. Owing to the characteristics of lesions in CXR images, including high similarity in appearance of the disease, varied sizes, and different occurrence locations, most existing convolutional neural network-based methods have insufficient feature extraction for thoracic lesions and struggle to adapt to changes in lesion size and location. To address these issues, this study proposes a high-resolution classification network with dynamic convolution and coordinate attention (HRCC-Net). In the method, this study suggests a parallel multi-resolution network in which a high-resolution branch acquires essential detailed features of the lesion and multi-resolution feature swapping and fusion to obtain multiple receptive fields to extract complicated disease features adequately. Furthermore, this study proposes dynamic convolution to enhance the network's ability to represent multi-scale information to accommodate lesions of diverse scales. In addition, this study introduces a coordinate attention mechanism, which enables automatic focus on pathologically relevant regions and capturing the variations in lesion location. The proposed method is evaluated on ChestX-ray14 and CheXpert datasets. The average AUC (area under ROC curve) values reach 0.845 and 0.913, respectively, indicating this method's advantages compared with the currently available methods. Meanwhile, with its specificity and sensitivity to measure the performance of medical diagnostic systems, the network can improve diagnostic efficiency while reducing the rate of misdiagnosis. The proposed algorithm has great potential for thoracic disease diagnosis and treatment.
Collapse
Affiliation(s)
- Qiang Li
- School of Microelectronics, Tianjin University, Tianjin 300072, China
| | - Mingyu Chen
- School of Microelectronics, Tianjin University, Tianjin 300072, China
| | - Jingjing Geng
- School of Microelectronics, Tianjin University, Tianjin 300072, China
| | | | - Xin Guan
- School of Microelectronics, Tianjin University, Tianjin 300072, China
| |
Collapse
|
15
|
Ahmad HK, Milne MR, Buchlak QD, Ektas N, Sanderson G, Chamtie H, Karunasena S, Chiang J, Holt X, Tang CHM, Seah JCY, Bottrell G, Esmaili N, Brotchie P, Jones C. Machine Learning Augmented Interpretation of Chest X-rays: A Systematic Review. Diagnostics (Basel) 2023; 13:diagnostics13040743. [PMID: 36832231 PMCID: PMC9955112 DOI: 10.3390/diagnostics13040743] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/19/2023] [Revised: 02/13/2023] [Accepted: 02/14/2023] [Indexed: 02/18/2023] Open
Abstract
Limitations of the chest X-ray (CXR) have resulted in attempts to create machine learning systems to assist clinicians and improve interpretation accuracy. An understanding of the capabilities and limitations of modern machine learning systems is necessary for clinicians as these tools begin to permeate practice. This systematic review aimed to provide an overview of machine learning applications designed to facilitate CXR interpretation. A systematic search strategy was executed to identify research into machine learning algorithms capable of detecting >2 radiographic findings on CXRs published between January 2020 and September 2022. Model details and study characteristics, including risk of bias and quality, were summarized. Initially, 2248 articles were retrieved, with 46 included in the final review. Published models demonstrated strong standalone performance and were typically as accurate, or more accurate, than radiologists or non-radiologist clinicians. Multiple studies demonstrated an improvement in the clinical finding classification performance of clinicians when models acted as a diagnostic assistance device. Device performance was compared with that of clinicians in 30% of studies, while effects on clinical perception and diagnosis were evaluated in 19%. Only one study was prospectively run. On average, 128,662 images were used to train and validate models. Most classified less than eight clinical findings, while the three most comprehensive models classified 54, 72, and 124 findings. This review suggests that machine learning devices designed to facilitate CXR interpretation perform strongly, improve the detection performance of clinicians, and improve the efficiency of radiology workflow. Several limitations were identified, and clinician involvement and expertise will be key to driving the safe implementation of quality CXR machine learning systems.
Collapse
Affiliation(s)
- Hassan K. Ahmad
- Annalise.ai, Sydney, NSW 2000, Australia
- Department of Emergency Medicine, Royal North Shore Hospital, Sydney, NSW 2065, Australia
- Correspondence:
| | | | - Quinlan D. Buchlak
- Annalise.ai, Sydney, NSW 2000, Australia
- School of Medicine, University of Notre Dame Australia, Sydney, NSW 2007, Australia
- Department of Neurosurgery, Monash Health, Melbourne, VIC 3168, Australia
| | | | | | | | | | - Jason Chiang
- Annalise.ai, Sydney, NSW 2000, Australia
- Department of General Practice, University of Melbourne, Melbourne, VIC 3010, Australia
- Westmead Applied Research Centre, University of Sydney, Sydney, NSW 2006, Australia
| | | | | | - Jarrel C. Y. Seah
- Annalise.ai, Sydney, NSW 2000, Australia
- Department of Radiology, Alfred Health, Melbourne, VIC 3004, Australia
| | | | - Nazanin Esmaili
- School of Medicine, University of Notre Dame Australia, Sydney, NSW 2007, Australia
- Faculty of Engineering and Information Technology, University of Technology Sydney, Sydney, NSW 2007, Australia
| | - Peter Brotchie
- Annalise.ai, Sydney, NSW 2000, Australia
- Department of Radiology, St Vincent’s Health Australia, Melbourne, VIC 3065, Australia
| | - Catherine Jones
- Annalise.ai, Sydney, NSW 2000, Australia
- I-MED Radiology Network, Brisbane, QLD 4006, Australia
- School of Public and Preventive Health, Monash University, Clayton, VIC 3800, Australia
- Department of Clinical Imaging Science, University of Sydney, Sydney, NSW 2006, Australia
| |
Collapse
|
16
|
Ibragimov B, Arzamasov K, Maksudov B, Kiselev S, Mongolin A, Mustafaev T, Ibragimova D, Evteeva K, Andreychenko A, Morozov S. A 178-clinical-center experiment of integrating AI solutions for lung pathology diagnosis. Sci Rep 2023; 13:1135. [PMID: 36670118 PMCID: PMC9859802 DOI: 10.1038/s41598-023-27397-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/02/2021] [Accepted: 01/02/2023] [Indexed: 01/22/2023] Open
Abstract
In 2020, an experiment testing AI solutions for lung X-ray analysis on a multi-hospital network was conducted. The multi-hospital network linked 178 Moscow state healthcare centers, where all chest X-rays from the network were redirected to a research facility, analyzed with AI, and returned to the centers. The experiment was formulated as a public competition with monetary awards for participating industrial and research teams. The task was to perform the binary detection of abnormalities from chest X-rays. For the objective real-life evaluation, no training X-rays were provided to the participants. This paper presents one of the top-performing AI frameworks from this experiment. First, the framework used two EfficientNets, histograms of gradients, Haar feature ensembles, and local binary patterns to recognize whether an input image represents an acceptable lung X-ray sample, meaning the X-ray is not grayscale inverted, is a frontal chest X-ray, and completely captures both lung fields. Second, the framework extracted the region with lung fields and then passed them to a multi-head DenseNet, where the heads recognized the patient's gender, age and the potential presence of abnormalities, and generated the heatmap with the abnormality regions highlighted. During one month of the experiment from 11.23.2020 to 12.25.2020, 17,888 cases have been analyzed by the framework with 11,902 cases having radiological reports with the reference diagnoses that were unequivocally parsed by the experiment organizers. The performance measured in terms of the area under receiving operator curve (AUC) was 0.77. The AUC for individual diseases ranged from 0.55 for herniation to 0.90 for pneumothorax.
Collapse
Affiliation(s)
- Bulat Ibragimov
- Department of Computer Science, University of Copenhagen, Copenhagen, Denmark.
| | - Kirill Arzamasov
- Research and Practical Clinical Center for Diagnostics and Telemedicine Technologies of the Moscow Healthcare Department, Moscow, Russia
| | - Bulat Maksudov
- School of Electronic Engineering, Dublin City University, Dublin, Ireland
| | | | - Alexander Mongolin
- Innopolis University, Innopolis, Russia
- Nova Information Management School, Universidade Nova de Lisboa, Lisbon, Portugal
| | - Tamerlan Mustafaev
- Innopolis University, Innopolis, Russia
- University Clinic Kazan State University, Kazan, Russia
| | | | - Ksenia Evteeva
- Research and Practical Clinical Center for Diagnostics and Telemedicine Technologies of the Moscow Healthcare Department, Moscow, Russia
| | - Anna Andreychenko
- Research and Practical Clinical Center for Diagnostics and Telemedicine Technologies of the Moscow Healthcare Department, Moscow, Russia
| | - Sergey Morozov
- Research and Practical Clinical Center for Diagnostics and Telemedicine Technologies of the Moscow Healthcare Department, Moscow, Russia
- Osimis SA, Liege, Belgium
| |
Collapse
|
17
|
Deep learning classification of active tuberculosis lung zones wise manifestations using chest X-rays: a multi label approach. Sci Rep 2023; 13:887. [PMID: 36650270 PMCID: PMC9845381 DOI: 10.1038/s41598-023-28079-0] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/27/2022] [Accepted: 01/12/2023] [Indexed: 01/19/2023] Open
Abstract
Chest X-rays are the most economically viable diagnostic imaging test for active pulmonary tuberculosis screening despite the high sensitivity and low specificity when interpreted by clinicians or radiologists. Computer aided detection (CAD) algorithms, especially convolution based deep learning architecture, have been proposed to facilitate the automation of radiography imaging modalities. Deep learning algorithms have found success in classifying various abnormalities in lung using chest X-ray. We fine-tuned, validated and tested EfficientNetB4 architecture and utilized the transfer learning methodology for multilabel approach to detect lung zone wise and image wise manifestations of active pulmonary tuberculosis using chest X-ray. We used Area Under Receiver Operating Characteristic (AUC), sensitivity and specificity along with 95% confidence interval as model evaluation metrics. We also utilized the visualisation capabilities of convolutional neural networks (CNN), Gradient-weighted Class Activation Mapping (Grad-CAM) as post-hoc attention method to investigate the model and visualisation of Tuberculosis abnormalities and discuss them from radiological perspectives. EfficientNetB4 trained network achieved remarkable AUC, sensitivity and specificity of various pulmonary tuberculosis manifestations in intramural test set and external test set from different geographical region. The grad-CAM visualisations and their ability to localize the abnormalities can aid the clinicians at primary care settings for screening and triaging of tuberculosis where resources are constrained or overburdened.
Collapse
|
18
|
Zhu X, Pang S, Zhang X, Huang J, Zhao L, Tang K, Feng Q. PCAN: Pixel-wise classification and attention network for thoracic disease classification and weakly supervised localization. Comput Med Imaging Graph 2022; 102:102137. [PMID: 36308870 DOI: 10.1016/j.compmedimag.2022.102137] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2022] [Revised: 09/23/2022] [Accepted: 10/05/2022] [Indexed: 11/06/2022]
Abstract
Automatic chest X-ray (CXR) disease classification has drawn increasing public attention as CXR is widely used in thoracic disease diagnosis. Existing classification networks typically employ a global average pooling layer to produce the final feature for the subsequent classifier. This limits the classification performance owing to the characteristics of lesions in CXR images, including small relative sizes, varied absolute sizes, and different occurrence locations. In this study, we propose a pixel-wise classification and attention network (PCAN) to simultaneously perform disease classification and weakly supervised localization, which provides interpretability for disease classification. The PCAN comprises a backbone network for extracting mid-level features, a pixel-wise classification branch (pc-branch) for generating pixel-wise diagnoses, and a pixel-wise attention branch (pa-branch) for producing pixel-wise weights. The pc-branch is capable of explicitly detecting small lesions, and the pa-branch is capable of adaptively focusing on different regions when classifying different thoracic diseases. Then, the pixel-wise diagnoses are multiplied with the pixel-wise weights to obtain the disease localization map, which provides the sizes and locations of lesions in a manner of weakly supervised learning. The final image-wise diagnosis is obtained by summing up the disease localization map at the spatial dimension. Comprehensive experiments conducted on the ChestX-ray14 and CheXpert datasets demonstrate the effectiveness of the proposed PCAN, which has great potential for thoracic disease diagnosis and treatment. The source codes are available at https://github.com/fzfs/PCAN.
Collapse
Affiliation(s)
- Xiongfeng Zhu
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China.
| | - Shumao Pang
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China
| | - Xiaoxuan Zhang
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China
| | - Junzhang Huang
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China
| | - Lei Zhao
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China
| | - Kai Tang
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China
| | - Qianjin Feng
- School of Biomedical Engineering, Southern Medical University, Guangzhou, 510515, China; Guangdong Provincial Key Laboratory of Medical Image Processing, Southern Medical University, Guangzhou, 510515, China; Guangdong Province Engineering Laboratory for Medical Imaging and Diagnostic Technology, Southern Medical University, Guangzhou, 510515, China.
| |
Collapse
|
19
|
Gandomkar Z, Khong PL, Punch A, Lewis S. Using Occlusion-Based Saliency Maps to Explain an Artificial Intelligence Tool in Lung Cancer Screening: Agreement Between Radiologists, Labels, and Visual Prompts. J Digit Imaging 2022; 35:1164-1175. [PMID: 35484439 PMCID: PMC9582174 DOI: 10.1007/s10278-022-00631-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/08/2021] [Revised: 03/03/2022] [Accepted: 04/04/2022] [Indexed: 11/29/2022] Open
Abstract
Occlusion-based saliency maps (OBSMs) are one of the approaches for interpreting decision-making process of an artificial intelligence (AI) system. This study explores the agreement among text responses from a cohort of radiologists to describe diagnostically relevant areas on low-dose CT (LDCT) images. It also explores if radiologists' descriptions of cases misclassified by the AI provide a rationale for ruling out the AI's output. The OBSM indicating the importance of different pixels on the final decision made by an AI were generated for 10 benign cases (3 misclassified by the AI tool as malignant) and 10 malignant cases (2 misclassified by the AI tool as benign). Thirty-six radiologists were asked to use radiological vocabulary, typical to reporting LDCT scans, to describe the mapped regions of interest (ROI). The radiologists' annotations were then grouped by using a clustering-based technique. Topics were extracted from the annotations and for each ROI, a percentage of annotations containing each topic were found. Radiologists annotated 17 and 24 unique ROIs on benign and malignant cases, respectively. Agreement on the main label (e.g., "vessel," "nodule") by radiologists was only seen in only in 12% of all areas (5/41 ROI). Topic analyses identified six descriptors which are commonly associated with a lower malignancy likelihood. Eight common topics related to a higher malignancy likelihood were also determined. Occlusion-based saliency maps were used to explain an AI decision-making process to radiologists, who in turn have provided insight into the level of agreement between the AI's decision and radiological lexicon.
Collapse
Affiliation(s)
- Ziba Gandomkar
- Discipline of Medical Imaging Science, Faculty of Medicine and Health, University of Sydney, Sydney, NSW, Australia
| | - Pek Lan Khong
- Clinical Imaging Research Center (CIRC), Department of Diagnostic Radiology, Yong Loo Lin School of Medicine, National University of Singapore, Singapore, Singapore
| | - Amanda Punch
- Discipline of Medical Imaging Science, Faculty of Medicine and Health, University of Sydney, Sydney, NSW, Australia
| | - Sarah Lewis
- Discipline of Medical Imaging Science, Faculty of Medicine and Health, University of Sydney, Sydney, NSW, Australia.
| |
Collapse
|
20
|
CheXGAT: A disease correlation-aware network for thorax disease diagnosis from chest X-ray images. Artif Intell Med 2022; 132:102382. [DOI: 10.1016/j.artmed.2022.102382] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/28/2021] [Revised: 08/07/2022] [Accepted: 08/19/2022] [Indexed: 11/23/2022]
|
21
|
Modeling global and local label correlation with graph convolutional networks for multi-label chest X-ray image classification. Med Biol Eng Comput 2022; 60:2567-2588. [DOI: 10.1007/s11517-022-02604-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/07/2021] [Accepted: 06/07/2022] [Indexed: 10/17/2022]
|
22
|
Albahli S, Nazir T. AI-CenterNet CXR: An artificial intelligence (AI) enabled system for localization and classification of chest X-ray disease. Front Med (Lausanne) 2022; 9:955765. [PMID: 36111113 PMCID: PMC9469020 DOI: 10.3389/fmed.2022.955765] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/29/2022] [Accepted: 07/21/2022] [Indexed: 12/03/2022] Open
Abstract
Machine learning techniques have lately attracted a lot of attention for their potential to execute expert-level clinical tasks, notably in the area of medical image analysis. Chest radiography is one of the most often utilized diagnostic imaging modalities in medical practice, and it necessitates timely coverage regarding the presence of probable abnormalities and disease diagnoses in the images. Computer-aided solutions for the identification of chest illness using chest radiography are being developed in medical imaging research. However, accurate localization and categorization of specific disorders in chest X-ray images is still a challenging problem due to the complex nature of radiographs, presence of different distortions, high inter-class similarities, and intra-class variations in abnormalities. In this work, we have presented an Artificial Intelligence (AI)-enabled fully automated approach using an end-to-end deep learning technique to improve the accuracy of thoracic illness diagnosis. We proposed AI-CenterNet CXR, a customized CenterNet model with an improved feature extraction network for the recognition of multi-label chest diseases. The enhanced backbone computes deep key points that improve the abnormality localization accuracy and, thus, overall disease classification performance. Moreover, the proposed architecture is lightweight and computationally efficient in comparison to the original CenterNet model. We have performed extensive experimentation to validate the effectiveness of the proposed technique using the National Institutes of Health (NIH) Chest X-ray dataset. Our method achieved an overall Area Under the Curve (AUC) of 0.888 and an average IOU of 0.801 to detect and classify the eight types of chest abnormalities. Both the qualitative and quantitative findings reveal that the suggested approach outperforms the existing methods, indicating the efficacy of our approach.
Collapse
Affiliation(s)
- Saleh Albahli
- Department of Information Technology, College of Computer, Qassim University, Buraydah, Saudi Arabia
| | - Tahira Nazir
- Faculty of Computing, Riphah International University, Islamabad, Pakistan
| |
Collapse
|
23
|
MXT: A New Variant of Pyramid Vision Transformer for Multi-label Chest X-ray Image Classification. Cognit Comput 2022. [DOI: 10.1007/s12559-022-10032-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/03/2022]
|
24
|
Mondal AK. COVID-19 prognosis using limited chest X-ray images. Appl Soft Comput 2022; 122:108867. [PMID: 35494338 PMCID: PMC9035620 DOI: 10.1016/j.asoc.2022.108867] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/09/2021] [Revised: 01/18/2022] [Accepted: 04/09/2022] [Indexed: 01/31/2023]
Abstract
The COrona VIrus Disease 2019 (COVID-19) pandemic is an ongoing global pandemic that has claimed millions of lives till date. Detecting COVID-19 and isolating affected patients at an early stage is crucial to contain its rapid spread. Although accurate, the primary viral test 'Reverse Transcription Polymerase Chain Reaction' (RT-PCR) for COVID-19 diagnosis has an elaborate test kit, and the turnaround time is high. This has motivated the research community to develop CXR based automated COVID-19 diagnostic methodologies. However, COVID-19 being a novel disease, there is no annotated large-scale CXR dataset for this particular disease. To address the issue of limited data, we propose to exploit a large-scale CXR dataset collected in the pre-COVID era and train a deep neural network in a self-supervised fashion to extract CXR specific features. Further, we compute attention maps between the global and the local features of the backbone convolutional network while finetuning using a limited COVID-19 CXR dataset. We empirically demonstrate the effectiveness of the proposed method. We provide a thorough ablation study to understand the effect of each proposed component. Finally, we provide visualizations highlighting the critical patches instrumental to the predictive decision made by our model. These saliency maps are not only a stepping stone towards explainable AI but also aids radiologists in localizing the infected area.
Collapse
|
25
|
Zhao J, Zhou X, Shi G, Xiao N, Song K, Zhao J, Hao R, Li K. Semantic consistency generative adversarial network for cross-modality domain adaptation in ultrasound thyroid nodule classification. APPL INTELL 2022; 52:10369-10383. [PMID: 35039715 PMCID: PMC8754560 DOI: 10.1007/s10489-021-03025-7] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 11/18/2021] [Indexed: 12/22/2022]
Abstract
Deep convolutional networks have been widely used for various medical image processing tasks. However, the performance of existing learning-based networks is still limited due to the lack of large training datasets. When a general deep model is directly deployed to a new dataset with heterogeneous features, the effect of domain shifts is usually ignored, and performance degradation problems occur. In this work, by designing the semantic consistency generative adversarial network (SCGAN), we propose a new multimodal domain adaptation method for medical image diagnosis. SCGAN performs cross-domain collaborative alignment of ultrasound images and domain knowledge. Specifically, we utilize a self-attention mechanism for adversarial learning between dual domains to overcome visual differences across modal data and preserve the domain invariance of the extracted semantic features. In particular, we embed nested metric learning in the semantic information space, thus enhancing the semantic consistency of cross-modal features. Furthermore, the adversarial learning of our network is guided by a discrepancy loss for encouraging the learning of semantic-level content and a regularization term for enhancing network generalization. We evaluate our method on a thyroid ultrasound image dataset for benign and malignant diagnosis of nodules. The experimental results of a comprehensive study show that the accuracy of the SCGAN method for the classification of thyroid nodules reaches 94.30%, and the AUC reaches 97.02%. These results are significantly better than the state-of-the-art methods.
Collapse
|
26
|
Virtual Healthcare Center for COVID-19 Patient Detection Based on Artificial Intelligence Approaches. CANADIAN JOURNAL OF INFECTIOUS DISEASES AND MEDICAL MICROBIOLOGY 2022; 2022:6786203. [PMID: 35069953 PMCID: PMC8767384 DOI: 10.1155/2022/6786203] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/19/2021] [Accepted: 12/22/2021] [Indexed: 11/18/2022]
Abstract
At the end of 2019, the infectious coronavirus disease (COVID-19) was reported for the first time in Wuhan, and, since then, it has become a public health issue in China and even worldwide. This pandemic has devastating effects on societies and economies around the world, and poor countries and continents are likely to face particularly serious and long-lasting damage, which could lead to large epidemic outbreaks because of the lack of financial and health resources. The increasing number of COVID-19 tests gives more information about the epidemic spread, and this can help contain the spread to avoid more infection. As COVID-19 keeps spreading, medical products, especially those needed to perform blood tests, will become scarce as a result of the high demand and insufficient supply and logistical means. However, technological tests based on deep learning techniques and medical images could be useful in fighting this pandemic. In this perspective, we propose a COVID-19 disease diagnosis (CDD) tool that implements a deep learning technique to provide automatic symptoms checking and COVID-19 detection. Our CDD scheme implements two main steps. First, the patient's symptoms are checked, and the infection probability is predicted. Then, based on the infection probability, the patient's lungs will be diagnosed by an automatic analysis of X-ray or computerized tomography (CT) images, and the presence of the infection will be accordingly confirmed or not. The numerical results prove the efficiency of the proposed scheme by achieving an accuracy value over 90% compared with the other schemes.
Collapse
|
27
|
Li Y, Wu X, Li C, Li X, Chen H, Sun C, Rahaman MM, Yao Y, Zhang Y, Jiang T. A hierarchical conditional random field-based attention mechanism approach for gastric histopathology image classification. APPL INTELL 2022. [DOI: 10.1007/s10489-021-02886-2] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/25/2022]
|
28
|
Mehrrotraa R, Ansari MA, Agrawal R, Tripathi P, Bin Heyat MB, Al-Sarem M, Muaad AYM, Nagmeldin WAE, Abdelmaboud A, Saeed F. Ensembling of Efficient Deep Convolutional Networks and Machine Learning Algorithms for Resource Effective Detection of Tuberculosis Using Thoracic (Chest) Radiography. IEEE ACCESS 2022; 10:85442-85458. [DOI: 10.1109/access.2022.3194152] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Affiliation(s)
- Rajat Mehrrotraa
- Department of Electrical and Electronics Engineering, G. L. Bajaj Institute of Technology & Management, Greater Noida, India
| | - M. A. Ansari
- Department of Electrical Engineering, School of Engineering, Gautam Buddha University, Greater Noida, India
| | - Rajeev Agrawal
- Department of Computer Science, Lloyd Institute of Engineering & Technology, Greater Noida, India
| | - Pragati Tripathi
- Department of Electrical Engineering, School of Engineering, Gautam Buddha University, Greater Noida, India
| | - Md Belal Bin Heyat
- IoT Research Center, College of Computer Science and Software Engineering, Shenzhen University, Shenzhen, China
| | - Mohammed Al-Sarem
- College of Computer Science and Engineering, Taibah University, Medina, Saudi Arabia
| | | | - Wamda Abdelrahman Elhag Nagmeldin
- Department of Information Systems, College of Computer Engineering and Sciences, Prince Sattam bin Abdulaziz University, Al-Kharj, Saudi Arabia
| | | | - Faisal Saeed
- Department of Computing and Data Science, DAAI Research Group, School of Computing and Digital Technology, Birmingham City University, Birmingham, U.K
| |
Collapse
|
29
|
Abdel-Basset M, Hawash H, Moustafa N, Elkomy OM. Two-Stage Deep Learning Framework for Discrimination between COVID-19 and Community-Acquired Pneumonia from Chest CT scans. Pattern Recognit Lett 2021; 152:311-319. [PMID: 34728870 PMCID: PMC8554046 DOI: 10.1016/j.patrec.2021.10.027] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2021] [Revised: 09/01/2021] [Accepted: 10/25/2021] [Indexed: 12/19/2022]
Abstract
COVID-19 stay threatening the health infrastructure worldwide. Computed tomography (CT) was demonstrated as an informative tool for the recognition, quantification, and diagnosis of this kind of disease. It is urgent to design efficient deep learning (DL) approach to automatically localize and discriminate COVID-19 from other comparable pneumonia on lung CT scans. Thus, this study introduces a novel two-stage DL framework for discriminating COVID-19 from community-acquired pneumonia (CAP) depending on the detected infection region within CT slices. Firstly, a novel U-shaped network is presented to segment the lung area where the infection appears. Then, the concept of transfer learning is applied to the feature extraction network to empower the network capabilities in learning the disease patterns. After that, multi-scale information is captured and pooled via an attention mechanism for powerful classification performance. Thirdly, we propose an infection prediction module that use the infection location to guide the classification decision and hence provides interpretable classification decision. Finally, the proposed model was evaluated on public datasets and achieved great segmentation and classification performance outperforming the cutting-edge studies.
Collapse
Affiliation(s)
- Mohamed Abdel-Basset
- Faculty of Computers and Informatics, Zagazig University, Zagazig, Sharqiyah, 44519, Egypt
| | - Hossam Hawash
- Faculty of Computers and Informatics, Zagazig University, Zagazig, Sharqiyah, 44519, Egypt
| | - Nour Moustafa
- School of Engineering and Information Technology, University of New South Wales @ ADFA, Canberra, ACT 2600, Australia
| | - Osama M Elkomy
- Faculty of Computers and Informatics, Zagazig University, Zagazig, Sharqiyah, 44519, Egypt
| |
Collapse
|
30
|
Liu XM, Hu J, Mwakapesa DS, Nanehkaran Y, Mao YM, Zhang RP, Chen ZG. A novel MapReduce-based deep convolutional neural network algorithm. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS 2021. [DOI: 10.3233/jifs-201790] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Deep convolutional neural networks (DCNNs), with their complex network structure and powerful feature learning and feature expression capabilities, have been remarkable successes in many large-scale recognition tasks. However, with the expectation of memory overhead and response time, along with the increasing scale of data, DCNN faces three non-rival challenges in a big data environment: excessive network parameters, slow convergence, and inefficient parallelism. To tackle these three problems, this paper develops a deep convolutional neural networks optimization algorithm (PDCNNO) in the MapReduce framework. The proposed method first pruned the network to obtain a compressed network in order to effectively reduce redundant parameters. Next, a conjugate gradient method based on modified secant equation (CGMSE) is developed in the Map phase to further accelerate the convergence of the network. Finally, a load balancing strategy based on regulate load rate (LBRLA) is proposed in the Reduce phase to quickly achieve equal grouping of data and thus improving the parallel performance of the system. We compared the PDCNNO algorithm with other algorithms on three datasets, including SVHN, EMNIST Digits, and ISLVRC2012. The experimental results show that our algorithm not only reduces the space and time overhead of network training but also obtains a well-performing speed-up ratio in a big data environment.
Collapse
Affiliation(s)
- Xiang-Min Liu
- School of Information Engineering, Jiangxi University of Science and Technology, Ganzhou, Jiangxi, China
| | - Jian Hu
- School of Information Engineering, Jiangxi University of Science and Technology, Ganzhou, Jiangxi, China
| | - Deborah Simon Mwakapesa
- School of Information Engineering, Jiangxi University of Science and Technology, Ganzhou, Jiangxi, China
| | - Y.A. Nanehkaran
- School of Information Engineering, Jiangxi University of Science and Technology, Ganzhou, Jiangxi, China
| | - Yi-Min Mao
- School of Information Engineering, Jiangxi University of Science and Technology, Ganzhou, Jiangxi, China
| | - Rui-Peng Zhang
- School of Information Engineering, Jiangxi University of Science and Technology, Ganzhou, Jiangxi, China
| | - Zhi-Gang Chen
- School of Computing, Central South University, Changsha, Hunan, China
| |
Collapse
|
31
|
Abstract
COVID-19, an infectious coronavirus disease, caused a pandemic with countless deaths. From the outset, clinical institutes have explored computed tomography as an effective and complementary screening tool alongside the reverse transcriptase-polymerase chain reaction. Deep learning techniques have shown promising results in similar medical tasks and, hence, may provide solutions to COVID-19 based on medical images of patients. We aim to contribute to the research in this field by: (i) Comparing different architectures on a public and extended reference dataset to find the most suitable; (ii) Proposing a patient-oriented investigation of the best performing networks; and (iii) Evaluating their robustness in a real-world scenario, represented by cross-dataset experiments. We exploited ten well-known convolutional neural networks on two public datasets. The results show that, on the reference dataset, the most suitable architecture is VGG19, which (i) Achieved 98.87% accuracy in the network comparison; (ii) Obtained 95.91% accuracy on the patient status classification, even though it misclassifies some patients that other networks classify correctly; and (iii) The cross-dataset experiments exhibit the limitations of deep learning approaches in a real-world scenario with 70.15% accuracy, which need further investigation to improve the robustness. Thus, VGG19 architecture showed promising performance in the classification of COVID-19 cases. Nonetheless, this architecture enables extensive improvements based on its modification, or even with preprocessing step in addition to it. Finally, the cross-dataset experiments exposed the critical weakness of classifying images from heterogeneous data sources, compatible with a real-world scenario.
Collapse
|
32
|
Qiu Z, Xu T, Langerman J, Das W, Wang C, Nair N, Aristizabal O, Mamou J, Turnbull DH, Ketterling JA, Wang Y. A Deep Learning Approach for Segmentation, Classification, and Visualization of 3-D High-Frequency Ultrasound Images of Mouse Embryos. IEEE TRANSACTIONS ON ULTRASONICS, FERROELECTRICS, AND FREQUENCY CONTROL 2021; 68:2460-2471. [PMID: 33755564 PMCID: PMC8274381 DOI: 10.1109/tuffc.2021.3068156] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/07/2023]
Abstract
Segmentation and mutant classification of high-frequency ultrasound (HFU) mouse embryo brain ventricle (BV) and body images can provide valuable information for developmental biologists. However, manual segmentation and identification of BV and body requires substantial time and expertise. This article proposes an accurate, efficient and explainable deep learning pipeline for automatic segmentation and classification of the BV and body. For segmentation, a two-stage framework is implemented. The first stage produces a low-resolution segmentation map, which is then used to crop a region of interest (ROI) around the target object and serve as the probability map of the autocontext input for the second-stage fine-resolution refinement network. The segmentation then becomes tractable on high-resolution 3-D images without time-consuming sliding windows. The proposed segmentation method significantly reduces inference time (102.36-0.09 s/volume ≈ 1000× faster) while maintaining high accuracy comparable to previous sliding-window approaches. Based on the BV and body segmentation map, a volumetric convolutional neural network (CNN) is trained to perform a mutant classification task. Through backpropagating the gradients of the predictions to the input BV and body segmentation map, the trained classifier is found to largely focus on the region where the Engrailed-1 (En1) mutation phenotype is known to manifest itself. This suggests that gradient backpropagation of deep learning classifiers may provide a powerful tool for automatically detecting unknown phenotypes associated with a known genetic mutation.
Collapse
|
33
|
Fang J, Xu Y, Zhao Y, Yan Y, Liu J, Liu J. Weighing features of lung and heart regions for thoracic disease classification. BMC Med Imaging 2021; 21:99. [PMID: 34112095 PMCID: PMC8194196 DOI: 10.1186/s12880-021-00627-y] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/15/2021] [Accepted: 05/24/2021] [Indexed: 11/10/2022] Open
Abstract
BACKGROUND Chest X-rays are the most commonly available and affordable radiological examination for screening thoracic diseases. According to the domain knowledge of screening chest X-rays, the pathological information usually lay on the lung and heart regions. However, it is costly to acquire region-level annotation in practice, and model training mainly relies on image-level class labels in a weakly supervised manner, which is highly challenging for computer-aided chest X-ray screening. To address this issue, some methods have been proposed recently to identify local regions containing pathological information, which is vital for thoracic disease classification. Inspired by this, we propose a novel deep learning framework to explore discriminative information from lung and heart regions. RESULT We design a feature extractor equipped with a multi-scale attention module to learn global attention maps from global images. To exploit disease-specific cues effectively, we locate lung and heart regions containing pathological information by a well-trained pixel-wise segmentation model to generate binarization masks. By introducing element-wise logical AND operator on the learned global attention maps and the binarization masks, we obtain local attention maps in which pixels are are 1 for lung and heart region and 0 for other regions. By zeroing features of non-lung and heart regions in attention maps, we can effectively exploit their disease-specific cues in lung and heart regions. Compared to existing methods fusing global and local features, we adopt feature weighting to avoid weakening visual cues unique to lung and heart regions. Our method with pixel-wise segmentation can help overcome the deviation of locating local regions. Evaluated by the benchmark split on the publicly available chest X-ray14 dataset, the comprehensive experiments show that our method achieves superior performance compared to the state-of-the-art methods. CONCLUSION We propose a novel deep framework for the multi-label classification of thoracic diseases in chest X-ray images. The proposed network aims to effectively exploit pathological regions containing the main cues for chest X-ray screening. Our proposed network has been used in clinic screening to assist the radiologists. Chest X-ray accounts for a significant proportion of radiological examinations. It is valuable to explore more methods for improving performance.
Collapse
Affiliation(s)
- Jiansheng Fang
- School of Computer Science and Technology, Harbin Institute of Technology, Harbin, China
- Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China
- CVTE Research, Guangzhou, China
| | - Yanwu Xu
- Cixi Institute of Biomedical Engineering, Chinese Academy of Sciences, Ningbo, China
| | - Yitian Zhao
- Cixi Institute of Biomedical Engineering, Chinese Academy of Sciences, Ningbo, China
| | - Yuguang Yan
- Department of Mathematics, University of Hong Kong, Hongkong, China
| | - Junling Liu
- Guangdong Armed Police Hospital, Guangzhou, China
| | - Jiang Liu
- Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China
- Cixi Institute of Biomedical Engineering, Chinese Academy of Sciences, Ningbo, China
| |
Collapse
|
34
|
Çallı E, Sogancioglu E, van Ginneken B, van Leeuwen KG, Murphy K. Deep learning for chest X-ray analysis: A survey. Med Image Anal 2021; 72:102125. [PMID: 34171622 DOI: 10.1016/j.media.2021.102125] [Citation(s) in RCA: 126] [Impact Index Per Article: 31.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/16/2021] [Revised: 05/17/2021] [Accepted: 05/27/2021] [Indexed: 12/14/2022]
Abstract
Recent advances in deep learning have led to a promising performance in many medical image analysis tasks. As the most commonly performed radiological exam, chest radiographs are a particularly important modality for which a variety of applications have been researched. The release of multiple, large, publicly available chest X-ray datasets in recent years has encouraged research interest and boosted the number of publications. In this paper, we review all studies using deep learning on chest radiographs published before March 2021, categorizing works by task: image-level prediction (classification and regression), segmentation, localization, image generation and domain adaptation. Detailed descriptions of all publicly available datasets are included and commercial systems in the field are described. A comprehensive discussion of the current state of the art is provided, including caveats on the use of public datasets, the requirements of clinically useful systems and gaps in the current literature.
Collapse
Affiliation(s)
- Erdi Çallı
- Radboud University Medical Center, Institute for Health Sciences, Department of Medical Imaging, Nijmegen, the Netherlands.
| | - Ecem Sogancioglu
- Radboud University Medical Center, Institute for Health Sciences, Department of Medical Imaging, Nijmegen, the Netherlands
| | - Bram van Ginneken
- Radboud University Medical Center, Institute for Health Sciences, Department of Medical Imaging, Nijmegen, the Netherlands
| | - Kicky G van Leeuwen
- Radboud University Medical Center, Institute for Health Sciences, Department of Medical Imaging, Nijmegen, the Netherlands
| | - Keelin Murphy
- Radboud University Medical Center, Institute for Health Sciences, Department of Medical Imaging, Nijmegen, the Netherlands
| |
Collapse
|
35
|
Zhang R, Yang F, Luo Y, Liu J, Li J, Wang C. Part-Aware Mask-Guided Attention for Thorax Disease Classification. ENTROPY 2021; 23:e23060653. [PMID: 34070982 PMCID: PMC8224595 DOI: 10.3390/e23060653] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/29/2021] [Revised: 05/20/2021] [Accepted: 05/21/2021] [Indexed: 11/24/2022]
Abstract
Thorax disease classification is a challenging task due to complex pathologies and subtle texture changes, etc. It has been extensively studied for years largely because of its wide application in computer-aided diagnosis. Most existing methods directly learn global feature representations from whole Chest X-ray (CXR) images, without considering in depth the richer visual cues lying around informative local regions. Thus, these methods often produce sub-optimal thorax disease classification performance because they ignore the very informative pathological changes around organs. In this paper, we propose a novel Part-Aware Mask-Guided Attention Network (PMGAN) that learns complementary global and local feature representations from all-organ region and multiple single-organ regions simultaneously for thorax disease classification. Specifically, multiple innovative soft attention modules are designed to progressively guide feature learning toward the global informative regions of whole CXR image. A mask-guided attention module is designed to further search for informative regions and visual cues within the all-organ or single-organ images, where attention is elegantly regularized by automatically generated organ masks and without introducing computation during the inference stage. In addition, a multi-task learning strategy is designed, which effectively maximizes the learning of complementary local and global representations. The proposed PMGAN has been evaluated on the ChestX-ray14 dataset and the experimental results demonstrate its superior thorax disease classification performance against the state-of-the-art methods.
Collapse
Affiliation(s)
- Ruihua Zhang
- School of Computer Science (National Pilot Software Engineering School), Beijing University of Posts and Telecommunications, Beijing 100876, China; (R.Z.); (Y.L.); (C.W.)
- Key Laboratory of Trustworthy Distributed Computing and Service, Beijing University of Posts and Telecommunications, Beijing 100876, China
| | - Fan Yang
- School of Electronics Engineering and Computer Science, Peking University, Beijing 100871, China;
| | - Yan Luo
- School of Computer Science (National Pilot Software Engineering School), Beijing University of Posts and Telecommunications, Beijing 100876, China; (R.Z.); (Y.L.); (C.W.)
- Key Laboratory of Trustworthy Distributed Computing and Service, Beijing University of Posts and Telecommunications, Beijing 100876, China
| | - Jianyi Liu
- School of Cyberspace Security, Beijing University of Posts and Telecommunications, Beijing 100876, China
- Correspondence:
| | - Jinbin Li
- Local Servive Center, National Population Health Data Center, Beijing 100005, China;
| | - Cong Wang
- School of Computer Science (National Pilot Software Engineering School), Beijing University of Posts and Telecommunications, Beijing 100876, China; (R.Z.); (Y.L.); (C.W.)
- Key Laboratory of Trustworthy Distributed Computing and Service, Beijing University of Posts and Telecommunications, Beijing 100876, China
| |
Collapse
|
36
|
Astuto B, Flament I, K. Namiri N, Shah R, Bharadwaj U, M. Link T, D. Bucknor M, Pedoia V, Majumdar S. Automatic Deep Learning-assisted Detection and Grading of Abnormalities in Knee MRI Studies. Radiol Artif Intell 2021; 3:e200165. [PMID: 34142088 PMCID: PMC8166108 DOI: 10.1148/ryai.2021200165] [Citation(s) in RCA: 44] [Impact Index Per Article: 11.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/09/2020] [Revised: 12/17/2020] [Accepted: 01/07/2021] [Indexed: 12/20/2022]
Abstract
PURPOSE To test the hypothesis that artificial intelligence (AI) techniques can aid in identifying and assessing lesion severity in the cartilage, bone marrow, meniscus, and anterior cruciate ligament (ACL) in the knee, improving overall MRI interreader agreement. MATERIALS AND METHODS This retrospective study was conducted on 1435 knee MRI studies (n = 294 patients; mean age, 43 years ± 15 [standard deviation]; 153 women) collected within three previous studies (from 2011 to 2014). All MRI studies were acquired using high-spatial-resolution three-dimensional fast-spin-echo CUBE sequence. Three-dimensional convolutional neural networks were developed to detect the regions of interest within MRI studies and grade abnormalities of the cartilage, bone marrow, menisci, and ACL. Evaluation included sensitivity, specificity, and Cohen linear-weighted ĸ. The impact of AI-aided grading in intergrader agreement was assessed on an external dataset. RESULTS Binary lesion sensitivity reported for all tissues was between 70% and 88%. Specificity ranged from 85% to 89%. The area under the receiver operating characteristic curve for all tissues ranged from 0.83 to 0.93. Deep learning-assisted intergrader Cohen ĸ agreement significantly improved in 10 of 16 comparisons among two attending physicians and two trainees for all tissues. CONCLUSION The three-dimensional convolutional neural network had high sensitivity, specificity, and accuracy for knee-lesion-severity scoring and also increased intergrader agreement when used on an external dataset.Supplemental material is available for this article. Keywords: Bone Marrow, Cartilage, Computer Aided Diagnosis (CAD), Computer Applications-3D, Computer Applications-Detection/Diagnosis, Knee, Ligaments, MR-Imaging, Neural Networks, Observer Performance, Segmentation, Statistics © RSNA, 2021See also the commentary by Li and Chang in this issue.: An earlier incorrect version of this article appeared online. This article was corrected on April 16, 2021.
Collapse
Affiliation(s)
- Bruno Astuto
- From the Center for Intelligent Imaging and Musculoskeletal and
Quantitative Imaging Research Group, Department of Radiology and Biomedical
Imaging (B.A., I.F., N.K.N., R.S., U.B., T.M.L., M.D.B., V.P., S.M.), and Center
of Digital Health Innovation (V.P., S.M.), University of California–San
Francisco, 1700 Fourth St, Suite 201, QB3 Building, San Francisco, CA
94107
| | - Io Flament
- From the Center for Intelligent Imaging and Musculoskeletal and
Quantitative Imaging Research Group, Department of Radiology and Biomedical
Imaging (B.A., I.F., N.K.N., R.S., U.B., T.M.L., M.D.B., V.P., S.M.), and Center
of Digital Health Innovation (V.P., S.M.), University of California–San
Francisco, 1700 Fourth St, Suite 201, QB3 Building, San Francisco, CA
94107
| | - Nikan K. Namiri
- From the Center for Intelligent Imaging and Musculoskeletal and
Quantitative Imaging Research Group, Department of Radiology and Biomedical
Imaging (B.A., I.F., N.K.N., R.S., U.B., T.M.L., M.D.B., V.P., S.M.), and Center
of Digital Health Innovation (V.P., S.M.), University of California–San
Francisco, 1700 Fourth St, Suite 201, QB3 Building, San Francisco, CA
94107
| | - Rutwik Shah
- From the Center for Intelligent Imaging and Musculoskeletal and
Quantitative Imaging Research Group, Department of Radiology and Biomedical
Imaging (B.A., I.F., N.K.N., R.S., U.B., T.M.L., M.D.B., V.P., S.M.), and Center
of Digital Health Innovation (V.P., S.M.), University of California–San
Francisco, 1700 Fourth St, Suite 201, QB3 Building, San Francisco, CA
94107
| | - Upasana Bharadwaj
- From the Center for Intelligent Imaging and Musculoskeletal and
Quantitative Imaging Research Group, Department of Radiology and Biomedical
Imaging (B.A., I.F., N.K.N., R.S., U.B., T.M.L., M.D.B., V.P., S.M.), and Center
of Digital Health Innovation (V.P., S.M.), University of California–San
Francisco, 1700 Fourth St, Suite 201, QB3 Building, San Francisco, CA
94107
| | - Thomas M. Link
- From the Center for Intelligent Imaging and Musculoskeletal and
Quantitative Imaging Research Group, Department of Radiology and Biomedical
Imaging (B.A., I.F., N.K.N., R.S., U.B., T.M.L., M.D.B., V.P., S.M.), and Center
of Digital Health Innovation (V.P., S.M.), University of California–San
Francisco, 1700 Fourth St, Suite 201, QB3 Building, San Francisco, CA
94107
| | - Matthew D. Bucknor
- From the Center for Intelligent Imaging and Musculoskeletal and
Quantitative Imaging Research Group, Department of Radiology and Biomedical
Imaging (B.A., I.F., N.K.N., R.S., U.B., T.M.L., M.D.B., V.P., S.M.), and Center
of Digital Health Innovation (V.P., S.M.), University of California–San
Francisco, 1700 Fourth St, Suite 201, QB3 Building, San Francisco, CA
94107
| | - Valentina Pedoia
- From the Center for Intelligent Imaging and Musculoskeletal and
Quantitative Imaging Research Group, Department of Radiology and Biomedical
Imaging (B.A., I.F., N.K.N., R.S., U.B., T.M.L., M.D.B., V.P., S.M.), and Center
of Digital Health Innovation (V.P., S.M.), University of California–San
Francisco, 1700 Fourth St, Suite 201, QB3 Building, San Francisco, CA
94107
| | - Sharmila Majumdar
- From the Center for Intelligent Imaging and Musculoskeletal and
Quantitative Imaging Research Group, Department of Radiology and Biomedical
Imaging (B.A., I.F., N.K.N., R.S., U.B., T.M.L., M.D.B., V.P., S.M.), and Center
of Digital Health Innovation (V.P., S.M.), University of California–San
Francisco, 1700 Fourth St, Suite 201, QB3 Building, San Francisco, CA
94107
| |
Collapse
|
37
|
Gündel S, Setio AAA, Ghesu FC, Grbic S, Georgescu B, Maier A, Comaniciu D. Robust classification from noisy labels: Integrating additional knowledge for chest radiography abnormality assessment. Med Image Anal 2021; 72:102087. [PMID: 34015595 DOI: 10.1016/j.media.2021.102087] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2020] [Revised: 02/24/2021] [Accepted: 04/16/2021] [Indexed: 12/29/2022]
Abstract
Chest radiography is the most common radiographic examination performed in daily clinical practice for the detection of various heart and lung abnormalities. The large amount of data to be read and reported, with more than 100 studies per day for a single radiologist, poses a challenge in consistently maintaining high interpretation accuracy. The introduction of large-scale public datasets has led to a series of novel systems for automated abnormality classification. However, the labels of these datasets were obtained using natural language processed medical reports, yielding a large degree of label noise that can impact the performance. In this study, we propose novel training strategies that handle label noise from such suboptimal data. Prior label probabilities were measured on a subset of training data re-read by 4 board-certified radiologists and were used during training to increase the robustness of the training model to the label noise. Furthermore, we exploit the high comorbidity of abnormalities observed in chest radiography and incorporate this information to further reduce the impact of label noise. Additionally, anatomical knowledge is incorporated by training the system to predict lung and heart segmentation, as well as spatial knowledge labels. To deal with multiple datasets and images derived from various scanners that apply different post-processing techniques, we introduce a novel image normalization strategy. Experiments were performed on an extensive collection of 297,541 chest radiographs from 86,876 patients, leading to a state-of-the-art performance level for 17 abnormalities from 2 datasets. With an average AUC score of 0.880 across all abnormalities, our proposed training strategies can be used to significantly improve performance scores.
Collapse
Affiliation(s)
- Sebastian Gündel
- Digital Technology and Inovation, Siemens Healthineers, Erlangen 91052, Germany; Pattern Recognition Lab, Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen 91058, Germany.
| | - Arnaud A A Setio
- Digital Technology and Inovation, Siemens Healthineers, Erlangen 91052, Germany
| | - Florin C Ghesu
- Digital Technology and Inovation, Siemens Healthineers, Princeton, NJ 08540, USA
| | - Sasa Grbic
- Digital Technology and Inovation, Siemens Healthineers, Princeton, NJ 08540, USA
| | - Bogdan Georgescu
- Digital Technology and Inovation, Siemens Healthineers, Princeton, NJ 08540, USA
| | - Andreas Maier
- Pattern Recognition Lab, Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen 91058, Germany
| | - Dorin Comaniciu
- Digital Technology and Inovation, Siemens Healthineers, Princeton, NJ 08540, USA
| |
Collapse
|
38
|
Zhang J, Xie Y, Pang G, Liao Z, Verjans J, Li W, Sun Z, He J, Li Y, Shen C, Xia Y. Viral Pneumonia Screening on Chest X-Rays Using Confidence-Aware Anomaly Detection. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:879-890. [PMID: 33245693 PMCID: PMC8544953 DOI: 10.1109/tmi.2020.3040950] [Citation(s) in RCA: 113] [Impact Index Per Article: 28.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/05/2020] [Revised: 11/10/2020] [Accepted: 11/22/2020] [Indexed: 05/24/2023]
Abstract
Clusters of viral pneumonia occurrences over a short period may be a harbinger of an outbreak or pandemic. Rapid and accurate detection of viral pneumonia using chest X-rays can be of significant value for large-scale screening and epidemic prevention, particularly when other more sophisticated imaging modalities are not readily accessible. However, the emergence of novel mutated viruses causes a substantial dataset shift, which can greatly limit the performance of classification-based approaches. In this paper, we formulate the task of differentiating viral pneumonia from non-viral pneumonia and healthy controls into a one-class classification-based anomaly detection problem. We therefore propose the confidence-aware anomaly detection (CAAD) model, which consists of a shared feature extractor, an anomaly detection module, and a confidence prediction module. If the anomaly score produced by the anomaly detection module is large enough, or the confidence score estimated by the confidence prediction module is small enough, the input will be accepted as an anomaly case (i.e., viral pneumonia). The major advantage of our approach over binary classification is that we avoid modeling individual viral pneumonia classes explicitly and treat all known viral pneumonia cases as anomalies to improve the one-class model. The proposed model outperforms binary classification models on the clinical X-VIRAL dataset that contains 5,977 viral pneumonia (no COVID-19) cases, 37,393 non-viral pneumonia or healthy cases. Moreover, when directly testing on the X-COVID dataset that contains 106 COVID-19 cases and 107 normal controls without any fine-tuning, our model achieves an AUC of 83.61% and sensitivity of 71.70%, which is comparable to the performance of radiologists reported in the literature.
Collapse
Affiliation(s)
- Jianpeng Zhang
- National Engineering Laboratory for Integrated Aero-Space-Ground-Ocean Big Data Application Technology, School of Computer Science and EngineeringNorthwestern Polytechnical UniversityXi’an710072China
| | - Yutong Xie
- National Engineering Laboratory for Integrated Aero-Space-Ground-Ocean Big Data Application Technology, School of Computer Science and EngineeringNorthwestern Polytechnical UniversityXi’an710072China
| | - Guansong Pang
- School of Computer ScienceThe University of AdelaideSA5005Australia
| | - Zhibin Liao
- School of Computer ScienceThe University of AdelaideSA5005Australia
| | - Johan Verjans
- School of Computer ScienceThe University of AdelaideSA5005Australia
| | | | - Zongji Sun
- National Engineering Laboratory for Integrated Aero-Space-Ground-Ocean Big Data Application Technology, School of Computer Science and EngineeringNorthwestern Polytechnical UniversityXi’an710072China
| | - Jian He
- Department of RadiologyNanjing Drum Tower Hospital-Affiliated Hospital, Medical SchoolNanjing UniversityNanjing210029China
| | - Yi Li
- GreyBird Ventures, LLCConcordMA01742USA
| | - Chunhua Shen
- School of Computer ScienceThe University of AdelaideSA5005Australia
| | - Yong Xia
- National Engineering Laboratory for Integrated Aero-Space-Ground-Ocean Big Data Application Technology, School of Computer Science and EngineeringNorthwestern Polytechnical UniversityXi’an710072China
- Research and Development Institute, Northwestern Polytechnical University in ShenzhenShenzhen518057China
| |
Collapse
|
39
|
Cheng CT, Wang Y, Chen HW, Hsiao PM, Yeh CN, Hsieh CH, Miao S, Xiao J, Liao CH, Lu L. A scalable physician-level deep learning algorithm detects universal trauma on pelvic radiographs. Nat Commun 2021; 12:1066. [PMID: 33594071 PMCID: PMC7887334 DOI: 10.1038/s41467-021-21311-3] [Citation(s) in RCA: 51] [Impact Index Per Article: 12.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/07/2020] [Accepted: 01/20/2021] [Indexed: 12/13/2022] Open
Abstract
Pelvic radiograph (PXR) is essential for detecting proximal femur and pelvis injuries in trauma patients, which is also the key component for trauma survey. None of the currently available algorithms can accurately detect all kinds of trauma-related radiographic findings on PXRs. Here, we show a universal algorithm can detect most types of trauma-related radiographic findings on PXRs. We develop a multiscale deep learning algorithm called PelviXNet trained with 5204 PXRs with weakly supervised point annotation. PelviXNet yields an area under the receiver operating characteristic curve (AUROC) of 0.973 (95% CI, 0.960-0.983) and an area under the precision-recall curve (AUPRC) of 0.963 (95% CI, 0.948-0.974) in the clinical population test set of 1888 PXRs. The accuracy, sensitivity, and specificity at the cutoff value are 0.924 (95% CI, 0.912-0.936), 0.908 (95% CI, 0.885-0.908), and 0.932 (95% CI, 0.919-0.946), respectively. PelviXNet demonstrates comparable performance with radiologists and orthopedics in detecting pelvic and hip fractures.
Collapse
Affiliation(s)
- Chi-Tung Cheng
- Department of Trauma and Emergency Surgery, Chang Gung Memorial Hospital, Linkou, Chang Gung University, Taoyuan, Taiwan
| | | | - Huan-Wu Chen
- Division of Emergency and Critical Care Radiology, Department of Medical Imaging and Intervention, Chang Gung Memorial Hospital, Chang Gung University College of Medicine, Taoyuan, Taiwan
| | - Po-Meng Hsiao
- New Taipei Municipal TuCheng Hospital, New Taipei city, Taiwan
| | - Chun-Nan Yeh
- Department of Surgery, Chang Gung Memorial Hospital, Linkou, Chang Gung University, Taoyuan, Taiwan
| | - Chi-Hsun Hsieh
- Department of Trauma and Emergency Surgery, Chang Gung Memorial Hospital, Linkou, Chang Gung University, Taoyuan, Taiwan
| | | | | | - Chien-Hung Liao
- Department of Trauma and Emergency Surgery, Chang Gung Memorial Hospital, Linkou, Chang Gung University, Taoyuan, Taiwan.
- Center for Artificial Intelligence in Medicine, Chang Gung Memorial hospital, Linkou, Taoyuan, Taiwan.
| | - Le Lu
- PAII Inc, Bethesda, MD, USA
| |
Collapse
|
40
|
Ayaz M, Shaukat F, Raja G. Ensemble learning based automatic detection of tuberculosis in chest X-ray images using hybrid feature descriptors. Phys Eng Sci Med 2021; 44:183-194. [PMID: 33459996 PMCID: PMC7812355 DOI: 10.1007/s13246-020-00966-0] [Citation(s) in RCA: 40] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/12/2020] [Accepted: 12/17/2020] [Indexed: 02/02/2023]
Abstract
Tuberculosis (TB) remains one of the major health problems in modern times with a high mortality rate. While efforts are being made to make early diagnosis accessible and more reliable in high burden TB countries, digital chest radiography has become a popular source for this purpose. However, the screening process requires expert radiologists which may be a potential barrier in developing countries. A fully automatic computer-aided diagnosis system can reduce the need of trained personnel for early diagnosis of TB using chest X-ray images. In this paper, we have proposed a novel TB detection technique that combines hand-crafted features with deep features (convolutional neural network-based) through Ensemble Learning. Handcrafted features were extracted via Gabor Filter and deep features were extracted via pre-trained deep learning models. Two publicly available datasets namely (i) Montgomery and (ii) Shenzhen were used to evaluate the proposed system. The proposed methodology was validated with a k-fold cross-validation scheme. The area under receiver operating characteristics curves of 0.99 and 0.97 were achieved for Shenzhen and Montgomery datasets respectively which shows the superiority of the proposed scheme.
Collapse
Affiliation(s)
- Muhammad Ayaz
- Faculty of Electronics & Electrical Engineering, University of Engineering & Technology, Taxila, 47080, Pakistan
| | - Furqan Shaukat
- Department of Electronics Engineering, University of Chakwal, Chakwal, Pakistan
| | - Gulistan Raja
- Faculty of Electronics & Electrical Engineering, University of Engineering & Technology, Taxila, 47080, Pakistan.
| |
Collapse
|
41
|
Biswas S, Sen D, Bhatia D, Phukan P, Mukherjee M. Chest X-Ray image and pathological data based artificial intelligence enabled dual diagnostic method for multi-stage classification of COVID-19 patients. AIMS BIOPHYSICS 2021. [DOI: 10.3934/biophy.2021028] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/08/2023] Open
Abstract
<abstract>
<p>The use of Artificial Intelligence (AI) in combination with Internet of Things (IoT) drastically reduces the need to test the COVID samples manually, saving not only time but money and ultimately lives. In this paper, the authors have proposed a novel methodology to identify the COVID-19 patients with an annotated stage to enable the medical staff to manually activate a geo-fence around the subject thus ensuring early detection and isolation. The use of radiography images with pathology data used for COVID-19 identification forms the first-ever contribution by any research group globally. The novelty lies in the correct stage classification of COVID-19 subjects as well. The present analysis would bring this AI Model on the edge to make the facility an IoT-enabled unit. The developed system has been compared and extensively verified thoroughly with those of clinical observations. The significance of radiography imaging for detecting and identification of COVID-19 subjects with severity score tag for stage classification is mathematically established. In a Nutshell, this entire algorithmic workflow can be used not only for predictive analytics but also for prescriptive analytics to complete the entire pipeline from the diagnostic viewpoint of a doctor. As a matter of fact, the authors have used a supervised based learning approach aided by a multiple hypothesis based decision fusion based technique to increase the overall system's accuracy and prediction. The end to end value chain has been put under an IoT based ecosystem to leverage the combined power of AI and IoT to not only detect but also to isolate the coronavirus affected individuals. To emphasize further, the developed AI model predicts the respective categories of a coronavirus affected patients and the IoT system helps the point of care facilities to isolate and prescribe the need of hospitalization for the COVID patients.</p>
</abstract>
Collapse
|
42
|
Wang H, Wang S, Qin Z, Zhang Y, Li R, Xia Y. Triple attention learning for classification of 14 thoracic diseases using chest radiography. Med Image Anal 2020; 67:101846. [PMID: 33129145 DOI: 10.1016/j.media.2020.101846] [Citation(s) in RCA: 54] [Impact Index Per Article: 10.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/26/2019] [Revised: 08/22/2020] [Accepted: 08/25/2020] [Indexed: 01/10/2023]
Abstract
Chest X-ray is the most common radiology examinations for the diagnosis of thoracic diseases. However, due to the complexity of pathological abnormalities and lack of detailed annotation of those abnormalities, computer-aided diagnosis (CAD) of thoracic diseases remains challenging. In this paper, we propose the triple-attention learning (A 3 Net) model for this CAD task. This model uses the pre-trained DenseNet-121 as the backbone network for feature extraction, and integrates three attention modules in a unified framework for channel-wise, element-wise, and scale-wise attention learning. Specifically, the channel-wise attention prompts the deep model to emphasize the discriminative channels of feature maps; the element-wise attention enables the deep model to focus on the regions of pathological abnormalities; the scale-wise attention facilitates the deep model to recalibrate the feature maps at different scales. The proposed model has been evaluated on 112,120images in the ChestX-ray14 dataset with the official patient-level data split. Compared to state-of-the-art deep learning models, our model achieves the highest per-class AUC in classifying 13 out of 14 thoracic diseases and the highest average per-class AUC of 0.826 over 14 thoracic diseases.
Collapse
Affiliation(s)
- Hongyu Wang
- National Engineering Laboratory for Integrated Aero-Space-Ground-Ocean Big Data Application Technology, School of Computer Science and Engineering, Northwestern Polytechnical University, Xi'an 710072, China; Research & Development Institute of Northwestern Polytechnical University in Shenzhen, Shenzhen 518057, China
| | - Shanshan Wang
- Paul C. Lauterbur Research Center for Biomedical Imaging, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
| | - Zibo Qin
- National Engineering Laboratory for Integrated Aero-Space-Ground-Ocean Big Data Application Technology, School of Computer Science and Engineering, Northwestern Polytechnical University, Xi'an 710072, China
| | - Yanning Zhang
- National Engineering Laboratory for Integrated Aero-Space-Ground-Ocean Big Data Application Technology, School of Computer Science and Engineering, Northwestern Polytechnical University, Xi'an 710072, China
| | - Ruijiang Li
- Department of Radiation Oncology, Stanford University School of Medicine, 1070 Arastradero Rd, Palo Alto, CA 94304, USA
| | - Yong Xia
- National Engineering Laboratory for Integrated Aero-Space-Ground-Ocean Big Data Application Technology, School of Computer Science and Engineering, Northwestern Polytechnical University, Xi'an 710072, China; Research & Development Institute of Northwestern Polytechnical University in Shenzhen, Shenzhen 518057, China.
| |
Collapse
|
43
|
Ma B, Li X, Xia Y, Zhang Y. Autonomous deep learning: A genetic DCNN designer for image classification. Neurocomputing 2020. [DOI: 10.1016/j.neucom.2019.10.007] [Citation(s) in RCA: 45] [Impact Index Per Article: 9.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|