1
|
Zeng X, Abdullah N, Sumari P. Self-supervised learning framework application for medical image analysis: a review and summary. Biomed Eng Online 2024; 23:107. [PMID: 39465395 PMCID: PMC11514943 DOI: 10.1186/s12938-024-01299-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/10/2024] [Accepted: 10/17/2024] [Indexed: 10/29/2024] Open
Abstract
Manual annotation of medical image datasets is labor-intensive and prone to biases. Moreover, the rate at which image data accumulates significantly outpaces the speed of manual annotation, posing a challenge to the advancement of machine learning, particularly in the realm of supervised learning. Self-supervised learning is an emerging field that capitalizes on unlabeled data for training, thereby circumventing the need for extensive manual labeling. This learning paradigm generates synthetic pseudo-labels through pretext tasks, compelling the network to acquire image representations in a pseudo-supervised manner and subsequently fine-tuning with a limited set of annotated data to achieve enhanced performance. This review begins with an overview of prevalent types and advancements in self-supervised learning, followed by an exhaustive and systematic examination of methodologies within the medical imaging domain from 2018 to September 2024. The review encompasses a range of medical image modalities, including CT, MRI, X-ray, Histology, and Ultrasound. It addresses specific tasks, such as Classification, Localization, Segmentation, Reduction of False Positives, Improvement of Model Performance, and Enhancement of Image Quality. The analysis reveals a descending order in the volume of related studies, with CT and MRI leading the list, followed by X-ray, Histology, and Ultrasound. Except for CT and MRI, there is a greater prevalence of studies focusing on contrastive learning methods over generative learning approaches. The performance of MRI/Ultrasound classification and all image types segmentation still has room for further exploration. Generally, this review can provide conceptual guidance for medical professionals to combine self-supervised learning with their research.
Collapse
Affiliation(s)
- Xiangrui Zeng
- School of Computer Sciences, Universiti Sains Malaysia, USM, 11800, Pulau Pinang, Malaysia.
| | - Nibras Abdullah
- Faculty of Computer Studies, Arab Open University, Jeddah, Saudi Arabia.
| | - Putra Sumari
- School of Computer Sciences, Universiti Sains Malaysia, USM, 11800, Pulau Pinang, Malaysia
| |
Collapse
|
2
|
Sogancioglu E, Ginneken BV, Behrendt F, Bengs M, Schlaefer A, Radu M, Xu D, Sheng K, Scalzo F, Marcus E, Papa S, Teuwen J, Scholten ET, Schalekamp S, Hendrix N, Jacobs C, Hendrix W, Sanchez CI, Murphy K. Nodule Detection and Generation on Chest X-Rays: NODE21 Challenge. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:2839-2853. [PMID: 38530714 DOI: 10.1109/tmi.2024.3382042] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/28/2024]
Abstract
Pulmonary nodules may be an early manifestation of lung cancer, the leading cause of cancer-related deaths among both men and women. Numerous studies have established that deep learning methods can yield high-performance levels in the detection of lung nodules in chest X-rays. However, the lack of gold-standard public datasets slows down the progression of the research and prevents benchmarking of methods for this task. To address this, we organized a public research challenge, NODE21, aimed at the detection and generation of lung nodules in chest X-rays. While the detection track assesses state-of-the-art nodule detection systems, the generation track determines the utility of nodule generation algorithms to augment training data and hence improve the performance of the detection systems. This paper summarizes the results of the NODE21 challenge and performs extensive additional experiments to examine the impact of the synthetically generated nodule training images on the detection algorithm performance.
Collapse
|
3
|
Zhou C, Wang J, Xiang S, Liu F, Huang H, Qian D. A Simple Normalization Technique Using Window Statistics to Improve the Out-of-Distribution Generalization on Medical Images. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:2086-2097. [PMID: 38224511 DOI: 10.1109/tmi.2024.3353800] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/17/2024]
Abstract
Since data scarcity and data heterogeneity are prevailing for medical images, well-trained Convolutional Neural Networks (CNNs) using previous normalization methods may perform poorly when deployed to a new site. However, a reliable model for real-world clinical applications should generalize well both on in-distribution (IND) and out-of-distribution (OOD) data (e.g., the new site data). In this study, we present a novel normalization technique called window normalization (WIN) to improve the model generalization on heterogeneous medical images, which offers a simple yet effective alternative to existing normalization methods. Specifically, WIN perturbs the normalizing statistics with the local statistics computed within a window. This feature-level augmentation technique regularizes the models well and improves their OOD generalization significantly. Leveraging its advantage, we propose a novel self-distillation method called WIN-WIN. WIN-WIN can be easily implemented with two forward passes and a consistency constraint, serving as a simple extension to existing methods. Extensive experimental results on various tasks (6 tasks) and datasets (24 datasets) demonstrate the generality and effectiveness of our methods.
Collapse
|
4
|
Malik M, Chong B, Fernandez J, Shim V, Kasabov NK, Wang A. Stroke Lesion Segmentation and Deep Learning: A Comprehensive Review. Bioengineering (Basel) 2024; 11:86. [PMID: 38247963 PMCID: PMC10813717 DOI: 10.3390/bioengineering11010086] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/18/2023] [Revised: 01/05/2024] [Accepted: 01/15/2024] [Indexed: 01/23/2024] Open
Abstract
Stroke is a medical condition that affects around 15 million people annually. Patients and their families can face severe financial and emotional challenges as it can cause motor, speech, cognitive, and emotional impairments. Stroke lesion segmentation identifies the stroke lesion visually while providing useful anatomical information. Though different computer-aided software are available for manual segmentation, state-of-the-art deep learning makes the job much easier. This review paper explores the different deep-learning-based lesion segmentation models and the impact of different pre-processing techniques on their performance. It aims to provide a comprehensive overview of the state-of-the-art models and aims to guide future research and contribute to the development of more robust and effective stroke lesion segmentation models.
Collapse
Affiliation(s)
- Mishaim Malik
- Auckland Bioengineering Institute, The University of Auckland, Auckland 1010, New Zealand; (M.M.); (B.C.); (N.K.K.)
| | - Benjamin Chong
- Auckland Bioengineering Institute, The University of Auckland, Auckland 1010, New Zealand; (M.M.); (B.C.); (N.K.K.)
- Faculty of Medical and Health Sciences, The University of Auckland, Auckland 1010, New Zealand
- Centre for Brain Research, The University of Auckland, Auckland 1010, New Zealand
| | - Justin Fernandez
- Auckland Bioengineering Institute, The University of Auckland, Auckland 1010, New Zealand; (M.M.); (B.C.); (N.K.K.)
- Centre for Brain Research, The University of Auckland, Auckland 1010, New Zealand
- Mātai Medical Research Institute, Gisborne 4010, New Zealand
| | - Vickie Shim
- Auckland Bioengineering Institute, The University of Auckland, Auckland 1010, New Zealand; (M.M.); (B.C.); (N.K.K.)
- Mātai Medical Research Institute, Gisborne 4010, New Zealand
| | - Nikola Kirilov Kasabov
- Auckland Bioengineering Institute, The University of Auckland, Auckland 1010, New Zealand; (M.M.); (B.C.); (N.K.K.)
- Knowledge Engineering and Discovery Research Innovation, School of Engineering, Computer and Mathematical Sciences, Auckland University of Technology, Auckland 1010, New Zealand
- Institute for Information and Communication Technologies, Bulgarian Academy of Sciences, 1113 Sofia, Bulgaria
- Knowledge Engineering Consulting Ltd., Auckland 1071, New Zealand
| | - Alan Wang
- Auckland Bioengineering Institute, The University of Auckland, Auckland 1010, New Zealand; (M.M.); (B.C.); (N.K.K.)
- Faculty of Medical and Health Sciences, The University of Auckland, Auckland 1010, New Zealand
- Centre for Brain Research, The University of Auckland, Auckland 1010, New Zealand
- Mātai Medical Research Institute, Gisborne 4010, New Zealand
- Medical Imaging Research Centre, The University of Auckland, Auckland 1010, New Zealand
- Centre for Co-Created Ageing Research, The University of Auckland, Auckland 1010, New Zealand
| |
Collapse
|
5
|
Behrendt F, Bengs M, Bhattacharya D, Krüger J, Opfer R, Schlaefer A. A systematic approach to deep learning-based nodule detection in chest radiographs. Sci Rep 2023; 13:10120. [PMID: 37344565 DOI: 10.1038/s41598-023-37270-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/25/2022] [Accepted: 06/19/2023] [Indexed: 06/23/2023] Open
Abstract
Lung cancer is a serious disease responsible for millions of deaths every year. Early stages of lung cancer can be manifested in pulmonary lung nodules. To assist radiologists in reducing the number of overseen nodules and to increase the detection accuracy in general, automatic detection algorithms have been proposed. Particularly, deep learning methods are promising. However, obtaining clinically relevant results remains challenging. While a variety of approaches have been proposed for general purpose object detection, these are typically evaluated on benchmark data sets. Achieving competitive performance for specific real-world problems like lung nodule detection typically requires careful analysis of the problem at hand and the selection and tuning of suitable deep learning models. We present a systematic comparison of state-of-the-art object detection algorithms for the task of lung nodule detection. In this regard, we address the critical aspect of class imbalance and and demonstrate a data augmentation approach as well as transfer learning to boost performance. We illustrate how this analysis and a combination of multiple architectures results in state-of-the-art performance for lung nodule detection, which is demonstrated by the proposed model winning the detection track of the Node21 competition. The code for our approach is available at https://github.com/FinnBehrendt/node21-submit.
Collapse
Affiliation(s)
- Finn Behrendt
- Institute of Medical Technology and Intelligent Systems, Hamburg University of Technology, 21073, Hamburg, Germany.
| | - Marcel Bengs
- Institute of Medical Technology and Intelligent Systems, Hamburg University of Technology, 21073, Hamburg, Germany
| | - Debayan Bhattacharya
- Institute of Medical Technology and Intelligent Systems, Hamburg University of Technology, 21073, Hamburg, Germany
| | | | | | - Alexander Schlaefer
- Institute of Medical Technology and Intelligent Systems, Hamburg University of Technology, 21073, Hamburg, Germany
| |
Collapse
|
6
|
Ghesu FC, Georgescu B, Mansoor A, Yoo Y, Neumann D, Patel P, Vishwanath RS, Balter JM, Cao Y, Grbic S, Comaniciu D. Contrastive self-supervised learning from 100 million medical images with optional supervision. J Med Imaging (Bellingham) 2022; 9:064503. [PMID: 36466078 PMCID: PMC9710476 DOI: 10.1117/1.jmi.9.6.064503] [Citation(s) in RCA: 14] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/25/2022] [Accepted: 11/14/2022] [Indexed: 12/05/2022] Open
Abstract
Purpose Building accurate and robust artificial intelligence systems for medical image assessment requires the creation of large sets of annotated training examples. However, constructing such datasets is very costly due to the complex nature of annotation tasks, which often require expert knowledge (e.g., a radiologist). To counter this limitation, we propose a method to learn from medical images at scale in a self-supervised way. Approach Our approach, based on contrastive learning and online feature clustering, leverages training datasets of over 100,000,000 medical images of various modalities, including radiography, computed tomography (CT), magnetic resonance (MR) imaging, and ultrasonography (US). We propose to use the learned features to guide model training in supervised and hybrid self-supervised/supervised regime on various downstream tasks. Results We highlight a number of advantages of this strategy on challenging image assessment problems in radiography, CT, and MR: (1) significant increase in accuracy compared to the state-of-the-art (e.g., area under the curve boost of 3% to 7% for detection of abnormalities from chest radiography scans and hemorrhage detection on brain CT); (2) acceleration of model convergence during training by up to 85% compared with using no pretraining (e.g., 83% when training a model for detection of brain metastases in MR scans); and (3) increase in robustness to various image augmentations, such as intensity variations, rotations or scaling reflective of data variation seen in the field. Conclusions The proposed approach enables large gains in accuracy and robustness on challenging image assessment problems. The improvement is significant compared with other state-of-the-art approaches trained on medical or vision images (e.g., ImageNet).
Collapse
Affiliation(s)
- Florin C. Ghesu
- Siemens Healthineers, Digital Technology and Innovation, Princeton, New Jersey, United States
| | - Bogdan Georgescu
- Siemens Healthineers, Digital Technology and Innovation, Princeton, New Jersey, United States
| | - Awais Mansoor
- Siemens Healthineers, Digital Technology and Innovation, Princeton, New Jersey, United States
| | - Youngjin Yoo
- Siemens Healthineers, Digital Technology and Innovation, Princeton, New Jersey, United States
| | - Dominik Neumann
- Siemens Healthineers, Digital Technology and Innovation, Erlangen, Germany
| | - Pragneshkumar Patel
- Siemens Healthineers, Digital Technology and Innovation, Princeton, New Jersey, United States
| | | | - James M. Balter
- University of Michigan, Department of Radiation Oncology, Ann Arbor, Michigan, United States
| | - Yue Cao
- University of Michigan, Department of Radiation Oncology, Ann Arbor, Michigan, United States
| | - Sasa Grbic
- Siemens Healthineers, Digital Technology and Innovation, Princeton, New Jersey, United States
| | - Dorin Comaniciu
- Siemens Healthineers, Digital Technology and Innovation, Princeton, New Jersey, United States
| |
Collapse
|
7
|
Maleki F, Muthukrishnan N, Ovens K, Reinhold C, Forghani R. Machine Learning Algorithm Validation: From Essentials to Advanced Applications and Implications for Regulatory Certification and Deployment. Neuroimaging Clin N Am 2021; 30:433-445. [PMID: 33038994 DOI: 10.1016/j.nic.2020.08.004] [Citation(s) in RCA: 40] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/13/2022]
Abstract
The deployment of machine learning (ML) models in the health care domain can increase the speed and accuracy of diagnosis and improve treatment planning and patient care. Translating academic research to applications that are deployable in clinical settings requires the ability to generalize and high reproducibility, which are contingent on a rigorous and sound methodology for the development and evaluation of ML models. This article describes the fundamental concepts and processes for ML model evaluation and highlights common workflows. It concludes with a discussion of the requirements for the deployment of ML models in clinical settings.
Collapse
Affiliation(s)
- Farhad Maleki
- Augmented Intelligence & Precision Health Laboratory (AIPHL), Department of Radiology & Research Institute of the McGill University Health Centre, 5252 Boulevard de Maisonneuve Ouest, Montreal, Quebec H4A 3S5, Canada
| | - Nikesh Muthukrishnan
- Augmented Intelligence & Precision Health Laboratory (AIPHL), Department of Radiology & Research Institute of the McGill University Health Centre, 5252 Boulevard de Maisonneuve Ouest, Montreal, Quebec H4A 3S5, Canada
| | - Katie Ovens
- Department of Computer Science, University of Saskatchewan, 176 Thorvaldson Bldg, 110 Science Place, Saskatoon S7N 5C9, Canada
| | - Caroline Reinhold
- Augmented Intelligence & Precision Health Laboratory (AIPHL), Department of Radiology & Research Institute of the McGill University Health Centre, 5252 Boulevard de Maisonneuve Ouest, Montreal, Quebec H4A 3S5, Canada; Department of Radiology, McGill University, 1650 Cedar Avenue, Montreal, Quebec H3G 1A4, Canada
| | - Reza Forghani
- Augmented Intelligence & Precision Health Laboratory (AIPHL), Department of Radiology & Research Institute of the McGill University Health Centre, 5252 Boulevard de Maisonneuve Ouest, Montreal, Quebec H4A 3S5, Canada; Department of Radiology, McGill University, 1650 Cedar Avenue, Montreal, Quebec H3G 1A4, Canada; Segal Cancer Centre, Lady Davis Institute for Medical Research, Jewish General Hospital, 3755 Cote Ste-Catherine Road, Montreal, Quebec H3T 1E2, Canada; Gerald Bronfman Department of Oncology, McGill University, Suite 720, 5100 Maisonneuve Boulevard West, Montreal, Quebec H4A3T2, Canada; Department of Otolaryngology - Head and Neck Surgery, Royal Victoria Hospital, McGill University Health Centre, 1001 boul. Decarie Boulevard, Montreal, Quebec H3A 3J1, Canada.
| |
Collapse
|
8
|
Mortani Barbosa EJ, Gefter WB, Ghesu FC, Liu S, Mailhe B, Mansoor A, Grbic S, Vogt S. Automated Detection and Quantification of COVID-19 Airspace Disease on Chest Radiographs: A Novel Approach Achieving Expert Radiologist-Level Performance Using a Deep Convolutional Neural Network Trained on Digital Reconstructed Radiographs From Computed Tomography-Derived Ground Truth. Invest Radiol 2021; 56:471-479. [PMID: 33481459 DOI: 10.1097/rli.0000000000000763] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
OBJECTIVES The aim of this study was to leverage volumetric quantification of airspace disease (AD) derived from a superior modality (computed tomography [CT]) serving as ground truth, projected onto digitally reconstructed radiographs (DRRs) to (1) train a convolutional neural network (CNN) to quantify AD on paired chest radiographs (CXRs) and CTs, and (2) compare the DRR-trained CNN to expert human readers in the CXR evaluation of patients with confirmed COVID-19. MATERIALS AND METHODS We retrospectively selected a cohort of 86 COVID-19 patients (with positive reverse transcriptase-polymerase chain reaction test results) from March to May 2020 at a tertiary hospital in the northeastern United States, who underwent chest CT and CXR within 48 hours. The ground-truth volumetric percentage of COVID-19-related AD (POv) was established by manual AD segmentation on CT. The resulting 3-dimensional masks were projected into 2-dimensional anterior-posterior DRR to compute area-based AD percentage (POa). A CNN was trained with DRR images generated from a larger-scale CT dataset of COVID-19 and non-COVID-19 patients, automatically segmenting lungs, AD, and quantifying POa on CXR. The CNN POa results were compared with POa quantified on CXR by 2 expert readers and to the POv ground truth, by computing correlations and mean absolute errors. RESULTS Bootstrap mean absolute error and correlations between POa and POv were 11.98% (11.05%-12.47%) and 0.77 (0.70-0.82) for average of expert readers and 9.56% to 9.78% (8.83%-10.22%) and 0.78 to 0.81 (0.73-0.85) for the CNN, respectively. CONCLUSIONS Our CNN trained with DRR using CT-derived airspace quantification achieved expert radiologist level of accuracy in the quantification of AD on CXR in patients with positive reverse transcriptase-polymerase chain reaction test results for COVID-19.
Collapse
Affiliation(s)
| | - Warren B Gefter
- From the Perelman School of Medicine, University of Pennsylvania, Philadelphia
| | - Florin C Ghesu
- Digital Technology and Innovation, Siemens Healthineers, Princeton, NJ
| | - Siqi Liu
- Digital Technology and Innovation, Siemens Healthineers, Princeton, NJ
| | - Boris Mailhe
- Digital Technology and Innovation, Siemens Healthineers, Princeton, NJ
| | - Awais Mansoor
- Digital Technology and Innovation, Siemens Healthineers, Princeton, NJ
| | - Sasa Grbic
- Digital Technology and Innovation, Siemens Healthineers, Princeton, NJ
| | | |
Collapse
|
9
|
Anatomic Point-Based Lung Region with Zone Identification for Radiologist Annotation and Machine Learning for Chest Radiographs. J Digit Imaging 2021; 34:922-931. [PMID: 34327625 DOI: 10.1007/s10278-021-00494-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/01/2020] [Revised: 06/02/2021] [Accepted: 07/05/2021] [Indexed: 10/20/2022] Open
Abstract
Our objective is to investigate the reliability and usefulness of anatomic point-based lung zone segmentation on chest radiographs (CXRs) as a reference standard framework and to evaluate the accuracy of automated point placement. Two hundred frontal CXRs were presented to two radiologists who identified five anatomic points: two at the lung apices, one at the top of the aortic arch, and two at the costophrenic angles. Of these 1000 anatomic points, 161 (16.1%) were obscured (mostly by pleural effusions). Observer variations were investigated. Eight anatomic zones then were automatically generated from the manually placed anatomic points, and a prototype algorithm was developed using the point-based lung zone segmentation to detect cardiomegaly and levels of diaphragm and pleural effusions. A trained U-Net neural network was used to automatically place these five points within 379 CXRs of an independent database. Intra- and inter-observer variation in mean distance between corresponding anatomic points was larger for obscured points (8.7 mm and 20 mm, respectively) than for visible points (4.3 mm and 7.6 mm, respectively). The computer algorithm using the point-based lung zone segmentation could diagnostically measure the cardiothoracic ratio and diaphragm position or pleural effusion. The mean distance between corresponding points placed by the radiologist and by the neural network was 6.2 mm. The network identified 95% of the radiologist-indicated points with only 3% of network-identified points being false-positives. In conclusion, a reliable anatomic point-based lung segmentation method for CXRs has been developed with expected utility for establishing reference standards for machine learning applications.
Collapse
|
10
|
Gündel S, Setio AAA, Ghesu FC, Grbic S, Georgescu B, Maier A, Comaniciu D. Robust classification from noisy labels: Integrating additional knowledge for chest radiography abnormality assessment. Med Image Anal 2021; 72:102087. [PMID: 34015595 DOI: 10.1016/j.media.2021.102087] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2020] [Revised: 02/24/2021] [Accepted: 04/16/2021] [Indexed: 12/29/2022]
Abstract
Chest radiography is the most common radiographic examination performed in daily clinical practice for the detection of various heart and lung abnormalities. The large amount of data to be read and reported, with more than 100 studies per day for a single radiologist, poses a challenge in consistently maintaining high interpretation accuracy. The introduction of large-scale public datasets has led to a series of novel systems for automated abnormality classification. However, the labels of these datasets were obtained using natural language processed medical reports, yielding a large degree of label noise that can impact the performance. In this study, we propose novel training strategies that handle label noise from such suboptimal data. Prior label probabilities were measured on a subset of training data re-read by 4 board-certified radiologists and were used during training to increase the robustness of the training model to the label noise. Furthermore, we exploit the high comorbidity of abnormalities observed in chest radiography and incorporate this information to further reduce the impact of label noise. Additionally, anatomical knowledge is incorporated by training the system to predict lung and heart segmentation, as well as spatial knowledge labels. To deal with multiple datasets and images derived from various scanners that apply different post-processing techniques, we introduce a novel image normalization strategy. Experiments were performed on an extensive collection of 297,541 chest radiographs from 86,876 patients, leading to a state-of-the-art performance level for 17 abnormalities from 2 datasets. With an average AUC score of 0.880 across all abnormalities, our proposed training strategies can be used to significantly improve performance scores.
Collapse
Affiliation(s)
- Sebastian Gündel
- Digital Technology and Inovation, Siemens Healthineers, Erlangen 91052, Germany; Pattern Recognition Lab, Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen 91058, Germany.
| | - Arnaud A A Setio
- Digital Technology and Inovation, Siemens Healthineers, Erlangen 91052, Germany
| | - Florin C Ghesu
- Digital Technology and Inovation, Siemens Healthineers, Princeton, NJ 08540, USA
| | - Sasa Grbic
- Digital Technology and Inovation, Siemens Healthineers, Princeton, NJ 08540, USA
| | - Bogdan Georgescu
- Digital Technology and Inovation, Siemens Healthineers, Princeton, NJ 08540, USA
| | - Andreas Maier
- Pattern Recognition Lab, Friedrich-Alexander-Universität Erlangen-Nürnberg, Erlangen 91058, Germany
| | - Dorin Comaniciu
- Digital Technology and Inovation, Siemens Healthineers, Princeton, NJ 08540, USA
| |
Collapse
|
11
|
Murphy K, Smits H, Knoops AJG, Korst MBJM, Samson T, Scholten ET, Schalekamp S, Schaefer-Prokop CM, Philipsen RHHM, Meijers A, Melendez J, van Ginneken B, Rutten M. COVID-19 on Chest Radiographs: A Multireader Evaluation of an Artificial Intelligence System. Radiology 2020; 296:E166-E172. [PMID: 32384019 PMCID: PMC7437494 DOI: 10.1148/radiol.2020201874] [Citation(s) in RCA: 127] [Impact Index Per Article: 25.4] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/28/2020] [Revised: 05/03/2020] [Accepted: 05/08/2020] [Indexed: 12/21/2022]
Abstract
Background Chest radiography may play an important role in triage for coronavirus disease 2019 (COVID-19), particularly in low-resource settings. Purpose To evaluate the performance of an artificial intelligence (AI) system for detection of COVID-19 pneumonia on chest radiographs. Materials and Methods An AI system (CAD4COVID-XRay) was trained on 24 678 chest radiographs, including 1540 used only for validation while training. The test set consisted of a set of continuously acquired chest radiographs (n = 454) obtained in patients suspected of having COVID-19 pneumonia between March 4 and April 6, 2020, at one center (223 patients with positive reverse transcription polymerase chain reaction [RT-PCR] results, 231 with negative RT-PCR results). Radiographs were independently analyzed by six readers and by the AI system. Diagnostic performance was analyzed with the receiver operating characteristic curve. Results For the test set, the mean age of patients was 67 years ± 14.4 (standard deviation) (56% male). With RT-PCR test results as the reference standard, the AI system correctly classified chest radiographs as COVID-19 pneumonia with an area under the receiver operating characteristic curve of 0.81. The system significantly outperformed each reader (P < .001 using the McNemar test) at their highest possible sensitivities. At their lowest sensitivities, only one reader significantly outperformed the AI system (P = .04). Conclusion The performance of an artificial intelligence system in the detection of coronavirus disease 2019 on chest radiographs was comparable with that of six independent readers. © RSNA, 2020.
Collapse
Affiliation(s)
- Keelin Murphy
- From the Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Groteplein 10, Nijmegen 6500 HB, the Netherlands (K.M., E.T.S., S.S., C.M.S., B.v.G.); Department of Radiology, Bernhoven Hospital, Uden, the Netherlands (H.S.); Department of Radiology, Jeroen Bosch Hospital, ‘s-Hertogenbosch, the Netherlands (A.J.G.K., M.B.J.M.K., T.S., M.R.); Department of Radiology, Meander Medisch Centrum, Amersfoort, the Netherlands (C.M.S.); and Thirona, Nijmegen, the Netherlands (R.H.H.M.P., A.M., J.M.)
| | - Henk Smits
- From the Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Groteplein 10, Nijmegen 6500 HB, the Netherlands (K.M., E.T.S., S.S., C.M.S., B.v.G.); Department of Radiology, Bernhoven Hospital, Uden, the Netherlands (H.S.); Department of Radiology, Jeroen Bosch Hospital, ‘s-Hertogenbosch, the Netherlands (A.J.G.K., M.B.J.M.K., T.S., M.R.); Department of Radiology, Meander Medisch Centrum, Amersfoort, the Netherlands (C.M.S.); and Thirona, Nijmegen, the Netherlands (R.H.H.M.P., A.M., J.M.)
| | - Arnoud J. G. Knoops
- From the Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Groteplein 10, Nijmegen 6500 HB, the Netherlands (K.M., E.T.S., S.S., C.M.S., B.v.G.); Department of Radiology, Bernhoven Hospital, Uden, the Netherlands (H.S.); Department of Radiology, Jeroen Bosch Hospital, ‘s-Hertogenbosch, the Netherlands (A.J.G.K., M.B.J.M.K., T.S., M.R.); Department of Radiology, Meander Medisch Centrum, Amersfoort, the Netherlands (C.M.S.); and Thirona, Nijmegen, the Netherlands (R.H.H.M.P., A.M., J.M.)
| | - Michael B. J. M. Korst
- From the Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Groteplein 10, Nijmegen 6500 HB, the Netherlands (K.M., E.T.S., S.S., C.M.S., B.v.G.); Department of Radiology, Bernhoven Hospital, Uden, the Netherlands (H.S.); Department of Radiology, Jeroen Bosch Hospital, ‘s-Hertogenbosch, the Netherlands (A.J.G.K., M.B.J.M.K., T.S., M.R.); Department of Radiology, Meander Medisch Centrum, Amersfoort, the Netherlands (C.M.S.); and Thirona, Nijmegen, the Netherlands (R.H.H.M.P., A.M., J.M.)
| | - Tijs Samson
- From the Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Groteplein 10, Nijmegen 6500 HB, the Netherlands (K.M., E.T.S., S.S., C.M.S., B.v.G.); Department of Radiology, Bernhoven Hospital, Uden, the Netherlands (H.S.); Department of Radiology, Jeroen Bosch Hospital, ‘s-Hertogenbosch, the Netherlands (A.J.G.K., M.B.J.M.K., T.S., M.R.); Department of Radiology, Meander Medisch Centrum, Amersfoort, the Netherlands (C.M.S.); and Thirona, Nijmegen, the Netherlands (R.H.H.M.P., A.M., J.M.)
| | - Ernst T. Scholten
- From the Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Groteplein 10, Nijmegen 6500 HB, the Netherlands (K.M., E.T.S., S.S., C.M.S., B.v.G.); Department of Radiology, Bernhoven Hospital, Uden, the Netherlands (H.S.); Department of Radiology, Jeroen Bosch Hospital, ‘s-Hertogenbosch, the Netherlands (A.J.G.K., M.B.J.M.K., T.S., M.R.); Department of Radiology, Meander Medisch Centrum, Amersfoort, the Netherlands (C.M.S.); and Thirona, Nijmegen, the Netherlands (R.H.H.M.P., A.M., J.M.)
| | - Steven Schalekamp
- From the Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Groteplein 10, Nijmegen 6500 HB, the Netherlands (K.M., E.T.S., S.S., C.M.S., B.v.G.); Department of Radiology, Bernhoven Hospital, Uden, the Netherlands (H.S.); Department of Radiology, Jeroen Bosch Hospital, ‘s-Hertogenbosch, the Netherlands (A.J.G.K., M.B.J.M.K., T.S., M.R.); Department of Radiology, Meander Medisch Centrum, Amersfoort, the Netherlands (C.M.S.); and Thirona, Nijmegen, the Netherlands (R.H.H.M.P., A.M., J.M.)
| | - Cornelia M. Schaefer-Prokop
- From the Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Groteplein 10, Nijmegen 6500 HB, the Netherlands (K.M., E.T.S., S.S., C.M.S., B.v.G.); Department of Radiology, Bernhoven Hospital, Uden, the Netherlands (H.S.); Department of Radiology, Jeroen Bosch Hospital, ‘s-Hertogenbosch, the Netherlands (A.J.G.K., M.B.J.M.K., T.S., M.R.); Department of Radiology, Meander Medisch Centrum, Amersfoort, the Netherlands (C.M.S.); and Thirona, Nijmegen, the Netherlands (R.H.H.M.P., A.M., J.M.)
| | - Rick H. H. M. Philipsen
- From the Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Groteplein 10, Nijmegen 6500 HB, the Netherlands (K.M., E.T.S., S.S., C.M.S., B.v.G.); Department of Radiology, Bernhoven Hospital, Uden, the Netherlands (H.S.); Department of Radiology, Jeroen Bosch Hospital, ‘s-Hertogenbosch, the Netherlands (A.J.G.K., M.B.J.M.K., T.S., M.R.); Department of Radiology, Meander Medisch Centrum, Amersfoort, the Netherlands (C.M.S.); and Thirona, Nijmegen, the Netherlands (R.H.H.M.P., A.M., J.M.)
| | - Annet Meijers
- From the Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Groteplein 10, Nijmegen 6500 HB, the Netherlands (K.M., E.T.S., S.S., C.M.S., B.v.G.); Department of Radiology, Bernhoven Hospital, Uden, the Netherlands (H.S.); Department of Radiology, Jeroen Bosch Hospital, ‘s-Hertogenbosch, the Netherlands (A.J.G.K., M.B.J.M.K., T.S., M.R.); Department of Radiology, Meander Medisch Centrum, Amersfoort, the Netherlands (C.M.S.); and Thirona, Nijmegen, the Netherlands (R.H.H.M.P., A.M., J.M.)
| | - Jaime Melendez
- From the Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Groteplein 10, Nijmegen 6500 HB, the Netherlands (K.M., E.T.S., S.S., C.M.S., B.v.G.); Department of Radiology, Bernhoven Hospital, Uden, the Netherlands (H.S.); Department of Radiology, Jeroen Bosch Hospital, ‘s-Hertogenbosch, the Netherlands (A.J.G.K., M.B.J.M.K., T.S., M.R.); Department of Radiology, Meander Medisch Centrum, Amersfoort, the Netherlands (C.M.S.); and Thirona, Nijmegen, the Netherlands (R.H.H.M.P., A.M., J.M.)
| | - Bram van Ginneken
- From the Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Groteplein 10, Nijmegen 6500 HB, the Netherlands (K.M., E.T.S., S.S., C.M.S., B.v.G.); Department of Radiology, Bernhoven Hospital, Uden, the Netherlands (H.S.); Department of Radiology, Jeroen Bosch Hospital, ‘s-Hertogenbosch, the Netherlands (A.J.G.K., M.B.J.M.K., T.S., M.R.); Department of Radiology, Meander Medisch Centrum, Amersfoort, the Netherlands (C.M.S.); and Thirona, Nijmegen, the Netherlands (R.H.H.M.P., A.M., J.M.)
| | - Matthieu Rutten
- From the Diagnostic Image Analysis Group, Radboud University Medical Center, Geert Groteplein 10, Nijmegen 6500 HB, the Netherlands (K.M., E.T.S., S.S., C.M.S., B.v.G.); Department of Radiology, Bernhoven Hospital, Uden, the Netherlands (H.S.); Department of Radiology, Jeroen Bosch Hospital, ‘s-Hertogenbosch, the Netherlands (A.J.G.K., M.B.J.M.K., T.S., M.R.); Department of Radiology, Meander Medisch Centrum, Amersfoort, the Netherlands (C.M.S.); and Thirona, Nijmegen, the Netherlands (R.H.H.M.P., A.M., J.M.)
| |
Collapse
|
12
|
Tang YX, Tang YB, Peng Y, Yan K, Bagheri M, Redd BA, Brandon CJ, Lu Z, Han M, Xiao J, Summers RM. Automated abnormality classification of chest radiographs using deep convolutional neural networks. NPJ Digit Med 2020; 3:70. [PMID: 32435698 PMCID: PMC7224391 DOI: 10.1038/s41746-020-0273-z] [Citation(s) in RCA: 86] [Impact Index Per Article: 17.2] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/03/2019] [Accepted: 04/03/2020] [Indexed: 02/06/2023] Open
Abstract
As one of the most ubiquitous diagnostic imaging tests in medical practice, chest radiography requires timely reporting of potential findings and diagnosis of diseases in the images. Automated, fast, and reliable detection of diseases based on chest radiography is a critical step in radiology workflow. In this work, we developed and evaluated various deep convolutional neural networks (CNN) for differentiating between normal and abnormal frontal chest radiographs, in order to help alert radiologists and clinicians of potential abnormal findings as a means of work list triaging and reporting prioritization. A CNN-based model achieved an AUC of 0.9824 ± 0.0043 (with an accuracy of 94.64 ± 0.45%, a sensitivity of 96.50 ± 0.36% and a specificity of 92.86 ± 0.48%) for normal versus abnormal chest radiograph classification. The CNN model obtained an AUC of 0.9804 ± 0.0032 (with an accuracy of 94.71 ± 0.32%, a sensitivity of 92.20 ± 0.34% and a specificity of 96.34 ± 0.31%) for normal versus lung opacity classification. Classification performance on the external dataset showed that the CNN model is likely to be highly generalizable, with an AUC of 0.9444 ± 0.0029. The CNN model pre-trained on cohorts of adult patients and fine-tuned on pediatric patients achieved an AUC of 0.9851 ± 0.0046 for normal versus pneumonia classification. Pretraining with natural images demonstrates benefit for a moderate-sized training image set of about 8500 images. The remarkable performance in diagnostic accuracy observed in this study shows that deep CNNs can accurately and effectively differentiate normal and abnormal chest radiographs, thereby providing potential benefits to radiology workflow and patient care.
Collapse
Affiliation(s)
- Yu-Xing Tang
- Imaging Biomarkers and Computer-Aided Diagnosis Laboratory, Radiology and Imaging Sciences, National Institutes of Health Clinical Center, Bethesda, MD 20892 USA
| | - You-Bao Tang
- Imaging Biomarkers and Computer-Aided Diagnosis Laboratory, Radiology and Imaging Sciences, National Institutes of Health Clinical Center, Bethesda, MD 20892 USA
| | - Yifan Peng
- National Center for Biotechnology Information, National Library of Medicine, National Institutes of Health, Bethesda, MD 20894 USA
| | - Ke Yan
- Imaging Biomarkers and Computer-Aided Diagnosis Laboratory, Radiology and Imaging Sciences, National Institutes of Health Clinical Center, Bethesda, MD 20892 USA
| | - Mohammadhadi Bagheri
- Clinical Image Processing Service, Radiology and Imaging Sciences, National Institutes of Health Clinical Center, Bethesda, MD 20892 USA
| | - Bernadette A. Redd
- Radiology and Imaging Sciences, National Institutes of Health Clinical Center, Bethesda, MD 20892 USA
| | | | - Zhiyong Lu
- National Center for Biotechnology Information, National Library of Medicine, National Institutes of Health, Bethesda, MD 20894 USA
| | - Mei Han
- PAII Inc, Palo Alto, CA 94306 USA
| | - Jing Xiao
- Ping An Technology, Shenzhen, Guangdong 518029 China
| | - Ronald M. Summers
- Imaging Biomarkers and Computer-Aided Diagnosis Laboratory, Radiology and Imaging Sciences, National Institutes of Health Clinical Center, Bethesda, MD 20892 USA
- Radiology and Imaging Sciences, National Institutes of Health Clinical Center, Bethesda, MD 20892 USA
| |
Collapse
|
13
|
Melendez J, Hogeweg L, Sánchez CI, Philipsen RHHM, Aldridge RW, Hayward AC, Abubakar I, van Ginneken B, Story A. Accuracy of an automated system for tuberculosis detection on chest radiographs in high-risk screening. Int J Tuberc Lung Dis 2019; 22:567-571. [PMID: 29663963 PMCID: PMC5905390 DOI: 10.5588/ijtld.17.0492] [Citation(s) in RCA: 18] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022] Open
Abstract
SETTING: Tuberculosis (TB) screening programmes can be optimised by reducing the number of chest radiographs (CXRs) requiring interpretation by human experts. OBJECTIVE: To evaluate the performance of computerised detection software in triaging CXRs in a high-throughput digital mobile TB screening programme. DESIGN: A retrospective evaluation of the software was performed on a database of 38 961 postero-anterior CXRs from unique individuals seen between 2005 and 2010, 87 of whom were diagnosed with TB. The software generated a TB likelihood score for each CXR. This score was compared with a reference standard for notified active pulmonary TB using receiver operating characteristic (ROC) curve and localisation ROC (LROC) curve analyses. RESULTS: On ROC curve analysis, software specificity was 55.71% (95%CI 55.21–56.20) and negative predictive value was 99.98% (95%CI 99.95–99.99), at a sensitivity of 95%. The area under the ROC curve was 0.90 (95%CI 0.86–0.93). Results of the LROC curve analysis were similar. CONCLUSION: The software could identify more than half of the normal images in a TB screening setting while maintaining high sensitivity, and may therefore be used for triage.
Collapse
Affiliation(s)
- J Melendez
- Diagnostic Image Analysis Group, Radboud University Medical Center, Nijmegen, Thirona, Nijmegen, The Netherlands
| | - L Hogeweg
- Diagnostic Image Analysis Group, Radboud University Medical Center, Nijmegen, The Netherlands
| | - C I Sánchez
- Diagnostic Image Analysis Group, Radboud University Medical Center, Nijmegen, The Netherlands
| | - R H H M Philipsen
- Diagnostic Image Analysis Group, Radboud University Medical Center, Nijmegen, The Netherlands; Thirona, Nijmegen, The Netherlands
| | - R W Aldridge
- Department of Infectious Disease Informatics, Institute of Health Informatics, University College London, London, UK
| | - A C Hayward
- Department of Infectious Disease Informatics, Institute of Health Informatics, University College London, London, UK; Institute of Epidemiology and Health Care, University College London, UK
| | - I Abubakar
- Institute for Global Health, University College London, UK
| | - B van Ginneken
- Diagnostic Image Analysis Group, Radboud University Medical Center, Nijmegen, The Netherlands; Thirona, Nijmegen, The Netherlands
| | - A Story
- Department of Infectious Disease Informatics, Institute of Health Informatics, University College London, London, UK
| |
Collapse
|
14
|
Accurate segmentation of inflammatory and abnormal regions using medical thermal imagery. AUSTRALASIAN PHYSICAL & ENGINEERING SCIENCES IN MEDICINE 2019; 42:647-657. [PMID: 30953251 DOI: 10.1007/s13246-019-00753-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/01/2018] [Accepted: 03/30/2019] [Indexed: 10/27/2022]
Abstract
Methodologies reported in the existing literature for identification of a region of interest (ROI) in medical thermograms suffer from over- and under-extraction of the abnormal and/or inflammatory region, thereby causing inaccurate diagnoses of the spread of an abnormality. We overcome this limitation by exploiting the advantages of a logarithmic transformation. Our algorithm extends the conventional region growing segmentation technique with a modified similarity criteria and a stopping rule. In this method, the ROI is generated by taking common information from two independent regions produced by two different versions of a region-growing algorithm that use different parameters. An automatic multi-seed selection procedure prevents missed segmentations in the proposed approach. We validate our technique by experimentation on various thermal images of the inflammation of affected knees and abnormal breasts. The images were obtained from three databases, namely the Knee joint dataset, the DBT-TU-JU dataset, and the DMR-IR dataset. The superiority of the proposed technique is established by comparison to the performance of state-of-the-art competing methodologies. This study performed temperature emitted inflammatory area segmentation on thermal images of knees and breasts. The proposed segmentation method is of potential value in thermal image processing applications that require expediency and automation.
Collapse
|
15
|
Chen C, Dou Q, Chen H, Heng PA. Semantic-Aware Generative Adversarial Nets for Unsupervised Domain Adaptation in Chest X-Ray Segmentation. MACHINE LEARNING IN MEDICAL IMAGING 2018. [DOI: 10.1007/978-3-030-00919-9_17] [Citation(s) in RCA: 80] [Impact Index Per Article: 11.4] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/14/2023]
|
16
|
Mansoor A, Cerrolaza JJ, Idrees R, Biggs E, Alsharid MA, Avery RA, Linguraru MG. Deep Learning Guided Partitioned Shape Model for Anterior Visual Pathway Segmentation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2016; 35:1856-65. [PMID: 26930677 DOI: 10.1109/tmi.2016.2535222] [Citation(s) in RCA: 25] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/23/2023]
Abstract
Analysis of cranial nerve systems, such as the anterior visual pathway (AVP), from MRI sequences is challenging due to their thin long architecture, structural variations along the path, and low contrast with adjacent anatomic structures. Segmentation of a pathologic AVP (e.g., with low-grade gliomas) poses additional challenges. In this work, we propose a fully automated partitioned shape model segmentation mechanism for AVP steered by multiple MRI sequences and deep learning features. Employing deep learning feature representation, this framework presents a joint partitioned statistical shape model able to deal with healthy and pathological AVP. The deep learning assistance is particularly useful in the poor contrast regions, such as optic tracts and pathological areas. Our main contributions are: 1) a fast and robust shape localization method using conditional space deep learning, 2) a volumetric multiscale curvelet transform-based intensity normalization method for robust statistical model, and 3) optimally partitioned statistical shape and appearance models based on regional shape variations for greater local flexibility. Our method was evaluated on MRI sequences obtained from 165 pediatric subjects. A mean Dice similarity coefficient of 0.779 was obtained for the segmentation of the entire AVP (optic nerve only =0.791 ) using the leave-one-out validation. Results demonstrated that the proposed localized shape and sparse appearance-based learning approach significantly outperforms current state-of-the-art segmentation approaches and is as robust as the manual segmentation.
Collapse
|