1
|
Irfan M, Haq IU, Malik KM, Muhammad K. One-shot learning for generalization in medical image classification across modalities. Comput Med Imaging Graph 2025; 122:102507. [PMID: 40049026 DOI: 10.1016/j.compmedimag.2025.102507] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/15/2024] [Revised: 01/22/2025] [Accepted: 01/29/2025] [Indexed: 03/24/2025]
Abstract
Generalizability is one of the biggest challenges hindering the advancement of medical sensing technologies across multiple imaging modalities. This issue is further impaired when the imaging data is limited in scope or of poor quality. To tackle this, we propose a generalized and robust, lightweight one-shot learning method for medical image classification across various imaging modalities, including X-ray, microscopic, and CT scans. Our model introduces a collaborative one-shot training (COST) approach, incorporating both meta-learning and metric-learning. This approach allows for effective training on only one image per class. To ensure generalization with fewer epochs, we employ gradient generalization at dense and fully connected layers, utilizing a lightweight Siamese network with triplet loss and shared parameters. The proposed model was evaluated on 12 medical image datasets from MedMNIST2D, achieving an average accuracy of 91.5 % and area under the curve (AUC) of 0.89, outperforming state-of-the-art models such as ResNet-50 and AutoML by over 10 % on certain datasets. Further, in the OCTMNIST dataset, our model achieved an AUC of 0.91 compared to ResNet-50's 0.77. Ablation studies further validate the superiority of our approach, with the COST method showing significant improvement in convergence speed and accuracy when compared to traditional one-shot learning setups. Additionally, our model's lightweight architecture requires only 0.15 million parameters, making it well-suited for deployment on resource-constrained devices.
Collapse
Affiliation(s)
- Muhammad Irfan
- SMILES LAB, College of Innovation & Technology, University of Michigan-Flint, Flint, MI 48502, USA
| | - Ijaz Ul Haq
- SMILES LAB, College of Innovation & Technology, University of Michigan-Flint, Flint, MI 48502, USA
| | - Khalid Mahmood Malik
- SMILES LAB, College of Innovation & Technology, University of Michigan-Flint, Flint, MI 48502, USA.
| | - Khan Muhammad
- VIS2KNOW Lab, Department of Applied Artificial Intelligence, School of Convergence, College of Computing and Informatics, Sungkyunkwan University, Seoul 03063, South Korea.
| |
Collapse
|
2
|
Guo S, Liu Z, Yang Z, Lee CH, Lv Q, Shen L. Multi-scale multi-object semi-supervised consistency learning for ultrasound image segmentation. Neural Netw 2025; 184:107095. [PMID: 39754842 DOI: 10.1016/j.neunet.2024.107095] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/17/2024] [Revised: 10/18/2024] [Accepted: 12/23/2024] [Indexed: 01/06/2025]
Abstract
Manual annotation of ultrasound images relies on expert knowledge and requires significant time and financial resources. Semi-supervised learning (SSL) exploits large amounts of unlabeled data to improve model performance under limited labeled data. However, it faces two challenges: fusion of contextual information at multiple scales and bias of spatial information between multiple objects. We propose a consistency learning-based multi-scale multi-object (MSMO) semi-supervised framework for ultrasound image segmentation. MSMO addresses these challenges by employing a contextual-aware encoder coupled with a multi-object semantic calibration and fusion decoder. First, the encoder extracts multi-scale multi-objects context-aware features, and introduces attention module to refine the feature map and enhance channel information interaction. Then, the decoder uses HConvLSTM to calibrate the output features of the current object by using the hidden state of the previous object, and recursively fuses multi-object semantics at different scales. Finally, MSMO further reduces variations among multiple decoders in different perturbations through consistency constraints, thereby producing consistent predictions for highly uncertain areas. Extensive experiments show that proposed MSMO outperforms the SSL baseline on four benchmark datasets, whether for single-object or multi-object ultrasound image segmentation. MSMO significantly reduces the burden of manual analysis of ultrasound images and holds great potential as a clinical tool. The source code is accessible to the public at: https://github.com/lol88/MSMO.
Collapse
Affiliation(s)
- Saidi Guo
- School of Cyber Science and Engineering, Zhengzhou University, Zhengzhou 450002, China; School of Computer and Artificial Intelligence, Zhengzhou University, Zhengzhou 450001, China
| | - Zhaoshan Liu
- Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore 117575, Singapore
| | - Ziduo Yang
- Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore 117575, Singapore; School of Intelligent Systems Engineering, Shenzhen Campus of Sun Yat-sen University, Shenzhen, Guangdong 518107, China
| | - Chau Hung Lee
- Department of Radiology, Tan Tock Seng Hospital, 11 Jalan Tan Tock Seng, Singapore 308433, Singapore
| | - Qiujie Lv
- School of Computer and Artificial Intelligence, Zhengzhou University, Zhengzhou 450001, China.
| | - Lei Shen
- Department of Mechanical Engineering, National University of Singapore, 9 Engineering Drive 1, Singapore 117575, Singapore.
| |
Collapse
|
3
|
Li H, Wang Y, Qiang Y. A semi-supervised domain adaptive medical image segmentation method based on dual-level multi-scale alignment. Sci Rep 2025; 15:8784. [PMID: 40082549 PMCID: PMC11906615 DOI: 10.1038/s41598-025-93824-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/17/2024] [Accepted: 03/10/2025] [Indexed: 03/16/2025] Open
Abstract
In the actual image segmentation tasks in the medical field, the phenomenon of limited labeled data accompanied by domain shifts often occurs and such domain shifts may exist in homologous or even heterologous data. In the study, a novel method was proposed to deal with this challenging phenomenon. Firstly, a model was trained with labeled data in source and target domains so as to adapt to unlabeled data. Then, the alignment at two main levels was realized. At the style level, based on multi-scale stylistic features, the alignment of unlabeled target images was maximized and unlabeled target image features were enhanced. At the inter-domain level, the similarity of the category centroids between target domain data and mixed image data was also maximized. Additionally, a fused supervised loss and alignment loss computation method was proposed. In validation experiments, two cross-domain medical image datasets were constructed: homologous and heterologous datasets. Experimental results showed that the proposed method had the more advantageous comprehensive performance than common semi-supervised and domain adaptation methods.
Collapse
Affiliation(s)
- Hualing Li
- School of Software, North University of China, Taiyuan, Shanxi, China.
| | - Yaodan Wang
- School of Software, North University of China, Taiyuan, Shanxi, China
| | - Yan Qiang
- School of Software, North University of China, Taiyuan, Shanxi, China
| |
Collapse
|
4
|
Hresko DJ, Drotar P, Ngo QC, Kumar DK. Enhanced Domain Adaptation for Foot Ulcer Segmentation Through Mixing Self-Trained Weak Labels. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025; 38:455-466. [PMID: 39020158 PMCID: PMC11810871 DOI: 10.1007/s10278-024-01193-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/28/2024] [Revised: 05/04/2024] [Accepted: 05/22/2024] [Indexed: 07/19/2024]
Abstract
Wound management requires the measurement of the wound parameters such as its shape and area. However, computerized analysis of the wound suffers the challenge of inexact segmentation of the wound images due to limited or inaccurate labels. It is a common scenario that the source domain provides an abundance of labeled data, while the target domain provides only limited labels. To overcome this, we propose a novel approach that combines self-training learning and mixup augmentation. The neural network is trained on the source domain to generate weak labels on the target domain via the self-training process. In the second stage, generated labels are mixed up with labels from the source domain to retrain the neural network and enhance generalization across diverse datasets. The efficacy of our approach was evaluated using the DFUC 2022, FUSeg, and RMIT datasets, demonstrating substantial improvements in segmentation accuracy and robustness across different data distributions. Specifically, in single-domain experiments, segmentation on the DFUC 2022 dataset scored a dice score of 0.711, while the score on the FUSeg dataset achieved 0.859. For domain adaptation, when these datasets were used as target datasets, the dice scores were 0.714 for DFUC 2022 and 0.561 for FUSeg.
Collapse
Affiliation(s)
- David Jozef Hresko
- IISLab, Technical University of Kosice, Letna 1/9, Kosice, 04200, Kosicky Kraj, Slovakia
| | - Peter Drotar
- IISLab, Technical University of Kosice, Letna 1/9, Kosice, 04200, Kosicky Kraj, Slovakia.
| | - Quoc Cuong Ngo
- School of Engineering, RMIT University, 80/445 Swanston St, Melbourne, 3000, VIC, Australia
| | - Dinesh Kant Kumar
- School of Engineering, RMIT University, 80/445 Swanston St, Melbourne, 3000, VIC, Australia
| |
Collapse
|
5
|
Chen J, Li X, Zhang H, Cho Y, Hwang SH, Gao Z, Yang G. Adaptive dynamic inference for few-shot left atrium segmentation. Med Image Anal 2024; 98:103321. [PMID: 39197302 DOI: 10.1016/j.media.2024.103321] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/06/2023] [Revised: 07/13/2024] [Accepted: 08/20/2024] [Indexed: 09/01/2024]
Abstract
Accurate segmentation of the left atrium (LA) from late gadolinium-enhanced cardiac magnetic resonance (LGE CMR) images is crucial for aiding the treatment of patients with atrial fibrillation. Few-shot learning holds significant potential for achieving accurate LA segmentation with low demand on high-cost labeled LGE CMR data and fast generalization across different centers. However, accurate LA segmentation with few-shot learning is a challenging task due to the low-intensity contrast between the LA and other neighboring organs in LGE CMR images. To address this issue, we propose an Adaptive Dynamic Inference Network (ADINet) that explicitly models the differences between the foreground and background. Specifically, ADINet leverages dynamic collaborative inference (DCI) and dynamic reverse inference (DRI) to adaptively allocate semantic-aware and spatial-specific convolution weights and indication information. These allocations are conditioned on the support foreground and background knowledge, utilizing pixel-wise correlations, for different spatial positions of query images. The convolution weights adapt to different visual patterns based on spatial positions, enabling effective encoding of differences between foreground and background regions. Meanwhile, the indication information adapts to the background visual pattern to reversely decode foreground LA regions, leveraging their spatial complementarity. To promote the learning of ADINet, we propose hierarchical supervision, which enforces spatial consistency and differences between the background and foreground regions through pixel-wise semantic supervision and pixel-pixel correlation supervision. We demonstrated the performance of ADINet on three LGE CMR datasets from different centers. Compared to state-of-the-art methods with ten available samples, ADINet yielded better segmentation performance in terms of four metrics.
Collapse
Affiliation(s)
- Jun Chen
- School of Computer Science and Communication Engineering, Jiangsu University, Zhenjiang, Jiangsu, PR China; School of Biomedical Engineering, Sun Yat-sen University, Shenzhen, Guangdong 518107, PR China
| | - Xuejiao Li
- School of Biomedical Engineering, Sun Yat-sen University, Shenzhen, Guangdong 518107, PR China
| | - Heye Zhang
- School of Biomedical Engineering, Sun Yat-sen University, Shenzhen, Guangdong 518107, PR China
| | - Yongwon Cho
- Department of Radiology, Korea University Anam Hospital, 73, Goryeodae-ro, Seoungbuk-gu, Seoul 02841, Republic of Korea
| | - Sung Ho Hwang
- Department of Radiology and the AI center, Korea University Anam Hospital, 73, Goryeodae-ro, Seoungbuk-gu, Seoul 02841, Republic of Korea
| | - Zhifan Gao
- School of Biomedical Engineering, Sun Yat-sen University, Shenzhen, Guangdong 518107, PR China.
| | - Guang Yang
- Bioengineering Department and Imperial-X, Imperial College London, London W12 7SL, UK; National Heart and Lung Institute, Imperial College London, London SW7 2AZ, UK; Cardiovascular Research Centre, Royal Brompton Hospital, London SW3 6NP, UK; School of Biomedical Engineering & Imaging Sciences, King's College London, London WC2R 2LS, UK
| |
Collapse
|
6
|
Wang G, Duan Q, Shen T, Zhang S. SenseCare: a research platform for medical image informatics and interactive 3D visualization. FRONTIERS IN RADIOLOGY 2024; 4:1460889. [PMID: 39639965 PMCID: PMC11617158 DOI: 10.3389/fradi.2024.1460889] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 07/07/2024] [Accepted: 11/06/2024] [Indexed: 12/07/2024]
Abstract
Introduction Clinical research on smart health has an increasing demand for intelligent and clinic-oriented medical image computing algorithms and platforms that support various applications. However, existing research platforms for medical image informatics have limited support for Artificial Intelligence (AI) algorithms and clinical applications. Methods To this end, we have developed SenseCare research platform, which is designed to facilitate translational research on intelligent diagnosis and treatment planning in various clinical scenarios. It has several appealing functions and features such as advanced 3D visualization, concurrent and efficient web-based access, fast data synchronization and high data security, multi-center deployment, support for collaborative research, etc. Results and discussion SenseCare provides a range of AI toolkits for different tasks, including image segmentation, registration, lesion and landmark detection from various image modalities ranging from radiology to pathology. It also facilitates the data annotation and model training processes, which makes it easier for clinical researchers to develop and deploy customized AI models. In addition, it is clinic-oriented and supports various clinical applications such as diagnosis and surgical planning for lung cancer, liver tumor, coronary artery disease, etc. By simplifying AI-based medical image analysis, SenseCare has a potential to promote clinical research in a wide range of disease diagnosis and treatment applications.
Collapse
Affiliation(s)
- Guotai Wang
- School of Mechanical and Electrical Engineering, University of Electronic Science and Technology of China, Chengdu, China
- SenseTime Research, Shanghai, China
| | - Qi Duan
- SenseTime Research, Shanghai, China
| | | | - Shaoting Zhang
- School of Mechanical and Electrical Engineering, University of Electronic Science and Technology of China, Chengdu, China
- SenseTime Research, Shanghai, China
| |
Collapse
|
7
|
Wang Y, Xiao B, Bi X, Li W, Gao X. Boundary-Aware Prototype in Semi-Supervised Medical Image Segmentation. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2024; 33:5456-5467. [PMID: 39316477 DOI: 10.1109/tip.2024.3463412] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/26/2024]
Abstract
The true label plays an important role in semi-supervised medical image segmentation (SSMIS) because it can provide the most accurate supervision information when the label is limited. The popular SSMIS method trains labeled and unlabeled data separately, and the unlabeled data cannot be directly supervised by the true label. This limits the contribution of labels to model training. Is there an interactive mechanism that can break the separation between two types of data training to maximize the utilization of true labels? Inspired by this, we propose a novel consistency learning framework based on the non-parametric distance metric of boundary-aware prototypes to alleviate this problem. This method combines CNN-based linear classification and nearest neighbor-based non-parametric classification into one framework, encouraging the two segmentation paradigms to have similar predictions for the same input. More importantly, the prototype can be clustered from both labeled and unlabeled data features so that it can be seen as a bridge for interactive training between labeled and unlabeled data. When the prototype-based prediction is supervised by the true label, the supervisory signal can simultaneously affect the feature extraction process of both data. In addition, boundary-aware prototypes can explicitly model the differences in boundaries and centers of adjacent categories, so pixel-prototype contrastive learning is introduced to further improve the discriminability of features and make them more suitable for non-parametric distance measurement. Experiments show that although our method uses a modified lightweight UNet as the backbone, it outperforms the comparison method using a 3D VNet with more parameters.
Collapse
|
8
|
Pachetti E, Colantonio S. A systematic review of few-shot learning in medical imaging. Artif Intell Med 2024; 156:102949. [PMID: 39178621 DOI: 10.1016/j.artmed.2024.102949] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/19/2023] [Revised: 07/16/2024] [Accepted: 08/13/2024] [Indexed: 08/26/2024]
Abstract
The lack of annotated medical images limits the performance of deep learning models, which usually need large-scale labelled datasets. Few-shot learning techniques can reduce data scarcity issues and enhance medical image analysis speed and robustness. This systematic review gives a comprehensive overview of few-shot learning methods for medical image analysis, aiming to establish a standard methodological pipeline for future research reference. With a particular emphasis on the role of meta-learning, we analysed 80 relevant articles published from 2018 to 2023, conducting a risk of bias assessment and extracting relevant information, especially regarding the employed learning techniques. From this, we delineated a comprehensive methodological pipeline shared among all studies. In addition, we performed a statistical analysis of the studies' results concerning the clinical task and the meta-learning method employed while also presenting supplemental information such as imaging modalities and model robustness evaluation techniques. We discussed the findings of our analysis, providing a deep insight into the limitations of the state-of-the-art methods and the most promising approaches. Drawing on our investigation, we yielded recommendations on potential future research directions aiming to bridge the gap between research and clinical practice.
Collapse
Affiliation(s)
- Eva Pachetti
- Institute of Information Science and Technologies "Alessandro Faedo", National Research Council of Italy (ISTI-CNR), via Giuseppe Moruzzi 1, Pisa, 56124, PI, Italy; Department of Information Engineering, University of Pisa, via Girolamo Caruso 16, Pisa, 56122, PI, Italy.
| | - Sara Colantonio
- Institute of Information Science and Technologies "Alessandro Faedo", National Research Council of Italy (ISTI-CNR), via Giuseppe Moruzzi 1, Pisa, 56124, PI, Italy.
| |
Collapse
|
9
|
Zwijnen AW, Watzema L, Ridwan Y, van Der Pluijm I, Smal I, Essers J. Self-adaptive deep learning-based segmentation for universal and functional clinical and preclinical CT image analysis. Comput Biol Med 2024; 179:108853. [PMID: 39013341 DOI: 10.1016/j.compbiomed.2024.108853] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2023] [Revised: 07/04/2024] [Accepted: 07/04/2024] [Indexed: 07/18/2024]
Abstract
BACKGROUND Methods to monitor cardiac functioning non-invasively can accelerate preclinical and clinical research into novel treatment options for heart failure. However, manual image analysis of cardiac substructures is resource-intensive and error-prone. While automated methods exist for clinical CT images, translating these to preclinical μCT data is challenging. We employed deep learning to automate the extraction of quantitative data from both CT and μCT images. METHODS We collected a public dataset of cardiac CT images of human patients, as well as acquired μCT images of wild-type and accelerated aging mice. The left ventricle, myocardium, and right ventricle were manually segmented in the μCT training set. After template-based heart detection, two separate segmentation neural networks were trained using the nnU-Net framework. RESULTS The mean Dice score of the CT segmentation results (0.925 ± 0.019, n = 40) was superior to those achieved by state-of-the-art algorithms. Automated and manual segmentations of the μCT training set were nearly identical. The estimated median Dice score (0.940) of the test set results was comparable to existing methods. The automated volume metrics were similar to manual expert observations. In aging mice, ejection fractions had significantly decreased, and myocardial volume increased by age 24 weeks. CONCLUSIONS With further optimization, automated data extraction expands the application of (μ)CT imaging, while reducing subjectivity and workload. The proposed method efficiently measures the left and right ventricular ejection fraction and myocardial mass. With uniform translation between image types, cardiac functioning in diastolic and systolic phases can be monitored in both animals and humans.
Collapse
Affiliation(s)
- Anne-Wietje Zwijnen
- Department of Molecular Genetics, Erasmus University Medical Center, Rotterdam, the Netherlands
| | | | - Yanto Ridwan
- AMIE Core Facility, Erasmus Medical Center, Rotterdam, the Netherlands
| | - Ingrid van Der Pluijm
- Department of Molecular Genetics, Erasmus University Medical Center, Rotterdam, the Netherlands; Department of Vascular Surgery, Erasmus University Medical Center, Rotterdam, the Netherlands
| | - Ihor Smal
- Department of Cell Biology, Erasmus University Medical Center, Rotterdam, the Netherlands
| | - Jeroen Essers
- Department of Molecular Genetics, Erasmus University Medical Center, Rotterdam, the Netherlands; Department of Vascular Surgery, Erasmus University Medical Center, Rotterdam, the Netherlands; Department of Radiotherapy, Erasmus University Medical Center, Rotterdam, the Netherlands.
| |
Collapse
|
10
|
Wang W, Xia Q, Yan Z, Hu Z, Chen Y, Zheng W, Wang X, Nie S, Metaxas D, Zhang S. AVDNet: Joint coronary artery and vein segmentation with topological consistency. Med Image Anal 2024; 91:102999. [PMID: 37862866 DOI: 10.1016/j.media.2023.102999] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2022] [Revised: 09/28/2023] [Accepted: 10/10/2023] [Indexed: 10/22/2023]
Abstract
Coronary CT angiography (CCTA) is an effective and non-invasive method for coronary artery disease diagnosis. Extracting an accurate coronary artery tree from CCTA image is essential for centerline extraction, plaque detection, and stenosis quantification. In practice, data quality varies. Sometimes, the arteries and veins have similar intensities and locate closely, which may confuse segmentation algorithms, even deep learning based ones, to obtain accurate arteries. However, it is not always feasible to re-scan the patient for better image quality. In this paper, we propose an artery and vein disentanglement network (AVDNet) for robust and accurate segmentation by incorporating the coronary vein into the segmentation task. This is the first work to segment coronary artery and vein at the same time. The AVDNet consists of an image based vessel recognition network (IVRN) and a topology based vessel refinement network (TVRN). IVRN learns to segment the arteries and veins, while TVRN learns to correct the segmentation errors based on topology consistency. We also design a novel inverse distance weighted dice (IDD) loss function to recover more thin vessel branches and preserve the vascular boundaries. Extensive experiments are conducted on a multi-center dataset of 700 patients. Quantitative and qualitative results demonstrate the effectiveness of the proposed method by comparing it with state-of-the-art methods and different variants. Prediction results of the AVDNet on the Automated Segmentation of Coronary Artery Challenge dataset are avaliabel at https://github.com/WennyJJ/Coronary-Artery-Vein-Segmentation for follow-up research.
Collapse
Affiliation(s)
- Wenji Wang
- SenseTime Research, Beijing, 100080, China.
| | - Qing Xia
- SenseTime Research, Beijing, 100080, China.
| | | | | | - Yinan Chen
- SenseTime Research, Beijing, 100080, China
| | - Wen Zheng
- Center for Coronary Artery Disease, Beijing Anzhen Hospital, Capital Medical University, Beijing, 100029, China
| | - Xiao Wang
- Center for Coronary Artery Disease, Beijing Anzhen Hospital, Capital Medical University, Beijing, 100029, China
| | - Shaoping Nie
- Center for Coronary Artery Disease, Beijing Anzhen Hospital, Capital Medical University, Beijing, 100029, China
| | - Dimitris Metaxas
- Department of Computer Science, Rutgers University, NJ, 08854, USA
| | - Shaoting Zhang
- SenseTime Research, Beijing, 100080, China; Shanghai Artificial Intelligence Laboratory, Shanghai, 200032, China
| |
Collapse
|
11
|
Osman YBM, Li C, Huang W, Wang S. Sparse annotation learning for dense volumetric MR image segmentation with uncertainty estimation. Phys Med Biol 2023; 69:015009. [PMID: 38035374 DOI: 10.1088/1361-6560/ad111b] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/11/2023] [Accepted: 11/30/2023] [Indexed: 12/02/2023]
Abstract
Objective.Training neural networks for pixel-wise or voxel-wise image segmentation is a challenging task that requires a considerable amount of training samples with highly accurate and densely delineated ground truth maps. This challenge becomes especially prominent in the medical imaging domain, where obtaining reliable annotations for training samples is a difficult, time-consuming, and expert-dependent process. Therefore, developing models that can perform well under the conditions of limited annotated training data is desirable.Approach.In this study, we propose an innovative framework called the extremely sparse annotation neural network (ESA-Net) that learns with only the single central slice label for 3D volumetric segmentation which explores both intra-slice pixel dependencies and inter-slice image correlations with uncertainty estimation. Specifically, ESA-Net consists of four specially designed distinct components: (1) an intra-slice pixel dependency-guided pseudo-label generation module that exploits uncertainty in network predictions while generating pseudo-labels for unlabeled slices with temporal ensembling; (2) an inter-slice image correlation-constrained pseudo-label propagation module which propagates labels from the labeled central slice to unlabeled slices by self-supervised registration with rotation ensembling; (3) a pseudo-label fusion module that fuses the two sets of generated pseudo-labels with voxel-wise uncertainty guidance; and (4) a final segmentation network optimization module to make final predictions with scoring-based label quantification.Main results.Extensive experimental validations have been performed on two popular yet challenging magnetic resonance image segmentation tasks and compared to five state-of-the-art methods.Significance.Results demonstrate that our proposed ESA-Net can consistently achieve better segmentation performances even under the extremely sparse annotation setting, highlighting its effectiveness in exploiting information from unlabeled data.
Collapse
Affiliation(s)
- Yousuf Babiker M Osman
- Paul C. Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, People's Republic of China
- University of Chinese Academy of Sciences, Beijing 100049, People's Republic of China
| | - Cheng Li
- Paul C. Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, People's Republic of China
- Guangdong Provincial Key Laboratory of Artificial Intelligence in Medical Image Analysis and Application, Guangzhou 510080, People's Republic of China
| | - Weijian Huang
- Paul C. Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, People's Republic of China
- University of Chinese Academy of Sciences, Beijing 100049, People's Republic of China
- Peng Cheng Laboratory, Shenzhen 518066, People's Republic of China
| | - Shanshan Wang
- Paul C. Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, People's Republic of China
- Guangdong Provincial Key Laboratory of Artificial Intelligence in Medical Image Analysis and Application, Guangzhou 510080, People's Republic of China
- Peng Cheng Laboratory, Shenzhen 518066, People's Republic of China
| |
Collapse
|
12
|
Li Y, Fu Y, Gayo IJMB, Yang Q, Min Z, Saeed SU, Yan W, Wang Y, Noble JA, Emberton M, Clarkson MJ, Huisman H, Barratt DC, Prisacariu VA, Hu Y. Prototypical few-shot segmentation for cross-institution male pelvic structures with spatial registration. Med Image Anal 2023; 90:102935. [PMID: 37716198 DOI: 10.1016/j.media.2023.102935] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/07/2022] [Revised: 06/08/2023] [Accepted: 08/16/2023] [Indexed: 09/18/2023]
Abstract
The prowess that makes few-shot learning desirable in medical image analysis is the efficient use of the support image data, which are labelled to classify or segment new classes, a task that otherwise requires substantially more training images and expert annotations. This work describes a fully 3D prototypical few-shot segmentation algorithm, such that the trained networks can be effectively adapted to clinically interesting structures that are absent in training, using only a few labelled images from a different institute. First, to compensate for the widely recognised spatial variability between institutions in episodic adaptation of novel classes, a novel spatial registration mechanism is integrated into prototypical learning, consisting of a segmentation head and an spatial alignment module. Second, to assist the training with observed imperfect alignment, support mask conditioning module is proposed to further utilise the annotation available from the support images. Extensive experiments are presented in an application of segmenting eight anatomical structures important for interventional planning, using a data set of 589 pelvic T2-weighted MR images, acquired at seven institutes. The results demonstrate the efficacy in each of the 3D formulation, the spatial registration, and the support mask conditioning, all of which made positive contributions independently or collectively. Compared with the previously proposed 2D alternatives, the few-shot segmentation performance was improved with statistical significance, regardless whether the support data come from the same or different institutes.
Collapse
Affiliation(s)
- Yiwen Li
- Active Vision Laboratory, Department of Engineering Science, University of Oxford, Oxford, UK.
| | - Yunguan Fu
- Department of Medical Physics and Biomedical Engineering, UCL Centre for Medical Image Computing, and Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, London, UK; InstaDeep Ltd., London, UK
| | - Iani J M B Gayo
- Department of Medical Physics and Biomedical Engineering, UCL Centre for Medical Image Computing, and Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, London, UK
| | - Qianye Yang
- Department of Medical Physics and Biomedical Engineering, UCL Centre for Medical Image Computing, and Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, London, UK
| | - Zhe Min
- Department of Medical Physics and Biomedical Engineering, UCL Centre for Medical Image Computing, and Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, London, UK
| | - Shaheer U Saeed
- Department of Medical Physics and Biomedical Engineering, UCL Centre for Medical Image Computing, and Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, London, UK
| | - Wen Yan
- Department of Medical Physics and Biomedical Engineering, UCL Centre for Medical Image Computing, and Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, London, UK; Department of Electrical Engineering, City University of Hong Kong, Hong Kong, China
| | - Yipei Wang
- Institute of Biomedical Engineering, Department of Engineering Science, University of Oxford, Oxford, UK
| | - J Alison Noble
- Institute of Biomedical Engineering, Department of Engineering Science, University of Oxford, Oxford, UK
| | - Mark Emberton
- Division of Surgery & Interventional Science, University College London, London, UK
| | - Matthew J Clarkson
- Department of Medical Physics and Biomedical Engineering, UCL Centre for Medical Image Computing, and Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, London, UK
| | - Henkjan Huisman
- Department of Radiology, Radboud University Nijmegen Medical Centre, Nijmegen, The Netherlands
| | - Dean C Barratt
- Department of Medical Physics and Biomedical Engineering, UCL Centre for Medical Image Computing, and Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, London, UK
| | - Victor A Prisacariu
- Active Vision Laboratory, Department of Engineering Science, University of Oxford, Oxford, UK
| | - Yipeng Hu
- Department of Medical Physics and Biomedical Engineering, UCL Centre for Medical Image Computing, and Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, London, UK; Institute of Biomedical Engineering, Department of Engineering Science, University of Oxford, Oxford, UK
| |
Collapse
|
13
|
Hao D, Li H, Zhang Y, Zhang Q. MUE-CoT: multi-scale uncertainty entropy-aware co-training framework for left atrial segmentation. Phys Med Biol 2023; 68:215008. [PMID: 37567214 DOI: 10.1088/1361-6560/acef8e] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/27/2023] [Accepted: 08/11/2023] [Indexed: 08/13/2023]
Abstract
Objective.Accurate left atrial segmentation is the basis of the recognition and clinical analysis of atrial fibrillation. Supervised learning has achieved some competitive segmentation results, but the high annotation cost often limits its performance. Semi-supervised learning is implemented from limited labeled data and a large amount of unlabeled data and shows good potential in solving practical medical problems.Approach. In this study, we proposed a collaborative training framework for multi-scale uncertain entropy perception (MUE-CoT) and achieved efficient left atrial segmentation from a small amount of labeled data. Based on the pyramid feature network, learning is implemented from unlabeled data by minimizing the pyramid prediction difference. In addition, novel loss constraints are proposed for co-training in the study. The diversity loss is defined as a soft constraint so as to accelerate the convergence and a novel multi-scale uncertainty entropy calculation method and a consistency regularization term are proposed to measure the consistency between prediction results. The quality of pseudo-labels cannot be guaranteed in the pre-training period, so a confidence-dependent empirical Gaussian function is proposed to weight the pseudo-supervised loss.Main results.The experimental results of a publicly available dataset and an in-house clinical dataset proved that our method outperformed existing semi-supervised methods. For the two datasets with a labeled ratio of 5%, the Dice similarity coefficient scores were 84.94% ± 4.31 and 81.24% ± 2.4, the HD95values were 4.63 mm ± 2.13 and 3.94 mm ± 2.72, and the Jaccard similarity coefficient scores were 74.00% ± 6.20 and 68.49% ± 3.39, respectively.Significance.The proposed model effectively addresses the challenges of limited data samples and high costs associated with manual annotation in the medical field, leading to enhanced segmentation accuracy.
Collapse
Affiliation(s)
- Dechen Hao
- School of Software, North University of China, Taiyuan Shanxi, People's Republic of China
| | - Hualing Li
- School of Software, North University of China, Taiyuan Shanxi, People's Republic of China
| | - Yonglai Zhang
- School of Software, North University of China, Taiyuan Shanxi, People's Republic of China
| | - Qi Zhang
- Department of Cardiology, The Second Hospital of Shanxi Medical University, Taiyuan Shanxi, People's Republic of China
| |
Collapse
|
14
|
Li X, Luo G, Wang W, Wang K, Li S. Curriculum label distribution learning for imbalanced medical image segmentation. Med Image Anal 2023; 89:102911. [PMID: 37542795 DOI: 10.1016/j.media.2023.102911] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/28/2022] [Revised: 04/27/2023] [Accepted: 07/25/2023] [Indexed: 08/07/2023]
Abstract
Label distribution learning (LDL) has the potential to resolve boundary ambiguity in semantic segmentation tasks. However, existing LDL-based segmentation methods suffer from severe label distribution imbalance: the ambiguous label distributions contain a small fraction of the data, while the unambiguous label distributions occupy the majority of the data. The imbalanced label distributions induce model-biased distribution learning and make it challenging to accurately predict ambiguous pixels. In this paper, we propose a curriculum label distribution learning (CLDL) framework to address the above data imbalance problem by performing a novel task-oriented curriculum learning strategy. Firstly, the region label distribution learning (R-LDL) is proposed to construct more balanced label distributions and improves the imbalanced model learning. Secondly, a novel learning curriculum (TCL) is proposed to enable easy-to-hard learning in LDL-based segmentation by decomposing the segmentation task into multiple label distribution estimation tasks. Thirdly, the prior perceiving module (PPM) is proposed to effectively connect easy and hard learning stages based on the priors generated from easier stages. Benefiting from the balanced label distribution construction and prior perception, the proposed CLDL effectively conducts a curriculum learning-based LDL and significantly improves the imbalanced learning. We evaluated the proposed CLDL using the publicly available BRATS2018 and MM-WHS2017 datasets. The experimental results demonstrate that our method significantly improves different segmentation metrics compared to many state-of-the-art methods. The code will be available.1.
Collapse
Affiliation(s)
- Xiangyu Li
- School of Computer Science and Technology, Harbin Institute of Technology, Harbin, China
| | - Gongning Luo
- School of Computer Science and Technology, Harbin Institute of Technology, Harbin, China.
| | - Wei Wang
- School of Computer Science and Technology, Harbin Institute of Technology, Shenzhen, China.
| | - Kuanquan Wang
- School of Computer Science and Technology, Harbin Institute of Technology, Harbin, China
| | - Shuo Li
- Department of Biomedical Engineering, Case Western Reserve University, Cleveland, OH 44106, USA.
| |
Collapse
|
15
|
Chang Q, Yan Z, Zhou M, Qu H, He X, Zhang H, Baskaran L, Al'Aref S, Li H, Zhang S, Metaxas DN. Mining multi-center heterogeneous medical data with distributed synthetic learning. Nat Commun 2023; 14:5510. [PMID: 37679325 PMCID: PMC10484909 DOI: 10.1038/s41467-023-40687-y] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2022] [Accepted: 08/03/2023] [Indexed: 09/09/2023] Open
Abstract
Overcoming barriers on the use of multi-center data for medical analytics is challenging due to privacy protection and data heterogeneity in the healthcare system. In this study, we propose the Distributed Synthetic Learning (DSL) architecture to learn across multiple medical centers and ensure the protection of sensitive personal information. DSL enables the building of a homogeneous dataset with entirely synthetic medical images via a form of GAN-based synthetic learning. The proposed DSL architecture has the following key functionalities: multi-modality learning, missing modality completion learning, and continual learning. We systematically evaluate the performance of DSL on different medical applications using cardiac computed tomography angiography (CTA), brain tumor MRI, and histopathology nuclei datasets. Extensive experiments demonstrate the superior performance of DSL as a high-quality synthetic medical image provider by the use of an ideal synthetic quality metric called Dist-FID. We show that DSL can be adapted to heterogeneous data and remarkably outperforms the real misaligned modalities segmentation model by 55% and the temporal datasets segmentation model by 8%.
Collapse
Affiliation(s)
- Qi Chang
- Department of Computer Science, Rutgers University, Piscataway, NJ, USA
| | | | - Mu Zhou
- SenseBrain Research, Princeton, NJ, USA
- Shanghai Artificial Intelligence Laboratory, Shanghai, China
| | - Hui Qu
- Department of Computer Science, Rutgers University, Piscataway, NJ, USA
| | - Xiaoxiao He
- Department of Computer Science, Rutgers University, Piscataway, NJ, USA
| | - Han Zhang
- Department of Computer Science, Rutgers University, Piscataway, NJ, USA
| | - Lohendran Baskaran
- Department of Cardiovascular Medicine, National Heart Centre Singapore, and Duke-National University Of Singapore, Singapore, Singapore
| | - Subhi Al'Aref
- Department of Medicine, Division of Cardiology, University of Arkansas for Medical Sciences, Little Rock, AR, USA
| | - Hongsheng Li
- Chinese University of Hong Kong, Hong Kong SAR, China.
- Centre for Perceptual and Interactive Intelligence (CPII), Hong Kong SAR, China.
| | - Shaoting Zhang
- Shanghai Artificial Intelligence Laboratory, Shanghai, China.
- Centre for Perceptual and Interactive Intelligence (CPII), Hong Kong SAR, China.
- SenseTime, Shanghai, China.
| | - Dimitris N Metaxas
- Department of Computer Science, Rutgers University, Piscataway, NJ, USA.
| |
Collapse
|
16
|
Pandey P, Chasmai M, Sur T, Lall B. Robust Prototypical Few-Shot Organ Segmentation With Regularized Neural-ODEs. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:2490-2501. [PMID: 37030728 DOI: 10.1109/tmi.2023.3258069] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Despite the tremendous progress made by deep learning models in image semantic segmentation, they typically require large annotated examples, and increasing attention is being diverted to problem settings like Few-Shot Learning (FSL) where only a small amount of annotation is needed for generalisation to novel classes. This is especially seen in medical domains where dense pixel-level annotations are expensive to obtain. In this paper, we propose Regularized Prototypical Neural Ordinary Differential Equation (R-PNODE), a method that leverages intrinsic properties of Neural-ODEs, assisted and enhanced by additional cluster and consistency losses to perform Few-Shot Segmentation (FSS) of organs. R-PNODE constrains support and query features from the same classes to lie closer in the representation space thereby improving the performance over the existing Convolutional Neural Network (CNN) based FSS methods. We further demonstrate that while many existing Deep CNN-based methods tend to be extremely vulnerable to adversarial attacks, R-PNODE exhibits increased adversarial robustness for a wide array of these attacks. We experiment with three publicly available multi-organ segmentation datasets in both in-domain and cross-domain FSS settings to demonstrate the efficacy of our method. In addition, we perform experiments with seven commonly used adversarial attacks in various settings to demonstrate R-PNODE's robustness. R-PNODE outperforms the baselines for FSS by significant margins and also shows superior performance for a wide array of attacks varying in intensity and design.
Collapse
|
17
|
Li X, Zhu Z, Yin H, Zhao P, Lv H, Tang R, Qin Y, Zhuo L, Wang Z. Labyrinth morphological modeling and its application on unreferenced segmentation assessment. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2023.104891] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/30/2023]
|
18
|
Tahir AM, Mutlu O, Bensaali F, Ward R, Ghareeb AN, Helmy SMHA, Othman KT, Al-Hashemi MA, Abujalala S, Chowdhury MEH, Alnabti ARDMH, Yalcin HC. Latest Developments in Adapting Deep Learning for Assessing TAVR Procedures and Outcomes. J Clin Med 2023; 12:4774. [PMID: 37510889 PMCID: PMC10381346 DOI: 10.3390/jcm12144774] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2023] [Revised: 04/08/2023] [Accepted: 04/10/2023] [Indexed: 07/30/2023] Open
Abstract
Aortic valve defects are among the most prevalent clinical conditions. A severely damaged or non-functioning aortic valve is commonly replaced with a bioprosthetic heart valve (BHV) via the transcatheter aortic valve replacement (TAVR) procedure. Accurate pre-operative planning is crucial for a successful TAVR outcome. Assessment of computational fluid dynamics (CFD), finite element analysis (FEA), and fluid-solid interaction (FSI) analysis offer a solution that has been increasingly utilized to evaluate BHV mechanics and dynamics. However, the high computational costs and the complex operation of computational modeling hinder its application. Recent advancements in the deep learning (DL) domain can offer a real-time surrogate that can render hemodynamic parameters in a few seconds, thus guiding clinicians to select the optimal treatment option. Herein, we provide a comprehensive review of classical computational modeling approaches, medical imaging, and DL approaches for planning and outcome assessment of TAVR. Particularly, we focus on DL approaches in previous studies, highlighting the utilized datasets, deployed DL models, and achieved results. We emphasize the critical challenges and recommend several future directions for innovative researchers to tackle. Finally, an end-to-end smart DL framework is outlined for real-time assessment and recommendation of the best BHV design for TAVR. Ultimately, deploying such a framework in future studies will support clinicians in minimizing risks during TAVR therapy planning and will help in improving patient care.
Collapse
Affiliation(s)
- Anas M Tahir
- Electrical and Computer Engineering Department, The University of British Columbia, Vancouver, BC V6T 1Z4, Canada
- Biomedical Research Center, Qatar University, Doha 2713, Qatar
| | - Onur Mutlu
- Biomedical Research Center, Qatar University, Doha 2713, Qatar
| | - Faycal Bensaali
- Department of Electrical Engineering, Qatar University, Doha 2713, Qatar
| | - Rabab Ward
- Electrical and Computer Engineering Department, The University of British Columbia, Vancouver, BC V6T 1Z4, Canada
| | - Abdel Naser Ghareeb
- Heart Hospital, Hamad Medical Corporation, Doha 3050, Qatar
- Faculty of Medicine, Al Azhar University, Cairo 11884, Egypt
| | - Sherif M H A Helmy
- Noninvasive Cardiology Section, Cardiology Department, Heart Hospital, Hamad Medical Corporation, Doha 3050, Qatar
| | | | - Mohammed A Al-Hashemi
- Noninvasive Cardiology Section, Cardiology Department, Heart Hospital, Hamad Medical Corporation, Doha 3050, Qatar
| | | | | | | | - Huseyin C Yalcin
- Biomedical Research Center, Qatar University, Doha 2713, Qatar
- Department of Biomedical Science, College of Health Sciences, QU Health, Qatar University, Doha 2713, Qatar
| |
Collapse
|
19
|
Wang Y, Xia W, Yan Z, Zhao L, Bian X, Liu C, Qi Z, Zhang S, Tang Z. Root canal treatment planning by automatic tooth and root canal segmentation in dental CBCT with deep multi-task feature learning. Med Image Anal 2023; 85:102750. [PMID: 36682153 DOI: 10.1016/j.media.2023.102750] [Citation(s) in RCA: 19] [Impact Index Per Article: 9.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/29/2022] [Revised: 10/16/2022] [Accepted: 01/10/2023] [Indexed: 01/21/2023]
Abstract
Accurate and automatic segmentation of individual tooth and root canal from cone-beam computed tomography (CBCT) images is an essential but challenging step for dental surgical planning. In this paper, we propose a novel framework, which consists of two neural networks, DentalNet and PulpNet, for efficient, precise, and fully automatic tooth instance segmentation and root canal segmentation from CBCT images. We first use the proposed DentalNet to achieve tooth instance segmentation and identification. Then, the region of interest (ROI) of the affected tooth is extracted and fed into the PulpNet to obtain precise segmentation of the pulp chamber and the root canal space. These two networks are trained by multi-task feature learning and evaluated on two clinical datasets respectively and achieve superior performances to several comparing methods. In addition, we incorporate our method into an efficient clinical workflow to improve the surgical planning process. In two clinical case studies, our workflow took only 2 min instead of 6 h to obtain the 3D model of tooth and root canal effectively for the surgical planning, resulting in satisfying outcomes in difficult root canal treatments.
Collapse
Affiliation(s)
- Yiwei Wang
- Department of Endodontics, Shanghai Ninth People's Hospital, Shanghai Jiao Tong University School of Medicine; College of Stomatology, Shanghai Jiao Tong University; National Center for Stomatology; National Clinical Research Center for Oral Diseases; Shanghai Key Laboratory of Stomatology; Research Unit of Oral and Maxillofacial Regenerative Medicine, Chinese Academy of Medical Sciences, Shanghai 200011, China
| | - Wenjun Xia
- Shanghai Xuhui District Dental Center, Shanghai 200031, China
| | - Zhennan Yan
- SenseBrain Technology, Princeton, NJ 08540, USA
| | - Liang Zhao
- SenseTime Research, Shanghai 200233, China
| | - Xiaohe Bian
- Department of Endodontics, Shanghai Ninth People's Hospital, Shanghai Jiao Tong University School of Medicine; College of Stomatology, Shanghai Jiao Tong University; National Center for Stomatology; National Clinical Research Center for Oral Diseases; Shanghai Key Laboratory of Stomatology; Research Unit of Oral and Maxillofacial Regenerative Medicine, Chinese Academy of Medical Sciences, Shanghai 200011, China
| | - Chang Liu
- SenseTime Research, Shanghai 200233, China
| | - Zhengnan Qi
- Department of Endodontics, Shanghai Ninth People's Hospital, Shanghai Jiao Tong University School of Medicine; College of Stomatology, Shanghai Jiao Tong University; National Center for Stomatology; National Clinical Research Center for Oral Diseases; Shanghai Key Laboratory of Stomatology; Research Unit of Oral and Maxillofacial Regenerative Medicine, Chinese Academy of Medical Sciences, Shanghai 200011, China
| | - Shaoting Zhang
- Shanghai Artificial Intelligence Laboratory, Shanghai 200232, China; Centre for Perceptual and Interactive Intelligence (CPII), Hong Kong Special Administrative Region of China.
| | - Zisheng Tang
- Department of Endodontics, Shanghai Ninth People's Hospital, Shanghai Jiao Tong University School of Medicine; College of Stomatology, Shanghai Jiao Tong University; National Center for Stomatology; National Clinical Research Center for Oral Diseases; Shanghai Key Laboratory of Stomatology; Research Unit of Oral and Maxillofacial Regenerative Medicine, Chinese Academy of Medical Sciences, Shanghai 200011, China.
| |
Collapse
|
20
|
Byrne N, Clough JR, Valverde I, Montana G, King AP. A Persistent Homology-Based Topological Loss for CNN-Based Multiclass Segmentation of CMR. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:3-14. [PMID: 36044487 PMCID: PMC7614102 DOI: 10.1109/tmi.2022.3203309] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/10/2023]
Abstract
Multi-class segmentation of cardiac magnetic resonance (CMR) images seeks a separation of data into anatomical components with known structure and configuration. The most popular CNN-based methods are optimised using pixel wise loss functions, ignorant of the spatially extended features that characterise anatomy. Therefore, whilst sharing a high spatial overlap with the ground truth, inferred CNN-based segmentations can lack coherence, including spurious connected components, holes and voids. Such results are implausible, violating anticipated anatomical topology. In response, (single-class) persistent homology-based loss functions have been proposed to capture global anatomical features. Our work extends these approaches to the task of multi-class segmentation. Building an enriched topological description of all class labels and class label pairs, our loss functions make predictable and statistically significant improvements in segmentation topology using a CNN-based post-processing framework. We also present (and make available) a highly efficient implementation based on cubical complexes and parallel execution, enabling practical application within high resolution 3D data for the first time. We demonstrate our approach on 2D short axis and 3D whole heart CMR segmentation, advancing a detailed and faithful analysis of performance on two publicly available datasets.
Collapse
Affiliation(s)
- Nick Byrne
- School of BMEIS at KCL: London, SE1 7EH, UK. Nick Byrne and Isra Valverde are also with the Medical Physics and Paediatric Cardiology Departments respectively, both at GSTT: London, SE1 7EH, UK
| | - James R. Clough
- School of BMEIS at KCL: London, SE1 7EH, UK. Nick Byrne and Isra Valverde are also with the Medical Physics and Paediatric Cardiology Departments respectively, both at GSTT: London, SE1 7EH, UK
| | - Israel Valverde
- School of BMEIS at KCL: London, SE1 7EH, UK. Nick Byrne and Isra Valverde are also with the Medical Physics and Paediatric Cardiology Departments respectively, both at GSTT: London, SE1 7EH, UK
| | - Giovanni Montana
- Warwick Manufacturing Group at the University of Warwick: Coventry, CV4 7AL, UK
| | - Andrew P. King
- School of BMEIS at KCL: London, SE1 7EH, UK. Nick Byrne and Isra Valverde are also with the Medical Physics and Paediatric Cardiology Departments respectively, both at GSTT: London, SE1 7EH, UK
| |
Collapse
|
21
|
Bui V, Hsu LY, Chang LC, Sun AY, Tran L, Shanbhag SM, Zhou W, Mehta NN, Chen MY. DeepHeartCT: A fully automatic artificial intelligence hybrid framework based on convolutional neural network and multi-atlas segmentation for multi-structure cardiac computed tomography angiography image segmentation. Front Artif Intell 2022; 5:1059007. [PMID: 36483981 PMCID: PMC9723331 DOI: 10.3389/frai.2022.1059007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/30/2022] [Accepted: 11/03/2022] [Indexed: 01/25/2023] Open
Abstract
Cardiac computed tomography angiography (CTA) is an emerging imaging modality for assessing coronary artery as well as various cardiovascular structures. Recently, deep learning (DL) methods have been successfully applied to many applications of medical image analysis including cardiac CTA structure segmentation. However, DL requires a large amounts of data and high-quality labels for training which can be burdensome to obtain due to its labor-intensive nature. In this study, we aim to develop a fully automatic artificial intelligence (AI) system, named DeepHeartCT, for accurate and rapid cardiac CTA segmentation based on DL. The proposed system was trained using a large clinical dataset with computer-generated labels to segment various cardiovascular structures including left and right ventricles (LV, RV), left and right atria (LA, RA), and LV myocardium (LVM). This new system was trained directly using high-quality computer labels generated from our previously developed multi-atlas based AI system. In addition, a reverse ranking strategy was proposed to assess the segmentation quality in the absence of manual reference labels. This strategy allowed the new framework to assemble optimal computer-generated labels from a large dataset for effective training of a deep convolutional neural network (CNN). A large clinical cardiac CTA studies (n = 1,064) were used to train and validate our framework. The trained model was then tested on another independent dataset with manual labels (n = 60). The Dice score, Hausdorff distance and mean surface distance were used to quantify the segmentation accuracy. The proposed DeepHeartCT framework yields a high median Dice score of 0.90 [interquartile range (IQR), 0.90-0.91], a low median Hausdorff distance of 7 mm (IQR, 4-15 mm) and a low mean surface distance of 0.80 mm (IQR, 0.57-1.29 mm) across all segmented structures. An additional experiment was conducted to evaluate the proposed DL-based AI framework trained with a small vs. large dataset. The results show our framework also performed well when trained on a small optimal training dataset (n = 110) with a significantly reduced training time. These results demonstrated that the proposed DeepHeartCT framework provides accurate and rapid cardiac CTA segmentation that can be readily generalized for handling large-scale medical imaging applications.
Collapse
Affiliation(s)
- Vy Bui
- National Heart Lung and Blood Institute, National Institutes of Health, Bethesda, MD, United States
| | - Li-Yueh Hsu
- Radiology and Imaging Sciences, Clinical Center, National Institutes of Health, Bethesda, MD, United States,*Correspondence: Li-Yueh Hsu
| | - Lin-Ching Chang
- Department of Electrical Engineering and Computer Science, Catholic University of America, Washington, DC, United States
| | - An-Yu Sun
- National Heart Lung and Blood Institute, National Institutes of Health, Bethesda, MD, United States,Department of Electrical Engineering and Computer Science, Catholic University of America, Washington, DC, United States
| | - Loc Tran
- National Heart Lung and Blood Institute, National Institutes of Health, Bethesda, MD, United States,Department of Electrical Engineering and Computer Science, Catholic University of America, Washington, DC, United States
| | - Sujata M. Shanbhag
- National Heart Lung and Blood Institute, National Institutes of Health, Bethesda, MD, United States
| | - Wunan Zhou
- National Heart Lung and Blood Institute, National Institutes of Health, Bethesda, MD, United States
| | - Nehal N. Mehta
- National Heart Lung and Blood Institute, National Institutes of Health, Bethesda, MD, United States
| | - Marcus Y. Chen
- National Heart Lung and Blood Institute, National Institutes of Health, Bethesda, MD, United States
| |
Collapse
|
22
|
Khosravi B, Rouzrokh P, Faghani S, Moassefi M, Vahdati S, Mahmoudi E, Chalian H, Erickson BJ. Machine Learning and Deep Learning in Cardiothoracic Imaging: A Scoping Review. Diagnostics (Basel) 2022; 12:2512. [PMID: 36292201 PMCID: PMC9600598 DOI: 10.3390/diagnostics12102512] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/17/2022] [Revised: 10/14/2022] [Accepted: 10/15/2022] [Indexed: 01/17/2023] Open
Abstract
Machine-learning (ML) and deep-learning (DL) algorithms are part of a group of modeling algorithms that grasp the hidden patterns in data based on a training process, enabling them to extract complex information from the input data. In the past decade, these algorithms have been increasingly used for image processing, specifically in the medical domain. Cardiothoracic imaging is one of the early adopters of ML/DL research, and the COVID-19 pandemic resulted in more research focus on the feasibility and applications of ML/DL in cardiothoracic imaging. In this scoping review, we systematically searched available peer-reviewed medical literature on cardiothoracic imaging and quantitatively extracted key data elements in order to get a big picture of how ML/DL have been used in the rapidly evolving cardiothoracic imaging field. During this report, we provide insights on different applications of ML/DL and some nuances pertaining to this specific field of research. Finally, we provide general suggestions on how researchers can make their research more than just a proof-of-concept and move toward clinical adoption.
Collapse
Affiliation(s)
- Bardia Khosravi
- Radiology Informatics Lab (RIL), Department of Radiology, Mayo Clinic, Rochester, MN 55905, USA
- Orthopedic Surgery Artificial Intelligence Laboratory (OSAIL), Department of Orthopedic Surgery, Mayo Clinic, Rochester, MN 55905, USA
| | - Pouria Rouzrokh
- Radiology Informatics Lab (RIL), Department of Radiology, Mayo Clinic, Rochester, MN 55905, USA
- Orthopedic Surgery Artificial Intelligence Laboratory (OSAIL), Department of Orthopedic Surgery, Mayo Clinic, Rochester, MN 55905, USA
| | - Shahriar Faghani
- Radiology Informatics Lab (RIL), Department of Radiology, Mayo Clinic, Rochester, MN 55905, USA
| | - Mana Moassefi
- Radiology Informatics Lab (RIL), Department of Radiology, Mayo Clinic, Rochester, MN 55905, USA
| | - Sanaz Vahdati
- Radiology Informatics Lab (RIL), Department of Radiology, Mayo Clinic, Rochester, MN 55905, USA
| | - Elham Mahmoudi
- Radiology Informatics Lab (RIL), Department of Radiology, Mayo Clinic, Rochester, MN 55905, USA
| | - Hamid Chalian
- Department of Radiology, Cardiothoracic Imaging, University of Washington, Seattle, WA 98195, USA
| | - Bradley J. Erickson
- Radiology Informatics Lab (RIL), Department of Radiology, Mayo Clinic, Rochester, MN 55905, USA
| |
Collapse
|
23
|
Dikici E, Nguyen XV, Bigelow M, Ryu JL, Prevedello LM. Advancing Brain Metastases Detection in T1-Weighted Contrast-Enhanced 3D MRI Using Noisy Student-Based Training. Diagnostics (Basel) 2022; 12:2023. [PMID: 36010373 PMCID: PMC9407228 DOI: 10.3390/diagnostics12082023] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2022] [Revised: 08/17/2022] [Accepted: 08/19/2022] [Indexed: 11/17/2022] Open
Abstract
The detection of brain metastases (BM) in their early stages could have a positive impact on the outcome of cancer patients. The authors previously developed a framework for detecting small BM (with diameters of <15 mm) in T1-weighted contrast-enhanced 3D magnetic resonance images (T1c). This study aimed to advance the framework with a noisy-student-based self-training strategy to use a large corpus of unlabeled T1c data. Accordingly, a sensitivity-based noisy-student learning approach was formulated to provide high BM detection sensitivity with a reduced count of false positives. This paper (1) proposes student/teacher convolutional neural network architectures, (2) presents data and model noising mechanisms, and (3) introduces a novel pseudo-labeling strategy factoring in the sensitivity constraint. The evaluation was performed using 217 labeled and 1247 unlabeled exams via two-fold cross-validation. The framework utilizing only the labeled exams produced 9.23 false positives for 90% BM detection sensitivity, whereas the one using the introduced learning strategy led to ~9% reduction in false detections (i.e., 8.44). Significant reductions in false positives (>10%) were also observed in reduced labeled data scenarios (using 50% and 75% of labeled data). The results suggest that the introduced strategy could be utilized in existing medical detection applications with access to unlabeled datasets to elevate their performances.
Collapse
Affiliation(s)
- Engin Dikici
- Department of Radiology, The Ohio State University College of Medicine, Columbus, OH 43210, USA
| | - Xuan V. Nguyen
- Department of Radiology, The Ohio State University College of Medicine, Columbus, OH 43210, USA
| | - Matthew Bigelow
- Department of Radiology, The Ohio State University College of Medicine, Columbus, OH 43210, USA
| | | | - Luciano M. Prevedello
- Department of Radiology, The Ohio State University College of Medicine, Columbus, OH 43210, USA
| |
Collapse
|
24
|
Liu Y, He Q, Duan H, Shi H, Han A, He Y. Using Sparse Patch Annotation for Tumor Segmentation in Histopathological Images. SENSORS (BASEL, SWITZERLAND) 2022; 22:6053. [PMID: 36015814 PMCID: PMC9414209 DOI: 10.3390/s22166053] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 07/19/2022] [Revised: 08/05/2022] [Accepted: 08/10/2022] [Indexed: 06/15/2023]
Abstract
Tumor segmentation is a fundamental task in histopathological image analysis. Creating accurate pixel-wise annotations for such segmentation tasks in a fully-supervised training framework requires significant effort. To reduce the burden of manual annotation, we propose a novel weakly supervised segmentation framework based on sparse patch annotation, i.e., only small portions of patches in an image are labeled as 'tumor' or 'normal'. The framework consists of a patch-wise segmentation model called PSeger, and an innovative semi-supervised algorithm. PSeger has two branches for patch classification and image classification, respectively. This two-branch structure enables the model to learn more general features and thus reduce the risk of overfitting when learning sparsely annotated data. We incorporate the idea of consistency learning and self-training into the semi-supervised training strategy to take advantage of the unlabeled images. Trained on the BCSS dataset with only 25% of the images labeled (five patches for each labeled image), our proposed method achieved competitive performance compared to the fully supervised pixel-wise segmentation models. Experiments demonstrate that the proposed solution has the potential to reduce the burden of labeling histopathological images.
Collapse
Affiliation(s)
- Yiqing Liu
- Institute of Biopharmaceutical and Health Engineering, Tsinghua Shenzhen International Graduate School, Shenzhen 518055, China
| | - Qiming He
- Institute of Biopharmaceutical and Health Engineering, Tsinghua Shenzhen International Graduate School, Shenzhen 518055, China
| | - Hufei Duan
- Institute of Biopharmaceutical and Health Engineering, Tsinghua Shenzhen International Graduate School, Shenzhen 518055, China
| | - Huijuan Shi
- Department of Pathology, The First Affiliated Hospital, Sun Yat-sen University, Guangzhou 510080, China
| | - Anjia Han
- Department of Pathology, The First Affiliated Hospital, Sun Yat-sen University, Guangzhou 510080, China
| | - Yonghong He
- Institute of Biopharmaceutical and Health Engineering, Tsinghua Shenzhen International Graduate School, Shenzhen 518055, China
| |
Collapse
|
25
|
Zhou L, Liu H, Zou YX, Zhang G, Su B, Lu L, Chen YC, Yin X, Jiang HB. Clinical validation of an AI-based motion correction reconstruction algorithm in cerebral CT. Eur Radiol 2022; 32:8550-8559. [PMID: 35678857 DOI: 10.1007/s00330-022-08883-4] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/12/2022] [Revised: 04/25/2022] [Accepted: 05/13/2022] [Indexed: 11/04/2022]
Abstract
OBJECTIVES To evaluate the clinical performance of an artificial intelligence (AI)-based motion correction (MC) reconstruction algorithm for cerebral CT. METHODS A total of 53 cases, where motion artifacts were found in the first scan so that an immediate rescan was taken, were retrospectively enrolled. While the rescanned images were reconstructed with a hybrid iterative reconstruction (IR) algorithm (reference group), images of the first scan were reconstructed with both the hybrid IR (motion group) and the MC algorithm (MC group). Image quality was compared in terms of standard deviation (SD), signal-to-noise ratio (SNR), contrast-to-noise ratio (CNR), the mean squared error (MSE), peak signal-to-noise ratio (PSNR), structural similarity index (SSIM), and mutual information (MI), as well as subjective scores. The diagnostic performance for each case was evaluated accordingly by lesion detectability or the Alberta Stroke Program Early CT Score (ASPECTS) assessment. RESULTS Compared with the motion group, the SNR and CNR of the MC group were significantly increased. The MSE, PSNR, SSIM, and MI with respect to the reference group were improved by 44.1%, 15.8%, 7.4%, and 18.3%, respectively (all p < 0.001). Subjective image quality indicators were scored higher for the MC than the motion group (p < 0.05). Improved lesion detectability and higher AUC (0.817 vs 0.614) in the ASPECTS assessment were found for the MC to the motion group. CONCLUSIONS The AI-based MC reconstruction algorithm has been clinically validated for reducing motion artifacts and improving diagnostic performance of cerebral CT. KEY POINTS • An artificial intelligence-based motion correction (MC) reconstruction algorithm has been clinically validated in both qualitative and quantitative manner. • The MC algorithm reduces motion artifacts in cerebral CT and increases the diagnostic confidence for brain lesions. • The MC algorithm can help avoiding rescans caused by motion and improving the efficiency of cerebral CT in the emergency department.
Collapse
Affiliation(s)
- Leilei Zhou
- Department of Radiology, Nanjing First Hospital, Nanjing Medical University, No.68, Changle Road, Nanjing, 210006, China
| | - Hao Liu
- Department of Radiology, Nanjing First Hospital, Nanjing Medical University, No.68, Changle Road, Nanjing, 210006, China
| | - Yi-Xuan Zou
- United Imaging Healthcare Co., Ltd., Shanghai, China
| | - Guozhi Zhang
- United Imaging Healthcare Co., Ltd., Shanghai, China
| | - Bin Su
- United Imaging Healthcare Co., Ltd., Shanghai, China
| | - Liyan Lu
- Department of Radiology, Nanjing First Hospital, Nanjing Medical University, No.68, Changle Road, Nanjing, 210006, China
| | - Yu-Chen Chen
- Department of Radiology, Nanjing First Hospital, Nanjing Medical University, No.68, Changle Road, Nanjing, 210006, China.
| | - Xindao Yin
- Department of Radiology, Nanjing First Hospital, Nanjing Medical University, No.68, Changle Road, Nanjing, 210006, China
| | - Hong-Bing Jiang
- Department of Medical Equipment, Nanjing First Hospital, Nanjing Medical University, No.68, Changle Road, Nanjing, 210006, China. .,Nanjing Emergency Medical Center, No. 3 Zizhulin, Nanjing, 210003, China.
| |
Collapse
|