1
|
Ince S, Kunduracioglu I, Algarni A, Bayram B, Pacal I. Deep learning for cerebral vascular occlusion segmentation: A novel ConvNeXtV2 and GRN-integrated U-Net framework for diffusion-weighted imaging. Neuroscience 2025; 574:42-53. [PMID: 40204150 DOI: 10.1016/j.neuroscience.2025.04.010] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/27/2025] [Revised: 03/26/2025] [Accepted: 04/05/2025] [Indexed: 04/11/2025]
Abstract
Cerebral vascular occlusion is a serious condition that can lead to stroke and permanent neurological damage due to insufficient oxygen and nutrients reaching brain tissue. Early diagnosis and accurate segmentation are critical for effective treatment planning. Due to its high soft tissue contrast, Magnetic Resonance Imaging (MRI) is commonly used for detecting these occlusions such as ischemic stroke. However, challenges such as low contrast, noise, and heterogeneous lesion structures in MRI images complicate manual segmentation and often lead to misinterpretations. As a result, deep learning-based Computer-Aided Diagnosis (CAD) systems are essential for faster and more accurate diagnosis and treatment methods, although they can sometimes face challenges such as high computational costs and difficulties in segmenting small or irregular lesions. This study proposes a novel U-Net architecture enhanced with ConvNeXtV2 blocks and GRN-based Multi-Layer Perceptrons (MLP) to address these challenges in cerebral vascular occlusion segmentation. This is the first application of ConvNeXtV2 in this domain. The proposed model significantly improves segmentation accuracy, even in low-contrast regions, while maintaining high computational efficiency, which is crucial for real-world clinical applications. To reduce false positives and improve overall accuracy, small lesions (≤5 pixels) were removed in the preprocessing step with the support of expert clinicians. Experimental results on the ISLES 2022 dataset showed superior performance with an Intersection over Union (IoU) of 0.8015 and a Dice coefficient of 0.8894. Comparative analyses indicate that the proposed model achieves higher segmentation accuracy than existing U-Net variants and other methods, offering a promising solution for clinical use.
Collapse
Affiliation(s)
- Suat Ince
- Department of Radiology, University of Health Sciences, Van Education and Research Hospital, 65000 Van, Turkey.
| | - Ismail Kunduracioglu
- Department of Computer Engineering, Faculty of Engineering, Igdir University, 76000 Igdir, Turkey.
| | - Ali Algarni
- Department of Informatics and Computer Systems, College of Computer Science, King Khalid University, Abha 61421, Saudi Arabia.
| | - Bilal Bayram
- Department of Neurology, University of Health Sciences, Van Education and Research Hospital, 65000 Van, Turkey.
| | - Ishak Pacal
- Department of Computer Engineering, Faculty of Engineering, Igdir University, 76000 Igdir, Turkey; Department of Electronics and Information Technologies, Faculty of Architecture and Engineering, Nakhchivan State University, AZ 7012 Nakhchivan, Azerbaijan.
| |
Collapse
|
2
|
Bayram B, Kunduracioglu I, Ince S, Pacal I. A systematic review of deep learning in MRI-based cerebral vascular occlusion-based brain diseases. Neuroscience 2025; 568:76-94. [PMID: 39805420 DOI: 10.1016/j.neuroscience.2025.01.020] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2024] [Revised: 01/09/2025] [Accepted: 01/10/2025] [Indexed: 01/16/2025]
Abstract
Neurological disorders, including cerebral vascular occlusions and strokes, present a major global health challenge due to their high mortality rates and long-term disabilities. Early diagnosis, particularly within the first hours, is crucial for preventing irreversible damage and improving patient outcomes. Although neuroimaging techniques like magnetic resonance imaging (MRI) have advanced significantly, traditional methods often fail to fully capture the complexity of brain lesions. Deep learning has recently emerged as a powerful tool in medical imaging, offering high accuracy in detecting and segmenting brain anomalies. This review examines 61 MRI-based studies published between 2020 and 2024, focusing on the role of deep learning in diagnosing cerebral vascular occlusion-related conditions. It evaluates the successes and limitations of these studies, including the adequacy and diversity of datasets, and addresses challenges such as data privacy and algorithm explainability. Comparisons between convolutional neural network (CNN)-based and Vision Transformer (ViT)-based approaches reveal distinct advantages and limitations. The findings emphasize the importance of ethically secure frameworks, the inclusion of diverse datasets, and improved model interpretability. Advanced architectures like U-Net variants and transformer-based models are highlighted as promising tools to enhance reliability in clinical applications. By automating complex neuroimaging tasks and improving diagnostic accuracy, deep learning facilitates personalized treatment strategies. This review provides a roadmap for integrating technical advancements into clinical practice, underscoring the transformative potential of deep learning in managing neurological disorders and improving healthcare outcomes globally.
Collapse
Affiliation(s)
- Bilal Bayram
- Department of Neurology, University of Health Sciences, Van Education and Research Hospital, 65000, Van, Turkey.
| | - Ismail Kunduracioglu
- Department of Computer Engineering, Faculty of Engineering, Igdir University, 76000, Igdir, Turkey.
| | - Suat Ince
- Department of Radiology, University of Health Sciences, Van Education and Research Hospital, 65000, Van, Turkey.
| | - Ishak Pacal
- Department of Computer Engineering, Faculty of Engineering, Igdir University, 76000, Igdir, Turkey.
| |
Collapse
|
3
|
Lyu J, Bartlett PF, Nasrallah FA, Tang X. Masked Deformation Modeling for Volumetric Brain MRI Self-Supervised Pre-Training. IEEE TRANSACTIONS ON MEDICAL IMAGING 2025; 44:1596-1607. [PMID: 40030579 DOI: 10.1109/tmi.2024.3510922] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/05/2025]
Abstract
Self-supervised learning (SSL) has been proposed to alleviate neural networks' reliance on annotated data and to improve downstream tasks' performance, which has obtained substantial success in several volumetric medical image segmentation tasks. However, most existing approaches are designed and pre-trained on CT or MRI datasets of non-brain organs. The lack of brain prior limits those methods' performance on brain segmentation, especially on fine-grained brain parcellation. To overcome this limitation, we here propose a novel SSL strategy for MRI of the human brain, named Masked Deformation Modeling (MDM). MDM first conducts atlas-guided patch sampling on individual brain MRI scans (moving volumes) and an MNI152 template (a fixed volume). The sampled moving volumes are randomly masked in a feature-aligned manner, and then sent into a U-Net-based network to extract latent features. An intensity head and a deformation field head are used to decode the latent features, respectively restoring the masked volume and predicting the deformation field from the moving volume to the fixed volume. The proposed MDM is fine-tuned and evaluated on three brain parcellation datasets with different granularities (JHU, Mindboggle-101, CANDI), a brain lesion segmentation dataset (ATLAS2), and a brain tumor segmentation dataset (BraTS21). Results demonstrate that MDM outperforms various state-of-the-art medical SSL methods by considerable margins, and can effectively reduce the annotation effort by at least 40%. Codes and pre-trained weights will be released at https://github.com/CRazorback/MDM.
Collapse
|
4
|
Bajaj S, Bala M, Angurala M. A comparative analysis of different augmentations for brain images. Med Biol Eng Comput 2024; 62:3123-3150. [PMID: 38782880 DOI: 10.1007/s11517-024-03127-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2023] [Accepted: 05/10/2024] [Indexed: 05/25/2024]
Abstract
Deep learning (DL) requires a large amount of training data to improve performance and prevent overfitting. To overcome these difficulties, we need to increase the size of the training dataset. This can be done by augmentation on a small dataset. The augmentation approaches must enhance the model's performance during the learning period. There are several types of transformations that can be applied to medical images. These transformations can be applied to the entire dataset or to a subset of the data, depending on the desired outcome. In this study, we categorize data augmentation methods into four groups: Absent augmentation, where no modifications are made; basic augmentation, which includes brightness and contrast adjustments; intermediate augmentation, encompassing a wider array of transformations like rotation, flipping, and shifting in addition to brightness and contrast adjustments; and advanced augmentation, where all transformation layers are employed. We plan to conduct a comprehensive analysis to determine which group performs best when applied to brain CT images. This evaluation aims to identify the augmentation group that produces the most favorable results in terms of improving model accuracy, minimizing diagnostic errors, and ensuring the robustness of the model in the context of brain CT image analysis.
Collapse
Affiliation(s)
- Shilpa Bajaj
- Applied Sciences (Computer Applications), I.K. Gujral Punjab Technical University, Jalandhar, Kapurthala, India.
| | - Manju Bala
- Department of Computer Science and Engineering, Khalsa College of Engineering and Technology, Amritsar, India
| | - Mohit Angurala
- Apex Institute of Technology (CSE), Chandigarh University, Gharuan, Mohali, Punjab, India
| |
Collapse
|
5
|
Ahmed R, Al Shehhi A, Werghi N, Seghier ML. Segmentation of stroke lesions using transformers-augmented MRI analysis. Hum Brain Mapp 2024; 45:e26803. [PMID: 39119860 PMCID: PMC11310771 DOI: 10.1002/hbm.26803] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2024] [Revised: 05/20/2024] [Accepted: 07/17/2024] [Indexed: 08/10/2024] Open
Abstract
Accurate segmentation of chronic stroke lesions from mono-spectral magnetic resonance imaging scans (e.g., T1-weighted images) is a difficult task due to the arbitrary shape, complex texture, variable size and intensities, and varied locations of the lesions. Due to this inherent spatial heterogeneity, existing machine learning methods have shown moderate performance for chronic lesion delineation. In this study, we introduced: (1) a method that integrates transformers' deformable feature attention mechanism with convolutional deep learning architecture to improve the accuracy and generalizability of stroke lesion segmentation, and (2) an ecological data augmentation technique based on inserting real lesions into intact brain regions. Our combination of these two approaches resulted in a significant increase in segmentation performance, with a Dice index of 0.82 (±0.39), outperforming the existing methods trained and tested on the same Anatomical Tracings of Lesions After Stroke (ATLAS) 2022 dataset. Our method performed relatively well even for cases with small stroke lesions. We validated the robustness of our method through an ablation study and by testing it on new unseen brain scans from the Ischemic Stroke Lesion Segmentation (ISLES) 2015 dataset. Overall, our proposed approach of transformers with ecological data augmentation offers a robust way to delineate chronic stroke lesions with clinically relevant accuracy. Our method can be extended to other challenging tasks that require automated detection and segmentation of diverse brain abnormalities from clinical scans.
Collapse
Affiliation(s)
- Ramsha Ahmed
- Department of Biomedical Engineering and BiotechnologyKhalifa University of Science and TechnologyAbu DhabiUAE
| | - Aamna Al Shehhi
- Department of Biomedical Engineering and BiotechnologyKhalifa University of Science and TechnologyAbu DhabiUAE
- Healthcare Engineering Innovation Center (HEIC)Khalifa University of Science and TechnologyAbu DhabiUAE
| | - Naoufel Werghi
- Department of Computer ScienceKhalifa University of Science and TechnologyAbu DhabiUAE
| | - Mohamed L. Seghier
- Department of Biomedical Engineering and BiotechnologyKhalifa University of Science and TechnologyAbu DhabiUAE
- Healthcare Engineering Innovation Center (HEIC)Khalifa University of Science and TechnologyAbu DhabiUAE
| |
Collapse
|
6
|
Ma F, Wang S, Guo Y, Dai C, Meng J. Image segmentation of mouse eye in vivo with optical coherence tomography based on Bayesian classification. BIOMED ENG-BIOMED TE 2024; 69:307-315. [PMID: 38178615 DOI: 10.1515/bmt-2023-0266] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/16/2023] [Accepted: 12/22/2023] [Indexed: 01/06/2024]
Abstract
OBJECTIVES Optical coherence tomography (OCT) is a new imaging technology that uses an optical analog of ultrasound imaging for biological tissues. Image segmentation plays an important role in dealing with quantitative analysis of medical images. METHODS We have proposed a novel framework to deal with the low intensity problem, based on the labeled patches and Bayesian classification (LPBC) model. The proposed method includes training and testing phases. During the training phase, firstly, we manually select the sub-images of background and Region of Interest (ROI) from the training image, and then extract features by patches. Finally, we train the Bayesian model with the features. The segmentation threshold of each patch is computed by the learned Bayesian model. RESULTS In addition, we have collected a new dataset of mouse eyes in vivo with OCT, named MEVOCT, which can be found at URL https://17861318579.github.io/LPBC. MEVOCT consists of 20 high-resolution images. The resolution of every image is 2048 × 2048 pixels. CONCLUSIONS The experimental results demonstrate the effectiveness of the LPBC method on the new MEVOCT dataset. The ROI segmentation is of great importance for the distortion correction.
Collapse
Affiliation(s)
- Fei Ma
- School of Computer Science, Qufu Normal University, Rizhao, Shandong, China
| | - Shengbo Wang
- School of Computer Science, Qufu Normal University, Rizhao, Shandong, China
| | - Yanfei Guo
- School of Computer Science, Qufu Normal University, Rizhao, Shandong, China
| | - Cuixia Dai
- Department of College Science, Shanghai Institute of Technology, Shanghai, Shanghai, China
| | - Jing Meng
- School of Computer Science, Qufu Normal University, Rizhao, Shandong, China
| |
Collapse
|
7
|
Lin PJ, Li W, Zhai X, Sun J, Pan Y, Ji L, Li C. AM-EEGNet: An advanced multi-input deep learning framework for classifying stroke patient EEG task states. Neurocomputing 2024; 585:127622. [DOI: 10.1016/j.neucom.2024.127622] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/24/2024]
|
8
|
Luo J, Dai P, He Z, Huang Z, Liao S, Liu K. Deep learning models for ischemic stroke lesion segmentation in medical images: A survey. Comput Biol Med 2024; 175:108509. [PMID: 38677171 DOI: 10.1016/j.compbiomed.2024.108509] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/02/2023] [Revised: 02/09/2024] [Accepted: 04/21/2024] [Indexed: 04/29/2024]
Abstract
This paper provides a comprehensive review of deep learning models for ischemic stroke lesion segmentation in medical images. Ischemic stroke is a severe neurological disease and a leading cause of death and disability worldwide. Accurate segmentation of stroke lesions in medical images such as MRI and CT scans is crucial for diagnosis, treatment planning and prognosis. This paper first introduces common imaging modalities used for stroke diagnosis, discussing their capabilities in imaging lesions at different disease stages from the acute to chronic stage. It then reviews three major public benchmark datasets for evaluating stroke segmentation algorithms: ATLAS, ISLES and AISD, highlighting their key characteristics. The paper proceeds to provide an overview of foundational deep learning architectures for medical image segmentation, including CNN-based and transformer-based models. It summarizes recent innovations in adapting these architectures to the task of stroke lesion segmentation across the three datasets, analyzing their motivations, modifications and results. A survey of loss functions and data augmentations employed for this task is also included. The paper discusses various aspects related to stroke segmentation tasks, including prior knowledge, small lesions, and multimodal fusion, and then concludes by outlining promising future research directions. Overall, this comprehensive review covers critical technical developments in the field to support continued progress in automated stroke lesion segmentation.
Collapse
Affiliation(s)
- Jialin Luo
- School of Computer Science and Engineering, Central South University, Changsha, Hunan, China
| | - Peishan Dai
- School of Computer Science and Engineering, Central South University, Changsha, Hunan, China.
| | - Zhuang He
- School of Computer Science and Engineering, Central South University, Changsha, Hunan, China
| | - Zhongchao Huang
- Department of Biomedical Engineering, School of Basic Medical Science, Central South University, Changsha, Hunan, China
| | - Shenghui Liao
- School of Computer Science and Engineering, Central South University, Changsha, Hunan, China
| | - Kun Liu
- Brain Hospital of Hunan Province (The Second People's Hospital of Hunan Province), Changsha, Hunan, China
| |
Collapse
|
9
|
Wu Z, Zhang X, Li F, Wang S, Li J. A feature-enhanced network for stroke lesion segmentation from brain MRI images. Comput Biol Med 2024; 174:108326. [PMID: 38599066 DOI: 10.1016/j.compbiomed.2024.108326] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/02/2024] [Revised: 03/14/2024] [Accepted: 03/15/2024] [Indexed: 04/12/2024]
Abstract
Accurate and expeditious segmentation of stroke lesions can greatly assist physicians in making accurate medical diagnoses and administering timely treatments. However, there are two limitations to the current deep learning methods. On the one hand, the attention structure utilizes only local features, which misleads the subsequent segmentation; on the other hand, simple downsampling compromises task-relevant detailed semantic information. To address these challenges, we propose a novel feature refinement and protection network (FRPNet) for stroke lesion segmentation. FRPNet employs a symmetric encoding-decoding structure and incorporates twin attention gate (TAG) and multi-dimension attention pooling (MAP) modules. The TAG module leverages the self-attention mechanism and bi-directional attention to extract both global and local features of the lesion. On the other hand, the MAP module establishes multidimensional pooling attention to effectively mitigate the loss of features during the encoding process. Extensive comparative experiments show that, our method significantly outperforms the state-of-the-art approaches with 60.16% DSC, 36.20px HD and 85.72% DSC, 27.02px HD on two ischemic stroke datasets that contain all stroke stages and several sequences of stroke images. The excellent results that exceed those of existing methods illustrate the efficacy and generalizability of the proposed method. The source code is released on https://github.com/wu2ze2lin2/FRPNet.
Collapse
Affiliation(s)
- Zelin Wu
- College of Electronic Information and Optical Engineering, Taiyuan University of Technology, Taiyuan, 030024, China
| | - Xueying Zhang
- College of Electronic Information and Optical Engineering, Taiyuan University of Technology, Taiyuan, 030024, China.
| | - Fenglian Li
- College of Electronic Information and Optical Engineering, Taiyuan University of Technology, Taiyuan, 030024, China.
| | - Suzhe Wang
- College of Electronic Information and Optical Engineering, Taiyuan University of Technology, Taiyuan, 030024, China
| | - Jiaying Li
- The first clinical medical College, Shanxi Medical University, Taiyuan, 030024, China
| |
Collapse
|
10
|
Wu Z, Zhang X, Li F, Wang S, Li J. TransRender: a transformer-based boundary rendering segmentation network for stroke lesions. Front Neurosci 2023; 17:1259677. [PMID: 37901438 PMCID: PMC10601640 DOI: 10.3389/fnins.2023.1259677] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/17/2023] [Accepted: 09/26/2023] [Indexed: 10/31/2023] Open
Abstract
Vision transformer architectures attract widespread interest due to their robust representation capabilities of global features. Transformer-based methods as the encoder achieve superior performance compared to convolutional neural networks and other popular networks in many segmentation tasks for medical images. Due to the complex structure of the brain and the approximate grayscale of healthy tissue and lesions, lesion segmentation suffers from over-smooth boundaries or inaccurate segmentation. Existing methods, including the transformer, utilize stacked convolutional layers as the decoder to uniformly treat each pixel as a grid, which is convenient for feature computation. However, they often neglect the high-frequency features of the boundary and focus excessively on the region features. We propose an effective method for lesion boundary rendering called TransRender, which adaptively selects a series of important points to compute the boundary features in a point-based rendering way. The transformer-based method is selected to capture global information during the encoding stage. Several renders efficiently map the encoded features of different levels to the original spatial resolution by combining global and local features. Furthermore, the point-based function is employed to supervise the render module generating points, so that TransRender can continuously refine the uncertainty region. We conducted substantial experiments on different stroke lesion segmentation datasets to prove the efficiency of TransRender. Several evaluation metrics illustrate that our method can automatically segment the stroke lesion with relatively high accuracy and low calculation complexity.
Collapse
Affiliation(s)
- Zelin Wu
- College of Electronic Information and Optical Engineering, Taiyuan University of Technology, Taiyuan, China
| | - Xueying Zhang
- College of Electronic Information and Optical Engineering, Taiyuan University of Technology, Taiyuan, China
| | - Fenglian Li
- College of Electronic Information and Optical Engineering, Taiyuan University of Technology, Taiyuan, China
| | - Suzhe Wang
- College of Electronic Information and Optical Engineering, Taiyuan University of Technology, Taiyuan, China
| | - Jiaying Li
- The First Clinical Medical College, Shanxi Medical University, Taiyuan, China
| |
Collapse
|
11
|
Ahmed R, Al Shehhi A, Hassan B, Werghi N, Seghier ML. An appraisal of the performance of AI tools for chronic stroke lesion segmentation. Comput Biol Med 2023; 164:107302. [PMID: 37572443 DOI: 10.1016/j.compbiomed.2023.107302] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/19/2023] [Revised: 07/18/2023] [Accepted: 07/28/2023] [Indexed: 08/14/2023]
Abstract
Automated demarcation of stoke lesions from monospectral magnetic resonance imaging scans is extremely useful for diverse research and clinical applications, including lesion-symptom mapping to explain deficits and predict recovery. There is a significant surge of interest in the development of supervised artificial intelligence (AI) methods for that purpose, including deep learning, with a performance comparable to trained experts. Such AI-based methods, however, require copious amounts of data. Thanks to the availability of large datasets, the development of AI-based methods for lesion segmentation has immensely accelerated in the last decade. One of these datasets is the Anatomical Tracings of Lesions After Stroke (ATLAS) dataset which includes T1-weighted images from hundreds of chronic stroke survivors with their manually traced lesions. This systematic review offers an appraisal of the impact of the ATLAS dataset in promoting the development of AI-based segmentation of stroke lesions. An examination of all published studies, that used the ATLAS dataset to both train and test their methods, highlighted an overall moderate performance (median Dice index = 59.40%) and a huge variability across studies in terms of data preprocessing, data augmentation, AI architecture, and the mode of operation (two-dimensional versus three-dimensional methods). Perhaps most importantly, almost all AI tools were borrowed from existing AI architectures in computer vision, as 90% of all selected studies relied on conventional convolutional neural network-based architectures. Overall, current research has not led to the development of robust AI architectures than can handle spatially heterogenous lesion patterns. This review also highlights the difficulty of gauging the performance of AI tools in the presence of uncertainties in the definition of the ground truth.
Collapse
Affiliation(s)
- Ramsha Ahmed
- Department of Biomedical Engineering, Khalifa University of Science and Technology, Abu Dhabi, United Arab Emirates
| | - Aamna Al Shehhi
- Department of Biomedical Engineering, Khalifa University of Science and Technology, Abu Dhabi, United Arab Emirates; Healthcare Engineering Innovation Center (HEIC), Khalifa University of Science and Technology, Abu Dhabi, United Arab Emirates
| | - Bilal Hassan
- Department of Electrical Engineering and Computer Science, Khalifa University of Science and Technology, Abu Dhabi, United Arab Emirates
| | - Naoufel Werghi
- Healthcare Engineering Innovation Center (HEIC), Khalifa University of Science and Technology, Abu Dhabi, United Arab Emirates; Department of Electrical Engineering and Computer Science, Khalifa University of Science and Technology, Abu Dhabi, United Arab Emirates
| | - Mohamed L Seghier
- Department of Biomedical Engineering, Khalifa University of Science and Technology, Abu Dhabi, United Arab Emirates; Healthcare Engineering Innovation Center (HEIC), Khalifa University of Science and Technology, Abu Dhabi, United Arab Emirates.
| |
Collapse
|
12
|
Liu L, Chang J, Liu Z, Zhang P, Xu X, Shang H. Hybrid Contextual Semantic Network for Accurate Segmentation and Detection of Small-Size Stroke Lesions From MRI. IEEE J Biomed Health Inform 2023; 27:4062-4073. [PMID: 37155390 DOI: 10.1109/jbhi.2023.3273771] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/10/2023]
Abstract
Stroke is a cerebrovascular disease with high mortality and disability rates. The occurrence of the stroke typically produces lesions of different sizes, with the accurate segmentation and detection of small-size stroke lesions being closely related to the prognosis of patients. However, the large lesions are usually correctly identified, the small-size lesions are usually ignored. This article provides a hybrid contextual semantic network (HCSNet) that can accurately and simultaneously segment and detect small-size stroke lesions from magnetic resonance images. HCSNet inherits the advantages of the encoder-decoder architecture and applies a novel hybrid contextual semantic module that generates high-quality contextual semantic features from the spatial and channel contextual semantic features through the skip connection layer. Moreover, a mixing-loss function is proposed to optimize HCSNet for unbalanced small-size lesions. HCSNet is trained and evaluated on 2D magnetic resonance images produced from the Anatomical Tracings of Lesions After Stroke challenge (ATLAS R2.0). Extensive experiments demonstrate that HCSNet outperforms several other state-of-the-art methods in its ability to segment and detect small-size stroke lesions. Visualization and ablation experiments reveal that the hybrid semantic module improves the segmentation and detection performance of HCSNet.
Collapse
|
13
|
Liu L, Chang J, Liang G, Xiong S. Simulated Quantum Mechanics-Based Joint Learning Network for Stroke Lesion Segmentation and TICI Grading. IEEE J Biomed Health Inform 2023; 27:3372-3383. [PMID: 37104101 DOI: 10.1109/jbhi.2023.3270861] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/28/2023]
Abstract
Segmenting stroke lesions and assessing the thrombolysis in cerebral infarction (TICI) grade are two important but challenging prerequisites for an auxiliary diagnosis of the stroke. However, most previous studies have focused only on a single one of two tasks, without considering the relation between them. In our study, we propose a simulated quantum mechanics-based joint learning network (SQMLP-net) that simultaneously segments a stroke lesion and assesses the TICI grade. The correlation and heterogeneity between the two tasks are tackled with a single-input double-output hybrid network. SQMLP-net has a segmentation branch and a classification branch. These two branches share an encoder, which extracts and shares the spatial and global semantic information for the segmentation and classification tasks. Both tasks are optimized by a novel joint loss function that learns the intra- and inter-task weights between these two tasks. Finally, we evaluate SQMLP-net with a public stroke dataset (ATLAS R2.0). SQMLP-net obtains state-of-the-art metrics (Dice:70.98% and accuracy:86.78%) and outperforms single-task and existing advanced methods. An analysis found a negative correlation between the severity of TICI grading and the accuracy of stroke lesion segmentation.
Collapse
|
14
|
Yu W, Huang Z, Zhang J, Shan H. SAN-Net: Learning generalization to unseen sites for stroke lesion segmentation with self-adaptive normalization. Comput Biol Med 2023; 156:106717. [PMID: 36878125 DOI: 10.1016/j.compbiomed.2023.106717] [Citation(s) in RCA: 9] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/25/2022] [Revised: 01/31/2023] [Accepted: 02/26/2023] [Indexed: 03/06/2023]
Abstract
There are considerable interests in automatic stroke lesion segmentation on magnetic resonance (MR) images in the medical imaging field, as stroke is an important cerebrovascular disease. Although deep learning-based models have been proposed for this task, generalizing these models to unseen sites is difficult due to not only the large inter-site discrepancy among different scanners, imaging protocols, and populations, but also the variations in stroke lesion shape, size, and location. To tackle this issue, we introduce a self-adaptive normalization network, termed SAN-Net, to achieve adaptive generalization on unseen sites for stroke lesion segmentation. Motivated by traditional z-score normalization and dynamic network, we devise a masked adaptive instance normalization (MAIN) to minimize inter-site discrepancies, which standardizes input MR images from different sites into a site-unrelated style by dynamically learning affine parameters from the input; i.e., MAIN can affinely transform the intensity values. Then, we leverage a gradient reversal layer to force the U-net encoder to learn site-invariant representation with a site classifier, which further improves the model generalization in conjunction with MAIN. Finally, inspired by the "pseudosymmetry" of the human brain, we introduce a simple yet effective data augmentation technique, termed symmetry-inspired data augmentation (SIDA), that can be embedded within SAN-Net to double the sample size while halving memory consumption. Experimental results on the benchmark Anatomical Tracings of Lesions After Stroke (ATLAS) v1.2 dataset, which includes MR images from 9 different sites, demonstrate that under the "leave-one-site-out" setting, the proposed SAN-Net outperforms recently published methods in terms of quantitative metrics and qualitative comparisons.
Collapse
Affiliation(s)
- Weiyi Yu
- Institute of Science and Technology for Brain-inspired Intelligence and MOE Frontiers Center for Brain Science, Fudan University, Shanghai, 200433, China
| | - Zhizhong Huang
- Shanghai Key Lab of Intelligent Information Processing and the School of Computer Science, Fudan University, Shanghai 200433, China
| | - Junping Zhang
- Shanghai Key Lab of Intelligent Information Processing and the School of Computer Science, Fudan University, Shanghai 200433, China
| | - Hongming Shan
- Institute of Science and Technology for Brain-inspired Intelligence and MOE Frontiers Center for Brain Science, Fudan University, Shanghai, 200433, China; Shanghai Center for Brain Science and Brain-inspired Technology, Shanghai 201210, China.
| |
Collapse
|
15
|
Szeskin A, Rochman S, Weiss S, Lederman R, Sosna J, Joskowicz L. Liver lesion changes analysis in longitudinal CECT scans by simultaneous deep learning voxel classification with SimU-Net. Med Image Anal 2023; 83:102675. [PMID: 36334393 DOI: 10.1016/j.media.2022.102675] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/24/2022] [Revised: 07/28/2022] [Accepted: 10/27/2022] [Indexed: 11/05/2022]
Abstract
The identification and quantification of liver lesions changes in longitudinal contrast enhanced CT (CECT) scans is required to evaluate disease status and to determine treatment efficacy in support of clinical decision-making. This paper describes a fully automatic end-to-end pipeline for liver lesion changes analysis in consecutive (prior and current) abdominal CECT scans of oncology patients. The three key novelties are: (1) SimU-Net, a simultaneous multi-channel 3D R2U-Net model trained on pairs of registered scans of each patient that identifies the liver lesions and their changes based on the lesion and healthy tissue appearance differences; (2) a model-based bipartite graph lesions matching method for the analysis of lesion changes at the lesion level; (3) a method for longitudinal analysis of one or more of consecutive scans of a patient based on SimU-Net that handles major liver deformations and incorporates lesion segmentations from previous analysis. To validate our methods, five experimental studies were conducted on a unique dataset of 3491 liver lesions in 735 pairs from 218 clinical abdominal CECT scans of 71 patients with metastatic disease manually delineated by an expert radiologist. The pipeline with the SimU-Net model, trained and validated on 385 pairs and tested on 249 pairs, yields a mean lesion detection recall of 0.86±0.14, a precision of 0.74±0.23 and a lesion segmentation Dice of 0.82±0.14 for lesions > 5 mm. This outperforms a reference standalone 3D R2-UNet mdel that analyzes each scan individually by ∼50% in precision with similar recall and Dice score on the same training and test datasets. For lesions matching, the precision is 0.86±0.18 and the recall is 0.90±0.15. For lesion classification, the specificity is 0.97±0.07, the precision is 0.85±0.31, and the recall is 0.86±0.23. Our new methods provide accurate and comprehensive results that may help reduce radiologists' time and effort and improve radiological oncology evaluation.
Collapse
Affiliation(s)
- Adi Szeskin
- The Rachel and Selim Benin School of Computer Science and Engineering, The Hebrew University of Jerusalem, Edmond J. Safra Campus, Givat Ram, Jerusalem 9190401, Israel; The Alexander Grass Center for Bioengineering, The Hebrew University of Jerusalem, Israel
| | - Shalom Rochman
- The Rachel and Selim Benin School of Computer Science and Engineering, The Hebrew University of Jerusalem, Edmond J. Safra Campus, Givat Ram, Jerusalem 9190401, Israel
| | - Snir Weiss
- The Rachel and Selim Benin School of Computer Science and Engineering, The Hebrew University of Jerusalem, Edmond J. Safra Campus, Givat Ram, Jerusalem 9190401, Israel
| | - Richard Lederman
- Department of Radiology, Hadassah Hebrew University Medical Center, Jerusalem, Israel
| | - Jacob Sosna
- Department of Radiology, Hadassah Hebrew University Medical Center, Jerusalem, Israel
| | - Leo Joskowicz
- The Rachel and Selim Benin School of Computer Science and Engineering, The Hebrew University of Jerusalem, Edmond J. Safra Campus, Givat Ram, Jerusalem 9190401, Israel.
| |
Collapse
|
16
|
Krithika alias AnbuDevi M, Suganthi K. Review of Semantic Segmentation of Medical Images Using Modified Architectures of UNET. Diagnostics (Basel) 2022; 12:diagnostics12123064. [PMID: 36553071 PMCID: PMC9777361 DOI: 10.3390/diagnostics12123064] [Citation(s) in RCA: 23] [Impact Index Per Article: 7.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/08/2022] [Revised: 11/17/2022] [Accepted: 11/22/2022] [Indexed: 12/12/2022] Open
Abstract
In biomedical image analysis, information about the location and appearance of tumors and lesions is indispensable to aid doctors in treating and identifying the severity of diseases. Therefore, it is essential to segment the tumors and lesions. MRI, CT, PET, ultrasound, and X-ray are the different imaging systems to obtain this information. The well-known semantic segmentation technique is used in medical image analysis to identify and label regions of images. The semantic segmentation aims to divide the images into regions with comparable characteristics, including intensity, homogeneity, and texture. UNET is the deep learning network that segments the critical features. However, UNETs basic architecture cannot accurately segment complex MRI images. This review introduces the modified and improved models of UNET suitable for increasing segmentation accuracy.
Collapse
|
17
|
Zhao D, Ge W, Chen P, Hu Y, Dang Y, Liang R, Guo X. Feature Pyramid U-Net with Attention for Semantic Segmentation of Forward-Looking Sonar Images. SENSORS (BASEL, SWITZERLAND) 2022; 22:s22218468. [PMID: 36366169 PMCID: PMC9653894 DOI: 10.3390/s22218468] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/15/2022] [Revised: 11/01/2022] [Accepted: 11/02/2022] [Indexed: 06/12/2023]
Abstract
Forward-looking sonar is a technique widely used for underwater detection. However, most sonar images have underwater noise and low resolution due to their acoustic properties. In recent years, the semantic segmentation model U-Net has shown excellent segmentation performance, and it has great potential in forward-looking sonar image segmentation. However, forward-looking sonar images are affected by noise, which prevents the existing U-Net model from segmenting small objects effectively. Therefore, this study presents a forward-looking sonar semantic segmentation model called Feature Pyramid U-Net with Attention (FPUA). This model uses residual blocks to improve the training depth of the network. To improve the segmentation accuracy of the network for small objects, a feature pyramid module combined with an attention structure is introduced. This improves the model's ability to learn deep semantic and shallow detail information. First, the proposed model is compared against other deep learning models and on two datasets, of which one was collected in a tank environment and the other was collected in a real marine environment. To further test the validity of the model, a real forward-looking sonar system was devised and employed in the lake trials. The results show that the proposed model performs better than the other models for small-object and few-sample classes and that it is competitive in semantic segmentation of forward-looking sonar images.
Collapse
Affiliation(s)
- Dongdong Zhao
- The College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou 310023, China
| | - Weihao Ge
- The College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou 310023, China
| | - Peng Chen
- The College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou 310023, China
| | - Yingtian Hu
- The College of Information Engineering, Zhejiang University of Technology, Hangzhou 310023, China
| | - Yuanjie Dang
- The College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou 310023, China
| | - Ronghua Liang
- The College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou 310023, China
| | - Xinxin Guo
- The Institute of Deep-Sea Science and Engineering Chinese Academy of Sciences, Sanya 572000, China
| |
Collapse
|
18
|
Liu L, Zhang P, Liang G, Xiong S, Wang J, Zheng G. A spatiotemporal correlation deep learning network for brain penumbra disease. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.11.041] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/18/2022]
|
19
|
Automatic Segmentation and Quantitative Assessment of Stroke Lesions on MR Images. Diagnostics (Basel) 2022; 12:diagnostics12092055. [PMID: 36140457 PMCID: PMC9497525 DOI: 10.3390/diagnostics12092055] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2022] [Revised: 08/12/2022] [Accepted: 08/22/2022] [Indexed: 12/20/2022] Open
Abstract
Lesion studies are crucial in establishing brain-behavior relationships, and accurately segmenting the lesion represents the first step in achieving this. Manual lesion segmentation is the gold standard for chronic strokes. However, it is labor-intensive, subject to bias, and limits sample size. Therefore, our objective is to develop an automatic segmentation algorithm for chronic stroke lesions on T1-weighted MR images. Methods: To train our model, we utilized an open-source dataset: ATLAS v2.0 (Anatomical Tracings of Lesions After Stroke). We partitioned the dataset of 655 T1 images with manual segmentation labels into five subsets and performed a 5-fold cross-validation to avoid overfitting of the model. We used a deep neural network (DNN) architecture for model training. Results: To evaluate the model performance, we used three metrics that pertain to diverse aspects of volumetric segmentation, including shape, location, and size. The Dice similarity coefficient (DSC) compares the spatial overlap between manual and machine segmentation. The average DSC was 0.65 (0.61−0.67; 95% bootstrapped CI). Average symmetric surface distance (ASSD) measures contour distances between the two segmentations. ASSD between manual and automatic segmentation was 12 mm. Finally, we compared the total lesion volumes and the Pearson correlation coefficient (ρ) between the manual and automatically segmented lesion volumes, which was 0.97 (p-value < 0.001). Conclusions: We present the first automated segmentation model trained on a large multicentric dataset. This model will enable automated on-demand processing of MRI scans and quantitative chronic stroke lesion assessment.
Collapse
|
20
|
Liu X, Pang Y, Jin R, Liu Y, Wang Z. Dual-Domain Reconstruction Network with V-Net and K-Net for Fast MRI. Magn Reson Med 2022; 88:2694-2708. [PMID: 35942977 DOI: 10.1002/mrm.29400] [Citation(s) in RCA: 15] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/23/2022] [Revised: 07/05/2022] [Accepted: 07/08/2022] [Indexed: 11/10/2022]
Abstract
PURPOSE To introduce a dual-domain reconstruction network with V-Net and K-Net for accurate MR image reconstruction from undersampled k-space data. METHODS Most state-of-the-art reconstruction methods apply U-Net or cascaded U-Nets in the image domain and/or k-space domain. Nevertheless, these methods have the following problems: (1) directly applying U-Net in the k-space domain is not optimal for extracting features; (2) classical image-domain-oriented U-Net is heavyweighted and hence inefficient when cascaded many times to yield good reconstruction accuracy; (3) classical image-domain-oriented U-Net does not make full use of information of the encoder network for extracting features in the decoder network; and (4) existing methods are ineffective in simultaneously extracting and fusing features in the image domain and its dual k-space domain. To tackle these problems, we present 3 different methods: (1) V-Net, an image-domain encoder-decoder subnetwork that is more lightweight for cascading and effective in fully utilizing features in the encoder for decoding; (2) K-Net, a k-space domain subnetwork that is more suitable for extracting hierarchical features in the k-space domain, and (3) KV-Net, a dual-domain reconstruction network in which V-Nets and K-Nets are effectively combined and cascaded. RESULTS Extensive experimental results on the fastMRI dataset demonstrate that the proposed KV-Net can reconstruct high-quality images and outperform state-of-the-art approaches with fewer parameters. CONCLUSIONS To reconstruct images effectively and efficiently from incomplete k-space data, we have presented a dual-domain KV-Net to combine K-Nets and V-Nets. The KV-Net achieves better results with 9% and 5% parameters than comparable methods (XPD-Net and i-RIM).
Collapse
Affiliation(s)
- Xiaohan Liu
- Tianjin Key Lab. of Brain Inspired Intelligence Technology, School of Electrical and Information Engineering, Tianjin University, Tianjin, People's Republic of China
| | - Yanwei Pang
- Tianjin Key Lab. of Brain Inspired Intelligence Technology, School of Electrical and Information Engineering, Tianjin University, Tianjin, People's Republic of China
| | - Ruiqi Jin
- Tianjin Key Lab. of Brain Inspired Intelligence Technology, School of Electrical and Information Engineering, Tianjin University, Tianjin, People's Republic of China
| | - Yu Liu
- Tianjin Key Lab. of Brain Inspired Intelligence Technology, School of Electrical and Information Engineering, Tianjin University, Tianjin, People's Republic of China
| | - Zhenchang Wang
- Beijing Friendship Hospital, Capital Medical University, Beijing, People's Republic of China
| |
Collapse
|
21
|
SIP-UNet: Sequential Inputs Parallel UNet Architecture for Segmentation of Brain Tissues from Magnetic Resonance Images. MATHEMATICS 2022. [DOI: 10.3390/math10152755] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/04/2023]
Abstract
Proper analysis of changes in brain structure can lead to a more accurate diagnosis of specific brain disorders. The accuracy of segmentation is crucial for quantifying changes in brain structure. In recent studies, UNet-based architectures have outperformed other deep learning architectures in biomedical image segmentation. However, improving segmentation accuracy is challenging due to the low resolution of medical images and insufficient data. In this study, we present a novel architecture that combines three parallel UNets using a residual network. This architecture improves upon the baseline methods in three ways. First, instead of using a single image as input, we use three consecutive images. This gives our model the freedom to learn from neighboring images as well. Additionally, the images are individually compressed and decompressed using three different UNets, which prevents the model from merging the features of the images. Finally, following the residual network architecture, the outputs of the UNets are combined in such a way that the features of the image corresponding to the output are enhanced by a skip connection. The proposed architecture performed better than using a single conventional UNet and other UNet variants.
Collapse
|
22
|
Zhao Q, Jia Q, Chi T. Deep learning as a novel method for endoscopic diagnosis of chronic atrophic gastritis: a prospective nested case-control study. BMC Gastroenterol 2022; 22:352. [PMID: 35879649 PMCID: PMC9310473 DOI: 10.1186/s12876-022-02427-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 05/15/2022] [Accepted: 07/15/2022] [Indexed: 11/28/2022] Open
Abstract
Background and aims Chronic atrophic gastritis (CAG) is a precancerous disease that often leads to the development of gastric cancer (GC) and is positively correlated with GC morbidity. However, the sensitivity of the endoscopic diagnosis of CAG is only 42%. Therefore, we developed a real-time video monitoring model for endoscopic diagnosis of CAG based on U-Net deep learning (DL) and conducted a prospective nested case–control study to evaluate the diagnostic evaluation indices of the model and its consistency with pathological diagnosis.
Methods Our cohort consisted of 1539 patients undergoing gastroscopy from December 1, 2020, to July 1, 2021. Based on pathological diagnosis, patients in the cohort were divided into the CAG group or the chronic nonatrophic gastritis (CNAG) group, and we assessed the diagnostic evaluation indices of this model and its consistency with pathological diagnosis after propensity score matching (PSM) to minimize selection bias in the study. Results After matching, the diagnostic evaluation indices and consistency evaluation of the model were better than those of endoscopists [sensitivity (84.02% vs. 62.72%), specificity (97.04% vs. 81.95%), positive predictive value (96.60% vs. 77.66%), negative predictive value (85.86% vs. 68.73%), accuracy rate (90.53% vs. 72.34%), Youden index (81.06% vs. 44.67%), odd product (172.5 vs. 7.64), positive likelihood ratio (28.39 vs. 3.47), negative likelihood ratio (0.16 vs. 0.45), AUC (95% CI) [0.909 (0.884–0.934) vs. 0.740 (0.702–0.778)] and Kappa (0.852 vs. 0.558)]. Conclusions Our prospective nested case–control study proved that the diagnostic evaluation indices and consistency evaluation of the real-time video monitoring model for endoscopic diagnosis of CAG based on U-Net DL were superior to those of endoscopists. Trial registrationChiCTR2100044458, 18/03/2020.
Collapse
Affiliation(s)
- Quchuan Zhao
- Department of Gastroenterology, Xuanwu Hospital of Capital Medical University, 45 Chang-chun Street, Beijing, 100053, China
| | - Qing Jia
- Department of Anesthesiology, Guang'anmen Hospital China Academy of Chinese Medical Sciences, 5 North Court Street, Beijing, 100053, China.
| | - Tianyu Chi
- Department of Gastroenterology, Xuanwu Hospital of Capital Medical University, 45 Chang-chun Street, Beijing, 100053, China.
| |
Collapse
|
23
|
Huang W, Wang X, Huang Y, Lin F, Tang X. Multi-parametric Magnetic Resonance Imaging Fusion for Automatic Classification of Prostate Cancer. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2022; 2022:471-474. [PMID: 36085623 DOI: 10.1109/embc48229.2022.9871334] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Computer-aided diagnosis (CAD) of prostate cancer (PCa) using multi-parametric magnetic resonance imaging (mp-MRI) has recently gained great research interest. In this work, a fully automatic CAD pipeline of PCa using mp-MRI data is presented. In order to fully explore the mp-MRI data, we systematically investigate three multi-modal medical image fusion strategies in convolutional neural networks, namely input-level fusion, feature-level fusion, and decision-level fusion. Extensive experiments are conducted on two datasets with different PCa-related diagnostic tasks. We identify a pipeline that works relatively the best for both diagnostic tasks, two important components of which are stacking three adjacent slices as the input and performing decision-level fusion with specific loss weights. Clinical relevance- This work provides a practical method for automated diagnosis of PCa based on multi-parametric MRI.
Collapse
|
24
|
Liew SL, Lo BP, Donnelly MR, Zavaliangos-Petropulu A, Jeong JN, Barisano G, Hutton A, Simon JP, Juliano JM, Suri A, Wang Z, Abdullah A, Kim J, Ard T, Banaj N, Borich MR, Boyd LA, Brodtmann A, Buetefisch CM, Cao L, Cassidy JM, Ciullo V, Conforto AB, Cramer SC, Dacosta-Aguayo R, de la Rosa E, Domin M, Dula AN, Feng W, Franco AR, Geranmayeh F, Gramfort A, Gregory CM, Hanlon CA, Hordacre BG, Kautz SA, Khlif MS, Kim H, Kirschke JS, Liu J, Lotze M, MacIntosh BJ, Mataró M, Mohamed FB, Nordvik JE, Park G, Pienta A, Piras F, Redman SM, Revill KP, Reyes M, Robertson AD, Seo NJ, Soekadar SR, Spalletta G, Sweet A, Telenczuk M, Thielman G, Westlye LT, Winstein CJ, Wittenberg GF, Wong KA, Yu C. A large, curated, open-source stroke neuroimaging dataset to improve lesion segmentation algorithms. Sci Data 2022; 9:320. [PMID: 35710678 PMCID: PMC9203460 DOI: 10.1038/s41597-022-01401-7] [Citation(s) in RCA: 38] [Impact Index Per Article: 12.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/20/2021] [Accepted: 05/19/2022] [Indexed: 01/16/2023] Open
Abstract
Accurate lesion segmentation is critical in stroke rehabilitation research for the quantification of lesion burden and accurate image processing. Current automated lesion segmentation methods for T1-weighted (T1w) MRIs, commonly used in stroke research, lack accuracy and reliability. Manual segmentation remains the gold standard, but it is time-consuming, subjective, and requires neuroanatomical expertise. We previously released an open-source dataset of stroke T1w MRIs and manually-segmented lesion masks (ATLAS v1.2, N = 304) to encourage the development of better algorithms. However, many methods developed with ATLAS v1.2 report low accuracy, are not publicly accessible or are improperly validated, limiting their utility to the field. Here we present ATLAS v2.0 (N = 1271), a larger dataset of T1w MRIs and manually segmented lesion masks that includes training (n = 655), test (hidden masks, n = 300), and generalizability (hidden MRIs and masks, n = 316) datasets. Algorithm development using this larger sample should lead to more robust solutions; the hidden datasets allow for unbiased performance evaluation via segmentation challenges. We anticipate that ATLAS v2.0 will lead to improved algorithms, facilitating large-scale stroke research.
Collapse
Affiliation(s)
- Sook-Lei Liew
- Chan Division of Occupational Science and Occupational Therapy, University of Southern California, Los Angeles, CA, USA.
- Mark and Mary Stevens Neuroimaging and Informatics Institute, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA.
| | - Bethany P Lo
- Chan Division of Occupational Science and Occupational Therapy, University of Southern California, Los Angeles, CA, USA
| | - Miranda R Donnelly
- Chan Division of Occupational Science and Occupational Therapy, University of Southern California, Los Angeles, CA, USA
| | - Artemis Zavaliangos-Petropulu
- Mark and Mary Stevens Neuroimaging and Informatics Institute, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
| | - Jessica N Jeong
- Chan Division of Occupational Science and Occupational Therapy, University of Southern California, Los Angeles, CA, USA
| | - Giuseppe Barisano
- Laboratory of Neuroimaging, Mark and Mary Stevens Neuroimaging and Informatics Institutes, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
- Neuroscience Graduate Program, University of Southern California, Los Angeles, CA, USA
| | - Alexandre Hutton
- Chan Division of Occupational Science and Occupational Therapy, University of Southern California, Los Angeles, CA, USA
| | - Julia P Simon
- Mark and Mary Stevens Neuroimaging and Informatics Institute, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
| | - Julia M Juliano
- Neuroscience Graduate Program, University of Southern California, Los Angeles, CA, USA
| | - Anisha Suri
- Electrical and Computer Engineering, Swanson School of Engineering, University of Pittsburgh, Pittsburgh, PA, USA
| | - Zhizhuo Wang
- Chan Division of Occupational Science and Occupational Therapy, University of Southern California, Los Angeles, CA, USA
| | - Aisha Abdullah
- Chan Division of Occupational Science and Occupational Therapy, University of Southern California, Los Angeles, CA, USA
| | - Jun Kim
- Mark and Mary Stevens Neuroimaging and Informatics Institute, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
| | - Tyler Ard
- Mark and Mary Stevens Neuroimaging and Informatics Institute, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
| | - Nerisa Banaj
- Laboratory of Neuropsychiatry, IRCCS Santa Lucia Foundation, Rome, Italy
| | - Michael R Borich
- Department of Rehabilitation Medicine, Emory University School of Medicine, Atlanta, GA, USA
| | - Lara A Boyd
- Department of Physical Therapy & Djavad Mowafaghian Centre for Brain Health, University of British Columbia, Vancouver, British Columbia, Canada
| | - Amy Brodtmann
- Florey Institute of Neuroscience and Mental Health, University of Melbourne, Melbourne, Victoria, Australia
| | - Cathrin M Buetefisch
- Department of Rehabilitation Medicine, Emory University School of Medicine, Atlanta, GA, USA
- Department of Neurology, Emory University, Atlanta, GA, USA
| | - Lei Cao
- Center for the Developing Brain, Child Mind Institute, New York, NY, USA
| | - Jessica M Cassidy
- Department of Allied Health Sciences, University of North Carolina at Chapel Hill, Chapel Hill, NC, USA
| | - Valentina Ciullo
- Laboratory of Neuropsychiatry, IRCCS Santa Lucia Foundation, Rome, Italy
| | - Adriana B Conforto
- Hospital das Clínicas, São Paulo University, Sao Paulo, SP, Brazil
- Hospital Israelita Albert Einstein, Sao Paulo, SP, Brazil
| | - Steven C Cramer
- Department of Neurology, University of California Los Angeles and California Rehabilitation Institute, Los Angeles, CA, USA
| | - Rosalia Dacosta-Aguayo
- Department of Psychiatry and Clinical Psychobiology, University of Barcelona, Barcelona, Spain
| | - Ezequiel de la Rosa
- icometrix, Leuven, Belgium
- Department of Computer Science, Technical University of Munich, Munich, Germany
| | - Martin Domin
- Functional Imaging Unit, Department of Diagnostic Radiology and Neuroradiology, University of Greifswald, Greifswald, Germany
| | - Adrienne N Dula
- Departments of Neurology and Diagnostic Medicine, Dell Medical School at The University of Texas Austin, Austin, TX, USA
| | - Wuwei Feng
- Department of Neurology, Duke University School of Medicine, Durham, NC, USA
| | - Alexandre R Franco
- Center for the Developing Brain, Child Mind Institute, New York, NY, USA
- Center for Biomedical Imaging and Neuromodulation, Nathan Kline Institute for Psychiatric Research, Orangeburg, NY, USA
- Department of Psychiatry, NYU Grossman School of Medicine, New York, NY, USA
| | | | - Alexandre Gramfort
- Center for Data Science, Université Paris-Saclay, Inria, Palaiseau, France
| | - Chris M Gregory
- Department of Health Sciences & Research, Medical University of South Carolina, Charleston, SC, USA
| | - Colleen A Hanlon
- Cancer Biology, Wake Forest School of Medicine, Winston Salem, NC, USA
| | - Brenton G Hordacre
- Innovation, Implementation and Clinical Translation (IIMPACT) in Health, Allied Health and Human Performance, University of South Australia, Adelaide, South Australia, Australia
| | - Steven A Kautz
- Department of Health Sciences & Research, Medical University of South Carolina, Charleston, SC, USA
- Ralph H Johnson VA Medical Center, Charleston, SC, USA
| | - Mohamed Salah Khlif
- The Florey Institute of Neuroscience and Mental Health, Heidelberg, VIC, Australia
| | - Hosung Kim
- Mark and Mary Stevens Neuroimaging and Informatics Institute, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
| | - Jan S Kirschke
- Neuroradiology, School of Medicine, Technical University Munich, München, Germany
| | - Jingchun Liu
- Department of Radiology, Tianjin Medical University General Hospital, Tianjin, China
| | - Martin Lotze
- Functional Imaging Unit, Department of Diagnostic Radiology and Neuroradiology, University of Greifswald, Greifswald, Germany
| | - Bradley J MacIntosh
- Department of Medical Biophysics, University of Toronto, Toronto, Ontario, Canada
- Hurvitz Brain Sciences Program, Toronto, Ontario, Canada
| | - Maria Mataró
- Department of Clinical Psychology and Psychobiology, Institut de Neurociències, Universitat de Barcelona, Barcelona, Spain
- Institut de Recerca Sant Joan de Déu, 08950, Esplugues de Llobregat, Spain
| | - Feroze B Mohamed
- Jefferson Magnetic Resonance Imaging Center, Philadelphia, PA, USA
| | - Jan E Nordvik
- CatoSenteret Rehabilitation Center, SON, Norway
- Faculty of Health Sciences, Oslo Metropolitan University, Oslo, Norway
| | - Gilsoon Park
- Laboratory of Neuroimaging, Mark and Mary Stevens Neuroimaging and Informatics Institutes, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
| | - Amy Pienta
- Inter-university Consortium for Political and Social Research, University of Michigan, Ann Arbor, MI, USA
| | - Fabrizio Piras
- Laboratory of Neuropsychiatry, IRCCS Santa Lucia Foundation, Rome, Italy
| | - Shane M Redman
- Inter-university Consortium for Political and Social Research, University of Michigan, Ann Arbor, MI, USA
| | - Kate P Revill
- Facility for Education and Research in Neuroscience, Emory University, Atlanta, GA, USA
| | - Mauricio Reyes
- ARTORG Center for Biomedical Engineering Research, University of Bern, Bern, Switzerland
| | - Andrew D Robertson
- Schlegel-University of Waterloo Research Institute for Aging, University of Waterloo, Waterloo, Ontario, Canada
- Canadian Partnership for Stroke Recovery, Sunnybrook Research Institute, Toronto, Ontario, Canada
| | - Na Jin Seo
- Department of Health Sciences & Research, Medical University of South Carolina, Charleston, SC, USA
- Ralph H Johnson VA Medical Center, Charleston, SC, USA
- Department of Rehabilitation Sciences, Medical University of South Carolina, Charleston, SC, USA
| | - Surjo R Soekadar
- Clinical Neurotechnology Laboratory, Dept. of Psychiatry and Neurosciences (CCM), Charité - Universitätsmedizin Berlin, Berlin, Germany
| | - Gianfranco Spalletta
- Laboratory of Neuropsychiatry, IRCCS Santa Lucia Foundation, Rome, Italy
- Menninger Department of Psychiatry and Behavioral Sciences, Division of Neuropsychiatry, Baylor College of Medicine, Houston, TX, USA
| | - Alison Sweet
- Inter-university Consortium for Political and Social Research, University of Michigan, Ann Arbor, MI, USA
| | - Maria Telenczuk
- Center for Data Science, Université Paris-Saclay, Inria, Palaiseau, France
| | - Gregory Thielman
- Department of Physical Therapy and Neuroscience, Samson College of Health Sciences, St. Joseph's University, Philadelphia, PA, USA
| | - Lars T Westlye
- Department of Psychology, University of Oslo, Oslo, Norway
- NORMENT, Department of Mental Health and Addiction, Oslo University Hospital, Oslo, Norway
| | - Carolee J Winstein
- Division of Biokinesiology and Physical Therapy of the Herman Ostrow School of Dentistry, University of Southern California, Los Angeles, CA, USA
- Department of Neurology, Keck School of Medicine, University of Southern California, Los Angeles, CA, USA
| | - George F Wittenberg
- Geriatrics Research, Education and Clinical Center, HERL, Department of Veterans Affairs, Pittsburgh, PA, USA
- Departments of Neurology, PM&R, RNEL, CNBC, University of Pittsburgh, Pittsburgh, PA, USA
| | - Kristin A Wong
- Department of Physical Medicine & Rehabilitation, Dell Medical School, University of Texas at Austin, Austin, TX, USA
| | - Chunshui Yu
- Department of Radiology, Tianjin Medical University General Hospital, Tianjin, China
- Tianjin Key Laboratory of Functional Imaging, Tianjin Medical University General Hospital, Tianjin, China
| |
Collapse
|
25
|
Zhang G, Gao X, Zhu Z, Zhou F, Yu D. Determination of the location of the needle entry point based on an improved pruning algorithm. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2022; 19:7952-7977. [PMID: 35801452 DOI: 10.3934/mbe.2022372] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Since the emergence of new coronaviruses and their variant virus, a large number of medical resources around the world have been put into treatment. In this case, the purpose of this article is to develop a handback intravenous intelligence injection robot, which reduces the direct contact between medical staff and patients and reduces the risk of infection. The core technology of hand back intravenous intelligent robot is a handlet venous vessel detection and segmentation and the position of the needle point position decision. In this paper, an image processing algorithm based on U-Net improvement mechanism (AT-U-Net) is proposed for core technology. It is investigated using a self-built dorsal hand vein database and the results show that it performs well, with an F1-score of 93.91%. After the detection of a dorsal hand vein, this paper proposes a location decision method for the needle entry point based on an improved pruning algorithm (PT-Pruning). The extraction of the trunk line of the dorsal hand vein is realized through this algorithm. Considering the vascular cross-sectional area and bending of each vein injection point area, the optimal injection point of the dorsal hand vein is obtained via a comprehensive decision-making process. Using the self-built dorsal hand vein injection point database, the accuracy of the detection of the effective injection area reaches 96.73%. The accuracy for the detection of the injection area at the optimal needle entry point is 96.50%, which lays a foundation for subsequent mechanical automatic injection.
Collapse
Affiliation(s)
- Guangyuan Zhang
- School of Information Science and Electrical Engineering, Shan Dong Jiao Tong University, Jinan 250000, China
| | - Xiaonan Gao
- School of Information Science and Electrical Engineering, Shan Dong Jiao Tong University, Jinan 250000, China
| | - Zhenfang Zhu
- School of Information Science and Electrical Engineering, Shan Dong Jiao Tong University, Jinan 250000, China
| | - Fengyv Zhou
- School of Control Science and Engineering, Shandong University, Jinan 250000, China
| | - Dexin Yu
- Department of Radiology, Qilu Hospital of Shandong University, Jinan 250000, China
| |
Collapse
|
26
|
Wei J, Wu Z, Wang L, Bui TD, Qu L, Yap PT, Xia Y, Li G, Shen D. A cascaded nested network for 3T brain MR image segmentation guided by 7T labeling. PATTERN RECOGNITION 2022; 124:108420. [PMID: 38469076 PMCID: PMC10927017 DOI: 10.1016/j.patcog.2021.108420] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/13/2024]
Abstract
Accurate segmentation of the brain into gray matter, white matter, and cerebrospinal fluid using magnetic resonance (MR) imaging is critical for visualization and quantification of brain anatomy. Compared to 3T MR images, 7T MR images exhibit higher tissue contrast that is contributive to accurate tissue delineation for training segmentation models. In this paper, we propose a cascaded nested network (CaNes-Net) for segmentation of 3T brain MR images, trained by tissue labels delineated from the corresponding 7T images. We first train a nested network (Nes-Net) for a rough segmentation. The second Nes-Net uses tissue-specific geodesic distance maps as contextual information to refine the segmentation. This process is iterated to build CaNes-Net with a cascade of Nes-Net modules to gradually refine the segmentation. To alleviate the misalignment between 3T and corresponding 7T MR images, we incorporate a correlation coefficient map to allow well-aligned voxels to play a more important role in supervising the training process. We compared CaNes-Net with SPM and FSL tools, as well as four deep learning models on 18 adult subjects and the ADNI dataset. Our results indicate that CaNes-Net reduces segmentation errors caused by the misalignment and improves segmentation accuracy substantially over the competing methods.
Collapse
Affiliation(s)
- Jie Wei
- National Engineering Laboratory for Integrated Aero-Space-Ground-Ocean Big Data Application Technology, School of Computer Science and Engineering, Northwestern Polytechnical University, Xi’an 710072, China
- Department of Radiology and BRIC, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
| | - Zhengwang Wu
- Department of Radiology and BRIC, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
| | - Li Wang
- Department of Radiology and BRIC, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
| | - Toan Duc Bui
- Department of Radiology and BRIC, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
| | - Liangqiong Qu
- Department of Biomedical Data Science at Stanford University, Stanford, CA 94305, USA
| | - Pew-Thian Yap
- Department of Radiology and BRIC, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
| | - Yong Xia
- National Engineering Laboratory for Integrated Aero-Space-Ground-Ocean Big Data Application Technology, School of Computer Science and Engineering, Northwestern Polytechnical University, Xi’an 710072, China
| | - Gang Li
- Department of Radiology and BRIC, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA
| | - Dinggang Shen
- School of Biomedical Engineering, ShanghaiTech University, Shanghai 201210, China
- Shanghai United Imaging Intelligence Co., Ltd., Shanghai 200232, China
| |
Collapse
|
27
|
Sheng M, Xu W, Yang J, Chen Z. Cross-Attention and Deep Supervision UNet for Lesion Segmentation of Chronic Stroke. Front Neurosci 2022; 16:836412. [PMID: 35392415 PMCID: PMC8980944 DOI: 10.3389/fnins.2022.836412] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/17/2021] [Accepted: 01/26/2022] [Indexed: 12/04/2022] Open
Abstract
Stroke is an acute cerebrovascular disease with high incidence, high mortality, and high disability rate. Determining the location and volume of the disease in MR images promotes accurate stroke diagnosis and surgical planning. Therefore, the automatic recognition and segmentation of stroke lesions has important clinical significance for large-scale stroke imaging analysis. There are some problems in the segmentation of stroke lesions, such as imbalance of the front and back scenes, uncertainty of position, and unclear boundary. To meet this challenge, this paper proposes a cross-attention and deep supervision UNet (CADS-UNet) to segment chronic stroke lesions from T1-weighted MR images. Specifically, we propose a cross-spatial attention module, which is different from the usual self-attention module. The location information interactively selects encode features and decode features to enrich the lost spatial focus. At the same time, the channel attention mechanism is used to screen the channel characteristics. Finally, combined with deep supervision and mixed loss, the model is supervised more accurately. We compared and verified the model on the authoritative open dataset "Anatomical Tracings of Lesions After Stroke" (Atlas), which fully proved the effectiveness of our model.
Collapse
Affiliation(s)
- Manjin Sheng
- School of Informatics, Xiamen University, Xiamen, China
| | - Wenjie Xu
- School of Informatics, Xiamen University, Xiamen, China
| | - Jane Yang
- Department of Cognitive Science, University of California, San Diego, San Diego, CA, United States
| | - Zhongjie Chen
- Department of Neurology, Zhongshan Hospital, Xiamen University, Xiamen, China
| |
Collapse
|
28
|
Abstract
In magnetic resonance imaging (MRI) segmentation, conventional approaches utilize U-Net models with encoder–decoder structures, segmentation models using vision transformers, or models that combine a vision transformer with an encoder–decoder model structure. However, conventional models have large sizes and slow computation speed and, in vision transformer models, the computation amount sharply increases with the image size. To overcome these problems, this paper proposes a model that combines Swin transformer blocks and a lightweight U-Net type model that has an HarDNet blocks-based encoder–decoder structure. To maintain the features of the hierarchical transformer and shifted-windows approach of the Swin transformer model, the Swin transformer is used in the first skip connection layer of the encoder instead of in the encoder–decoder bottleneck. The proposed model, called STHarDNet, was evaluated by separating the anatomical tracings of lesions after stroke (ATLAS) dataset, which comprises 229 T1-weighted MRI images, into training and validation datasets. It achieved Dice, IoU, precision, and recall values of 0.5547, 0.4185, 0.6764, and 0.5286, respectively, which are better than those of the state-of-the-art models U-Net, SegNet, PSPNet, FCHarDNet, TransHarDNet, Swin Transformer, Swin UNet, X-Net, and D-UNet. Thus, STHarDNet improves the accuracy and speed of MRI image-based stroke diagnosis.
Collapse
|
29
|
Shin H, Agyeman R, Rafiq M, Chang MC, Choi GS. Automated segmentation of chronic stroke lesion using efficient U-Net architecture. Biocybern Biomed Eng 2022; 42:285-294. [DOI: 10.1016/j.bbe.2022.01.002] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/02/2023]
|
30
|
Zhang Y, Zhong P, Jie D, Wu J, Zeng S, Chu J, Liu Y, Wu EX, Tang X. Brain Tumor Segmentation From Multi-Modal MR Images via Ensembling UNets. FRONTIERS IN RADIOLOGY 2021; 1:704888. [PMID: 37492172 PMCID: PMC10365098 DOI: 10.3389/fradi.2021.704888] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/04/2021] [Accepted: 09/27/2021] [Indexed: 07/27/2023]
Abstract
Glioma is a type of severe brain tumor, and its accurate segmentation is useful in surgery planning and progression evaluation. Based on different biological properties, the glioma can be divided into three partially-overlapping regions of interest, including whole tumor (WT), tumor core (TC), and enhancing tumor (ET). Recently, UNet has identified its effectiveness in automatically segmenting brain tumor from multi-modal magnetic resonance (MR) images. In this work, instead of network architecture, we focus on making use of prior knowledge (brain parcellation), training and testing strategy (joint 3D+2D), ensemble and post-processing to improve the brain tumor segmentation performance. We explore the accuracy of three UNets with different inputs, and then ensemble the corresponding three outputs, followed by post-processing to achieve the final segmentation. Similar to most existing works, the first UNet uses 3D patches of multi-modal MR images as the input. The second UNet uses brain parcellation as an additional input. And the third UNet is inputted by 2D slices of multi-modal MR images, brain parcellation, and probability maps of WT, TC, and ET obtained from the second UNet. Then, we sequentially unify the WT segmentation from the third UNet and the fused TC and ET segmentation from the first and the second UNets as the complete tumor segmentation. Finally, we adopt a post-processing strategy by labeling small ET as non-enhancing tumor to correct some false-positive ET segmentation. On one publicly-available challenge validation dataset (BraTS2018), the proposed segmentation pipeline yielded average Dice scores of 91.03/86.44/80.58% and average 95% Hausdorff distances of 3.76/6.73/2.51 mm for WT/TC/ET, exhibiting superior segmentation performance over other state-of-the-art methods. We then evaluated the proposed method on the BraTS2020 training data through five-fold cross validation, with similar performance having also been observed. The proposed method was finally evaluated on 10 in-house data, the effectiveness of which has been established qualitatively by professional radiologists.
Collapse
Affiliation(s)
- Yue Zhang
- Department of Electrical and Electronic Engineering, Southern University of Science and Technology, Shenzhen, China
- Laboratory of Biomedical Imaging and Signal Processing, The University of Hong Kong, Hong Kong, Hong Kong SAR, China
- Department of Electrical and Electronic Engineering, The University of Hong Kong, Hong Kong, Hong Kong SAR, China
- Tencent Music Entertainment, Shenzhen, China
| | - Pinyuan Zhong
- Department of Electrical and Electronic Engineering, Southern University of Science and Technology, Shenzhen, China
| | - Dabin Jie
- Department of Electrical and Electronic Engineering, Southern University of Science and Technology, Shenzhen, China
| | - Jiewei Wu
- School of Electronics and Information Technology, Sun Yat-Sen University, Guangzhou, China
| | - Shanmei Zeng
- Department of Radiology, The First Affiliated Hospital, Sun Yat-Sen University, Guangzhou, China
| | - Jianping Chu
- Department of Radiology, The First Affiliated Hospital, Sun Yat-Sen University, Guangzhou, China
| | - Yilong Liu
- Laboratory of Biomedical Imaging and Signal Processing, The University of Hong Kong, Hong Kong, Hong Kong SAR, China
- Department of Electrical and Electronic Engineering, The University of Hong Kong, Hong Kong, Hong Kong SAR, China
| | - Ed X. Wu
- Laboratory of Biomedical Imaging and Signal Processing, The University of Hong Kong, Hong Kong, Hong Kong SAR, China
- Department of Electrical and Electronic Engineering, The University of Hong Kong, Hong Kong, Hong Kong SAR, China
| | - Xiaoying Tang
- Department of Electrical and Electronic Engineering, Southern University of Science and Technology, Shenzhen, China
| |
Collapse
|
31
|
Bao Q, Mi S, Gang B, Yang W, Chen J, Liao Q. MDAN: Mirror Difference Aware Network for Brain Stroke Lesion Segmentation. IEEE J Biomed Health Inform 2021; 26:1628-1639. [PMID: 34543208 DOI: 10.1109/jbhi.2021.3113460] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Brain stroke lesion segmentation is of great importance for stroke rehabilitation neuroimaging analysis. Due to the large variance of stroke lesion shapes and similarities of tissue intensity distribution, it remains a challenging task. To help detect abnormalities, the anatomical symmetries of brain magnetic resonance (MR) images have been widely used as visual cues for clinical practices. However, most methods do not fully utilize structural symmetry information in brain images segmentation. This paper presents a novel mirror difference aware network (MDAN) for stroke lesion segmentation in an encoder-decoder architecture, aiming at holistically exploiting the symmetries of image features. Specifically, a differential feature augmentation (DFA) module is developed in the encoding path to highlight the semantically pathological asymmetries of the features in abnormalities. In the DFA module, a Siamese contrastive supervised loss is designed to enhance discriminative features, and a mirror position-based difference augmentation (MDA) module is used to further magnify the discrepancy information. Moreover, mirror feature fusion (MFF) modules are applied to fuse and transfer the information both of the original input and the horizontally flipped features to the decoding path. Extensive experiments on the Anatomical Tracings of Lesions After Stroke (ATLAS) dataset show the proposed MDAN outperforms the state-of-the-art methods.
Collapse
|
32
|
Fan Z, Li J, Zhang L, Zhu G, Li P, Lu X, Shen P, Shah SAA, Bennamoun M, Hua T, Wei W. U-net based analysis of MRI for Alzheimer’s disease diagnosis. Neural Comput Appl 2021. [DOI: 10.1007/s00521-021-05983-y] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/08/2023]
|