1
|
Zhu H, Huang J, Chen K, Ying X, Qian Y. multiPI-TransBTS: A multi-path learning framework for brain tumor image segmentation based on multi-physical information. Comput Biol Med 2025; 191:110148. [PMID: 40215867 DOI: 10.1016/j.compbiomed.2025.110148] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/18/2024] [Revised: 04/01/2025] [Accepted: 04/04/2025] [Indexed: 04/29/2025]
Abstract
Brain Tumor Segmentation (BraTS) plays a critical role in clinical diagnosis, treatment planning, and monitoring the progression of brain tumors. However, due to the variability in tumor appearance, size, and intensity across different MRI modalities, automated segmentation remains a challenging task. In this study, we propose a novel Transformer-based framework, multiPI-TransBTS, which integrates multi-physical information to enhance segmentation accuracy. The model leverages spatial information, semantic information, and multi-modal imaging data, addressing the inherent heterogeneity in brain tumor characteristics. The multiPI-TransBTS framework consists of an encoder, an Adaptive Feature Fusion (AFF) module, and a multi-source, multi-scale feature decoder. The encoder incorporates a multi-branch architecture to separately extract modality-specific features from different MRI sequences. The AFF module fuses information from multiple sources using channel-wise and element-wise attention, ensuring effective feature recalibration. The decoder combines both common and task-specific features through a Task-Specific Feature Introduction (TSFI) strategy, producing accurate segmentation outputs for Whole Tumor (WT), Tumor Core (TC), and Enhancing Tumor (ET) regions. Comprehensive evaluations on the BraTS2019 and BraTS2020 datasets demonstrate the superiority of multiPI-TransBTS over the state-of-the-art methods. The model consistently achieves better Dice coefficients, Hausdorff distances, and Sensitivity scores, highlighting its effectiveness in addressing the BraTS challenges. Our results also indicate the need for further exploration of the balance between precision and recall in the ET segmentation task. The proposed framework represents a significant advancement in BraTS, with potential implications for improving clinical outcomes for brain tumor patients.
Collapse
Affiliation(s)
- Hongjun Zhu
- School of Software Engineering, Chongqing University of Posts and Telecommunications, Chongqing, 400065, China; Chongqing Engineering Research Center of Software Quality Assurance, Testing and Assessment, Chongqing, 400065, China; Key Laboratory of Big Data Intelligent Computing, Chongqing University of Posts and Telecommunications, Chongqing, 400065, China.
| | - Jiaohang Huang
- School of Software Engineering, Chongqing University of Posts and Telecommunications, Chongqing, 400065, China
| | - Kuo Chen
- School of Software Engineering, Chongqing University of Posts and Telecommunications, Chongqing, 400065, China; Chongqing Engineering Research Center of Software Quality Assurance, Testing and Assessment, Chongqing, 400065, China
| | - Xuehui Ying
- School of Software Engineering, Chongqing University of Posts and Telecommunications, Chongqing, 400065, China; Chongqing Engineering Research Center of Software Quality Assurance, Testing and Assessment, Chongqing, 400065, China
| | - Ying Qian
- School of Software Engineering, Chongqing University of Posts and Telecommunications, Chongqing, 400065, China; Chongqing Engineering Research Center of Software Quality Assurance, Testing and Assessment, Chongqing, 400065, China
| |
Collapse
|
2
|
Jiang M, Wang S, Chan KH, Sun Y, Xu Y, Zhang Z, Gao Q, Gao Z, Tong T, Chang HC, Tan T. Multimodal Cross Global Learnable Attention Network for MR images denoising with arbitrary modal missing. Comput Med Imaging Graph 2025; 121:102497. [PMID: 39904265 DOI: 10.1016/j.compmedimag.2025.102497] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/23/2024] [Revised: 12/10/2024] [Accepted: 01/22/2025] [Indexed: 02/06/2025]
Abstract
Magnetic Resonance Imaging (MRI) generates medical images of multiple sequences, i.e., multimodal, from different contrasts. However, noise will reduce the quality of MR images, and then affect the doctor's diagnosis of diseases. Existing filtering methods, transform-domain methods, statistical methods and Convolutional Neural Network (CNN) methods main aim to denoise individual sequences of images without considering the relationships between multiple different sequences. They cannot balance the extraction of high-dimensional and low-dimensional features in MR images, and hard to maintain a good balance between preserving image texture details and denoising strength. To overcome these challenges, this work proposes a controllable Multimodal Cross-Global Learnable Attention Network (MMCGLANet) for MR image denoising with Arbitrary Modal Missing. Specifically, Encoder is employed to extract the shallow features of the image which share weight module, and Convolutional Long Short-Term Memory(ConvLSTM) is employed to extract the associated features between different frames within the same modal. Cross Global Learnable Attention Network(CGLANet) is employed to extract and fuse image features between multimodal and within the same modality. In addition, sequence code is employed to label missing modalities, which allows for Arbitrary Modal Missing during model training, validation, and testing. Experimental results demonstrate that our method has achieved good denoising results on different public and real MR image dataset.
Collapse
Affiliation(s)
- Mingfu Jiang
- Faculty of Applied Sciences, Macao Polytechnic University, R. de Luís Gonzaga Gomes, Macao, 999078, Macao Special Administrative Region of China; College of Information Engineering, Xinyang Agriculture and Forestry University, No. 1 North Ring Road, Pingqiao District, Xinyang, 464000, Henan, China
| | - Shuai Wang
- School of Cyberspace, Hangzhou Dianzi University, No. 65 Wen Yi Road, Hangzhou, 310018, Zhejiang, China
| | - Ka-Hou Chan
- Faculty of Applied Sciences, Macao Polytechnic University, R. de Luís Gonzaga Gomes, Macao, 999078, Macao Special Administrative Region of China
| | - Yue Sun
- Faculty of Applied Sciences, Macao Polytechnic University, R. de Luís Gonzaga Gomes, Macao, 999078, Macao Special Administrative Region of China
| | - Yi Xu
- Shanghai Key Lab of Digital Media Processing and Transmission, Shanghai Jiao Tong University MoE Key Lab of Artificial Intelligence, Shanghai Jiao Tong University, No. 800 Dongchuan Road, Minhang District, Shanghai, 200030, China
| | - Zhuoneng Zhang
- Faculty of Applied Sciences, Macao Polytechnic University, R. de Luís Gonzaga Gomes, Macao, 999078, Macao Special Administrative Region of China
| | - Qinquan Gao
- College of Physics and Information Engineering, Fuzhou University, No. 2 Wulongjiang Avenue, Fuzhou, 350108, Fujian, China
| | - Zhifan Gao
- School of Biomedical Engineering, Sun Yat-sen University, No. 66 Gongchang Road, Guangming District, Shenzhen, 518107, Guangdong, China
| | - Tong Tong
- College of Physics and Information Engineering, Fuzhou University, No. 2 Wulongjiang Avenue, Fuzhou, 350108, Fujian, China
| | - Hing-Chiu Chang
- Department of Biomedical Engineering, Chinese University of Hong Kong, Sha Tin District, 999077, Hong Kong, China
| | - Tao Tan
- Faculty of Applied Sciences, Macao Polytechnic University, R. de Luís Gonzaga Gomes, Macao, 999078, Macao Special Administrative Region of China.
| |
Collapse
|
3
|
Zhou J, Wang S, Wang H, Li Y, Li X. Multi-Modality Fusion and Tumor Sub-Component Relationship Ensemble Network for Brain Tumor Segmentation. Bioengineering (Basel) 2025; 12:159. [PMID: 40001679 PMCID: PMC11851405 DOI: 10.3390/bioengineering12020159] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/31/2024] [Revised: 01/21/2025] [Accepted: 01/27/2025] [Indexed: 02/27/2025] Open
Abstract
Deep learning technology has been widely used in brain tumor segmentation with multi-modality magnetic resonance imaging, helping doctors achieve faster and more accurate diagnoses. Previous studies have demonstrated that the weighted fusion segmentation method effectively extracts modality importance, laying a solid foundation for multi-modality magnetic resonance imaging segmentation. However, the challenge of fusing multi-modality features with single-modality features remains unresolved, which motivated us to explore an effective fusion solution. We propose a multi-modality and single-modality feature recalibration network for magnetic resonance imaging brain tumor segmentation. Specifically, we designed a dual recalibration module that achieves accurate feature calibration by integrating the complementary features of multi-modality with the specific features of a single modality. Experimental results on the BraTS 2018 dataset showed that the proposed method outperformed existing multi-modal network methods across multiple evaluation metrics, with spatial recalibration significantly improving the results, including Dice score increases of 1.7%, 0.5%, and 1.6% for the enhanced tumor core, whole tumor, and tumor core regions, respectively.
Collapse
Affiliation(s)
- Jinyan Zhou
- Basic Medical College, Heilongjiang University of Chinese Medicine, Harbin 150040, China; (J.Z.); (S.W.)
| | - Shuwen Wang
- Basic Medical College, Heilongjiang University of Chinese Medicine, Harbin 150040, China; (J.Z.); (S.W.)
| | - Hao Wang
- Department of Control Science and Engineering, Harbin Institute of Technology, Harbin 150001, China;
| | - Yaxue Li
- Basic Medical College, Heilongjiang University of Chinese Medicine, Harbin 150040, China; (J.Z.); (S.W.)
| | - Xiang Li
- Department of Control Science and Engineering, Harbin Institute of Technology, Harbin 150001, China;
- College of Information Science and Engineering, Northeastern University, Shenyang 110819, China
- Hebei Key Laboratory of Micro-Nano Precision Optical Sensing and Measurement Technology, Qinhuangdao 066004, China
| |
Collapse
|
4
|
Shen C, Li W, Chen H, Wang X, Zhu F, Li Y, Wang X, Jin B. Complementary information mutual learning for multimodality medical image segmentation. Neural Netw 2024; 180:106670. [PMID: 39299035 DOI: 10.1016/j.neunet.2024.106670] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/25/2024] [Revised: 07/10/2024] [Accepted: 08/26/2024] [Indexed: 09/22/2024]
Abstract
Radiologists must utilize medical images of multiple modalities for tumor segmentation and diagnosis due to the limitations of medical imaging technology and the diversity of tumor signals. This has led to the development of multimodal learning in medical image segmentation. However, the redundancy among modalities creates challenges for existing subtraction-based joint learning methods, such as misjudging the importance of modalities, ignoring specific modal information, and increasing cognitive load. These thorny issues ultimately decrease segmentation accuracy and increase the risk of overfitting. This paper presents the complementary information mutual learning (CIML) framework, which can mathematically model and address the negative impact of inter-modal redundant information. CIML adopts the idea of addition and removes inter-modal redundant information through inductive bias-driven task decomposition and message passing-based redundancy filtering. CIML first decomposes the multimodal segmentation task into multiple subtasks based on expert prior knowledge, minimizing the information dependence between modalities. Furthermore, CIML introduces a scheme in which each modality can extract information from other modalities additively through message passing. To achieve non-redundancy of extracted information, the redundant filtering is transformed into complementary information learning inspired by the variational information bottleneck. The complementary information learning procedure can be efficiently solved by variational inference and cross-modal spatial attention. Numerical results from the verification task and standard benchmarks indicate that CIML efficiently removes redundant information between modalities, outperforming SOTA methods regarding validation accuracy and segmentation effect. To emphasize, message-passing-based redundancy filtering allows neural network visualization techniques to visualize the knowledge relationship among different modalities, which reflects interpretability.
Collapse
Affiliation(s)
- Chuyun Shen
- School of Computer Science and Technology, East China Normal University, Shanghai 200062, China.
| | - Wenhao Li
- School of Data Science, The Chinese University of Hong Kong, Shenzhen Shenzhen Institute of Artificial Intelligence and Robotics for Society, Shenzhen 518172, China.
| | - Haoqing Chen
- School of Computer Science and Technology, East China Normal University, Shanghai 200062, China.
| | - Xiaoling Wang
- School of Computer Science and Technology, East China Normal University, Shanghai 200062, China.
| | - Fengping Zhu
- Huashan Hospital Fudan University, Shanghai 200040, China.
| | - Yuxin Li
- Huashan Hospital Fudan University, Shanghai 200040, China.
| | - Xiangfeng Wang
- School of Computer Science and Technology, East China Normal University, Shanghai 200062, China.
| | - Bo Jin
- School of Software Engineering, Shanghai Research Institute for Intelligent Autonomous Systems, Tongji University, Shanghai 200092, China.
| |
Collapse
|
5
|
Zhang Q, Hang Y, Qiu J, Chen H. Application of U-Net Network Utilizing Multiattention Gate for MRI Segmentation of Brain Tumors. J Comput Assist Tomogr 2024; 48:991-997. [PMID: 39190714 DOI: 10.1097/rct.0000000000001641] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 08/29/2024]
Abstract
BACKGROUND Studies have shown that the type of low-grade glioma is associated with its shape. The traditional diagnostic method involves extraction of the tumor shape from MRIs and diagnosing the type of glioma based on corresponding relationship between the glioma shape and type. This method is affected by the MRI background, tumor pixel size, and doctors' professional level, leading to misdiagnoses and missed diagnoses. With the help of deep learning algorithms, the shape of a glioma can be automatically segmented, thereby assisting doctors to focus more on the diagnosis of glioma and improving diagnostic efficiency. The segmentation of glioma MRIs using traditional deep learning algorithms exhibits limited accuracy, thereby impeding the effectiveness of assisting doctors in the diagnosis. The primary objective of our research is to facilitate the segmentation of low-grade glioma MRIs for medical practitioners through the utilization of deep learning algorithms. METHODS In this study, a UNet glioma segmentation network that incorporates multiattention gates was proposed to address this limitation. The UNet-based algorithm in the coding part integrated the attention gate into the hierarchical structure of the network to suppress the features of irrelevant regions and reduce the feature redundancy. In the decoding part, by adding attention gates in the fusion process of low- and high-level features, important feature information was highlighted, model parameters were reduced, and model sensitivity and accuracy were improved. RESULTS The network model performed image segmentation on the glioma MRI dataset, and the accuracy and average intersection ratio (mIoU) of the algorithm segmentation reached 99.7%, 87.3%, 99.7%, and 87.6%. CONCLUSIONS Compared with the UNet, PSPNet, and Attention UNet network models, this network model has obvious advantages in accuracy, mIoU, and loss convergence. It can serve as a standard for assisting doctors in diagnosis.
Collapse
Affiliation(s)
- Qiong Zhang
- From the College of Computer and Information Engineering, Nantong Institute of Technology
| | - Yiliu Hang
- From the College of Computer and Information Engineering, Nantong Institute of Technology
| | | | - Hao Chen
- From the College of Computer and Information Engineering, Nantong Institute of Technology
| |
Collapse
|
6
|
Zhao Y, Wang X, Phan J, Chen X, Lee A, Yu C, Huang K, Court LE, Pan T, Wang H, Wahid KA, Mohamed ASR, Naser M, Fuller CD, Yang J. Multi-modal segmentation with missing image data for automatic delineation of gross tumor volumes in head and neck cancers. Med Phys 2024; 51:7295-7307. [PMID: 38896829 PMCID: PMC11479854 DOI: 10.1002/mp.17260] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/28/2023] [Revised: 05/22/2024] [Accepted: 06/05/2024] [Indexed: 06/21/2024] Open
Abstract
BACKGROUND Head and neck (HN) gross tumor volume (GTV) auto-segmentation is challenging due to the morphological complexity and low image contrast of targets. Multi-modality images, including computed tomography (CT) and positron emission tomography (PET), are used in the routine clinic to assist radiation oncologists for accurate GTV delineation. However, the availability of PET imaging may not always be guaranteed. PURPOSE To develop a deep learning segmentation framework for automated GTV delineation of HN cancers using a combination of PET/CT images, while addressing the challenge of missing PET data. METHODS Two datasets were included for this study: Dataset I: 524 (training) and 359 (testing) oropharyngeal cancer patients from different institutions with their PET/CT pairs provided by the HECKTOR Challenge; Dataset II: 90 HN patients(testing) from a local institution with their planning CT, PET/CT pairs. To handle potentially missing PET images, a model training strategy named the "Blank Channel" method was implemented. To simulate the absence of a PET image, a blank array with the same dimensions as the CT image was generated to meet the dual-channel input requirement of the deep learning model. During the model training process, the model was randomly presented with either a real PET/CT pair or a blank/CT pair. This allowed the model to learn the relationship between the CT image and the corresponding GTV delineation based on available modalities. As a result, our model had the ability to handle flexible inputs during prediction, making it suitable for cases where PET images are missing. To evaluate the performance of our proposed model, we trained it using training patients from Dataset I and tested it with Dataset II. We compared our model (Model 1) with two other models which were trained for specific modality segmentations: Model 2 trained with only CT images, and Model 3 trained with real PET/CT pairs. The performance of the models was evaluated using quantitative metrics, including Dice similarity coefficient (DSC), mean surface distance (MSD), and 95% Hausdorff Distance (HD95). In addition, we evaluated our Model 1 and Model 3 using the 359 test cases in Dataset I. RESULTS Our proposed model(Model 1) achieved promising results for GTV auto-segmentation using PET/CT images, with the flexibility of missing PET images. Specifically, when assessed with only CT images in Dataset II, Model 1 achieved DSC of 0.56 ± 0.16, MSD of 3.4 ± 2.1 mm, and HD95 of 13.9 ± 7.6 mm. When the PET images were included, the performance of our model was improved to DSC of 0.62 ± 0.14, MSD of 2.8 ± 1.7 mm, and HD95 of 10.5 ± 6.5 mm. These results are comparable to those achieved by Model 2 and Model 3, illustrating Model 1's effectiveness in utilizing flexible input modalities. Further analysis using the test dataset from Dataset I showed that Model 1 achieved an average DSC of 0.77, surpassing the overall average DSC of 0.72 among all participants in the HECKTOR Challenge. CONCLUSIONS We successfully refined a multi-modal segmentation tool for accurate GTV delineation for HN cancer. Our method addressed the issue of missing PET images by allowing flexible data input, thereby providing a practical solution for clinical settings where access to PET imaging may be limited.
Collapse
Affiliation(s)
- Yao Zhao
- Department of Radiation Physics, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
- UTHealth Houston Graduate School of Biomedical Sciences, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
| | - Xin Wang
- Department of Radiation Physics, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
- UTHealth Houston Graduate School of Biomedical Sciences, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
| | - Jack Phan
- Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
| | - Xinru Chen
- Department of Radiation Physics, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
- UTHealth Houston Graduate School of Biomedical Sciences, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
| | - Anna Lee
- Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
| | - Cenji Yu
- Department of Radiation Physics, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
- UTHealth Houston Graduate School of Biomedical Sciences, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
| | - Kai Huang
- Department of Radiation Physics, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
- UTHealth Houston Graduate School of Biomedical Sciences, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
| | - Laurence E. Court
- Department of Radiation Physics, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
- UTHealth Houston Graduate School of Biomedical Sciences, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
| | - Tinsu Pan
- UTHealth Houston Graduate School of Biomedical Sciences, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
| | - He Wang
- Department of Radiation Physics, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
- UTHealth Houston Graduate School of Biomedical Sciences, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
| | - Kareem Abdul Wahid
- Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
| | - Abdalah S R Mohamed
- Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
| | - Mohamed Naser
- Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
| | - Clifton D Fuller
- Department of Radiation Oncology, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
| | - Jinzhong Yang
- Department of Radiation Physics, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
- UTHealth Houston Graduate School of Biomedical Sciences, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
| |
Collapse
|
7
|
Nazir M, Shakil S, Khurshid K. End-to-End Multi-task Learning Architecture for Brain Tumor Analysis with Uncertainty Estimation in MRI Images. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024; 37:2149-2172. [PMID: 38565728 PMCID: PMC11522262 DOI: 10.1007/s10278-024-01009-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/24/2023] [Revised: 11/25/2023] [Accepted: 11/28/2023] [Indexed: 04/04/2024]
Abstract
Brain tumors are a threat to life for every other human being, be it adults or children. Gliomas are one of the deadliest brain tumors with an extremely difficult diagnosis. The reason is their complex and heterogenous structure which gives rise to subjective as well as objective errors. Their manual segmentation is a laborious task due to their complex structure and irregular appearance. To cater to all these issues, a lot of research has been done and is going on to develop AI-based solutions that can help doctors and radiologists in the effective diagnosis of gliomas with the least subjective and objective errors, but an end-to-end system is still missing. An all-in-one framework has been proposed in this research. The developed end-to-end multi-task learning (MTL) architecture with a feature attention module can classify, segment, and predict the overall survival of gliomas by leveraging task relationships between similar tasks. Uncertainty estimation has also been incorporated into the framework to enhance the confidence level of healthcare practitioners. Extensive experimentation was performed by using combinations of MRI sequences. Brain tumor segmentation (BraTS) challenge datasets of 2019 and 2020 were used for experimental purposes. Results of the best model with four sequences show 95.1% accuracy for classification, 86.3% dice score for segmentation, and a mean absolute error (MAE) of 456.59 for survival prediction on the test data. It is evident from the results that deep learning-based MTL models have the potential to automate the whole brain tumor analysis process and give efficient results with least inference time without human intervention. Uncertainty quantification confirms the idea that more data can improve the generalization ability and in turn can produce more accurate results with less uncertainty. The proposed model has the potential to be utilized in a clinical setup for the initial screening of glioma patients.
Collapse
Affiliation(s)
- Maria Nazir
- Medical Imaging and Diagnostics Lab, NCAI COMSATS University Islamabad, Islamabad, Pakistan.
- iVision Lab, Department of Electrical Engineering, Institute of Space Technology, Islamabad, Pakistan.
- BiCoNeS Lab, Department of Electrical Engineering, Institute of Space Technology, Islamabad, Pakistan.
| | - Sadia Shakil
- Department of Biomedical Engineering, The Chinese University of Hong Kong, Hong Kong, China
| | - Khurram Khurshid
- iVision Lab, Department of Electrical Engineering, Institute of Space Technology, Islamabad, Pakistan
| |
Collapse
|
8
|
Jiang N, Wang G, Ye C, Liu T, Yan T. Multi-Task Collaborative Pre-Training and Adaptive Token Selection: A Unified Framework for Brain Representation Learning. IEEE J Biomed Health Inform 2024; 28:5528-5539. [PMID: 38889024 DOI: 10.1109/jbhi.2024.3416038] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/20/2024]
Abstract
Structural magnetic resonance imaging (sMRI) reveals the structural organization of the brain. Learning general brain representations from sMRI is an enduring topic in neuroscience. Previous deep learning models neglect that the brain, as the core of cognition, is distinct from other organs whose primary attribute is anatomy. Capturing the high-level representation associated with inter-individual cognitive variability is key to appropriately represent the brain. Given that this cognition-related information is subtle, mixed, and distributed in the brain structure, sMRI-based models need to both capture fine-grained details and understand how they relate to the overall global structure. Additionally, it is also necessary to explicitly express the cognitive information that implicitly embedded in local-global image features. Therefore, we propose MCPATS, a brain representation learning framework that combines Multi-task Collaborative Pre-training (MCP) and Adaptive Token Selection (ATS). First, we develop MCP, including mask-reconstruction to understand global context, distort-restoration to capture fine-grained local details, adversarial learning to integrate features at different granularities, and age-prediction, using age as a surrogate for cognition to explicitly encode cognition-related information from local-global image features. This co-training allows progressive learning of implicit and explicit cognition-related representations. Then, we develop ATS based on mutual attention for downstream use of the learned representation. During fine-tuning, the ATS highlights discriminative features and reduces the impact of irrelevant information. MCPATS was validated on three different public datasets for brain disease diagnosis, outperforming competing methods and achieving accurate diagnosis. Further, we performed detailed analysis to confirm that the MCPATS-learned representation captures cognition-related information.
Collapse
|
9
|
Wang J, Tang Y, Xiao Y, Zhou JT, Fang Z, Yang F. GREnet: Gradually REcurrent Network With Curriculum Learning for 2-D Medical Image Segmentation. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:10018-10032. [PMID: 37022080 DOI: 10.1109/tnnls.2023.3238381] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Medical image segmentation is a vital stage in medical image analysis. Numerous deep-learning methods are booming to improve the performance of 2-D medical image segmentation, owing to the fast growth of the convolutional neural network. Generally, the manually defined ground truth is utilized directly to supervise models in the training phase. However, direct supervision of the ground truth often results in ambiguity and distractors as complex challenges appear simultaneously. To alleviate this issue, we propose a gradually recurrent network with curriculum learning, which is supervised by gradual information of the ground truth. The whole model is composed of two independent networks. One is the segmentation network denoted as GREnet, which formulates 2-D medical image segmentation as a temporal task supervised by pixel-level gradual curricula in the training phase. The other is a curriculum-mining network. To a certain degree, the curriculum-mining network provides curricula with an increasing difficulty in the ground truth of the training set by progressively uncovering hard-to-segmentation pixels via a data-driven manner. Given that segmentation is a pixel-level dense-prediction challenge, to the best of our knowledge, this is the first work to function 2-D medical image segmentation as a temporal task with pixel-level curriculum learning. In GREnet, the naive UNet is adopted as the backbone, while ConvLSTM is used to establish the temporal link between gradual curricula. In the curriculum-mining network, UNet++ supplemented by transformer is designed to deliver curricula through the outputs of the modified UNet++ at different layers. Experimental results have demonstrated the effectiveness of GREnet on seven datasets, i.e., three lesion segmentation datasets in dermoscopic images, an optic disc and cup segmentation dataset and a blood vessel segmentation dataset in retinal images, a breast lesion segmentation dataset in ultrasound images, and a lung segmentation dataset in computed tomography (CT).
Collapse
|
10
|
Hu X, Wang L, Wang L, Chen Q, Zheng L, Zhu Y. Glioma segmentation based on dense contrastive learning and multimodal features recalibration. Phys Med Biol 2024; 69:095016. [PMID: 38537288 DOI: 10.1088/1361-6560/ad387f] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/09/2023] [Accepted: 03/27/2024] [Indexed: 04/23/2024]
Abstract
Accurate segmentation of different regions of gliomas from multimodal magnetic resonance (MR) images is crucial for glioma grading and precise diagnosis, but many existing segmentation methods are difficult to effectively utilize multimodal MR image information to recognize accurately the lesion regions with small size, low contrast and irregular shape. To address this issue, this work proposes a novel 3D glioma segmentation model DCL-MANet. DCL-MANet has an architecture of multiple encoders and one single decoder. Each encoder is used to extract MR image features of a given modality. To overcome the entangle problems of multimodal semantic features, a dense contrastive learning (DCL) strategy is presented to extract the modality-specific and common features. Following that, feature recalibration block (RFB) based on modality-wise attention is used to recalibrate the semantic features of each modality, enabling the model to focus on the features that are beneficial for glioma segmentation. These recalibrated features are input into the decoder to obtain the segmentation results. To verify the superiority of the proposed method, we compare it with several state-of-the-art (SOTA) methods in terms of Dice, average symmetric surface distance (ASSD), HD95 and volumetric similarity (Vs). The comparison results show that the average Dice, ASSD, HD95 and Vs of DCL-MANet on all tumor regions are improved at least by 0.66%, 3.47%, 8.94% and 1.07% respectively. For small enhance tumor (ET) region, the corresponding improvement can be up to 0.37%, 7.83%, 11.32%, and 1.35%, respectively. In addition, the ablation results demonstrate the effectiveness of the proposed DCL and RFB, and combining them can significantly increase Dice (1.59%) and Vs (1.54%) while decreasing ASSD (40.51%) and HD95 (45.16%) on ET region. The proposed DCL-MANet could disentangle multimodal features and enhance the semantics of modality-dependent features, providing a potential means to accurately segment small lesion regions in gliomas.
Collapse
Affiliation(s)
- Xubin Hu
- Engineering Research Center of Text Computing & Cognitive Intelligence, Ministry of Education, Key Laboratory of Intelligent Medical Image Analysis and Precise Diagnosis of Guizhou Province, State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang 550025, People's Republic of China
| | - Lihui Wang
- Engineering Research Center of Text Computing & Cognitive Intelligence, Ministry of Education, Key Laboratory of Intelligent Medical Image Analysis and Precise Diagnosis of Guizhou Province, State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang 550025, People's Republic of China
| | - Li Wang
- Engineering Research Center of Text Computing & Cognitive Intelligence, Ministry of Education, Key Laboratory of Intelligent Medical Image Analysis and Precise Diagnosis of Guizhou Province, State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang 550025, People's Republic of China
| | - Qijian Chen
- Engineering Research Center of Text Computing & Cognitive Intelligence, Ministry of Education, Key Laboratory of Intelligent Medical Image Analysis and Precise Diagnosis of Guizhou Province, State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang 550025, People's Republic of China
| | - Licheng Zheng
- Engineering Research Center of Text Computing & Cognitive Intelligence, Ministry of Education, Key Laboratory of Intelligent Medical Image Analysis and Precise Diagnosis of Guizhou Province, State Key Laboratory of Public Big Data, College of Computer Science and Technology, Guizhou University, Guiyang 550025, People's Republic of China
| | - Yuemin Zhu
- University Lyon, INSA Lyon, CNRS, Inserm, IRP Metislab CREATIS UMR5220, U1206, Lyon F-69621, France
| |
Collapse
|
11
|
Liu H, Huang J, Li Q, Guan X, Tseng M. A deep convolutional neural network for the automatic segmentation of glioblastoma brain tumor: Joint spatial pyramid module and attention mechanism network. Artif Intell Med 2024; 148:102776. [PMID: 38325925 DOI: 10.1016/j.artmed.2024.102776] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2023] [Revised: 12/20/2023] [Accepted: 01/14/2024] [Indexed: 02/09/2024]
Abstract
This study proposes a deep convolutional neural network for the automatic segmentation of glioblastoma brain tumors, aiming sat replacing the manual segmentation method that is both time-consuming and labor-intensive. There are many challenges for automatic segmentation to finely segment sub-regions from multi-sequence magnetic resonance images because of the complexity and variability of glioblastomas, such as the loss of boundary information, misclassified regions, and subregion size. To overcome these challenges, this study introduces a spatial pyramid module and attention mechanism to the automatic segmentation algorithm, which focuses on multi-scale spatial details and context information. The proposed method has been tested in the public benchmarks BraTS 2018, BraTS 2019, BraTS 2020 and BraTS 2021 datasets. The Dice score on the enhanced tumor, whole tumor, and tumor core were respectively 79.90 %, 89.63 %, and 85.89 % on the BraTS 2018 dataset, respectively 77.14 %, 89.58 %, and 83.33 % on the BraTS 2019 dataset, and respectively 77.80 %, 90.04 %, and 83.18 % on the BraTS 2020 dataset, and respectively 83.48 %, 90.70 %, and 88.94 % on the BraTS 2021 dataset offering performance on par with that of state-of-the-art methods with only 1.90 M parameters. In addition, our approach significantly reduced the requirements for experimental equipment, and the average time taken to segment one case was only 1.48 s; these two benefits rendered the proposed network intensely competitive for clinical practice.
Collapse
Affiliation(s)
- Hengxin Liu
- School of Microelectronics, Tianjin University, Tianjin, China
| | - Jingteng Huang
- School of Microelectronics, Tianjin University, Tianjin, China
| | - Qiang Li
- School of Microelectronics, Tianjin University, Tianjin, China
| | - Xin Guan
- School of Microelectronics, Tianjin University, Tianjin, China.
| | - Minglang Tseng
- Institute of Innovation and Circular Economy, Asia University, Taichung, Taiwan; Department of Medical Research, China Medical University Hospital, China Medical University, Taichung, Taiwan; UKM-Graduate School of Business, Universiti Kebangsaan Malaysia, 43000 Bangi, Selangor, Malaysia; Department of Industrial Engineering, Khon Kaen University, 40002, Thailand.
| |
Collapse
|
12
|
Xie Y, Zhang J, Xia Y, Shen C. Learning From Partially Labeled Data for Multi-Organ and Tumor Segmentation. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2023; 45:14905-14919. [PMID: 37672381 DOI: 10.1109/tpami.2023.3312587] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/08/2023]
Abstract
Medical image benchmarks for the segmentation of organs and tumors suffer from the partially labeling issue due to its intensive cost of labor and expertise. Current mainstream approaches follow the practice of one network solving one task. With this pipeline, not only the performance is limited by the typically small dataset of a single task, but also the computation cost linearly increases with the number of tasks. To address this, we propose a Transformer based dynamic on-demand network (TransDoDNet) that learns to segment organs and tumors on multiple partially labeled datasets. Specifically, TransDoDNet has a hybrid backbone that is composed of the convolutional neural network and Transformer. A dynamic head enables the network to accomplish multiple segmentation tasks flexibly. Unlike existing approaches that fix kernels after training, the kernels in the dynamic head are generated adaptively by the Transformer, which employs the self-attention mechanism to model long-range organ-wise dependencies and decodes the organ embedding that can represent each organ. We create a large-scale partially labeled Multi-Organ and Tumor Segmentation benchmark, termed MOTS, and demonstrate the superior performance of our TransDoDNet over other competitors on seven organ and tumor segmentation tasks. This study also provides a general 3D medical image segmentation model, which has been pre-trained on the large-scale MOTS benchmark and has demonstrated advanced performance over current predominant self-supervised learning methods.
Collapse
|
13
|
Ahamed MF, Hossain MM, Nahiduzzaman M, Islam MR, Islam MR, Ahsan M, Haider J. A review on brain tumor segmentation based on deep learning methods with federated learning techniques. Comput Med Imaging Graph 2023; 110:102313. [PMID: 38011781 DOI: 10.1016/j.compmedimag.2023.102313] [Citation(s) in RCA: 9] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/13/2023] [Revised: 11/13/2023] [Accepted: 11/13/2023] [Indexed: 11/29/2023]
Abstract
Brain tumors have become a severe medical complication in recent years due to their high fatality rate. Radiologists segment the tumor manually, which is time-consuming, error-prone, and expensive. In recent years, automated segmentation based on deep learning has demonstrated promising results in solving computer vision problems such as image classification and segmentation. Brain tumor segmentation has recently become a prevalent task in medical imaging to determine the tumor location, size, and shape using automated methods. Many researchers have worked on various machine and deep learning approaches to determine the most optimal solution using the convolutional methodology. In this review paper, we discuss the most effective segmentation techniques based on the datasets that are widely used and publicly available. We also proposed a survey of federated learning methodologies to enhance global segmentation performance and ensure privacy. A comprehensive literature review is suggested after studying more than 100 papers to generalize the most recent techniques in segmentation and multi-modality information. Finally, we concentrated on unsolved problems in brain tumor segmentation and a client-based federated model training strategy. Based on this review, future researchers will understand the optimal solution path to solve these issues.
Collapse
Affiliation(s)
- Md Faysal Ahamed
- Department of Computer Science & Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Md Munawar Hossain
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Md Nahiduzzaman
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Md Rabiul Islam
- Department of Computer Science & Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Md Robiul Islam
- Department of Electrical & Computer Engineering, Rajshahi University of Engineering & Technology, Rajshahi 6204, Bangladesh
| | - Mominul Ahsan
- Department of Computer Science, University of York, Deramore Lane, Heslington, York YO10 5GH, UK
| | - Julfikar Haider
- Department of Engineering, Manchester Metropolitan University, Chester St, Manchester M1 5GD, UK.
| |
Collapse
|
14
|
Hou Q, Peng Y, Wang Z, Wang J, Jiang J. MFD-Net: Modality Fusion Diffractive Network for Segmentation of Multimodal Brain Tumor Image. IEEE J Biomed Health Inform 2023; 27:5958-5969. [PMID: 37747864 DOI: 10.1109/jbhi.2023.3318640] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/27/2023]
Abstract
Automatic brain tumor segmentation using multi-parametric magnetic resonance imaging (mpMRI) holds substantial importance for brain diagnosis, monitoring, and therapeutic strategy planning. Given the constraints inherent to manual segmentation, adopting deep learning networks for accomplishing accurate and automated segmentation emerges as an essential advancement. In this article, we propose a modality fusion diffractive network (MFD-Net) composed of diffractive blocks and modality feature extractors for the automatic and accurate segmentation of brain tumors. The diffractive block, designed based on Fraunhofer's single-slit diffraction principle, emphasizes neighboring high-confidence feature points and suppresses low-quality or isolated feature points, enhancing the interrelation of features. Adopting a global passive reception mode overcomes the issue of fixed receptive fields. Through a self-supervised approach, the modality feature extractor effectively utilizes the inherent generalization information of each modality, enabling the main segmentation branch to focus more on multimodal fusion feature information. We apply the diffractive block on nn-UNet in the MICCAI BraTS 2022 challenge, ranked first in the pediatric population data and third in the BraTS continuous evaluation data, proving the superior generalizability of our network. We also train separately on the BraTS 2018, 2019, and 2021 datasets. Experiments demonstrate that the proposed network outperforms state-of-the-art methods.
Collapse
|
15
|
Li X, Luo G, Wang W, Wang K, Li S. Curriculum label distribution learning for imbalanced medical image segmentation. Med Image Anal 2023; 89:102911. [PMID: 37542795 DOI: 10.1016/j.media.2023.102911] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/28/2022] [Revised: 04/27/2023] [Accepted: 07/25/2023] [Indexed: 08/07/2023]
Abstract
Label distribution learning (LDL) has the potential to resolve boundary ambiguity in semantic segmentation tasks. However, existing LDL-based segmentation methods suffer from severe label distribution imbalance: the ambiguous label distributions contain a small fraction of the data, while the unambiguous label distributions occupy the majority of the data. The imbalanced label distributions induce model-biased distribution learning and make it challenging to accurately predict ambiguous pixels. In this paper, we propose a curriculum label distribution learning (CLDL) framework to address the above data imbalance problem by performing a novel task-oriented curriculum learning strategy. Firstly, the region label distribution learning (R-LDL) is proposed to construct more balanced label distributions and improves the imbalanced model learning. Secondly, a novel learning curriculum (TCL) is proposed to enable easy-to-hard learning in LDL-based segmentation by decomposing the segmentation task into multiple label distribution estimation tasks. Thirdly, the prior perceiving module (PPM) is proposed to effectively connect easy and hard learning stages based on the priors generated from easier stages. Benefiting from the balanced label distribution construction and prior perception, the proposed CLDL effectively conducts a curriculum learning-based LDL and significantly improves the imbalanced learning. We evaluated the proposed CLDL using the publicly available BRATS2018 and MM-WHS2017 datasets. The experimental results demonstrate that our method significantly improves different segmentation metrics compared to many state-of-the-art methods. The code will be available.1.
Collapse
Affiliation(s)
- Xiangyu Li
- School of Computer Science and Technology, Harbin Institute of Technology, Harbin, China
| | - Gongning Luo
- School of Computer Science and Technology, Harbin Institute of Technology, Harbin, China.
| | - Wei Wang
- School of Computer Science and Technology, Harbin Institute of Technology, Shenzhen, China.
| | - Kuanquan Wang
- School of Computer Science and Technology, Harbin Institute of Technology, Harbin, China
| | - Shuo Li
- Department of Biomedical Engineering, Case Western Reserve University, Cleveland, OH 44106, USA.
| |
Collapse
|
16
|
Wu J, Guo D, Wang L, Yang S, Zheng Y, Shapey J, Vercauteren T, Bisdas S, Bradford R, Saeed S, Kitchen N, Ourselin S, Zhang S, Wang G. TISS-net: Brain tumor image synthesis and segmentation using cascaded dual-task networks and error-prediction consistency. Neurocomputing 2023; 544:None. [PMID: 37528990 PMCID: PMC10243514 DOI: 10.1016/j.neucom.2023.126295] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2022] [Revised: 03/15/2023] [Accepted: 04/30/2023] [Indexed: 08/03/2023]
Abstract
Accurate segmentation of brain tumors from medical images is important for diagnosis and treatment planning, and it often requires multi-modal or contrast-enhanced images. However, in practice some modalities of a patient may be absent. Synthesizing the missing modality has a potential for filling this gap and achieving high segmentation performance. Existing methods often treat the synthesis and segmentation tasks separately or consider them jointly but without effective regularization of the complex joint model, leading to limited performance. We propose a novel brain Tumor Image Synthesis and Segmentation network (TISS-Net) that obtains the synthesized target modality and segmentation of brain tumors end-to-end with high performance. First, we propose a dual-task-regularized generator that simultaneously obtains a synthesized target modality and a coarse segmentation, which leverages a tumor-aware synthesis loss with perceptibility regularization to minimize the high-level semantic domain gap between synthesized and real target modalities. Based on the synthesized image and the coarse segmentation, we further propose a dual-task segmentor that predicts a refined segmentation and error in the coarse segmentation simultaneously, where a consistency between these two predictions is introduced for regularization. Our TISS-Net was validated with two applications: synthesizing FLAIR images for whole glioma segmentation, and synthesizing contrast-enhanced T1 images for Vestibular Schwannoma segmentation. Experimental results showed that our TISS-Net largely improved the segmentation accuracy compared with direct segmentation from the available modalities, and it outperformed state-of-the-art image synthesis-based segmentation methods.
Collapse
Affiliation(s)
- Jianghao Wu
- School of Mechanical and Electrical Engineering, University of Electronic Science and Technology of China, Chengdu, China
| | - Dong Guo
- School of Mechanical and Electrical Engineering, University of Electronic Science and Technology of China, Chengdu, China
| | - Lu Wang
- School of Mechanical and Electrical Engineering, University of Electronic Science and Technology of China, Chengdu, China
| | - Shuojue Yang
- School of Mechanical and Electrical Engineering, University of Electronic Science and Technology of China, Chengdu, China
| | - Yuanjie Zheng
- School of Information Science and Engineering, Shandong Normal University, Jinan, China
| | - Jonathan Shapey
- School of Biomedical Engineering & Imaging Sciences, King’s College London, London, UK
- Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, London, UK
- Department of Neurosurgery, National Hospital for Neurology and Neurosurgery, London, UK
| | - Tom Vercauteren
- School of Biomedical Engineering & Imaging Sciences, King’s College London, London, UK
| | - Sotirios Bisdas
- Department of Neuroradiology, National Hospital for Neurology and Neurosurgery, London, UK
| | - Robert Bradford
- Department of Neurosurgery, National Hospital for Neurology and Neurosurgery, London, UK
| | - Shakeel Saeed
- Department of Neurosurgery, National Hospital for Neurology and Neurosurgery, London, UK
| | - Neil Kitchen
- Department of Neurosurgery, National Hospital for Neurology and Neurosurgery, London, UK
| | - Sebastien Ourselin
- School of Biomedical Engineering & Imaging Sciences, King’s College London, London, UK
| | - Shaoting Zhang
- School of Mechanical and Electrical Engineering, University of Electronic Science and Technology of China, Chengdu, China
- SenseTime Research, Shanghai, China
| | - Guotai Wang
- School of Mechanical and Electrical Engineering, University of Electronic Science and Technology of China, Chengdu, China
- Shanghai Artificial Intelligence Laboratory, Shanghai, China
| |
Collapse
|
17
|
GMetaNet: Multi-scale ghost convolutional neural network with auxiliary MetaFormer decoding path for brain tumor segmentation. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2023.104694] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/22/2023]
|
18
|
Li X, Jiang Y, Li M, Zhang J, Yin S, Luo H. MSFR-Net: Multi-modality and single-modality feature recalibration network for brain tumor segmentation. Med Phys 2023; 50:2249-2262. [PMID: 35962724 DOI: 10.1002/mp.15933] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2022] [Revised: 05/16/2022] [Accepted: 06/14/2022] [Indexed: 11/11/2022] Open
Abstract
BACKGROUND Accurate and automated brain tumor segmentation from multi-modality MR images plays a significant role in tumor treatment. However, the existing approaches mainly focus on the fusion of multi-modality while ignoring the correlation between single-modality and tumor subcomponents. For example, T2-weighted images show good visualization of edema, and T1-contrast images have a good contrast between enhancing tumor core and necrosis. In the actual clinical process, professional physicians also label tumors according to these characteristics. We design a method for brain tumors segmentation that utilizes both multi-modality fusion and single-modality characteristics. METHODS A multi-modality and single-modality feature recalibration network (MSFR-Net) is proposed for brain tumor segmentation from MR images. Specifically, multi-modality information and single-modality information are assigned to independent pathways. Multi-modality network explicitly learns the relationship between all modalities and all tumor sub-components. Single-modality network learns the relationship between single-modality and its highly correlated tumor subcomponents. Then, a dual recalibration module (DRM) is designed to connect the parallel single-modality network and multi-modality network at multiple stages. The function of the DRM is to unify the two types of features into the same feature space. RESULTS Experiments on BraTS 2015 dataset and BraTS 2018 dataset show that the proposed method is competitive and superior to other state-of-the-art methods. The proposed method achieved the segmentation results with Dice coefficients of 0.86 and Hausdorff distance of 4.82 on BraTS 2018 dataset, with dice coefficients of 0.80, positive predictive value of 0.76, and sensitivity of 0.78 on BraTS 2015 dataset. CONCLUSIONS This work combines the manual labeling process of doctors and introduces the correlation between single-modality and the tumor subcomponents into the segmentation network. The method improves the segmentation performance of brain tumors and can be applied in the clinical practice. The code of the proposed method is available at: https://github.com/xiangQAQ/MSFR-Net.
Collapse
Affiliation(s)
- Xiang Li
- Department of Control Science and Engineering, Harbin Institute of Technology, Harbin, China
| | - Yuchen Jiang
- Department of Control Science and Engineering, Harbin Institute of Technology, Harbin, China
| | - Minglei Li
- Department of Control Science and Engineering, Harbin Institute of Technology, Harbin, China
| | - Jiusi Zhang
- Department of Control Science and Engineering, Harbin Institute of Technology, Harbin, China
| | - Shen Yin
- Department of Mechanical and Industrial Engineering, Faculty of Engineering, Norwegian University of Science and Technology, Trondheim, Norway
| | - Hao Luo
- Department of Control Science and Engineering, Harbin Institute of Technology, Harbin, China
| |
Collapse
|
19
|
Balasundaram A, Kavitha MS, Pratheepan Y, Akshat D, Kaushik MV. A Foreground Prototype-Based One-Shot Segmentation of Brain Tumors. Diagnostics (Basel) 2023; 13:diagnostics13071282. [PMID: 37046500 PMCID: PMC10093064 DOI: 10.3390/diagnostics13071282] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2023] [Revised: 03/07/2023] [Accepted: 03/16/2023] [Indexed: 03/30/2023] Open
Abstract
The potential for enhancing brain tumor segmentation with few-shot learning is enormous. While several deep learning networks (DNNs) show promising segmentation results, they all take a substantial amount of training data in order to yield appropriate results. Moreover, a prominent problem for most of these models is to perform well in unseen classes. To overcome these challenges, we propose a one-shot learning model to segment brain tumors on brain magnetic resonance images (MRI) based on a single prototype similarity score. With the use of recently developed few-shot learning techniques, where training and testing are carried out utilizing support and query sets of images, we attempt to acquire a definitive tumor region by focusing on slices containing foreground classes. It is unlike other recent DNNs that employed the entire set of images. The training of this model is carried out in an iterative manner where in each iteration, random slices containing foreground classes of randomly sampled data are selected as the query set, along with a different random slice from the same sample as the support set. In order to differentiate query images from class prototypes, we used a metric learning-based approach based on non-parametric thresholds. We employed the multimodal Brain Tumor Image Segmentation (BraTS) 2021 dataset with 60 training images and 350 testing images. The effectiveness of the model is evaluated using the mean dice score and mean IoU score. The experimental results provided a dice score of 83.42 which was greater than other works in the literature. Additionally, the proposed one-shot segmentation model outperforms the conventional methods in terms of computational time, memory usage, and the number of data.
Collapse
Affiliation(s)
- Ananthakrishnan Balasundaram
- School of Computer Science and Engineering, Vellore Institute of Technology, Chennai 600127, Tamil Nadu, India; (A.B.)
| | - Muthu Subash Kavitha
- School of Information and Data Sciences, Nagasaki University, Nagasaki 852-8521, Japan
- Correspondence:
| | - Yogarajah Pratheepan
- School of Computing, Engineering and Intelligent System, Ulster University, Londonderry BT48 7JL, UK;
| | - Dhamale Akshat
- School of Computer Science and Engineering, Vellore Institute of Technology, Chennai 600127, Tamil Nadu, India; (A.B.)
| | - Maddirala Venkata Kaushik
- School of Computer Science and Engineering, Vellore Institute of Technology, Chennai 600127, Tamil Nadu, India; (A.B.)
| |
Collapse
|
20
|
Tong J, Wang C. A dual tri-path CNN system for brain tumor segmentation. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104411] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|
21
|
Rath A, Mohanty DK, Mishra BSP, Bagal DK. A Bibliometric Review: Brain Tumor Magnetic Resonance Imagings Using Different Convolutional Neural Network Architectures. World Neurosurg 2023; 170:e681-e694. [PMID: 36442778 DOI: 10.1016/j.wneu.2022.11.091] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/08/2022] [Revised: 11/20/2022] [Accepted: 11/21/2022] [Indexed: 11/27/2022]
Abstract
BACKGROUND Numerous scientists and researchers have been developing advanced procedures and methods for diagnosing the kind and phase of a human tumor. Brain tumors, which are neoplastic and abnormal developments of brain cells, are one of the most prominent causes of death. Brain tumors, also known as lesions or neoplasia, may be roughly classified as either primary or metastatic. Primary brain tumors arise from brain tissue and its surrounding environment. The recognition of brain tumors using magnetic resonance images via a deep learning technique such as convolutional neural network (CNN) has garnered significant academic interest over the last few decades. METHODS In this study, a detailed evaluation based on bibliometrics is considered in order to synthesize and organize the available academic literature and to identify current research trends and hotspots. We used bibliometric methodologies and a literature review for the CNN-based brain tumor to synthesize and evaluate prior studies. RESULTS For this bibliometric analysis, we applied the Visualization of Similarity Viewer program to classify the major publications, notable journals, financial sponsors, and affiliations. CONCLUSIONS In conclusion, we suggest that one of the next paths of study will be the incorporation of other databases to advance CNN-based brain tumor identification from magnetic resonance images. No drug dosages are applied in this work.
Collapse
Affiliation(s)
- Arati Rath
- School of Computer Engineering, KIIT Deemed to be University, Odisha, India.
| | | | | | | |
Collapse
|
22
|
Zhao Y, Wang X, Che T, Bao G, Li S. Multi-task deep learning for medical image computing and analysis: A review. Comput Biol Med 2023; 153:106496. [PMID: 36634599 DOI: 10.1016/j.compbiomed.2022.106496] [Citation(s) in RCA: 24] [Impact Index Per Article: 12.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/21/2022] [Revised: 12/06/2022] [Accepted: 12/27/2022] [Indexed: 12/29/2022]
Abstract
The renaissance of deep learning has provided promising solutions to various tasks. While conventional deep learning models are constructed for a single specific task, multi-task deep learning (MTDL) that is capable to simultaneously accomplish at least two tasks has attracted research attention. MTDL is a joint learning paradigm that harnesses the inherent correlation of multiple related tasks to achieve reciprocal benefits in improving performance, enhancing generalizability, and reducing the overall computational cost. This review focuses on the advanced applications of MTDL for medical image computing and analysis. We first summarize four popular MTDL network architectures (i.e., cascaded, parallel, interacted, and hybrid). Then, we review the representative MTDL-based networks for eight application areas, including the brain, eye, chest, cardiac, abdomen, musculoskeletal, pathology, and other human body regions. While MTDL-based medical image processing has been flourishing and demonstrating outstanding performance in many tasks, in the meanwhile, there are performance gaps in some tasks, and accordingly we perceive the open challenges and the perspective trends. For instance, in the 2018 Ischemic Stroke Lesion Segmentation challenge, the reported top dice score of 0.51 and top recall of 0.55 achieved by the cascaded MTDL model indicate further research efforts in high demand to escalate the performance of current models.
Collapse
Affiliation(s)
- Yan Zhao
- Beijing Advanced Innovation Center for Biomedical Engineering, School of Biological Science and Medical Engineering, Beihang University, Beijing, 100083, China
| | - Xiuying Wang
- School of Computer Science, The University of Sydney, Sydney, NSW, 2008, Australia.
| | - Tongtong Che
- Beijing Advanced Innovation Center for Biomedical Engineering, School of Biological Science and Medical Engineering, Beihang University, Beijing, 100083, China
| | - Guoqing Bao
- School of Computer Science, The University of Sydney, Sydney, NSW, 2008, Australia
| | - Shuyu Li
- State Key Laboratory of Cognitive Neuroscience and Learning, Beijing Normal University, Beijing, 100875, China.
| |
Collapse
|
23
|
Yang D, Li Y, Yu J. Multi-task thyroid tumor segmentation based on the joint loss function. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104249] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
|
24
|
Deng Y, Wang H, Hou Y, Liang S, Zeng D. LFU-Net: A Lightweight U-Net with Full Skip Connections for Medical Image Segmentation. Curr Med Imaging 2023; 19:347-360. [PMID: 35733312 DOI: 10.2174/1573405618666220622154853] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/02/2022] [Revised: 04/21/2022] [Accepted: 04/26/2022] [Indexed: 11/22/2022]
Abstract
BACKGROUND In the series of improved versions of U-Net, while the segmentation accuracy continues to improve, the number of parameters does not change, which makes the hardware required for training expensive, thus affecting the speed of training convergence. OBJECTIVE The objective of this study is to propose a lightweight U-Net to balance the relationship between the parameters and the segmentation accuracy. METHODS A lightweight U-Net with full skip connections and deep supervision (LFU-Net) was proposed. The full skip connections include skip connections from shallow encoders, deep decoders, and sub-networks, while the deep supervision learns hierarchical representations from full-resolution feature representations in outputs of sub-networks. The key lightweight design is that the number of output channels is based on 8 rather than 64 or 32. Its pruning scheme was designed to further reduce parameters. The code is available at: https://github.com/dengdy22/U-Nets. RESULTS For the ISBI LiTS 2017 Challenge validation dataset, the LFU-Net with no pruning received a Dice value of 0.9699, which achieved equal or better performance with a mere about 1% of the parameters of existing networks. For the BraTS 2018 validation dataset, its Dice values were 0.8726, 0.9363, 0.8699 and 0.8116 on average, WT, TC and ET, respectively, and its Hausdorff95 distances values were 3.9514, 4.3960, 3.0607 and 4.3975, respectively, which was not inferior to the existing networks and showed that it can achieve balanced recognition of each region. CONCLUSION LFU-Net can be used as a lightweight and effective method in the segmentation tasks of two and multiple classification medical imaging datasets.
Collapse
Affiliation(s)
- Yunjiao Deng
- School of Mechanical Engineering, Yanshan University, Qinhuangdao 066004, China
| | - Hui Wang
- School of Mechanical Engineering, Yanshan University, Qinhuangdao 066004, China
| | - Yulei Hou
- School of Mechanical Engineering, Yanshan University, Qinhuangdao 066004, China
| | - Shunpan Liang
- School of Information Science and Engineering, Yanshan University, Qinhuangdao 066004, China
| | - Daxing Zeng
- School of Mechanical Engineering, Dongguan University of Technology, Dongguan 523015, China
| |
Collapse
|
25
|
Rehman MU, Ryu J, Nizami IF, Chong KT. RAAGR2-Net: A brain tumor segmentation network using parallel processing of multiple spatial frames. Comput Biol Med 2023; 152:106426. [PMID: 36565485 DOI: 10.1016/j.compbiomed.2022.106426] [Citation(s) in RCA: 16] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2022] [Revised: 11/16/2022] [Accepted: 12/13/2022] [Indexed: 12/24/2022]
Abstract
Brain tumors are one of the most fatal cancers. Magnetic Resonance Imaging (MRI) is a non-invasive method that provides multi-modal images containing important information regarding the tumor. Many contemporary techniques employ four modalities: T1-weighted (T1), T1-weighted with contrast (T1c), T2-weighted (T2), and fluid-attenuation-inversion-recovery (FLAIR), each of which provides unique and important characteristics for the location of each tumor. Although several modern procedures provide decent segmentation results on the multimodal brain tumor image segmentation benchmark (BraTS) dataset, they lack performance when evaluated simultaneously on all the regions of MRI images. Furthermore, there is still room for improvement due to parameter limitations and computational complexity. Therefore, in this work, a novel encoder-decoder-based architecture is proposed for the effective segmentation of brain tumor regions. Data pre-processing is performed by applying N4 bias field correction, z-score, and 0 to 1 resampling to facilitate model training. To minimize the loss of location information in different modules, a residual spatial pyramid pooling (RASPP) module is proposed. RASPP is a set of parallel layers using dilated convolution. In addition, an attention gate (AG) module is used to efficiently emphasize and restore the segmented output from extracted feature maps. The proposed modules attempt to acquire rich feature representations by combining knowledge from diverse feature maps and retaining their local information. The performance of the proposed deep network based on RASPP, AG, and recursive residual (R2) block termed RAAGR2-Net is evaluated on the BraTS benchmarks. The experimental results show that the suggested network outperforms existing networks that exhibit the usefulness of the proposed modules for "fine" segmentation. The code for this work is made available online at: https://github.com/Rehman1995/RAAGR2-Net.
Collapse
Affiliation(s)
- Mobeen Ur Rehman
- Department of Electronics and Information Engineering, Jeonbuk National University, Jeonju 54896, South Korea.
| | - Jihyoung Ryu
- Electronics and Telecommunications Research Institute, 176-11 Cheomdan Gwagi-ro, Buk-gu, Gwangju 61012, Republic of Korea.
| | - Imran Fareed Nizami
- Department of Electrical Engineering, Bahria University, Islamabad, Pakistan.
| | - Kil To Chong
- Department of Electronics and Information Engineering, Jeonbuk National University, Jeonju 54896, South Korea; Advances Electronics and Information Research Center, Jeonbuk National University, Jeonju 54896, South Korea.
| |
Collapse
|
26
|
Wang J, Zheng Y, Ma J, Li X, Wang C, Gee J, Wang H, Huang W. Information bottleneck-based interpretable multitask network for breast cancer classification and segmentation. Med Image Anal 2023; 83:102687. [PMID: 36436356 DOI: 10.1016/j.media.2022.102687] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2021] [Revised: 09/19/2022] [Accepted: 11/07/2022] [Indexed: 11/13/2022]
Abstract
Breast cancer is one of the most common causes of death among women worldwide. Early signs of breast cancer can be an abnormality depicted on breast images (e.g., mammography or breast ultrasonography). However, reliable interpretation of breast images requires intensive labor and physicians with extensive experience. Deep learning is evolving breast imaging diagnosis by introducing a second opinion to physicians. However, most deep learning-based breast cancer analysis algorithms lack interpretability because of their black box nature, which means that domain experts cannot understand why the algorithms predict a label. In addition, most deep learning algorithms are formulated as a single-task-based model that ignores correlations between different tasks (e.g., tumor classification and segmentation). In this paper, we propose an interpretable multitask information bottleneck network (MIB-Net) to accomplish simultaneous breast tumor classification and segmentation. MIB-Net maximizes the mutual information between the latent representations and class labels while minimizing information shared by the latent representations and inputs. In contrast from existing models, our MIB-Net generates a contribution score map that offers an interpretable aid for physicians to understand the model's decision-making process. In addition, MIB-Net implements multitask learning and further proposes a dual prior knowledge guidance strategy to enhance deep task correlation. Our evaluations are carried out on three breast image datasets in different modalities. Our results show that the proposed framework is not only able to help physicians better understand the model's decisions but also improve breast tumor classification and segmentation accuracy over representative state-of-the-art models. Our code is available at https://github.com/jxw0810/MIB-Net.
Collapse
Affiliation(s)
- Junxia Wang
- School of Information Science and Engineering, Shandong Normal University, No. 1 Daxue Road, Changqing District, Jinan 250358, China
| | - Yuanjie Zheng
- School of Information Science and Engineering, Shandong Normal University, No. 1 Daxue Road, Changqing District, Jinan 250358, China; Shanghai AI Laboratory, No. 701 Yunjin Road, Xuhui District, Shanghai, 200433, China.
| | - Jun Ma
- School of Cyber Science and Engineering, Southeast University, No. 2 Southeast University Road, Jiangning District, Nanjing, 211189, China
| | - Xinmeng Li
- School of Information Science and Engineering, Shandong Normal University, No. 1 Daxue Road, Changqing District, Jinan 250358, China
| | - Chongjing Wang
- China Academy of Information and Communications Technology, No. 52 Huayuan North Road, Haidian District, Beijing 100191, China
| | - James Gee
- Penn Image Computing and Science Laboratory, University of Pennsylvania, PA 19104, USA
| | - Haipeng Wang
- Institute of Information Fusion, Naval Aviation University, Erma Road Yantai Shandong, Yantai 264001, China.
| | - Wenhui Huang
- School of Information Science and Engineering, Shandong Normal University, No. 1 Daxue Road, Changqing District, Jinan 250358, China.
| |
Collapse
|
27
|
Gao H, Miao Q, Ma D, Liua R. Deep Mutual Learning for Brain Tumor Segmentation with the Fusion Network. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.11.038] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|
28
|
AboElenein NM, Piao S, Zhang Z. Encoder–Decoder Network with Depthwise Atrous Spatial Pyramid Pooling for Automatic Brain Tumor Segmentation. Neural Process Lett 2022. [DOI: 10.1007/s11063-022-10959-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
29
|
Kihira S, Mei X, Mahmoudi K, Liu Z, Dogra S, Belani P, Tsankova N, Hormigo A, Fayad ZA, Doshi A, Nael K. U-Net Based Segmentation and Characterization of Gliomas. Cancers (Basel) 2022; 14:4457. [PMID: 36139616 PMCID: PMC9496685 DOI: 10.3390/cancers14184457] [Citation(s) in RCA: 13] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2022] [Revised: 09/12/2022] [Accepted: 09/13/2022] [Indexed: 11/18/2022] Open
Abstract
(1) Background: Gliomas are the most common primary brain neoplasms accounting for roughly 40−50% of all malignant primary central nervous system tumors. We aim to develop a deep learning-based framework for automated segmentation and prediction of biomarkers and prognosis in patients with gliomas. (2) Methods: In this retrospective two center study, patients were included if they (1) had a diagnosis of glioma with known surgical histopathology and (2) had preoperative MRI with FLAIR sequence. The entire tumor volume including FLAIR hyperintense infiltrative component and necrotic and cystic components was segmented. Deep learning-based U-Net framework was developed based on symmetric architecture from the 512 × 512 segmented maps from FLAIR as the ground truth mask. (3) Results: The final cohort consisted of 208 patients with mean ± standard deviation of age (years) of 56 ± 15 with M/F of 130/78. DSC of the generated mask was 0.93. Prediction for IDH-1 and MGMT status had a performance of AUC 0.88 and 0.62, respectively. Survival prediction of <18 months demonstrated AUC of 0.75. (4) Conclusions: Our deep learning-based framework can detect and segment gliomas with excellent performance for the prediction of IDH-1 biomarker status and survival.
Collapse
Affiliation(s)
- Shingo Kihira
- Department of Radiology, Icahn School of Medicine at Mount Sinai, New York, NY 10029, USA
- Department of Radiological Sciences, David Geffen School of Medicine at University of California Los Angeles, Los Angeles, CA 90033, USA
| | - Xueyan Mei
- Biomedical Engineering and Imaging Institute, Icahn School of Medicine at Mount Sinai, New York, NY 10029, USA
| | - Keon Mahmoudi
- Department of Radiological Sciences, David Geffen School of Medicine at University of California Los Angeles, Los Angeles, CA 90033, USA
| | - Zelong Liu
- Biomedical Engineering and Imaging Institute, Icahn School of Medicine at Mount Sinai, New York, NY 10029, USA
| | - Siddhant Dogra
- Department of Radiology, Icahn School of Medicine at Mount Sinai, New York, NY 10029, USA
| | - Puneet Belani
- Department of Radiology, Icahn School of Medicine at Mount Sinai, New York, NY 10029, USA
| | - Nadejda Tsankova
- Department of Pathology, Icahn School of Medicine at Mount Sinai, New York, NY 10029, USA
| | - Adilia Hormigo
- Department of Pathology, Icahn School of Medicine at Mount Sinai, New York, NY 10029, USA
| | - Zahi A. Fayad
- Department of Radiology, Icahn School of Medicine at Mount Sinai, New York, NY 10029, USA
- Biomedical Engineering and Imaging Institute, Icahn School of Medicine at Mount Sinai, New York, NY 10029, USA
| | - Amish Doshi
- Department of Radiology, Icahn School of Medicine at Mount Sinai, New York, NY 10029, USA
| | - Kambiz Nael
- Department of Radiology, Icahn School of Medicine at Mount Sinai, New York, NY 10029, USA
- Department of Radiological Sciences, David Geffen School of Medicine at University of California Los Angeles, Los Angeles, CA 90033, USA
| |
Collapse
|
30
|
Liu Y, Mu F, Shi Y, Cheng J, Li C, Chen X. Brain tumor segmentation in multimodal MRI via pixel-level and feature-level image fusion. Front Neurosci 2022; 16:1000587. [PMID: 36188482 PMCID: PMC9515796 DOI: 10.3389/fnins.2022.1000587] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/22/2022] [Accepted: 08/18/2022] [Indexed: 11/30/2022] Open
Abstract
Brain tumor segmentation in multimodal MRI volumes is of great significance to disease diagnosis, treatment planning, survival prediction and other relevant tasks. However, most existing brain tumor segmentation methods fail to make sufficient use of multimodal information. The most common way is to simply stack the original multimodal images or their low-level features as the model input, and many methods treat each modality data with equal importance to a given segmentation target. In this paper, we introduce multimodal image fusion technique including both pixel-level fusion and feature-level fusion for brain tumor segmentation, aiming to achieve more sufficient and finer utilization of multimodal information. At the pixel level, we present a convolutional network named PIF-Net for 3D MR image fusion to enrich the input modalities of the segmentation model. The fused modalities can strengthen the association among different types of pathological information captured by multiple source modalities, leading to a modality enhancement effect. At the feature level, we design an attention-based modality selection feature fusion (MSFF) module for multimodal feature refinement to address the difference among multiple modalities for a given segmentation target. A two-stage brain tumor segmentation framework is accordingly proposed based on the above components and the popular V-Net model. Experiments are conducted on the BraTS 2019 and BraTS 2020 benchmarks. The results demonstrate that the proposed components on both pixel-level and feature-level fusion can effectively improve the segmentation accuracy of brain tumors.
Collapse
Affiliation(s)
- Yu Liu
- Department of Biomedical Engineering, Hefei University of Technology, Hefei, China
- Anhui Province Key Laboratory of Measuring Theory and Precision Instrument, Hefei University of Technology, Hefei, China
| | - Fuhao Mu
- Department of Biomedical Engineering, Hefei University of Technology, Hefei, China
| | - Yu Shi
- Department of Biomedical Engineering, Hefei University of Technology, Hefei, China
| | - Juan Cheng
- Department of Biomedical Engineering, Hefei University of Technology, Hefei, China
- Anhui Province Key Laboratory of Measuring Theory and Precision Instrument, Hefei University of Technology, Hefei, China
| | - Chang Li
- Department of Biomedical Engineering, Hefei University of Technology, Hefei, China
- Anhui Province Key Laboratory of Measuring Theory and Precision Instrument, Hefei University of Technology, Hefei, China
| | - Xun Chen
- Department of Electronic Engineering and Information Science, University of Science and Technology of China, Hefei, China
- *Correspondence: Xun Chen
| |
Collapse
|
31
|
|
32
|
Mazumdar I, Mukherjee J. Fully automatic MRI brain tumor segmentation using efficient spatial attention convolutional networks with composite loss. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.05.050] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|
33
|
Agnostic multimodal brain anomalies detection using a novel single-structured framework for better patient diagnosis and therapeutic planning in clinical oncology. Biomed Signal Process Control 2022. [DOI: 10.1016/j.bspc.2022.103786] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022]
|
34
|
Abstract
AbstractBrain tumor segmentation is one of the most challenging problems in medical image analysis. The goal of brain tumor segmentation is to generate accurate delineation of brain tumor regions. In recent years, deep learning methods have shown promising performance in solving various computer vision problems, such as image classification, object detection and semantic segmentation. A number of deep learning based methods have been applied to brain tumor segmentation and achieved promising results. Considering the remarkable breakthroughs made by state-of-the-art technologies, we provide this survey with a comprehensive study of recently developed deep learning based brain tumor segmentation techniques. More than 150 scientific papers are selected and discussed in this survey, extensively covering technical aspects such as network architecture design, segmentation under imbalanced conditions, and multi-modality processes. We also provide insightful discussions for future development directions.
Collapse
|
35
|
Niyas S, Pawan S, Anand Kumar M, Rajan J. Medical image segmentation with 3D convolutional neural networks: A survey. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.04.065] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/17/2022]
|
36
|
Kafiyan-Safari M, Rouhani M. Adaptive one-pass passive-aggressive radial basis function for classification problems. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.03.047] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
|
37
|
Zhou J, Ye J, Liang Y, Zhao J, Wu Y, Luo S, Lai X, Wang J. scSE-NL V-Net: A Brain Tumor Automatic Segmentation Method Based on Spatial and Channel "Squeeze-and-Excitation" Network With Non-local Block. Front Neurosci 2022; 16:916818. [PMID: 35712454 PMCID: PMC9197379 DOI: 10.3389/fnins.2022.916818] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/10/2022] [Accepted: 04/27/2022] [Indexed: 11/23/2022] Open
Abstract
Intracranial tumors are commonly known as brain tumors, which can be life-threatening in severe cases. Magnetic resonance imaging (MRI) is widely used in diagnosing brain tumors because of its harmless to the human body and high image resolution. Due to the heterogeneity of brain tumor height, MRI imaging is exceptionally irregular. How to accurately and quickly segment brain tumor MRI images is still one of the hottest topics in the medical image analysis community. However, according to the brain tumor segmentation algorithms, we could find now, most segmentation algorithms still stay in two-dimensional (2D) image segmentation, which could not obtain the spatial dependence between features effectively. In this study, we propose a brain tumor automatic segmentation method called scSE-NL V-Net. We try to use three-dimensional (3D) data as the model input and process the data by 3D convolution to get some relevance between dimensions. Meanwhile, we adopt non-local block as the self-attention block, which can reduce inherent image noise interference and make up for the lack of spatial dependence due to convolution. To improve the accuracy of convolutional neural network (CNN) image recognition, we add the "Spatial and Channel Squeeze-and-Excitation" Network (scSE-Net) to V-Net. The dataset used in this paper is from the brain tumor segmentation challenge 2020 database. In the test of the official BraTS2020 verification set, the Dice similarity coefficient is 0.65, 0.82, and 0.76 for the enhanced tumor (ET), whole tumor (WT), and tumor core (TC), respectively. Thereby, our model can make an auxiliary effect on the diagnosis of brain tumors established.
Collapse
Affiliation(s)
- Juhua Zhou
- School of Medical Technology and Information Engineering, Zhejiang Chinese Medical University, Hangzhou, China
| | - Jianming Ye
- The First Affiliated Hospital, Gannan Medical University, Ganzhou, China
| | - Yu Liang
- School of Medical Technology and Information Engineering, Zhejiang Chinese Medical University, Hangzhou, China
| | - Jialu Zhao
- School of Medical Technology and Information Engineering, Zhejiang Chinese Medical University, Hangzhou, China
| | - Yan Wu
- School of Medical Technology and Information Engineering, Zhejiang Chinese Medical University, Hangzhou, China
| | - Siyuan Luo
- School of Medical Technology and Information Engineering, Zhejiang Chinese Medical University, Hangzhou, China
| | - Xiaobo Lai
- School of Medical Technology and Information Engineering, Zhejiang Chinese Medical University, Hangzhou, China
| | - Jianqing Wang
- School of Medical Technology and Information Engineering, Zhejiang Chinese Medical University, Hangzhou, China
| |
Collapse
|
38
|
Kouli O, Hassane A, Badran D, Kouli T, Hossain-Ibrahim K, Steele JD. Automated brain tumour identification using magnetic resonance imaging: a systematic review and meta-analysis. Neurooncol Adv 2022; 4:vdac081. [PMID: 35769411 PMCID: PMC9234754 DOI: 10.1093/noajnl/vdac081] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/23/2022] Open
Abstract
Background Automated brain tumor identification facilitates diagnosis and treatment planning. We evaluate the performance of traditional machine learning (TML) and deep learning (DL) in brain tumor detection and segmentation, using MRI. Methods A systematic literature search from January 2000 to May 8, 2021 was conducted. Study quality was assessed using the Checklist for Artificial Intelligence in Medical Imaging (CLAIM). Detection meta-analysis was performed using a unified hierarchical model. Segmentation studies were evaluated using a random effects model. Sensitivity analysis was performed for externally validated studies. Results Of 224 studies included in the systematic review, 46 segmentation and 38 detection studies were eligible for meta-analysis. In detection, DL achieved a lower false positive rate compared to TML; 0.018 (95% CI, 0.011 to 0.028) and 0.048 (0.032 to 0.072) (P < .001), respectively. In segmentation, DL had a higher dice similarity coefficient (DSC), particularly for tumor core (TC); 0.80 (0.77 to 0.83) and 0.63 (0.56 to 0.71) (P < .001), persisting on sensitivity analysis. Both manual and automated whole tumor (WT) segmentation had “good” (DSC ≥ 0.70) performance. Manual TC segmentation was superior to automated; 0.78 (0.69 to 0.86) and 0.64 (0.53 to 0.74) (P = .014), respectively. Only 30% of studies reported external validation. Conclusions The comparable performance of automated to manual WT segmentation supports its integration into clinical practice. However, manual outperformance for sub-compartmental segmentation highlights the need for further development of automated methods in this area. Compared to TML, DL provided superior performance for detection and sub-compartmental segmentation. Improvements in the quality and design of studies, including external validation, are required for the interpretability and generalizability of automated models.
Collapse
Affiliation(s)
- Omar Kouli
- School of Medicine, University of Dundee , Dundee UK
- NHS Greater Glasgow and Clyde , Dundee UK
| | | | | | - Tasnim Kouli
- School of Medicine, University of Dundee , Dundee UK
| | | | - J Douglas Steele
- Division of Imaging Science and Technology, School of Medicine, University of Dundee , UK
| |
Collapse
|
39
|
Siddiqi MH, Alsayat A, Alhwaiti Y, Azad M, Alruwaili M, Alanazi S, Kamruzzaman MM, Khan A. A Precise Medical Imaging Approach for Brain MRI Image Classification. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2022; 2022:6447769. [PMID: 35548099 PMCID: PMC9085323 DOI: 10.1155/2022/6447769] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/12/2022] [Accepted: 04/12/2022] [Indexed: 11/22/2022]
Abstract
Magnetic resonance imaging (MRI) is an accurate and noninvasive method employed for the diagnosis of various kinds of diseases in medical imaging. Most of the existing systems showed significant performances on small MRI datasets, while their performances decrease against large MRI datasets. Hence, the goal was to design an efficient and robust classification system that sustains a high recognition rate against large MRI dataset. Accordingly, in this study, we have proposed the usage of a novel feature extraction technique that has the ability to extract and select the prominent feature from MRI image. The proposed algorithm selects the best features from the MRI images of various diseases. Further, this approach discriminates various classes based on recursive values such as partial Z-value. The proposed approach only extracts a minor feature set through, respectively, forward and backward recursion models. The most interrelated features are nominated in the forward regression model that depends on the values of partial Z-test, while the minimum interrelated features are diminished from the corresponding feature space under the presence of the backward model. In both cases, the values of Z-test are estimated through the defined labels of the diseases. The proposed model is efficiently looking the localized features, which is one of the benefits of this method. After extracting and selecting the best features, the model is trained by utilizing support vector machine (SVM) to provide the predicted labels to the corresponding MRI images. To show the significance of the proposed model, we utilized a publicly available standard dataset such as Harvard Medical School and Open Access Series of Imaging Studies (OASIS), which contains 24 various brain diseases including normal. The proposed approach achieved the best classification accuracy against existing state-of-the-art systems.
Collapse
Affiliation(s)
| | - Ahmed Alsayat
- College of Computer and Information Sciences, Jouf University, Sakaka, Aljouf, Saudi Arabia
| | - Yousef Alhwaiti
- College of Computer and Information Sciences, Jouf University, Sakaka, Aljouf, Saudi Arabia
| | - Mohammad Azad
- College of Computer and Information Sciences, Jouf University, Sakaka, Aljouf, Saudi Arabia
| | - Madallah Alruwaili
- College of Computer and Information Sciences, Jouf University, Sakaka, Aljouf, Saudi Arabia
| | - Saad Alanazi
- College of Computer and Information Sciences, Jouf University, Sakaka, Aljouf, Saudi Arabia
| | - M. M. Kamruzzaman
- College of Computer and Information Sciences, Jouf University, Sakaka, Aljouf, Saudi Arabia
| | - Asfandyar Khan
- Institute of Computer Science & IT, The University of Agriculture Peshawar, Peshawar, Pakistan
| |
Collapse
|
40
|
Wu L, Hu S, Liu C. MR brain segmentation based on DE-ResUnet combining texture features and background knowledge. Biomed Signal Process Control 2022. [DOI: 10.1016/j.bspc.2022.103541] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/21/2022]
|
41
|
Zhang J, Jiang Z, Liu D, Sun Q, Hou Y, Liu B. 3D asymmetric expectation-maximization attention network for brain tumor segmentation. NMR IN BIOMEDICINE 2022; 35:e4657. [PMID: 34859922 DOI: 10.1002/nbm.4657] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/19/2021] [Revised: 10/23/2021] [Accepted: 11/02/2021] [Indexed: 06/13/2023]
Abstract
Automatic brain tumor segmentation on MRI is a prerequisite to provide a quantitative and intuitive assistance for clinical diagnosis and treatment. Meanwhile, 3D deep neural network related brain tumor segmentation models have demonstrated considerable accuracy improvement over corresponding 2D methodologies. However, 3D brain tumor segmentation models generally suffer from high computation cost. Motivated by a recently proposed 3D dilated multi-fiber network (DMF-Net) architecture that pays more attention to reduction of computation cost, we present in this work a novel encoder-decoder neural network, ie a 3D asymmetric expectation-maximization attention network (AEMA-Net), to automatically segment brain tumors. We modify DMF-Net by introducing an asymmetric convolution block into a multi-fiber unit and a dilated multi-fiber unit to capture more powerful deep features for the brain tumor segmentation. In addition, AEMA-Net further incorporates an expectation-maximization attention (EMA) module into the DMF-Net by embedding the EMA block in the third stage of skip connection, which focuses on capturing the long-range dependence of context. We extensively evaluate AEMA-Net on three MRI brain tumor segmentation benchmarks of BraTS 2018, 2019 and 2020 datasets. Experimental results demonstrate that AEMA-Net outperforms both 3D U-Net and DMF-Net, and it achieves competitive performance compared with the state-of-the-art brain tumor segmentation methods.
Collapse
Affiliation(s)
- Jianxin Zhang
- School of Computer Science and Engineering, Dalian Minzu University, Dalian, China
- Key Lab of Advanced Design and Intelligent Computing (Ministry of Education), Dalian University, Dalian, China
| | - Zongkang Jiang
- Key Lab of Advanced Design and Intelligent Computing (Ministry of Education), Dalian University, Dalian, China
| | - Dongwei Liu
- School of Computer Science and Engineering, Dalian Minzu University, Dalian, China
| | - Qiule Sun
- Faculty of Electronic Information and Electrical Engineering, Dalian University of Technology, Dalian, China
| | - Yaqing Hou
- Faculty of Electronic Information and Electrical Engineering, Dalian University of Technology, Dalian, China
| | - Bin Liu
- International School of Information Science and Engineering (DUT-RUISE), Dalian University of Technology, Dalian, China
| |
Collapse
|
42
|
Li J, Yu H, Chen C, Ding M, Zha S. Category guided attention network for brain tumor segmentation in MRI. Phys Med Biol 2022; 67. [DOI: 10.1088/1361-6560/ac628a] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/21/2021] [Accepted: 03/30/2022] [Indexed: 12/26/2022]
Abstract
Abstract
Objective. Magnetic resonance imaging (MRI) has been widely used for the analysis and diagnosis of brain diseases. Accurate and automatic brain tumor segmentation is of paramount importance for radiation treatment. However, low tissue contrast in tumor regions makes it a challenging task. Approach. We propose a novel segmentation network named Category Guided Attention U-Net (CGA U-Net). In this model, we design a Supervised Attention Module (SAM) based on the attention mechanism, which can capture more accurate and stable long-range dependency in feature maps without introducing much computational cost. Moreover, we propose an intra-class update approach to reconstruct feature maps by aggregating pixels of the same category. Main results. Experimental results on the BraTS 2019 datasets show that the proposed method outperformers the state-of-the-art algorithms in both segmentation performance and computational complexity. Significance. The CGA U-Net can effectively capture the global semantic information in the MRI image by using the SAM module, while significantly reducing the computational cost. Code is available at https://github.com/delugewalker/CGA-U-Net.
Collapse
|
43
|
Li S, Liu J, Song Z. Brain tumor segmentation based on region of interest-aided localization and segmentation U-Net. INT J MACH LEARN CYB 2022; 13:2435-2445. [PMID: 35378734 PMCID: PMC8967694 DOI: 10.1007/s13042-022-01536-4] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/10/2021] [Accepted: 02/24/2022] [Indexed: 11/30/2022]
Abstract
Since magnetic resonance imaging (MRI) has superior soft tissue contrast, contouring (brain) tumor accurately by MRI images is essential in medical image processing. Segmenting tumor accurately is immensely challenging, since tumor and normal tissues are often inextricably intertwined in the brain. It is also extremely time consuming manually. Late deep learning techniques start to show reasonable success in brain tumor segmentation automatically. The purpose of this study is to develop a new region-of-interest-aided (ROI-aided) deep learning technique for automatic brain tumor MRI segmentation. The method consists of two major steps. Step one is to use a 2D network with U-Net architecture to localize the tumor ROI, which is to reduce the impact of normal tissue’s disturbance. Then a 3D U-Net is performed in step 2 for tumor segmentation within identified ROI. The proposed method is validated on MICCAI BraTS 2015 Challenge with 220 high Gliomas grade (HGG) and 54 low Gliomas grade (LGG) patients’ data. The Dice similarity coefficient and the Hausdorff distance between the manual tumor contour and that segmented by the proposed method are 0.876 ±0.068 and 3.594±1.347 mm, respectively. These numbers are indications that our proposed method is an effective ROI-aided deep learning strategy for brain MRI tumor segmentation, and a valid and useful tool in medical image processing.
Collapse
Affiliation(s)
- Shidong Li
- Department of Mathematics, San Francisco University, San Francisco, CA 94132 USA
| | - Jianwei Liu
- School of Mathematics, Tianjin University, Tianjin, 300354 China
- Tianjin Key Laboratory of Brain-Inspired Intelligence Technology, Tianjin, 300072 China
| | - Zhanjie Song
- School of Mathematics, Tianjin University, Tianjin, 300354 China
- Tianjin Key Laboratory of Brain-Inspired Intelligence Technology, Tianjin, 300072 China
| |
Collapse
|
44
|
Zhang TC, Zhang J, Chen SC, Saada B. A Novel Prediction Model for Brain Glioma Image Segmentation Based on the Theory of Bose-Einstein Condensate. Front Med (Lausanne) 2022; 9:794125. [PMID: 35372409 PMCID: PMC8971582 DOI: 10.3389/fmed.2022.794125] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2021] [Accepted: 01/14/2022] [Indexed: 11/26/2022] Open
Abstract
Background The input image of a blurry glioma image segmentation is, usually, very unclear. It is difficult to obtain the accurate contour line of image segmentation. The main challenge facing the researchers is to correctly determine the area where the points on the contour line belong to the glioma image. This article highlights the mechanism of formation of glioma and provides an image segmentation prediction model to assist in the accurate division of glioma contour points. The proposed prediction model of segmentation associated with the process of the formation of glioma is innovative and challenging. Bose-Einstein Condensate (BEC) is a microscopic quantum phenomenon in which atoms condense to the ground state of energy as the temperature approaches absolute zero. In this article, we propose a BEC kernel function and a novel prediction model based on the BEC kernel to detect the relationship between the process of the BEC and the formation of a brain glioma. Furthermore, the theoretical derivation and proof of the prediction model are given from micro to macro through quantum mechanics, wave, oscillation of glioma, and statistical distribution of laws. The prediction model is a distinct segmentation model that is guided by BEC theory for blurry glioma image segmentation. Results Our approach is based on five tests. The first three tests aimed at confirming the measuring range of T and μ in the BEC kernel. The results are extended from −10 to 10, approximating the standard range to T ≤ 0, and μ from 0 to 6.7. Tests 4 and 5 are comparison tests. The comparison in Test 4 was based on various established cluster methods. The results show that our prediction model in image evaluation parameters of P, R, and F is the best amongst all the existent ten forms except for only one reference with the mean value of F that is between 0.88 and 0.93, while our approach returns between 0.85 and 0.99. Test 5 aimed to further compare our results, especially with CNN (Convolutional Neural Networks) methods, by challenging Brain Tumor Segmentation (BraTS) and clinic patient datasets. Our results were also better than all reference tests. In addition, the proposed prediction model with the BEC kernel is feasible and has a comparative validity in glioma image segmentation. Conclusions Theoretical derivation and experimental verification show that the prediction model based on the BEC kernel can solve the problem of accurate segmentation of blurry glioma images. It demonstrates that the BEC kernel is a more feasible, valid, and accurate approach than a lot of the recent year segmentation methods. It is also an advanced and innovative model of prediction deducing from micro BEC theory to macro glioma image segmentation.
Collapse
Affiliation(s)
- Tian Chi Zhang
- School of Information Science and Engineering, Chongqing Jiaotong University, Chongqing, China
| | - Jing Zhang
- School of Information Science and Engineering, University of Jinan, Jinan, China
- Shandong Provincial Key Laboratory of Network-Based Intelligent Computing, Jinan, China
- *Correspondence: Jing Zhang
| | - Shou Cun Chen
- School of Information Science and Engineering, University of Jinan, Jinan, China
- Shandong Provincial Key Laboratory of Network-Based Intelligent Computing, Jinan, China
| | - Bacem Saada
- Cancer Institute, Eighth Affiliated Hospital of Sun Yat-sen University, Shenzhen, China
- Department of Animal Biosciences, University of Guelph, Guelph, ON, Canada
| |
Collapse
|
45
|
Preethi Saroj S, Gurunathan P. Cascaded layer-coalescing convolution network for brain tumor segmentation. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS 2022. [DOI: 10.3233/jifs-220167] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Accurate segmentation of brain tumor regions from magnetic resonance images continues to be one of the active topics of research due to the high usability levels of the automation process. Faster processing helps clinicians in identification at initial stage of tumor and hence saves valuable time taken for manual image analysis. This work proposes a Cascaded Layer-Coalescing (CLC) model using convolution neural networks for brain tumor segmentation. The process includes three layers of convolution networks, each with cascading inputs from the previous layer and provides multiple outputs segmenting complete, core and enhancing tumor regions. The initial layer identifies complete tumor, coalesces the discriminative features and the input data, and passes it to the core tumor detection layer. The core tumor detection layer in- turn passes discriminative features to the enhancing tumor identification layer. The information injection through data coalescing voxels results in enhanced predictions and also in effective handling of data imbalance, which is a major contributor in model viewpoint. Experiments were performed with Brain Tumor Segmentation (BraTS) 2015 data. A comparison with existing literature works indicate improvements up to35% in sensitivity, 27% in PPV and 28% in Dice Score, indicating improvement in the segmentation process.
Collapse
Affiliation(s)
- S. Preethi Saroj
- Department of Computer Science and Engineering, Anna University, Chennai, India
| | - Pradeep Gurunathan
- Department of Computer Applications, A.V.C College of Engineering, Mayiladuthurai, India
| |
Collapse
|
46
|
Sambath Kumar K, Rajendran A. An automatic brain tumor segmentation using modified inception module based U-Net model. JOURNAL OF INTELLIGENT & FUZZY SYSTEMS 2022. [DOI: 10.3233/jifs-211879] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
Abstract
Manual segmentation of brain tumor is not only a tedious task that may bring human mistakes. An automatic segmentation gives results faster, and it extends the survival rate with an earlier treatment plan. So, an automatic brain tumor segmentation model, modified inception module based U-Net (IMU-Net) proposed. It takes Magnetic resonance (MR) images from the BRATS 2017 training dataset with four modalities (FLAIR, T1, T1ce, and T2). The concatenation of two series 3×3 kernels, one 5×5, and one 1×1 convolution kernels are utilized to extract the whole tumor (WT), core tumor (CT), and enhance tumor (ET). The modified inception module (IM) collects all the relevant features and provides better segmentation results. The proposed deep learning model contains 40 convolution layers and utilizes intensity normalization and data augmentation operation for further improvement. It achieved the mean dice similarity coefficient (DSC) of 0.90, 0.77, 0.74, and the mean Intersection over Union (IOU) of 0.79, 0.70, 0.70 for WT, CT, and ET during the evaluation.
Collapse
Affiliation(s)
- K. Sambath Kumar
- Department of Electronics and Communication Engineering, Vel Tech Rangarajan Dr. Sagunthala R&D Institute of Science and Technology, Avadi, Chennai, Tamilnadu, India
| | - A. Rajendran
- Department of Electronics and Communication Engineering, Karpagam College of Engineering, Myleripalayam Village, Othakalmandapam, Coimbatore, Tamilnadu, India
| |
Collapse
|
47
|
Guan X, Yang G, Ye J, Yang W, Xu X, Jiang W, Lai X. 3D AGSE-VNet: an automatic brain tumor MRI data segmentation framework. BMC Med Imaging 2022; 22:6. [PMID: 34986785 PMCID: PMC8734251 DOI: 10.1186/s12880-021-00728-8] [Citation(s) in RCA: 40] [Impact Index Per Article: 13.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/16/2021] [Accepted: 07/26/2021] [Indexed: 12/02/2022] Open
Abstract
BACKGROUND Glioma is the most common brain malignant tumor, with a high morbidity rate and a mortality rate of more than three percent, which seriously endangers human health. The main method of acquiring brain tumors in the clinic is MRI. Segmentation of brain tumor regions from multi-modal MRI scan images is helpful for treatment inspection, post-diagnosis monitoring, and effect evaluation of patients. However, the common operation in clinical brain tumor segmentation is still manual segmentation, lead to its time-consuming and large performance difference between different operators, a consistent and accurate automatic segmentation method is urgently needed. With the continuous development of deep learning, researchers have designed many automatic segmentation algorithms; however, there are still some problems: (1) The research of segmentation algorithm mostly stays on the 2D plane, this will reduce the accuracy of 3D image feature extraction to a certain extent. (2) MRI images have gray-scale offset fields that make it difficult to divide the contours accurately. METHODS To meet the above challenges, we propose an automatic brain tumor MRI data segmentation framework which is called AGSE-VNet. In our study, the Squeeze and Excite (SE) module is added to each encoder, the Attention Guide Filter (AG) module is added to each decoder, using the channel relationship to automatically enhance the useful information in the channel to suppress the useless information, and use the attention mechanism to guide the edge information and remove the influence of irrelevant information such as noise. RESULTS We used the BraTS2020 challenge online verification tool to evaluate our approach. The focus of verification is that the Dice scores of the whole tumor, tumor core and enhanced tumor are 0.68, 0.85 and 0.70, respectively. CONCLUSION Although MRI images have different intensities, AGSE-VNet is not affected by the size of the tumor, and can more accurately extract the features of the three regions, it has achieved impressive results and made outstanding contributions to the clinical diagnosis and treatment of brain tumor patients.
Collapse
Affiliation(s)
- Xi Guan
- School of Medical Technology and Information Engineering, Zhejiang Chinese Medical University, Hangzhou, 310053, China
| | - Guang Yang
- Cardiovascular Research Centre, Royal Brompton Hospital, London, SW3 6NP, UK.
- National Heart and Lung Institute, Imperial College London, London, SW7 2AZ, UK.
| | - Jianming Ye
- First Affiliated Hospital, Gannan Medical University, Ganzhou, 341000, China
| | - Weiji Yang
- College of Life Science, Zhejiang Chinese Medical University, Hangzhou, 310053, China
| | - Xiaomei Xu
- School of Medical Technology and Information Engineering, Zhejiang Chinese Medical University, Hangzhou, 310053, China
| | - Weiwei Jiang
- School of Medical Technology and Information Engineering, Zhejiang Chinese Medical University, Hangzhou, 310053, China
| | - Xiaobo Lai
- School of Medical Technology and Information Engineering, Zhejiang Chinese Medical University, Hangzhou, 310053, China.
| |
Collapse
|
48
|
Abstract
AbstractBrain tumor occurs owing to uncontrolled and rapid growth of cells. If not treated at an initial phase, it may lead to death. Despite many significant efforts and promising outcomes in this domain, accurate segmentation and classification remain a challenging task. A major challenge for brain tumor detection arises from the variations in tumor location, shape, and size. The objective of this survey is to deliver a comprehensive literature on brain tumor detection through magnetic resonance imaging to help the researchers. This survey covered the anatomy of brain tumors, publicly available datasets, enhancement techniques, segmentation, feature extraction, classification, and deep learning, transfer learning and quantum machine learning for brain tumors analysis. Finally, this survey provides all important literature for the detection of brain tumors with their advantages, limitations, developments, and future trends.
Collapse
|
49
|
A novel method for image segmentation: two-stage decoding network with boundary attention. INT J MACH LEARN CYB 2021. [DOI: 10.1007/s13042-021-01459-6] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/03/2023]
|
50
|
Zhou T, Canu S, Vera P, Ruan S. Feature-enhanced generation and multi-modality fusion based deep neural network for brain tumor segmentation with missing MR modalities. Neurocomputing 2021. [DOI: 10.1016/j.neucom.2021.09.032] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/15/2022]
|