1
|
Abinaya B, Malleswaran M. Segmentation and severity classification of scar tissues in LGE-CMR images using HDResC-Net with Flamingo gannet search optimization. Health Inf Sci Syst 2025; 13:21. [PMID: 39980632 PMCID: PMC11836265 DOI: 10.1007/s13755-025-00340-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2024] [Accepted: 01/18/2025] [Indexed: 02/22/2025] Open
Abstract
Late gadolinium enhanced-cardiac magnetic resonance (LGE-CMR) images play a critical role in evaluating cardiac pathology, where scar tissue serves as a vital indicator impacting prognosis and treatment decisions. However, accurately segmenting scar tissues and assessing their severity present challenges due to complex tissue composition and imaging artifacts. Existing methods often lack precision and robustness, limiting their clinical applicability. This work proposes a novel methodology that integrates the optimal segmentation algorithm (OSA) for segmentation and Flamingo Gannet search optimization-enabled hybrid deep residual convolutional network (FGSO-HDResC-Net) for severity classification of scar tissues in LGE-CMR images. Initially, the input image is pre-processed by using the adaptive Gabor Kuwahara filter. Then, the approach combines myocardium segmentation via region-based convolutional neural network and scar segmentation using OSA. Subsequently, FGSO-HDResC-Net integrates feature extraction and classification while optimizing hyperparameters through Flamingo Gannet search optimization. The feature extraction stage introduces two sets of techniques: localization features with texture analysis and spatial/temporal features using a deep residual network, complemented by feature fusion using the fractional concept. These features are inputted into a customized 1D convolutional neural network model for severity classification. Through comprehensive evaluation, the effectiveness of FGSO-HDResC-Net in accurately classifying scar tissue severity is demonstrated, offering improved disease assessment and treatment planning for cardiac patients. Moreover, the proposed FGSO-HDResC-Net model demonstrated superior performance, achieving an accuracy of 96.45%, a true positive rate of 95.42%, a true negative rate of 96.48%, a positive predictive value of 94.20%, and a negative predictive value of 94.18%. The accuracy of the devised model is 14.50%, 12.99%, 10.74%, 9.75%, 12.79%, and 11.26% improved than the traditional models.
Collapse
Affiliation(s)
- B. Abinaya
- Department of ECE, Easwari Engineering College, Ramapuram, Chennai, Tamilnadu 603201 India
| | - M. Malleswaran
- Department of ECE, University College of Engineering Kancheepuram, Ponnerikkarai, Tamilnadu 631552 India
| |
Collapse
|
2
|
Zhan Y, Hao Y, Wang X, Guo D. Advances of artificial intelligence in clinical application and scientific research of neuro-oncology: Current knowledge and future perspectives. Crit Rev Oncol Hematol 2025; 209:104682. [PMID: 40032186 DOI: 10.1016/j.critrevonc.2025.104682] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/01/2024] [Revised: 02/16/2025] [Accepted: 02/25/2025] [Indexed: 03/05/2025] Open
Abstract
Brain tumors refer to the abnormal growths that occur within the brain's tissue, comprising both primary neoplasms and metastatic lesions. Timely detection, precise staging, suitable treatment, and standardized management are of significant clinical importance for extending the survival rates of brain tumor patients. Artificial intelligence (AI), a discipline within computer science, is leveraging its robust capacity for information identification and combination to revolutionize traditional paradigms of oncology care, offering substantial potential for precision medicine. This article provides an overview of the current applications of AI in brain tumors, encompassing the primary AI technologies, their working mechanisms and working workflow, the contributions of AI to brain tumor diagnosis and treatment, as well as the role of AI in brain tumor scientific research, particularly in drug innovation and revealing tumor microenvironment. Finally, the paper addresses the existing challenges, potential solutions, and the future application prospects. This review aims to enhance our understanding of the application of AI in brain tumors and provide valuable insights for forthcoming clinical applications and scientific inquiries.
Collapse
Affiliation(s)
- Yankun Zhan
- First People's Hospital of Linping District; Linping Campus, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 311100, China
| | - Yanying Hao
- First People's Hospital of Linping District; Linping Campus, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 311100, China
| | - Xiang Wang
- First People's Hospital of Linping District; Linping Campus, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 311100, China.
| | - Duancheng Guo
- Cancer Institute, Fudan University Shanghai Cancer Center; Department of Oncology, Shanghai Medical College, Fudan University, Shanghai 200032, China.
| |
Collapse
|
3
|
Al Hasan S, Mahim SM, Hossen ME, Hasan MO, Islam MK, Livreri P, Khan SU, Alibakhshikenari M, Miah MS. DSIT UNet a dual stream iterative transformer based UNet architecture for segmenting brain tumors from FLAIR MRI images. Sci Rep 2025; 15:13815. [PMID: 40259039 PMCID: PMC12012032 DOI: 10.1038/s41598-025-98464-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/11/2024] [Accepted: 04/11/2025] [Indexed: 04/23/2025] Open
Abstract
Brain tumor segmentation remains challenging in medical imaging with conventional therapies and rehabilitation owing to the complex morphology and heterogeneous nature of tumors. Although convolutional neural networks (CNNs) have advanced medical image segmentation, they struggle with long-range dependencies because of their limited receptive fields. We propose Dual-Stream Iterative Transformer UNet (DSIT-UNet), a novel framework that combines Iterative Transformer (IT) modules with a dual-stream encoder-decoder architecture. Our model incorporates a transformed spatial-hybrid attention optimization (TSHAO) module to enhance multiscale feature interactions and balance local details with the global context. We evaluated DSIT-UNet using three benchmark datasets: The Cancer Imaging Archive (TCIA) from The Cancer Genome Atlas (TCGA), BraTS2020, and BraTS2021. On TCIA, our model achieved a Mean Intersection over Union of 95.21%, mean Dice Coefficient of 96.23%, precision of 95.91%, and recall of 96.55%. BraTS2020 attained a Mean IoU of 95.88%, mDice of 96.32%, precision of 96.21%, and recall of 96.44%, surpassing the performance of the existing methods. The superior results of DSIT-UNet demonstrate its effectiveness in capturing tumor boundaries and improving segmentation robustness through hierarchical attention mechanisms and multiscale feature extraction. This architecture advances automated brain tumor segmentation, with potential applications in clinical neuroimaging and future extensions to 3D volumetric segmentation.
Collapse
Affiliation(s)
- Shakib Al Hasan
- Department of Biomedical Engineering, Islamic University, Kushtia, 7003, Bangladesh
- Bio-Imaging Research Lab, BME, Islamic University, Kushtia, 7003, Bangladesh
| | - S M Mahim
- Department of Biomedical Engineering, Islamic University, Kushtia, 7003, Bangladesh
- Bio-Imaging Research Lab, BME, Islamic University, Kushtia, 7003, Bangladesh
| | - Md Emamul Hossen
- Department of Biomedical Engineering, Islamic University, Kushtia, 7003, Bangladesh
- Bio-Imaging Research Lab, BME, Islamic University, Kushtia, 7003, Bangladesh
| | - Md Olid Hasan
- Department of Biomedical Engineering, Islamic University, Kushtia, 7003, Bangladesh
- Bio-Imaging Research Lab, BME, Islamic University, Kushtia, 7003, Bangladesh
| | - Md Khairul Islam
- Department of Biomedical Engineering, Islamic University, Kushtia, 7003, Bangladesh.
- Bio-Imaging Research Lab, BME, Islamic University, Kushtia, 7003, Bangladesh.
| | - Patrizia Livreri
- Department of Engineering, University of Palermo, 90128, Palermo, Italy.
| | - Salah Uddin Khan
- Sustainable Energy Technologies Center, College of Engineering, College of Engineering, King Saud University, 11421, Riyadh, Saudi Arabia
- King Salman Center for Disability Research, 11614, Riyadh, Saudi Arabia
| | | | - Md Sipon Miah
- Department of Information and Communication Technology, Islamic University, Kushtia, 7003, Bangladesh.
- Machine Learning-aided Wireless Communications (WCML) Research Laboratory, Islamic University, Kushtia, 7003, Bangladesh.
| |
Collapse
|
4
|
Kanthaswamy SP, GnanaPrakasam RNK. Brain tumour classification and survival prediction using a novel hybrid deep learning model using MRI image. NETWORK (BRISTOL, ENGLAND) 2025:1-37. [PMID: 40243150 DOI: 10.1080/0954898x.2025.2486206] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/08/2024] [Revised: 01/22/2025] [Accepted: 03/23/2025] [Indexed: 04/18/2025]
Abstract
Brain Tumor (BT) is an irregular growth of cells in the brain or in the tissues surrounding it. Detecting and predicting tumours is essential in today's world, yet managing these diseases poses a considerable challenge. Among the various modalities, Magnetic Resonance Imaging (MRI) has been extensively exploited for diagnosing tumours. The traditional methods for predicting survival are based on handcrafted features from MRI and clinical information, which is generally subjective and laborious. This paper devises a new method named, Deep Residual PyramidNet (DRP_Net) for BT classification and survival prediction. The input MRI image is primarily derived from the BraTS dataset. Then, image enhancement is done to improve the quality of images using homomorphic filtering. Next, deep joint segmentation is used to process the tumourtumour region segmentation. Consequently, Haar wavelet and Local Directional Number Pattern (LDNP) based feature extraction is mined. Afterward, BT classification is achieved through DRP_Net, which is a fusion of Deep Residual Network (DRN) and PyramidNet. At last, the survival prediction is accomplished by employing the Deep Recurrent Neural Network (DRNN). Furthermore, DRP_Net has attained superior performance with a True Negative Rate (TNR) of 91.99%, an accuracy of 90.18%, and True Positive Rate (TPR) of 91.08%.
Collapse
Affiliation(s)
- Shanmuga Priya Kanthaswamy
- Department of Computer Science and Engineering, Bharath Institute of Higher Education and research, Chennai, India
| | | |
Collapse
|
5
|
Lim H, Gi Y, Ko Y, Jo Y, Hong J, Kim J, Ahn SH, Park HC, Kim H, Chung K, Yoon M. A device-dependent auto-segmentation method based on combined generalized and single-device datasets. Med Phys 2025; 52:2375-2383. [PMID: 39699056 DOI: 10.1002/mp.17570] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/08/2024] [Revised: 11/12/2024] [Accepted: 11/26/2024] [Indexed: 12/20/2024] Open
Abstract
BACKGROUND Although generalized-dataset-based auto-segmentation models that consider various computed tomography (CT) scanners have shown great clinical potential, their application to medical images from unseen scanners remains challenging because of device-dependent image features. PURPOSE This study aims to investigate the performance of a device-dependent auto-segmentation model based on a combined dataset of a generalized dataset and single CT scanner dataset. METHOD We constructed two training datasets for 21 chest and abdominal organs. The generalized dataset comprised 1203 publicly available multi-scanner data. The device-dependent dataset comprised 1253 data, including the 1203 multi-CT scanner data and 50 single CT scanner data. Using these datasets, the generalized-dataset-based model (GDSM) and the device-dependent-dataset-based model (DDSM) were trained. The models were trained using nnU-Net and tested on ten data samples from a single CT scanner. The evaluation metrics included the Dice similarity coefficient (DSC), the Hausdorff distance (HD), and the average symmetric surface distance (ASSD), which were used to assess the overall performance of the models. In addition, DSCdiff, HDratio, and ASSDratio, which are variations of the three existing metrics, were used to compare the performance of the models across different organs. RESULT For the average DSC, the GDSM and DDSM had values of 0.9251 and 0.9323, respectively. For the average HD, the GDSM and DDSM had values of 10.66 and 9.139 mm, respectively; for the average ASSD, the GDSM and DDSM had values of 0.8318 and 0.6656 mm, respectively. Compared with the GDSM, the DDSM showed consistent performance improvements of 0.78%, 14%, and 20% for the DSC, HD, and ASSD metrics, respectively. In addition, compared with the GDSM, the DDSM had better DSCdiff values in 14 of 21 tested organs, better HDratio values in 13 of 21 tested organs, and better ASSDratio values in 14 of 21 tested organs. The three averages of the variant metrics were all better for the DDSM than for the GDSM. CONCLUSION The results suggest that combining the generalized dataset with a single scanner dataset resulted in an overall improvement in model performance for that device image.
Collapse
Affiliation(s)
- Hyeongjin Lim
- Department of Bio-medical Engineering, Korea University, Seoul, Republic of Korea
| | - Yongha Gi
- Department of Bio-medical Engineering, Korea University, Seoul, Republic of Korea
| | - Yousun Ko
- Department of Bio-medical Engineering, Korea University, Seoul, Republic of Korea
| | - Yunhui Jo
- Institute of Global Health Technology (IGHT), Korea University, Seoul, Republic of Korea
| | - Jinyoung Hong
- Department of Bio-medical Engineering, Korea University, Seoul, Republic of Korea
| | | | - Sung Hwan Ahn
- Department of Radiation Oncology, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul, Republic of Korea
| | - Hee-Chul Park
- Department of Radiation Oncology, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul, Republic of Korea
| | - Haeyoung Kim
- Department of Radiation Oncology, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul, Republic of Korea
| | - Kwangzoo Chung
- Department of Radiation Oncology, Samsung Medical Center, Sungkyunkwan University School of Medicine, Seoul, Republic of Korea
| | - Myonggeun Yoon
- Department of Bio-medical Engineering, Korea University, Seoul, Republic of Korea
- FieldCure Ltd, Seoul, Republic of Korea
| |
Collapse
|
6
|
Huang X, Qin M, Fang M, Wang Z, Hu C, Zhao T, Qin Z, Zhu H, Wu L, Yu G, De Cobelli F, Xie X, Palumbo D, Tian J, Dong D. The application of artificial intelligence in upper gastrointestinal cancers. JOURNAL OF THE NATIONAL CANCER CENTER 2025; 5:113-131. [PMID: 40265096 PMCID: PMC12010392 DOI: 10.1016/j.jncc.2024.12.006] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/13/2024] [Revised: 09/17/2024] [Accepted: 12/20/2024] [Indexed: 04/24/2025] Open
Abstract
Upper gastrointestinal cancers, mainly comprising esophageal and gastric cancers, are among the most prevalent cancers worldwide. There are many new cases of upper gastrointestinal cancers annually, and the survival rate tends to be low. Therefore, timely screening, precise diagnosis, appropriate treatment strategies, and effective prognosis are crucial for patients with upper gastrointestinal cancers. In recent years, an increasing number of studies suggest that artificial intelligence (AI) technology can effectively address clinical tasks related to upper gastrointestinal cancers. These studies mainly focus on four aspects: screening, diagnosis, treatment, and prognosis. In this review, we focus on the application of AI technology in clinical tasks related to upper gastrointestinal cancers. Firstly, the basic application pipelines of radiomics and deep learning in medical image analysis were introduced. Furthermore, we separately reviewed the application of AI technology in the aforementioned aspects for both esophageal and gastric cancers. Finally, the current limitations and challenges faced in the field of upper gastrointestinal cancers were summarized, and explorations were conducted on the selection of AI algorithms in various scenarios, the popularization of early screening, the clinical applications of AI, and large multimodal models.
Collapse
Affiliation(s)
- Xiaoying Huang
- CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China
| | - Minghao Qin
- CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China
- University of Science and Technology Beijing, Beijing, China
| | - Mengjie Fang
- Beijing Advanced Innovation Center for Big Data-Based Precision Medicine, Beihang University, Beijing, China
- Key Laboratory of Big Data-Based Precision Medicine, Beihang University, Ministry of Industry and Information Technology, Beijing, China
| | - Zipei Wang
- CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China
| | - Chaoen Hu
- CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China
| | - Tongyu Zhao
- CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China
- University of Science and Technology of China, Hefei, China
| | - Zhuyuan Qin
- Beijing Institute of Genomics, Chinese Academy of Sciences, Beijing, China
- Beijing University of Chinese Medicine, Beijing, China
| | | | - Ling Wu
- KiangWu Hospital, Macau, China
| | | | | | | | - Diego Palumbo
- Department of Radiology, IRCCS San Raffaele Scientific Institute, Milan, Italy
| | - Jie Tian
- CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing, China
- Beijing Advanced Innovation Center for Big Data-Based Precision Medicine, Beihang University, Beijing, China
- Key Laboratory of Big Data-Based Precision Medicine, Beihang University, Ministry of Industry and Information Technology, Beijing, China
| | - Di Dong
- CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing, China
| |
Collapse
|
7
|
Goodkin O, Wu J, Pemberton H, Prados F, Vos SB, Thust S, Thornton J, Yousry T, Bisdas S, Barkhof F. Structured reporting of gliomas based on VASARI criteria to improve report content and consistency. BMC Med Imaging 2025; 25:99. [PMID: 40128670 PMCID: PMC11934815 DOI: 10.1186/s12880-025-01603-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/02/2024] [Accepted: 02/18/2025] [Indexed: 03/26/2025] Open
Abstract
PURPOSE Gliomas are the commonest malignant brain tumours. Baseline characteristics on structural MRI, such as size, enhancement proportion and eloquent brain involvement inform grading and treatment planning. Currently, free-text imaging reports depend on the individual style and experience of the radiologist. Standardisation may increase consistency of feature reporting. METHODS We compared 100 baseline free-text reports for glioma MRI scans with a structured feature list based on VASARI criteria and performed a full second read to document which VASARI features were in the baseline report. RESULTS We found that quantitative features including tumour size and proportion of necrosis and oedema/infiltration were commonly not included in free-text reports. Thirty-three percent of reports gave a description of size only, and 38% of reports did not refer to tumour size at all. Detailed information about tumour location including involvement of eloquent areas and infiltration of deep white matter was also missing from the majority of free-text reports. Overall, we graded 6% of reports as having omitted some key VASARI features that would alter patient management. CONCLUSIONS Tumour size and anatomical information is often omitted by neuroradiologists. Comparison with a structured report identified key features that would benefit from standardisation and/or quantification. Structured reporting may improve glioma reporting consistency, clinical communication, and treatment decisions.
Collapse
Affiliation(s)
- Olivia Goodkin
- Queen Square Multiple Sclerosis Centre, Department of Neuroinflammation, UCL Queen Square Institute of Neurology, Faculty of Brain Sciences, University College London, London, UK
- Centre for Medical Image Computing (CMIC), Department of Medical Physics and Biomedical Engineering, University College London, London, UK
- Neuroradiological Academic Unit, UCL Queen Square Institute of Neurology, University College London, London, UK
- National Institute for Health Research (NIHR), University College London Hospitals (UCLH) Biomedical Research Centre (BRC), London, UK
| | - Jiaming Wu
- Centre for Medical Image Computing (CMIC), Department of Medical Physics and Biomedical Engineering, University College London, London, UK
| | - Hugh Pemberton
- Centre for Medical Image Computing (CMIC), Department of Medical Physics and Biomedical Engineering, University College London, London, UK
- Neuroradiological Academic Unit, UCL Queen Square Institute of Neurology, University College London, London, UK
- GE Healthcare, Amersham, UK
| | - Ferran Prados
- Queen Square Multiple Sclerosis Centre, Department of Neuroinflammation, UCL Queen Square Institute of Neurology, Faculty of Brain Sciences, University College London, London, UK
- Centre for Medical Image Computing (CMIC), Department of Medical Physics and Biomedical Engineering, University College London, London, UK
- National Institute for Health Research (NIHR), University College London Hospitals (UCLH) Biomedical Research Centre (BRC), London, UK
- E-Health Center, Universitat Oberta de Catalunya, Barcelona, Spain
| | - Sjoerd B Vos
- Centre for Medical Image Computing (CMIC), Department of Medical Physics and Biomedical Engineering, University College London, London, UK
- Neuroradiological Academic Unit, UCL Queen Square Institute of Neurology, University College London, London, UK
- Centre for Microscopy, Characterisation and Analysis, University of Western Australia, Perth, Australia
| | - Stefanie Thust
- Neuroradiological Academic Unit, UCL Queen Square Institute of Neurology, University College London, London, UK
- National Institute for Health Research (NIHR), University College London Hospitals (UCLH) Biomedical Research Centre (BRC), London, UK
- Nottingham NIHR Biomedical Research Centre, Nottingham, UK
- Radiological Sciences, School of Medicine, Mental Health and Neurosciences, University of Nottingham, Nottingham, UK
- Lysholm Department of Neuroradiology, National Hospital for Neurology and Neurosurgery, UCLH NHS Foundation Trust, London, UK
| | - John Thornton
- Neuroradiological Academic Unit, UCL Queen Square Institute of Neurology, University College London, London, UK
- National Institute for Health Research (NIHR), University College London Hospitals (UCLH) Biomedical Research Centre (BRC), London, UK
- Lysholm Department of Neuroradiology, National Hospital for Neurology and Neurosurgery, UCLH NHS Foundation Trust, London, UK
| | - Tarek Yousry
- Neuroradiological Academic Unit, UCL Queen Square Institute of Neurology, University College London, London, UK
- National Institute for Health Research (NIHR), University College London Hospitals (UCLH) Biomedical Research Centre (BRC), London, UK
- Lysholm Department of Neuroradiology, National Hospital for Neurology and Neurosurgery, UCLH NHS Foundation Trust, London, UK
| | - Sotirios Bisdas
- Neuroradiological Academic Unit, UCL Queen Square Institute of Neurology, University College London, London, UK
- Lysholm Department of Neuroradiology, National Hospital for Neurology and Neurosurgery, UCLH NHS Foundation Trust, London, UK
| | - Frederik Barkhof
- Queen Square Multiple Sclerosis Centre, Department of Neuroinflammation, UCL Queen Square Institute of Neurology, Faculty of Brain Sciences, University College London, London, UK.
- Centre for Medical Image Computing (CMIC), Department of Medical Physics and Biomedical Engineering, University College London, London, UK.
- Neuroradiological Academic Unit, UCL Queen Square Institute of Neurology, University College London, London, UK.
- National Institute for Health Research (NIHR), University College London Hospitals (UCLH) Biomedical Research Centre (BRC), London, UK.
- Lysholm Department of Neuroradiology, National Hospital for Neurology and Neurosurgery, UCLH NHS Foundation Trust, London, UK.
- Department of Radiology and Nuclear Medicine, VU Medical Centre, Amsterdam, Netherlands.
| |
Collapse
|
8
|
Wen L, Sun H, Liang G, Yu Y. A deep ensemble learning framework for glioma segmentation and grading prediction. Sci Rep 2025; 15:4448. [PMID: 39910114 PMCID: PMC11799385 DOI: 10.1038/s41598-025-87127-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/13/2024] [Accepted: 01/16/2025] [Indexed: 02/07/2025] Open
Abstract
The segmentation and risk grade prediction of gliomas based on preoperative multimodal magnetic resonance imaging (MRI) are crucial tasks in computer-aided diagnosis. Due to the significant heterogeneity between and within tumors, existing methods mainly rely on single-task approaches, overlooking the inherent correlation between segmentation and grading tasks. Furthermore, the limited availability of glioma grading data presents further challenges. To address these issues, we propose a deep-ensemble learning framework based on multimodal MRI and the U-Net model, which simultaneously performs glioma segmentation and risk grade prediction. We introduce asymmetric convolution and dual-domain attention in the encoder, fully integrating effective information from different modalities, enhancing the extraction of features from critical regions, and constructing a dual-branch decoder that combines spatial features and global semantic information for both segmentation and grading. In addition, we propose a weighted composite adaptive loss function to balance the optimization objectives of the two tasks. Our experimental results on the BraTS dataset demonstrate that our method outperforms state-of-the-art methods, yielding superior segmentation accuracy and precise risk grade prediction.
Collapse
Affiliation(s)
- Liang Wen
- General Hospital of Northern Theater Command, Shenyang, 110122, China.
- China Medical University, Shenyang, 110122, China.
| | - Hui Sun
- College of Information Science and Engineering, Northeastern University, Shenyang, 110819, China
| | - Guobiao Liang
- General Hospital of Northern Theater Command, Shenyang, 110122, China
- China Medical University, Shenyang, 110122, China
| | - Yue Yu
- College of Information Science and Engineering, Northeastern University, Shenyang, 110819, China
| |
Collapse
|
9
|
Bao N, Zhang J, Li Z, Wei S, Zhang J, Greenwald SE, Onofrey JA, Lu Y, Xu L. CT-Less Whole-Body Bone Segmentation of PET Images Using a Multimodal Deep Learning Network. IEEE J Biomed Health Inform 2025; 29:1151-1164. [PMID: 40030243 DOI: 10.1109/jbhi.2024.3501386] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/06/2025]
Abstract
In bone cancer imaging, positron emission tomography (PET) is ideal for the diagnosis and staging of bone cancers due to its high sensitivity to malignant tumors. The diagnosis of bone cancer requires tumor analysis and localization, where accurate and automated wholebody bone segmentation (WBBS) is often needed. Current WBBS for PET imaging is based on paired Computed Tomography (CT) images. However, mismatches between CT and PET images often occur due to patient motion, which leads to erroneous bone segmentation and thus, to inaccurate tumor analysis. Furthermore, there are some instances where CT images are unavailable for WBBS. In this work, we propose a novel multimodal fusion network (MMF-Net) for WBBS of PET images, without the need for CT images. Specifically, the tracer activity ($\lambda$-MLAA), attenuation map ($\mu$-MLAA), and synthetic attenuation map ($\mu$-DL) images are introduced into the training data. We first design a multi-encoder structure employed to fully learn modalityspecific encoding representations of the three PET modality images through independent encoding branches. Then, we propose a multimodal fusion module in the decoder to further integrate the complementary information across the three modalities. Additionally, we introduce revised convolution units, SE (Squeeze-and-Excitation) Normalization and deep supervision to improve segmentation performance. Extensive comparisons and ablation experiments, using 130 whole-body PET image datasets, show promising results. We conclude that the proposed method can achieve WBBS with moderate to high accuracy using PET information only, which potentially can be used to overcome the current limitations of CT-based approaches, while minimizing exposure to ionizing radiation.
Collapse
|
10
|
Abirami S, Ramesh K, Lalitha VaniSree K. Classification and Pixel Change Detection of Brain Tumor Using Adam Kookaburra Optimization-Based Shepard Convolutional Neural Network. NMR IN BIOMEDICINE 2025; 38:e5307. [PMID: 39832921 DOI: 10.1002/nbm.5307] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/23/2024] [Revised: 11/22/2024] [Accepted: 11/27/2024] [Indexed: 01/22/2025]
Abstract
The uncommon growth of cells in the brain is termed as brain tumor. To identify chronic nerve problems, like strokes, brain tumors, multiple sclerosis, and dementia, brain magnetic resonance imaging (MRI) is normally utilized. Identifying the tumor on early stage can improve the patient's survival rate. However, it is difficult to identify the exact tumor region with less computational complexity. Also, the tumors can vary significantly in shape, size, and appearance, which complicates the task of correctly classifying tumor types and detecting subtle pixel changes over time. Hence, an Adam kookaburra optimization-based Shepard convolutional neural network (AKO-based Shepard CNN) is established in this study for the classification and pixel change detection of brain tumor. The Adam kookaburra optimization (AKO) is established by integrating the kookaburra optimization algorithm (KOA) and Adam. Here, the pre- and post-operative MRIs are pre-processed and then segmented by U-Net++. The tuning of U-Net++ is done by the bald Border collie firefly optimization algorithm (BBCFO). The bald eagle search (BES), firefly algorithm (FA), and Border collie optimization (BCO) are combined to form the BBCFO. The next operation is the feature extraction and the classification is conducted at last using ShCNN. The AKO is utilized to tune the ShCNN for obtaining effective classification results. Unlike conventional optimization algorithms, AKO offers faster convergence and higher accuracy in classification. The highest negative predictive value (NPV), true negative rate (TNR), true positive rate (TPR), positive predictive value (PPV), and accuracy produced by the AKO-based ShCNN are 89.91%, 92.26%, 93.78%, and 93.60%, respectively, using Brain Images of Tumors for Evaluation database (BITE).
Collapse
Affiliation(s)
- S Abirami
- Department of Computer Science and Engineering, Sri Krishna College of Engineering and Technology, Coimbatore, Tamilnadu, India
| | - K Ramesh
- Department of Computer Science and Engineering, Sri Krishna College of Engineering and Technology, Coimbatore, Tamilnadu, India
| | - K Lalitha VaniSree
- Department of Computer Science and Engineering Koneru Lakshmaiah Education Foundation, Vijayawada, India
| |
Collapse
|
11
|
Alhajim D, Ansari-Asl K, Akbarizadeh G, Soorki MN. Improved lung nodule segmentation with a squeeze excitation dilated attention based residual UNet. Sci Rep 2025; 15:3770. [PMID: 39885263 PMCID: PMC11782676 DOI: 10.1038/s41598-025-85199-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/05/2024] [Accepted: 01/01/2025] [Indexed: 02/01/2025] Open
Abstract
The diverse types and sizes, proximity to non-nodule structures, identical shape characteristics, and varying sizes of nodules make them challenging for segmentation methods. Although many efforts have been made in automatic lung nodule segmentation, most of them have not sufficiently addressed the challenges related to the type and size of nodules, such as juxta-pleural and juxta-vascular nodules. The current research introduces a Squeeze-Excitation Dilated Attention-based Residual U-Net (SEDARU-Net) with a robust intensity normalization technique to address the challenges related to different types and sizes of lung nodules and to achieve an improved lung nodule segmentation. After preprocessing the images with the intensity normalization method and extracting the Regions of Interest by YOLOv3, they are fed into the SEDARU-Net with dilated convolutions in the encoder part. Then, the extracted features are given to the decoder part, which involves transposed convolutions, Squeeze-Excitation Dilated Residual blocks, and skip connections equipped with an Attention Gate, to decode the feature maps and construct the segmentation mask. The proposed model was evaluated using the publicly available Lung Nodule Analysis 2016 (LUNA16) dataset, achieving a Dice Similarity Coefficient of 97.86%, IoU of 96.40%, sensitivity of 96.54%, and precision of 98.84%. Finally, it was shown that each added component to the U-Net's structure and the intensity normalization technique increased the Dice Similarity Coefficient by more than 2%. The proposed method suggests a potential clinical tool to address challenges related to the segmentation of lung nodules with different types located in the proximity of non-nodule structures.
Collapse
Affiliation(s)
- Dhafer Alhajim
- Department of Electrical Engineering, Faculty of Engineering, Shahid Chamran University of Ahvaz, Ahvaz, Iran
| | - Karim Ansari-Asl
- Department of Electrical Engineering, Faculty of Engineering, Shahid Chamran University of Ahvaz, Ahvaz, Iran.
| | - Gholamreza Akbarizadeh
- Department of Electrical Engineering, Faculty of Engineering, Shahid Chamran University of Ahvaz, Ahvaz, Iran
| | - Mehdi Naderi Soorki
- Department of Electrical Engineering, Faculty of Engineering, Shahid Chamran University of Ahvaz, Ahvaz, Iran
| |
Collapse
|
12
|
Zhang WJ, Chen WT, Liu CH, Chen SW, Lai YH, You SD. Feasibility Study of Detecting and Segmenting Small Brain Tumors in a Small MRI Dataset with Self-Supervised Learning. Diagnostics (Basel) 2025; 15:249. [PMID: 39941179 PMCID: PMC11817956 DOI: 10.3390/diagnostics15030249] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/29/2024] [Revised: 01/07/2025] [Accepted: 01/15/2025] [Indexed: 02/16/2025] Open
Abstract
Objectives: This paper studies the segmentation and detection of small metastatic brain tumors. This study aims to evaluate the feasibility of training a deep neural network for the segmentation and detection of metastatic brain tumors in MRI using a very small dataset of 33 cases, by leveraging large public datasets of primary tumors; Methods: This study explores various methods, including supervised learning, two transfer learning approaches, and self-supervised learning, utilizing U-net and Swin UNETR models; Results: The self-supervised learning approach utilizing the Swin UNETR model yielded the best performance. The Dice score for small brain tumors was approximately 0.19. Sensitivity reached 100%, while specificity was 54.5%. When excluding subjects with hyperintensities, the specificity improved to 80.0%; Conclusions: It is feasible to train a model using self-supervised learning and a small dataset for the segmentation and detection of small brain tumors.
Collapse
Affiliation(s)
- Wei-Jun Zhang
- Department of Computer Science and Information Engineering, National Taipei University of Technology, Taipei 106, Taiwan; (W.-J.Z.); (C.-H.L.)
| | - Wei-Teing Chen
- Division of Thoracic Medicine, Department of Medicine, Cheng Hsin General Hospital, Taipei 112, Taiwan
- Department of Internal Medicine, Tri-Service General Hospital, National Defense Medical Center, Taipei 114, Taiwan
| | - Chien-Hung Liu
- Department of Computer Science and Information Engineering, National Taipei University of Technology, Taipei 106, Taiwan; (W.-J.Z.); (C.-H.L.)
| | - Shiuan-Wen Chen
- Department of Electrical and Computer Engineering (ECE), University of Toronto, Toronto, ON M5S 1A1, Canada;
| | - Yu-Hua Lai
- Division of Neurology, Department of Medicine, Cheng Hsin General Hospital, Taipei 112, Taiwan;
| | - Shingchern D. You
- Department of Computer Science and Information Engineering, National Taipei University of Technology, Taipei 106, Taiwan; (W.-J.Z.); (C.-H.L.)
| |
Collapse
|
13
|
Gencer K, Gencer G. Hybrid deep learning approach for brain tumor classification using EfficientNetB0 and novel quantum genetic algorithm. PeerJ Comput Sci 2025; 11:e2556. [PMID: 39896007 PMCID: PMC11784816 DOI: 10.7717/peerj-cs.2556] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2024] [Accepted: 11/06/2024] [Indexed: 02/04/2025]
Abstract
One of the most complex and life-threatening pathologies of the central nervous system is brain tumors. Correct diagnosis of these tumors plays an important role in determining the treatment plans of patients. Traditional classification methods often rely on manual assessments, which can be prone to error. Therefore, multiple classification of brain tumors has gained significant interest in recent years in both the medical and computer science fields. The use of artificial intelligence and machine learning, especially in the automatic classification of brain tumors, is increasing significantly. Deep learning models can achieve high accuracy when trained on datasets in diagnosis and classification. This study examined deep learning-based approaches for automatic multi-class classification of brain tumors, and a new approach combining deep learning and quantum genetic algorithms (QGA) was proposed. The powerful feature extraction ability of the pre-trained EfficientNetB0 was utilized and combined with this quantum genetic algorithms, a new approach was proposed. It is aimed to develop the feature selection method. With this hybrid method, high reliability and accuracy in brain tumor classification was achieved. The proposed model achieved high accuracy of 98.36% and 98.25%, respectively, with different data sets and significantly outperformed traditional methods. As a result, the proposed method offers a robust and scalable solution that will help classify brain tumors in early and accurate diagnosis and contribute to the field of medical imaging with patient outcomes.
Collapse
Affiliation(s)
- Kerem Gencer
- Afyon Kocatepe University, Faculty of Engineering, Department of Computer Engineering, Afyonkarahisar, Turkey
| | - Gülcan Gencer
- Afyonkarahisar Health Sciences University, Faculty of Medicine, Department of Biostatistics and Medical Informatics, Afyonkarahisar, Turkey
| |
Collapse
|
14
|
Rastogi D, Johri P, Donelli M, Kadry S, Khan AA, Espa G, Feraco P, Kim J. Deep learning-integrated MRI brain tumor analysis: feature extraction, segmentation, and Survival Prediction using Replicator and volumetric networks. Sci Rep 2025; 15:1437. [PMID: 39789043 PMCID: PMC11718254 DOI: 10.1038/s41598-024-84386-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/21/2024] [Accepted: 12/23/2024] [Indexed: 01/12/2025] Open
Abstract
The most prevalent form of malignant tumors that originate in the brain are known as gliomas. In order to diagnose, treat, and identify risk factors, it is crucial to have precise and resilient segmentation of the tumors, along with an estimation of the patients' overall survival rate. Therefore, we have introduced a deep learning approach that employs a combination of MRI scans to accurately segment brain tumors and predict survival in patients with gliomas. To ensure strong and reliable tumor segmentation, we employ 2D volumetric convolution neural network architectures that utilize a majority rule. This method helps to significantly decrease model bias and improve performance. Additionally, in order to predict survival rates, we extract radiomic features from the tumor regions that have been segmented, and then use a Deep Learning Inspired 3D replicator neural network to identify the most effective features. The model presented in this study was successful in segmenting brain tumors and predicting the outcome of enhancing tumor and real enhancing tumor. The model was evaluated using the BRATS2020 benchmarks dataset, and the obtained results are quite satisfactory and promising.
Collapse
Affiliation(s)
- Deependra Rastogi
- School of Computer Science and Engineering, IILM University, Greater Noida, Noida, 201306, UP, India
| | - Prashant Johri
- SCSE, Galgotias University, Greater Noida, Noida, 203201, UP, India
| | - Massimo Donelli
- Department of Civil, Environmental, Mechanical Engineering University of Trento, Trento, 38100, Italy
- Radiomics Laboratory, Department of Economy and Management, University of Trento, Trento, 38100, Italy
| | - Seifedine Kadry
- Department of Computer Science and Mathematics, Lebanese American University, Beirut, Lebanon
- Noroff University College, Kristiansand, 4612, Norway
| | - Arfat Ahmad Khan
- Department of Engineering, Simpson University, California, 96003, USA.
| | - Giuseppe Espa
- Radiomics Laboratory, Department of Economy and Management, University of Trento, Trento, 38100, Italy
| | - Paola Feraco
- Neuroradiology Unit, Santa Chiara Hospital, Azienda provinciale per i servizi sanitari, Trento, 38100, Italy
| | - Jungeun Kim
- Department of Computer Engineering, Inha University, Incheon, Republic of Korea.
| |
Collapse
|
15
|
Rohilla S, Jain S. Detection of Brain Tumor Employing Residual Network-based Optimized Deep Learning. Curr Comput Aided Drug Des 2025; 21:15-27. [PMID: 37587819 DOI: 10.2174/1573409920666230816090626] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/18/2022] [Revised: 05/19/2023] [Accepted: 05/30/2023] [Indexed: 08/18/2023]
Abstract
BACKGROUND Diagnosis and treatment planning play a very vital role in improving the survival of oncological patients. However, there is high variability in the shape, size, and structure of the tumor, making automatic segmentation difficult. The automatic and accurate detection and segmentation methods for brain tumors are proposed in this paper. METHODS A modified ResNet50 model was used for tumor detection, and a ResUNetmodel-based convolutional neural network for segmentation is proposed in this paper. The detection and segmentation were performed on the same dataset consisting of pre-contrast, FLAIR, and postcontrast MRI images of 110 patients collected from the cancer imaging archive. Due to the use of residual networks, the authors observed improvement in evaluation parameters, such as accuracy for tumor detection and dice similarity coefficient for tumor segmentation. RESULTS The accuracy of tumor detection and dice similarity coefficient achieved by the segmentation model were 96.77% and 0.893, respectively, for the TCIA dataset. The results were compared based on manual segmentation and existing segmentation techniques. The tumor mask was also individually compared to the ground truth using the SSIM value. The proposed detection and segmentation models were validated on BraTS2015 and BraTS2017 datasets, and the results were consensus. CONCLUSION The use of residual networks in both the detection and the segmentation model resulted in improved accuracy and DSC score. DSC score was increased by 5.9% compared to the UNet model, and the accuracy of the model was increased from 92% to 96.77% for the test set.
Collapse
Affiliation(s)
- Saransh Rohilla
- Department of Electronics and Communication Engineering, Jaypee University of Information Technology, Solan, Himachal Pradesh, India
| | - Shruti Jain
- Department of Electronics and Communication Engineering, Jaypee University of Information Technology, Solan, Himachal Pradesh, India
| |
Collapse
|
16
|
Lam VK, Fischer E, Jawad K, Tabaie S, Cleary K, Anwar SM. An automated framework for pediatric hip surveillance and severity assessment using radiographs. Int J Comput Assist Radiol Surg 2025; 20:203-211. [PMID: 39283409 DOI: 10.1007/s11548-024-03254-4] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/13/2024] [Accepted: 08/12/2024] [Indexed: 01/25/2025]
Abstract
PURPOSE Hip dysplasia is the second most common orthopedic condition in children with cerebral palsy (CP) and may result in disability and pain. The migration percentage (MP) is a widely used metric in hip surveillance, calculated based on an anterior-posterior pelvis radiograph. However, manual quantification of MP values using hip X-ray scans in current standard practice has challenges including being time-intensive, requiring expert knowledge, and not considering human bias. The purpose of this study is to develop a machine learning algorithm to automatically quantify MP values using a hip X-ray scan, and hence provide an assessment for severity, which then can be used for surveillance, treatment planning, and management. METHODS X-ray scans from 210 patients were curated, pre-processed, and manually annotated at our clinical center. Several machine learning models were trained using pre-trained weights from Inception ResNet-V2, VGG-16, and VGG-19, with different strategies (pre-processing, with and without region of interest (ROI) detection, with and without data augmentation) to find an optimal model for automatic hip landmarking. The predicted landmarks were then used by our geometric algorithm to quantify the MP value for the input hip X-ray scan. RESULTS The pre-trained VGG-19 model, fine-tuned with additional custom layers, outputted the lowest mean squared error values for both train and test data, when ROI cropped images were used along with data augmentation for model training. The MP value calculated by the algorithm was compared to manual ground truth labels from our orthopedic fellows using the hip screen application for benchmarking. CONCLUSION The results showed the feasibility of the machine learning model in automatic hip landmark detection for reliably quantifying MP value from hip X-ray scans. The algorithm could be used as an accurate and reliable tool in orthopedic care for diagnosing, severity assessment, and hence treatment and surgical planning for hip displacement.
Collapse
Affiliation(s)
- Van Khanh Lam
- Sheikh Zayed Institute for Pediatric Surgical Innovation, Childrens National Hospital, Washington, DC, 20008, USA
| | - Elizabeth Fischer
- Sheikh Zayed Institute for Pediatric Surgical Innovation, Childrens National Hospital, Washington, DC, 20008, USA
| | - Kochai Jawad
- Sheikh Zayed Institute for Pediatric Surgical Innovation, Childrens National Hospital, Washington, DC, 20008, USA
| | - Sean Tabaie
- Sheikh Zayed Institute for Pediatric Surgical Innovation, Childrens National Hospital, Washington, DC, 20008, USA
| | - Kevin Cleary
- Sheikh Zayed Institute for Pediatric Surgical Innovation, Childrens National Hospital, Washington, DC, 20008, USA
| | - Syed Muhammad Anwar
- Sheikh Zayed Institute for Pediatric Surgical Innovation, Childrens National Hospital, Washington, DC, 20008, USA.
- School of Medicine and Health Sciences, George Washington University, Washington, DC, 20052, USA.
| |
Collapse
|
17
|
Omari Alaoui A, Boutahir MK, El Bahi O, Hessane A, Farhaoui Y, El Allaoui A. Accelerating deep learning model development—towards scalable automated architecture generation for optimal model design. MULTIMEDIA TOOLS AND APPLICATIONS 2024; 84:3053-3069. [DOI: 10.1007/s11042-024-20481-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/25/2024] [Revised: 10/14/2024] [Accepted: 11/21/2024] [Indexed: 01/15/2025]
|
18
|
Su J, Luo Z, Wang C, Lian S, Lin X, Li S. Reconstruct incomplete relation for incomplete modality brain tumor segmentation. Neural Netw 2024; 180:106657. [PMID: 39186839 DOI: 10.1016/j.neunet.2024.106657] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/28/2023] [Revised: 08/14/2024] [Accepted: 08/19/2024] [Indexed: 08/28/2024]
Abstract
Different brain tumor magnetic resonance imaging (MRI) modalities provide diverse tumor-specific information. Previous works have enhanced brain tumor segmentation performance by integrating multiple MRI modalities. However, multi-modal MRI data are often unavailable in clinical practice. An incomplete modality leads to missing tumor-specific information, which degrades the performance of existing models. Various strategies have been proposed to transfer knowledge from a full modality network (teacher) to an incomplete modality one (student) to address this issue. However, they neglect the fact that brain tumor segmentation is a structural prediction problem that requires voxel semantic relations. In this paper, we propose a Reconstruct Incomplete Relation Network (RIRN) that transfers voxel semantic relational knowledge from the teacher to the student. Specifically, we propose two types of voxel relations to incorporate structural knowledge: Class-relative relations (CRR) and Class-agnostic relations (CAR). The CRR groups voxels into different tumor regions and constructs a relation between them. The CAR builds a global relation between all voxel features, complementing the local inter-region relation. Moreover, we use adversarial learning to align the holistic structural prediction between the teacher and the student. Extensive experimentation on both the BraTS 2018 and BraTS 2020 datasets establishes that our method outperforms all state-of-the-art approaches.
Collapse
Affiliation(s)
- Jiawei Su
- School of Computer Engineering, Jimei University, Xiamen, China; The Department of Artificial Intelligence, Xiamen University, Fujian, China
| | - Zhiming Luo
- The Department of Artificial Intelligence, Xiamen University, Fujian, China.
| | - Chengji Wang
- The School of Computer Science, Central China Normal University, Wuhan, China
| | - Sheng Lian
- The College of Computer and Data Science, Fuzhou University, Fujian, China
| | - Xuejuan Lin
- The Department of Traditional Chinese Medicine, Fujian University of Traditional Chinese Medicine, Fujian, China
| | - Shaozi Li
- The Department of Artificial Intelligence, Xiamen University, Fujian, China
| |
Collapse
|
19
|
Seshimo H, Rashed EA. Segmentation of Low-Grade Brain Tumors Using Mutual Attention Multimodal MRI. SENSORS (BASEL, SWITZERLAND) 2024; 24:7576. [PMID: 39686112 DOI: 10.3390/s24237576] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/24/2024] [Revised: 11/18/2024] [Accepted: 11/26/2024] [Indexed: 12/18/2024]
Abstract
Early detection and precise characterization of brain tumors play a crucial role in improving patient outcomes and extending survival rates. Among neuroimaging modalities, magnetic resonance imaging (MRI) is the gold standard for brain tumor diagnostics due to its ability to produce high-contrast images across a variety of sequences, each highlighting distinct tissue characteristics. This study focuses on enabling multimodal MRI sequences to advance the automatic segmentation of low-grade astrocytomas, a challenging task due to their diffuse and irregular growth patterns. A novel mutual-attention deep learning framework is proposed, which integrates complementary information from multiple MRI sequences, including T2-weighted and fluid-attenuated inversion recovery (FLAIR) sequences, to enhance the segmentation accuracy. Unlike conventional segmentation models, which treat each modality independently or simply concatenate them, our model introduces mutual attention mechanisms. This allows the network to dynamically focus on salient features across modalities by jointly learning interdependencies between imaging sequences, leading to more precise boundary delineations even in regions with subtle tumor signals. The proposed method is validated using the UCSF-PDGM dataset, which consists of 35 astrocytoma cases, presenting a realistic and clinically challenging dataset. The results demonstrate that T2w/FLAIR modalities contribute most significantly to the segmentation performance. The mutual-attention model achieves an average Dice coefficient of 0.87. This study provides an innovative pathway toward improving segmentation of low-grade tumors by enabling context-aware fusion across imaging sequences. Furthermore, the study showcases the clinical relevance of integrating AI with multimodal MRI, potentially improving non-invasive tumor characterization and guiding future research in radiological diagnostics.
Collapse
Affiliation(s)
- Hiroyuki Seshimo
- Graduate School of Information Science, University of Hyogo, Kobe 650-0047, Japan
| | - Essam A Rashed
- Graduate School of Information Science, University of Hyogo, Kobe 650-0047, Japan
- Advanced Medical Engineering Research Institute, University of Hyogo, Himeji 670-0836, Japan
| |
Collapse
|
20
|
Yin S, Li H, Teng L, Laghari AA, Almadhor A, Gregus M, Sampedro GA. Brain CT image classification based on mask RCNN and attention mechanism. Sci Rep 2024; 14:29300. [PMID: 39905102 PMCID: PMC11794603 DOI: 10.1038/s41598-024-78566-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/14/2024] [Accepted: 10/31/2024] [Indexed: 02/06/2025] Open
Abstract
Along with the computer application technology progress, machine learning, and block-chain techniques have been applied comprehensively in various fields. The application of machine learning, and block-chain techniques into medical image retrieval, classification and auxiliary diagnosis has become one of the research hotspots at present. Brain tumor is one of the major diseases threatening human life. The number of deaths caused by these diseases is increasing dramatically every year in the world. Aiming at the classification problem of brain CT images in healthcare. We propose a Mask RCNN with attention mechanism method in this research. First, the ResNet-10 is utilized as the backbone model to extract local features of the input brain CT images. In the partial residual module, the standard convolution is substituted by deformable convolution. Then, the spatial attention mechanism and the channel attention mechanism are connected in parallel. The deformable convolution is embedded to the two modules to extract global features. Finally, the loss function is improved to further optimize the precision of target edge segmentation in the Mask RCNN branch. Finally, we make experiments on public brain CT data set, the results show that the proposed image classification fragrance can effectively refine the edge features, increase the degree of separation between target and background, and improve the classification effect.
Collapse
Affiliation(s)
- Shoulin Yin
- Software College, Shenyang Normal University, Shenyang, China.
| | - Hang Li
- Software College, Shenyang Normal University, Shenyang, China.
| | - Lin Teng
- Software College, Shenyang Normal University, Shenyang, China.
| | | | - Ahmad Almadhor
- Department of Computer Engineering and Networks, College of Computer and Information Sciences, Jouf University, 72388, Sakaka, Saudi Arabia
| | - Michal Gregus
- Faculty of Management, Comenius University in Bratislava, Bratislava, Slovakia.
| | - Gabriel Avelino Sampedro
- Department of Computer Science, University of the Philippines Diliman, 1101, Quezon City, Philippines
| |
Collapse
|
21
|
Gasmi K, Ben Aoun N, Alsalem K, Ltaifa IB, Alrashdi I, Ammar LB, Mrabet M, Shehab A. Enhanced brain tumor diagnosis using combined deep learning models and weight selection technique. Front Neuroinform 2024; 18:1444650. [PMID: 39659489 PMCID: PMC11628532 DOI: 10.3389/fninf.2024.1444650] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/06/2024] [Accepted: 10/21/2024] [Indexed: 12/12/2024] Open
Abstract
Brain tumor classification is a critical task in medical imaging, as accurate diagnosis directly influences treatment planning and patient outcomes. Traditional methods often fall short in achieving the required precision due to the complex and heterogeneous nature of brain tumors. In this study, we propose an innovative approach to brain tumor multi-classification by leveraging an ensemble learning method that combines advanced deep learning models with an optimal weighting strategy. Our methodology integrates Vision Transformers (ViT) and EfficientNet-V2 models, both renowned for their powerful feature extraction capabilities in medical imaging. This model enhances the feature extraction step by capturing both global and local features, thanks to the combination of different deep learning models with the ViT model. These models are then combined using a weighted ensemble approach, where each model's prediction is assigned a weight. To optimize these weights, we employ a genetic algorithm, which iteratively selects the best weight combinations to maximize classification accuracy. We trained and validated our ensemble model using a well-curated dataset comprising labeled brain MRI images. The model's performance was benchmarked against standalone ViT and EfficientNet-V2 models, as well as other traditional classifiers. The ensemble approach achieved a notable improvement in classification accuracy, precision, recall, and F1-score compared to individual models. Specifically, our model attained an accuracy rate of 95%, significantly outperforming existing methods. This study underscores the potential of combining advanced deep learning models with a genetic algorithm-optimized weighting strategy to tackle complex medical classification tasks. The enhanced diagnostic precision offered by our ensemble model can lead to better-informed clinical decisions, ultimately improving patient outcomes. Furthermore, our approach can be generalized to other medical imaging classification problems, paving the way for broader applications of AI in healthcare. This advancement in brain tumor classification contributes valuable insights to the field of medical AI, supporting the ongoing efforts to integrate advanced computational tools in clinical practice.
Collapse
Affiliation(s)
- Karim Gasmi
- Department of Computer Science, College of Computer and Information Sciences, Jouf University, Sakkaka, Saudi Arabia
| | - Najib Ben Aoun
- College of Computing and Information, Al-Baha University, Alaqiq, Saudi Arabia
- REGIM-Lab: Research Groups in Intelligent Machines, National School of Engineers of Sfax (ENIS), University of Sfax, Sfax, Tunisia
| | - Khalaf Alsalem
- Department of Information Systems, College of Computer and Information Sciences, Jouf University, Sakaka, Saudi Arabia
| | - Ibtihel Ben Ltaifa
- STIH: Sens Texte Informatique Histoire, Sorbonne University, Paris, France
| | - Ibrahim Alrashdi
- Department of Computer Science, College of Computer and Information Sciences, Jouf University, Sakkaka, Saudi Arabia
| | | | - Manel Mrabet
- Prince Sattam bin Abdulaziz University, Al-Kharj, Saudi Arabia
| | - Abdulaziz Shehab
- Department of Information Systems, College of Computer and Information Sciences, Jouf University, Sakaka, Saudi Arabia
| |
Collapse
|
22
|
Pan Y, Yong H, Lu W, Li G, Cong J. Brain tumor segmentation by combining MultiEncoder UNet with wavelet fusion. J Appl Clin Med Phys 2024; 25:e14527. [PMID: 39284311 PMCID: PMC11540057 DOI: 10.1002/acm2.14527] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/26/2024] [Revised: 04/28/2024] [Accepted: 08/07/2024] [Indexed: 11/07/2024] Open
Abstract
BACKGROUND AND OBJECTIVE Accurate segmentation of brain tumors from multimodal magnetic resonance imaging (MRI) holds significant importance in clinical diagnosis and surgical intervention, while current deep learning methods cope with situations of multimodal MRI by an early fusion strategy that implicitly assumes that the modal relationships are linear, which tends to ignore the complementary information between modalities, negatively impacting the model's performance. Meanwhile, long-range relationships between voxels cannot be captured due to the localized character of the convolution procedure. METHOD Aiming at this problem, we propose a multimodal segmentation network based on a late fusion strategy that employs multiple encoders and a decoder for the segmentation of brain tumors. Each encoder is specialized for processing distinct modalities. Notably, our framework includes a feature fusion module based on a 3D discrete wavelet transform aimed at extracting complementary features among the encoders. Additionally, a 3D global context-aware module was introduced to capture the long-range dependencies of tumor voxels at a high level of features. The decoder combines fused and global features to enhance the network's segmentation performance. RESULT Our proposed model is experimented on the publicly available BraTS2018 and BraTS2021 datasets. The experimental results show competitiveness with state-of-the-art methods. CONCLUSION The results demonstrate that our approach applies a novel concept for multimodal fusion within deep neural networks and delivers more accurate and promising brain tumor segmentation, with the potential to assist physicians in diagnosis.
Collapse
Affiliation(s)
- Yuheng Pan
- Computer and Information Engineering DepartmentTianjin Chengjian UniversityTianjinChina
| | - Haohan Yong
- Computer and Information Engineering DepartmentTianjin Chengjian UniversityTianjinChina
| | - Weijia Lu
- Computer and Information Engineering DepartmentTianjin Chengjian UniversityTianjinChina
| | - Guoyan Li
- Computer and Information Engineering DepartmentTianjin Chengjian UniversityTianjinChina
| | - Jia Cong
- Computer and Information Engineering DepartmentTianjin Chengjian UniversityTianjinChina
| |
Collapse
|
23
|
Lin J, Xie W, Kang L, Wu H. Dynamic-Guided Spatiotemporal Attention for Echocardiography Video Segmentation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:3843-3855. [PMID: 38771692 DOI: 10.1109/tmi.2024.3403687] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/23/2024]
Abstract
Left ventricle (LV) endocardium segmentation in echocardiography video has received much attention as an important step in quantifying LV ejection fraction. Most existing methods are dedicated to exploiting temporal information on top of 2D convolutional networks. In addition to single appearance semantic learning, some research attempted to introduce motion cues through the optical flow estimation (OFE) task to enhance temporal consistency modeling. However, OFE in these methods is tightly coupled to LV endocardium segmentation, resulting in noisy inter-frame flow prediction, and post-optimization based on these flows accumulates errors. To address these drawbacks, we propose dynamic-guided spatiotemporal attention (DSA) for semi-supervised echocardiography video segmentation. We first fine-tune the off-the-shelf OFE network RAFT on echocardiography data to provide dynamic information. Taking inter-frame flows as additional input, we use a dual-encoder structure to extract motion and appearance features separately. Based on the connection between dynamic continuity and semantic consistency, we propose a bilateral feature calibration module to enhance both features. For temporal consistency modeling, the DSA is proposed to aggregate neighboring frame context using deformable attention that is realized by offsets grid attention. Dynamic information is introduced into DSA through a bilateral offset estimation module to effectively combine with appearance semantics and predict attention offsets, thereby guiding semantic-based spatiotemporal attention. We evaluated our method on two popular echocardiography datasets, CAMUS and EchoNet-Dynamic, and achieved state-of-the-art.
Collapse
|
24
|
Gopinath K, Hoopes A, Alexander DC, Arnold SE, Balbastre Y, Billot B, Casamitjana A, Cheng Y, Chua RYZ, Edlow BL, Fischl B, Gazula H, Hoffmann M, Keene CD, Kim S, Kimberly WT, Laguna S, Larson KE, Van Leemput K, Puonti O, Rodrigues LM, Rosen MS, Tregidgo HFJ, Varadarajan D, Young SI, Dalca AV, Iglesias JE. Synthetic data in generalizable, learning-based neuroimaging. IMAGING NEUROSCIENCE (CAMBRIDGE, MASS.) 2024; 2:1-22. [PMID: 39850547 PMCID: PMC11752692 DOI: 10.1162/imag_a_00337] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 01/19/2024] [Revised: 09/20/2024] [Accepted: 09/20/2024] [Indexed: 01/25/2025]
Abstract
Synthetic data have emerged as an attractive option for developing machine-learning methods in human neuroimaging, particularly in magnetic resonance imaging (MRI)-a modality where image contrast depends enormously on acquisition hardware and parameters. This retrospective paper reviews a family of recently proposed methods, based on synthetic data, for generalizable machine learning in brain MRI analysis. Central to this framework is the concept of domain randomization, which involves training neural networks on a vastly diverse array of synthetically generated images with random contrast properties. This technique has enabled robust, adaptable models that are capable of handling diverse MRI contrasts, resolutions, and pathologies, while working out-of-the-box, without retraining. We have successfully applied this method to tasks such as whole-brain segmentation (SynthSeg), skull-stripping (SynthStrip), registration (SynthMorph, EasyReg), super-resolution, and MR contrast transfer (SynthSR). Beyond these applications, the paper discusses other possible use cases and future work in our methodology. Neural networks trained with synthetic data enable the analysis of clinical MRI, including large retrospective datasets, while greatly alleviating (and sometimes eliminating) the need for substantial labeled datasets, and offer enormous potential as robust tools to address various research goals.
Collapse
Affiliation(s)
- Karthik Gopinath
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | - Andrew Hoopes
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
- Massachusetts Institute of Technology, Cambridge, MA, United States
| | | | - Steven E. Arnold
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | - Yael Balbastre
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | - Benjamin Billot
- Computer Science & Artificial Intelligence Laboratory, Massachusetts Institute of Technology, Cambridge, MA, United States
| | | | - You Cheng
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | - Russ Yue Zhi Chua
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
- Massachusetts Institute of Technology, Cambridge, MA, United States
| | - Brian L. Edlow
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | - Bruce Fischl
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | | | - Malte Hoffmann
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | - C. Dirk Keene
- University of Washington, Seattle, WA, United States
| | | | - W. Taylor Kimberly
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | | | - Kathleen E. Larson
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | - Koen Van Leemput
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | - Oula Puonti
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
- Copenhagen University Hospital, København, Denmark
| | - Livia M. Rodrigues
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
- Universidade Estadual de Campinas, São Paulo, Brazil
| | - Matthew S. Rosen
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | | | - Divya Varadarajan
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
| | - Sean I. Young
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
- Massachusetts Institute of Technology, Cambridge, MA, United States
| | - Adrian V. Dalca
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
- Massachusetts Institute of Technology, Cambridge, MA, United States
| | - Juan Eugenio Iglesias
- Massachusetts General Hospital, Harvard Medical School, Boston, MA, United States
- Massachusetts Institute of Technology, Cambridge, MA, United States
- University College London, London, England
| |
Collapse
|
25
|
Loosen AM, Kato A, Gu X. Revisiting the role of computational neuroimaging in the era of integrative neuroscience. Neuropsychopharmacology 2024; 50:103-113. [PMID: 39242921 PMCID: PMC11525590 DOI: 10.1038/s41386-024-01946-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 03/05/2024] [Revised: 07/12/2024] [Accepted: 07/17/2024] [Indexed: 09/09/2024]
Abstract
Computational models have become integral to human neuroimaging research, providing both mechanistic insights and predictive tools for human cognition and behavior. However, concerns persist regarding the ecological validity of lab-based neuroimaging studies and whether their spatiotemporal resolution is not sufficient for capturing neural dynamics. This review aims to re-examine the utility of computational neuroimaging, particularly in light of the growing prominence of alternative neuroscientific methods and the growing emphasis on more naturalistic behaviors and paradigms. Specifically, we will explore how computational modeling can both enhance the analysis of high-dimensional imaging datasets and, conversely, how neuroimaging, in conjunction with other data modalities, can inform computational models through the lens of neurobiological plausibility. Collectively, this evidence suggests that neuroimaging remains critical for human neuroscience research, and when enhanced by computational models, imaging can serve an important role in bridging levels of analysis and understanding. We conclude by proposing key directions for future research, emphasizing the development of standardized paradigms and the integrative use of computational modeling across neuroimaging techniques.
Collapse
Affiliation(s)
- Alisa M Loosen
- Department of Psychiatry, Icahn School of Medicine at Mount Sinai, New York, NY, USA.
- Center for Computational Psychiatry, Icahn School of Medicine at Mount Sinai, New York, NY, USA.
- Nash Family Department of Neuroscience, Icahn School of Medicine at Mount Sinai, New York, NY, USA.
| | - Ayaka Kato
- Department of Psychiatry, Icahn School of Medicine at Mount Sinai, New York, NY, USA.
- Center for Computational Psychiatry, Icahn School of Medicine at Mount Sinai, New York, NY, USA.
- Nash Family Department of Neuroscience, Icahn School of Medicine at Mount Sinai, New York, NY, USA.
| | - Xiaosi Gu
- Department of Psychiatry, Icahn School of Medicine at Mount Sinai, New York, NY, USA
- Center for Computational Psychiatry, Icahn School of Medicine at Mount Sinai, New York, NY, USA
- Nash Family Department of Neuroscience, Icahn School of Medicine at Mount Sinai, New York, NY, USA
| |
Collapse
|
26
|
O'Sullivan NJ, Temperley HC, Horan MT, Kamran W, Corr A, O'Gorman C, Saadeh F, Meaney JM, Kelly ME. Role of radiomics as a predictor of disease recurrence in ovarian cancer: a systematic review. Abdom Radiol (NY) 2024; 49:3540-3547. [PMID: 38744703 PMCID: PMC11390851 DOI: 10.1007/s00261-024-04330-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/27/2023] [Revised: 04/04/2024] [Accepted: 04/05/2024] [Indexed: 05/16/2024]
Abstract
Ovarian cancer is associated with high cancer-related mortality rate attributed to late-stage diagnosis, limited treatment options, and frequent disease recurrence. As a result, careful patient selection is important especially in setting of radical surgery. Radiomics is an emerging field in medical imaging, which may help provide vital prognostic evaluation and help patient selection for radical treatment strategies. This systematic review aims to assess the role of radiomics as a predictor of disease recurrence in ovarian cancer. A systematic search was conducted in Medline, EMBASE, and Web of Science databases. Studies meeting inclusion criteria investigating the use of radiomics to predict post-operative recurrence in ovarian cancer were included in our qualitative analysis. Study quality was assessed using the QUADAS-2 and Radiomics Quality Score tools. Six retrospective studies met the inclusion criteria, involving a total of 952 participants. Radiomic-based signatures demonstrated consistent performance in predicting disease recurrence, as evidenced by satisfactory area under the receiver operating characteristic curve values (AUC range 0.77-0.89). Radiomic-based signatures appear to good prognosticators of disease recurrence in ovarian cancer as estimated by AUC. The reviewed studies consistently reported the potential of radiomic features to enhance risk stratification and personalise treatment decisions in this complex cohort of patients. Further research is warranted to address limitations related to feature reliability, workflow heterogeneity, and the need for prospective validation studies.
Collapse
Affiliation(s)
- Niall J O'Sullivan
- Department of Radiology, St. James's Hospital, Dublin, Ireland.
- School of Medicine, Trinity College Dublin, Dublin, Ireland.
- The National Centre for Advanced Medical Imaging (CAMI), St. James's Hospital, Dublin, Ireland.
| | | | - Michelle T Horan
- Department of Radiology, St. James's Hospital, Dublin, Ireland
- The National Centre for Advanced Medical Imaging (CAMI), St. James's Hospital, Dublin, Ireland
| | - Waseem Kamran
- Department of Gynaecology, St. James's Hospital, Dublin, Ireland
| | - Alison Corr
- Department of Radiology, St. James's Hospital, Dublin, Ireland
| | | | - Feras Saadeh
- Department of Gynaecology, St. James's Hospital, Dublin, Ireland
| | - James M Meaney
- Department of Radiology, St. James's Hospital, Dublin, Ireland
- School of Medicine, Trinity College Dublin, Dublin, Ireland
- The National Centre for Advanced Medical Imaging (CAMI), St. James's Hospital, Dublin, Ireland
| | - Michael E Kelly
- Department of Radiology, St. James's Hospital, Dublin, Ireland
- Department of Surgery, St. James's Hospital, Dublin, Ireland
| |
Collapse
|
27
|
Kudus K, Wagner M, Ertl-Wagner BB, Khalvati F. Applications of machine learning to MR imaging of pediatric low-grade gliomas. Childs Nerv Syst 2024; 40:3027-3035. [PMID: 38972953 DOI: 10.1007/s00381-024-06522-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/25/2024] [Accepted: 06/21/2024] [Indexed: 07/09/2024]
Abstract
INTRODUCTION Machine learning (ML) shows promise for the automation of routine tasks related to the treatment of pediatric low-grade gliomas (pLGG) such as tumor grading, typing, and segmentation. Moreover, it has been shown that ML can identify crucial information from medical images that is otherwise currently unattainable. For example, ML appears to be capable of preoperatively identifying the underlying genetic status of pLGG. METHODS In this chapter, we reviewed, to the best of our knowledge, all published works that have used ML techniques for the imaging-based evaluation of pLGGs. Additionally, we aimed to provide some context on what it will take to go from the exploratory studies we reviewed to clinically deployed models. RESULTS Multiple studies have demonstrated that ML can accurately grade, type, and segment and detect the genetic status of pLGGs. We compared the approaches used between the different studies and observed a high degree of variability throughout the methodologies. Standardization and cooperation between the numerous groups working on these approaches will be key to accelerating the clinical deployment of these models. CONCLUSION The studies reviewed in this chapter detail the potential for ML techniques to transform the treatment of pLGG. However, there are still challenges that need to be overcome prior to clinical deployment.
Collapse
Affiliation(s)
- Kareem Kudus
- Neurosciences & Mental Health Research Program, The Hospital for Sick Children, Toronto, Canada
- Institute of Medical Science, University of Toronto, Toronto, Canada
| | - Matthias Wagner
- Department of Diagnostic & Interventional Radiology, The Hospital for Sick Children, Toronto, Canada
- Department of Diagnostic and Interventional Neuroradiology, University Hospital Augsburg, Augsburg, Germany
| | - Birgit Betina Ertl-Wagner
- Neurosciences & Mental Health Research Program, The Hospital for Sick Children, Toronto, Canada
- Institute of Medical Science, University of Toronto, Toronto, Canada
- Department of Diagnostic & Interventional Radiology, The Hospital for Sick Children, Toronto, Canada
- Department of Medical Imaging, University of Toronto, Toronto, Canada
| | - Farzad Khalvati
- Neurosciences & Mental Health Research Program, The Hospital for Sick Children, Toronto, Canada.
- Institute of Medical Science, University of Toronto, Toronto, Canada.
- Department of Diagnostic & Interventional Radiology, The Hospital for Sick Children, Toronto, Canada.
- Department of Medical Imaging, University of Toronto, Toronto, Canada.
- Department of Computer Science, University of Toronto, Toronto, Canada.
- Department of Mechanical and Industrial Engineering, University of Toronto, Toronto, Canada.
| |
Collapse
|
28
|
Bian C, Hu C, Cao N. Exploiting K-Space in Magnetic Resonance Imaging Diagnosis: Dual-Path Attention Fusion for K-Space Global and Image Local Features. Bioengineering (Basel) 2024; 11:958. [PMID: 39451334 PMCID: PMC11504126 DOI: 10.3390/bioengineering11100958] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/05/2024] [Revised: 09/04/2024] [Accepted: 09/21/2024] [Indexed: 10/26/2024] Open
Abstract
Magnetic resonance imaging (MRI) diagnosis, enhanced by deep learning methods, plays a crucial role in medical image processing, facilitating precise clinical diagnosis and optimal treatment planning. Current methodologies predominantly focus on feature extraction from the image domain, which often results in the loss of global features during down-sampling processes. However, the unique global representational capacity of MRI K-space is often overlooked. In this paper, we present a novel MRI K-space-based global feature extraction and dual-path attention fusion network. Our proposed method extracts global features from MRI K-space data and fuses them with local features from the image domain using a dual-path attention mechanism, thereby achieving accurate MRI segmentation for diagnosis. Specifically, our method consists of four main components: an image-domain feature extraction module, a K-space domain feature extraction module, a dual-path attention feature fusion module, and a decoder. We conducted ablation studies and comprehensive comparisons on the Brain Tumor Segmentation (BraTS) MRI dataset to validate the effectiveness of each module. The results demonstrate that our method exhibits superior performance in segmentation diagnostics, outperforming state-of-the-art methods with improvements of up to 63.82% in the HD95 distance evaluation metric. Furthermore, we performed generalization testing and complexity analysis on the Automated Cardiac Diagnosis Challenge (ACDC) MRI cardiac segmentation dataset. The findings indicate robust performance across different datasets, highlighting strong generalizability and favorable algorithmic complexity. Collectively, these results suggest that our proposed method holds significant potential for practical clinical applications.
Collapse
Affiliation(s)
- Congchao Bian
- College of Information Science and Engineering, Hohai University, Nanjing 210098, China;
| | - Can Hu
- College of Computer Science and Software Engineering, Hohai University, Nanjing 210098, China;
| | - Ning Cao
- College of Information Science and Engineering, Hohai University, Nanjing 210098, China;
| |
Collapse
|
29
|
Rao KN, Khalaf OI, Krishnasree V, Kumar AS, Alsekait DM, Priyanka SS, Alattas AS, AbdElminaam DS. An efficient brain tumor detection and classification using pre-trained convolutional neural network models. Heliyon 2024; 10:e36773. [PMID: 39281484 PMCID: PMC11401085 DOI: 10.1016/j.heliyon.2024.e36773] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2023] [Revised: 08/21/2024] [Accepted: 08/22/2024] [Indexed: 09/18/2024] Open
Abstract
In cases of brain tumors, some brain cells experience abnormal and rapid growth, leading to the development of tumors. Brain tumors represent a significant source of illness affecting the brain. Magnetic Resonance Imaging (MRI) stands as a well-established and coherent diagnostic method for brain cancer detection. However, the resulting MRI scans produce a vast number of images, which require thorough examination by radiologists. Manual assessment of these images consumes considerable time and may result in inaccuracies in cancer detection. Recently, deep learning has emerged as a reliable tool for decision-making tasks across various domains, including finance, medicine, cybersecurity, agriculture, and forensics. In the context of brain cancer diagnosis, Deep Learning and Machine Learning algorithms applied to MRI data enable rapid prognosis. However, achieving higher accuracy is crucial for providing appropriate treatment to patients and facilitating prompt decision-making by radiologists. To address this, we propose the use of Convolutional Neural Networks (CNN) for brain tumor detection. Our approach utilizes a dataset consisting of two classes: three representing different tumor types and one representing non-tumor samples. We present a model that leverages pre-trained CNNs to categorize brain cancer cases. Additionally, data augmentation techniques are employed to augment the dataset size. The effectiveness of our proposed CNN model is evaluated through various metrics, including validation loss, confusion matrix, and overall loss. The proposed approach employing ResNet50 and EfficientNet demonstrated higher levels of accuracy, precision, and recall in detecting brain tumors.
Collapse
Affiliation(s)
- K Nishanth Rao
- Department of ECE, MLR Institute of Technology, Hyderabad, India
| | - Osamah Ibrahim Khalaf
- Department of Solar, Al-Nahrain Research Center for Renewable Energy, Al-Nahrain University, Jadriya, Baghdad, Iraq
| | - V Krishnasree
- Department of ECE, VNR Vignana Jyothi Institute of Engineering and Technology, Telangana, India
| | - Aruru Sai Kumar
- Department of ECE, VNR Vignana Jyothi Institute of Engineering and Technology, Telangana, India
| | - Deema Mohammed Alsekait
- Department of Computer Science and Information Technology, Applied College, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
| | - S Siva Priyanka
- Department of ECE, Chaitanya Bharathi Institute of Technology, Telangana, India
| | - Ahmed Saleh Alattas
- Information Science Department, Faculty of Arts & Humanities, King Abdul-Aziz University, Jeddah, Saudi Arabia
| | - Diaa Salama AbdElminaam
- MEU Research Unit, Middle East University, Amman 11831, Jordan
- Jadara Research Center, Jadara University, Irbid, 21110, Jordan
| |
Collapse
|
30
|
Guo X, Xiang Y, Yang Y, Ye C, Yu Y, Ma T. Accelerating denoising diffusion probabilistic model via truncated inverse processes for medical image segmentation. Comput Biol Med 2024; 180:108933. [PMID: 39096612 DOI: 10.1016/j.compbiomed.2024.108933] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2023] [Revised: 06/07/2024] [Accepted: 07/19/2024] [Indexed: 08/05/2024]
Abstract
Medical image segmentation demands precise accuracy and the capability to assess segmentation uncertainty for informed clinical decision-making. Denoising Diffusion Probability Models (DDPMs), with their advancements in image generation, can treat segmentation as a conditional generation task, providing accurate segmentation and uncertainty estimation. However, current DDPMs used in medical image segmentation suffer from low inference efficiency and prediction errors caused by excessive noise at the end of the forward process. To address this issue, we propose an accelerated denoising diffusion probabilistic model via truncated inverse processes (ADDPM) that is specifically designed for medical image segmentation. The inverse process of ADDPM starts from a non-Gaussian distribution and terminates early once a prediction with relatively low noise is obtained after multiple iterations of denoising. We employ a separate powerful segmentation network to obtain pre-segmentation and construct the non-Gaussian distribution of the segmentation based on the forward diffusion rule. By further adopting a separate denoising network, the final segmentation can be obtained with just one denoising step from the predictions with low noise. ADDPM greatly reduces the number of denoising steps to approximately one-tenth of that in vanilla DDPMs. Our experiments on four segmentation tasks demonstrate that ADDPM outperforms both vanilla DDPMs and existing representative accelerating DDPMs methods. Moreover, ADDPM can be easily integrated with existing advanced segmentation models to improve segmentation performance and provide uncertainty estimation. Implementation code: https://github.com/Guoxt/ADDPM.
Collapse
Affiliation(s)
- Xutao Guo
- School of Electronics and Information Engineering, Harbin Institute of Technology (Shenzhen), Shenzhen, Guangdong, China; The Peng Cheng Laboratory, Shenzhen, Guangdong, China
| | - Yang Xiang
- The Peng Cheng Laboratory, Shenzhen, Guangdong, China
| | - Yanwu Yang
- School of Electronics and Information Engineering, Harbin Institute of Technology (Shenzhen), Shenzhen, Guangdong, China; The Peng Cheng Laboratory, Shenzhen, Guangdong, China
| | - Chenfei Ye
- The International Research Institute for Artificial Intelligence, Harbin Institute of Technology (Shenzhen), Shenzhen, Guangdong, China
| | - Yue Yu
- The Peng Cheng Laboratory, Shenzhen, Guangdong, China
| | - Ting Ma
- School of Electronics and Information Engineering, Harbin Institute of Technology (Shenzhen), Shenzhen, Guangdong, China; The Peng Cheng Laboratory, Shenzhen, Guangdong, China; Guangdong Provincial Key Laboratory of Aerospace Communication and Networking Technology, Harbin Institute of Technology (Shenzhen), Shenzhen, Guangdong, China.
| |
Collapse
|
31
|
Kaur J, Kaur P. A systematic literature analysis of multi-organ cancer diagnosis using deep learning techniques. Comput Biol Med 2024; 179:108910. [PMID: 39032244 DOI: 10.1016/j.compbiomed.2024.108910] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/13/2024] [Revised: 07/14/2024] [Accepted: 07/15/2024] [Indexed: 07/23/2024]
Abstract
Cancer is becoming the most toxic ailment identified among individuals worldwide. The mortality rate has been increasing rapidly every year, which causes progression in the various diagnostic technologies to handle this illness. The manual procedure for segmentation and classification with a large set of data modalities can be a challenging task. Therefore, a crucial requirement is to significantly develop the computer-assisted diagnostic system intended for the initial cancer identification. This article offers a systematic review of Deep Learning approaches using various image modalities to detect multi-organ cancers from 2012 to 2023. It emphasizes the detection of five supreme predominant tumors, i.e., breast, brain, lung, skin, and liver. Extensive review has been carried out by collecting research and conference articles and book chapters from reputed international databases, i.e., Springer Link, IEEE Xplore, Science Direct, PubMed, and Wiley that fulfill the criteria for quality evaluation. This systematic review summarizes the overview of convolutional neural network model architectures and datasets used for identifying and classifying the diverse categories of cancer. This study accomplishes an inclusive idea of ensemble deep learning models that have achieved better evaluation results for classifying the different images into cancer or healthy cases. This paper will provide a broad understanding to the research scientists within the domain of medical imaging procedures of which deep learning technique perform best over which type of dataset, extraction of features, different confrontations, and their anticipated solutions for the complex problems. Lastly, some challenges and issues which control the health emergency have been discussed.
Collapse
Affiliation(s)
- Jaspreet Kaur
- Department of Computer Engineering & Technology, Guru Nanak Dev University, Amritsar, Punjab, India.
| | - Prabhpreet Kaur
- Department of Computer Engineering & Technology, Guru Nanak Dev University, Amritsar, Punjab, India.
| |
Collapse
|
32
|
Konar D, Bhattacharyya S, Gandhi TK, Panigrahi BK, Jiang R. 3-D Quantum-Inspired Self-Supervised Tensor Network for Volumetric Segmentation of Medical Images. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2024; 35:10312-10325. [PMID: 37022399 DOI: 10.1109/tnnls.2023.3240238] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
This article introduces a novel shallow 3-D self-supervised tensor neural network in quantum formalism for volumetric segmentation of medical images with merits of obviating training and supervision. The proposed network is referred to as the 3-D quantum-inspired self-supervised tensor neural network (3-D-QNet). The underlying architecture of 3-D-QNet is composed of a trinity of volumetric layers, viz., input, intermediate, and output layers interconnected using an S -connected third-order neighborhood-based topology for voxelwise processing of 3-D medical image data, suitable for semantic segmentation. Each of the volumetric layers contains quantum neurons designated by qubits or quantum bits. The incorporation of tensor decomposition in quantum formalism leads to faster convergence of network operations to preclude the inherent slow convergence problems faced by the classical supervised and self-supervised networks. The segmented volumes are obtained once the network converges. The suggested 3-D-QNet is tailored and tested on the BRATS 2019 Brain MR image dataset and the Liver Tumor Segmentation Challenge (LiTS17) dataset extensively in our experiments. The 3-D-QNet has achieved promising dice similarity (DS) as compared with the time-intensive supervised convolutional neural network (CNN)-based models, such as 3-D-UNet, voxelwise residual network (VoxResNet), Dense-Res-Inception Net (DRINet), and 3-D-ESPNet, thereby showing a potential advantage of our self-supervised shallow network on facilitating semantic segmentation.
Collapse
|
33
|
Wu D, Tao J, Qin Z, Mumtaz RA, Qin J, Yu L, Courtney J. LightNet: A Novel Lightweight Convolutional Network for Brain Tumor Segmentation in Healthcare. IEEE J Biomed Health Inform 2024; 28:4471-4482. [PMID: 37498760 DOI: 10.1109/jbhi.2023.3297227] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/29/2023]
Abstract
Diagnosis, treatment planning, surveillance, and the monitoring of clinical trials for brain diseases all benefit greatly from neuroimaging-based tumor segmentation. Recently, Convolutional Neural Networks (CNNs) have demonstrated promising results in enhancing the efficiency of image-based brain tumor segmentation. Most current work on CNNs, however, is devoted to creating increasingly complicated convolution modules to improve performance, which in turn raises the computing cost of the model. This work proposes a simple and effective feed-forward CNN, LightNet (Light Network). Based on multi-path and multi-level, it replaces traditional convolutional methods with light operations, which reduces network parameters and redundant feature maps. In the up-sampling stage, a light channel attention module is added to achieve richer multi-scale and spatial semantic feature information extraction of brain tumor. The performance of the network is evaluated in the Multimodal Brain Tumor Segmentation Challenge (BraTS 2015) dataset, and results are presented here alongside other high-performing CNNs. Results show comparable accuracy with other methods but with increased efficiency, segmentation performance, and reduced redundancy and computational complexity. The result is a high-performing network with a balance between efficiency and accuracy, allowing, for example, better energy performance on mobile devices.
Collapse
|
34
|
Abidin ZU, Naqvi RA, Haider A, Kim HS, Jeong D, Lee SW. Recent deep learning-based brain tumor segmentation models using multi-modality magnetic resonance imaging: a prospective survey. Front Bioeng Biotechnol 2024; 12:1392807. [PMID: 39104626 PMCID: PMC11298476 DOI: 10.3389/fbioe.2024.1392807] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2024] [Accepted: 06/14/2024] [Indexed: 08/07/2024] Open
Abstract
Radiologists encounter significant challenges when segmenting and determining brain tumors in patients because this information assists in treatment planning. The utilization of artificial intelligence (AI), especially deep learning (DL), has emerged as a useful tool in healthcare, aiding radiologists in their diagnostic processes. This empowers radiologists to understand the biology of tumors better and provide personalized care to patients with brain tumors. The segmentation of brain tumors using multi-modal magnetic resonance imaging (MRI) images has received considerable attention. In this survey, we first discuss multi-modal and available magnetic resonance imaging modalities and their properties. Subsequently, we discuss the most recent DL-based models for brain tumor segmentation using multi-modal MRI. We divide this section into three parts based on the architecture: the first is for models that use the backbone of convolutional neural networks (CNN), the second is for vision transformer-based models, and the third is for hybrid models that use both convolutional neural networks and transformer in the architecture. In addition, in-depth statistical analysis is performed of the recent publication, frequently used datasets, and evaluation metrics for segmentation tasks. Finally, open research challenges are identified and suggested promising future directions for brain tumor segmentation to improve diagnostic accuracy and treatment outcomes for patients with brain tumors. This aligns with public health goals to use health technologies for better healthcare delivery and population health management.
Collapse
Affiliation(s)
- Zain Ul Abidin
- Department of Intelligent Mechatronics Engineering, Sejong University, Seoul, Republic of Korea
| | - Rizwan Ali Naqvi
- Department of Intelligent Mechatronics Engineering, Sejong University, Seoul, Republic of Korea
| | - Amir Haider
- Department of Intelligent Mechatronics Engineering, Sejong University, Seoul, Republic of Korea
| | - Hyung Seok Kim
- Department of Intelligent Mechatronics Engineering, Sejong University, Seoul, Republic of Korea
| | - Daesik Jeong
- College of Convergence Engineering, Sangmyung University, Seoul, Republic of Korea
| | - Seung Won Lee
- School of Medicine, Sungkyunkwan University, Suwon, Republic of Korea
| |
Collapse
|
35
|
Sarah P, Krishnapriya S, Saladi S, Karuna Y, Bavirisetti DP. A novel approach to brain tumor detection using K-Means++, SGLDM, ResNet50, and synthetic data augmentation. Front Physiol 2024; 15:1342572. [PMID: 39077759 PMCID: PMC11284281 DOI: 10.3389/fphys.2024.1342572] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2023] [Accepted: 06/24/2024] [Indexed: 07/31/2024] Open
Abstract
Introduction: Brain tumors are abnormal cell growths in the brain, posing significant treatment challenges. Accurate early detection using non-invasive methods is crucial for effective treatment. This research focuses on improving the early detection of brain tumors in MRI images through advanced deep-learning techniques. The primary goal is to identify the most effective deep-learning model for classifying brain tumors from MRI data, enhancing diagnostic accuracy and reliability. Methods: The proposed method for brain tumor classification integrates segmentation using K-means++, feature extraction from the Spatial Gray Level Dependence Matrix (SGLDM), and classification with ResNet50, along with synthetic data augmentation to enhance model robustness. Segmentation isolates tumor regions, while SGLDM captures critical texture information. The ResNet50 model then classifies the tumors accurately. To further improve the interpretability of the classification results, Grad-CAM is employed, providing visual explanations by highlighting influential regions in the MRI images. Result: In terms of accuracy, sensitivity, and specificity, the evaluation on the Br35H::BrainTumorDetection2020 dataset showed superior performance of the suggested method compared to existing state-of-the-art approaches. This indicates its effectiveness in achieving higher precision in identifying and classifying brain tumors from MRI data, showcasing advancements in diagnostic reliability and efficacy. Discussion: The superior performance of the suggested method indicates its robustness in accurately classifying brain tumors from MRI images, achieving higher accuracy, sensitivity, and specificity compared to existing methods. The method's enhanced sensitivity ensures a greater detection rate of true positive cases, while its improved specificity reduces false positives, thereby optimizing clinical decision-making and patient care in neuro-oncology.
Collapse
Affiliation(s)
- Ponuku Sarah
- School of Electronics Engineering, Vellore Institute of Technology, Vellore, India
| | - Srigiri Krishnapriya
- School of Electronics Engineering, Vellore Institute of Technology, Vellore, India
| | - Saritha Saladi
- School of Electronics Engineering, VIT-AP University, Amaravati, India
| | - Yepuganti Karuna
- School of Electronics Engineering, VIT-AP University, Amaravati, India
| | - Durga Prasad Bavirisetti
- Department of Computer Science, Norwegian University of Science and Technology, Trondheim, Norway
| |
Collapse
|
36
|
D S CS, Christopher Clement J. Enhancing brain tumor segmentation in MRI images using the IC-net algorithm framework. Sci Rep 2024; 14:15660. [PMID: 38977779 PMCID: PMC11231217 DOI: 10.1038/s41598-024-66314-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/01/2024] [Accepted: 07/01/2024] [Indexed: 07/10/2024] Open
Abstract
Brain tumors, often referred to as intracranial tumors, are abnormal tissue masses that arise from rapidly multiplying cells. During medical imaging, it is essential to separate brain tumors from healthy tissue. The goal of this paper is to improve the accuracy of separating tumorous regions from healthy tissues in medical imaging, specifically for brain tumors in MRI images which is difficult in the field of medical image analysis. In our research work, we propose IC-Net (Inverted-C), a novel semantic segmentation architecture that combines elements from various models to provide effective and precise results. The architecture includes Multi-Attention (MA) blocks, Feature Concatenation Networks (FCN), Attention-blocks which performs crucial tasks in improving brain tumor segmentation. MA-block aggregates multi-attention features to adapt to different tumor sizes and shapes. Attention-block is focusing on key regions, resulting in more effective segmentation in complex images. FCN-block captures diverse features, making the model more robust to various characteristics of brain tumor images. Our proposed architecture is used to accelerate the training process and also to address the challenges posed by the diverse nature of brain tumor images, ultimately leads to potentially improved segmentation performance. IC-Net significantly outperforms the typical U-Net architecture and other contemporary effective segmentation techniques. On the BraTS 2020 dataset, our IC-Net design obtained notable outcomes in Accuracy, Loss, Specificity, Sensitivity as 99.65, 0.0159, 99.44, 99.86 and DSC (core, whole, and enhancing tumors as 0.998717, 0.888930, 0.866183) respectively.
Collapse
Affiliation(s)
- Chandra Sekaran D S
- School of Electronics Engineering, Vellore Institute of Technology, Vellore, Tamilnadu, 632014, India
| | - J Christopher Clement
- School of Electronics Engineering, Vellore Institute of Technology, Vellore, Tamilnadu, 632014, India.
| |
Collapse
|
37
|
Li Y, El Habib Daho M, Conze PH, Zeghlache R, Le Boité H, Tadayoni R, Cochener B, Lamard M, Quellec G. A review of deep learning-based information fusion techniques for multimodal medical image classification. Comput Biol Med 2024; 177:108635. [PMID: 38796881 DOI: 10.1016/j.compbiomed.2024.108635] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/05/2023] [Revised: 03/18/2024] [Accepted: 05/18/2024] [Indexed: 05/29/2024]
Abstract
Multimodal medical imaging plays a pivotal role in clinical diagnosis and research, as it combines information from various imaging modalities to provide a more comprehensive understanding of the underlying pathology. Recently, deep learning-based multimodal fusion techniques have emerged as powerful tools for improving medical image classification. This review offers a thorough analysis of the developments in deep learning-based multimodal fusion for medical classification tasks. We explore the complementary relationships among prevalent clinical modalities and outline three main fusion schemes for multimodal classification networks: input fusion, intermediate fusion (encompassing single-level fusion, hierarchical fusion, and attention-based fusion), and output fusion. By evaluating the performance of these fusion techniques, we provide insight into the suitability of different network architectures for various multimodal fusion scenarios and application domains. Furthermore, we delve into challenges related to network architecture selection, handling incomplete multimodal data management, and the potential limitations of multimodal fusion. Finally, we spotlight the promising future of Transformer-based multimodal fusion techniques and give recommendations for future research in this rapidly evolving field.
Collapse
Affiliation(s)
- Yihao Li
- LaTIM UMR 1101, Inserm, Brest, France; University of Western Brittany, Brest, France
| | - Mostafa El Habib Daho
- LaTIM UMR 1101, Inserm, Brest, France; University of Western Brittany, Brest, France.
| | | | - Rachid Zeghlache
- LaTIM UMR 1101, Inserm, Brest, France; University of Western Brittany, Brest, France
| | - Hugo Le Boité
- Sorbonne University, Paris, France; Ophthalmology Department, Lariboisière Hospital, AP-HP, Paris, France
| | - Ramin Tadayoni
- Ophthalmology Department, Lariboisière Hospital, AP-HP, Paris, France; Paris Cité University, Paris, France
| | - Béatrice Cochener
- LaTIM UMR 1101, Inserm, Brest, France; University of Western Brittany, Brest, France; Ophthalmology Department, CHRU Brest, Brest, France
| | - Mathieu Lamard
- LaTIM UMR 1101, Inserm, Brest, France; University of Western Brittany, Brest, France
| | | |
Collapse
|
38
|
Shunmugavel G, Suriyan K, Arumugam J. Magnetic Resonance Imaging Images Based Brain Tumor Extraction, Segmentation and Detection Using Convolutional Neural Network and VGC 16 Model. Am J Clin Oncol 2024; 47:339-349. [PMID: 38632686 DOI: 10.1097/coc.0000000000001097] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/19/2024]
Abstract
OBJECTIVES In this paper, we look at how to design and build a system to find tumors using 2 Convolutional Neural Network (CNN) models. With the help of digital image processing and deep Learning, we can make a system that automatically diagnoses and finds different diseases and abnormalities. The tumor detection system may include image enhancement, segmentation, data enhancement, feature extraction, and classification. These options are set up so that the CNN model can give the best results. METHODS During the training phase, the learning rate is used to change the weights and bias. The learning rate also changes the weights. One Epoch is when all of the training images are shown to the model. As the training data may be very large, the data in each epoch are split into batches. Every epoch has a training session and a test session. After each epoch, the weights are changed based on how fast the CNN is learning. This is done with the help of optimization algorithms. The suggested technique uses the anticipated mean intersection over union value to identify failure instances in addition to forecasting the mean intersection over union. RESULTS This paper talks about how to separate brain tumors from magnetic resonance images of patients taken from "Brain web." Using basic ideas of digital image processing, magnetic resonance images are used to extract and find tumors using a hybrid method. In this paper, the proposed algorithm is applied with the help of MATLAB. In medical image processing, brain tumor segmentation is an important task. The goal of this paper is to look at different ways to divide brain tumors using magnetic resonance imaging. Recently, automatic segmentation using deep learning methods has become popular because these methods get the best results and are better at solving this problem than others. Deep learning methods can also be used to process and evaluate large amounts of magnetic resonance imaging image data quickly and objectively. CONCLUSIONS A classification method based on a convolution neural network is also added to the proposed scheme to make it more accurate and cut down on the amount of time it takes to do the calculations. Also, the results of the classification are given as images of a tumor or a healthy brain. The training is 98.5% correct. In the same way, both the validation accuracy and validation loss are high.
Collapse
Affiliation(s)
| | - Kannadhasan Suriyan
- Department of Electronics and Communication Engineering, Study World College of Engineering
| | - Jayachandran Arumugam
- Department of Computer Science and Engineering, PSN College of Engineering and Technology, Tirunelveli, TN, India
| |
Collapse
|
39
|
Kumar R, Bernard CM, Ullah A, Khan RU, Kumar J, Kulevome DKB, Yunbo R, Zeng S. Privacy-preserving blockchain-based federated learning for brain tumor segmentation. Comput Biol Med 2024; 177:108646. [PMID: 38824788 DOI: 10.1016/j.compbiomed.2024.108646] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2023] [Revised: 05/15/2024] [Accepted: 05/21/2024] [Indexed: 06/04/2024]
Abstract
Improved data sharing between healthcare providers can lead to a higher probability of accurate diagnosis, more effective treatments, and enhanced capabilities of healthcare organizations. One critical area of focus is brain tumor segmentation, a complex task due to the heterogeneous appearance, irregular shape, and variable location of tumors. Accurate segmentation is essential for proper diagnosis and effective treatment planning, yet current techniques often fall short due to these complexities. However, the sensitive nature of health data often prohibits its sharing. Moreover, the healthcare industry faces significant issues, including preserving the privacy of the model and instilling trust in the model. This paper proposes a framework to address these privacy and trust issues by introducing a mechanism for training the global model using federated learning and sharing the encrypted learned parameters via a permissioned blockchain. The blockchain-federated learning algorithm we designed aggregates gradients in the permissioned blockchain to decentralize the global model, while the introduced masking approach retains the privacy of the model parameters. Unlike traditional raw data sharing, this approach enables hospitals or medical research centers to contribute to a globally learned model, thereby enhancing the performance of the central model for all participating medical entities. As a result, the global model can learn about several specific diseases and benefit each contributor with new disease diagnosis tasks, leading to improved treatment options. The proposed algorithm ensures the quality of model data when aggregating the local model, using an asynchronous federated learning procedure to evaluate the shared model's quality. The experimental results demonstrate the efficacy of the proposed scheme for the critical and challenging task of brain tumor segmentation. Specifically, our method achieved a 1.99% improvement in Dice similarity coefficient for enhancing tumors and a 19.08% reduction in Hausdorff distance for whole tumors compared to the baseline methods, highlighting the significant advancement in segmentation performance and reliability.
Collapse
Affiliation(s)
- Rajesh Kumar
- Yangtze Delta Region Institute (Huzhou), University of Electronic Science and Technology of China, Huzhou 313001, China
| | - Cobbinah M Bernard
- Yangtze Delta Region Institute (Huzhou), University of Electronic Science and Technology of China, Huzhou 313001, China.
| | - Aman Ullah
- Yangtze Delta Region Institute (Huzhou), University of Electronic Science and Technology of China, Huzhou 313001, China
| | - Riaz Ullah Khan
- Yangtze Delta Region Institute (Huzhou), University of Electronic Science and Technology of China, Huzhou 313001, China
| | - Jay Kumar
- Institute for Big Data Analytics, Dalhousie University, Halifax, NS, Canada
| | - Delanyo K B Kulevome
- Department of Intelligent Manufacturing and Control Engineering, Qilu Institute of Technology, Jinan, Shangdong, 250200, China
| | - Rao Yunbo
- Yangtze Delta Region Institute (Huzhou), University of Electronic Science and Technology of China, Huzhou 313001, China
| | - Shaoning Zeng
- Yangtze Delta Region Institute (Huzhou), University of Electronic Science and Technology of China, Huzhou 313001, China
| |
Collapse
|
40
|
Balraj K, Ramteke M, Mittal S, Bhargava R, Rathore AS. MADR-Net: multi-level attention dilated residual neural network for segmentation of medical images. Sci Rep 2024; 14:12699. [PMID: 38830932 PMCID: PMC11148105 DOI: 10.1038/s41598-024-63538-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/03/2023] [Accepted: 05/29/2024] [Indexed: 06/05/2024] Open
Abstract
Medical image segmentation has made a significant contribution towards delivering affordable healthcare by facilitating the automatic identification of anatomical structures and other regions of interest. Although convolution neural networks have become prominent in the field of medical image segmentation, they suffer from certain limitations. In this study, we present a reliable framework for producing performant outcomes for the segmentation of pathological structures of 2D medical images. Our framework consists of a novel deep learning architecture, called deep multi-level attention dilated residual neural network (MADR-Net), designed to improve the performance of medical image segmentation. MADR-Net uses a U-Net encoder/decoder backbone in combination with multi-level residual blocks and atrous pyramid scene parsing pooling. To improve the segmentation results, channel-spatial attention blocks were added in the skip connection to capture both the global and local features and superseded the bottleneck layer with an ASPP block. Furthermore, we introduce a hybrid loss function that has an excellent convergence property and enhances the performance of the medical image segmentation task. We extensively validated the proposed MADR-Net on four typical yet challenging medical image segmentation tasks: (1) Left ventricle, left atrium, and myocardial wall segmentation from Echocardiogram images in the CAMUS dataset, (2) Skin cancer segmentation from dermoscopy images in ISIC 2017 dataset, (3) Electron microscopy in FIB-SEM dataset, and (4) Fluid attenuated inversion recovery abnormality from MR images in LGG segmentation dataset. The proposed algorithm yielded significant results when compared to state-of-the-art architectures such as U-Net, Residual U-Net, and Attention U-Net. The proposed MADR-Net consistently outperformed the classical U-Net by 5.43%, 3.43%, and 3.92% relative improvement in terms of dice coefficient, respectively, for electron microscopy, dermoscopy, and MRI. The experimental results demonstrate superior performance on single and multi-class datasets and that the proposed MADR-Net can be utilized as a baseline for the assessment of cross-dataset and segmentation tasks.
Collapse
Affiliation(s)
- Keerthiveena Balraj
- Yardi School of Artificial Intelligence, Indian Institute of Technology Delhi, Hauz Khas, New Delhi, 110016, India
| | - Manojkumar Ramteke
- Yardi School of Artificial Intelligence, Indian Institute of Technology Delhi, Hauz Khas, New Delhi, 110016, India
- Department of Chemical Engineering, Indian Institute of Technology Delhi, Hauz Khas, New Delhi, India
| | - Shachi Mittal
- Department of Laboratory Medicine and Pathology, School of Medicine, University of Washington, Seattle, WA, USA
| | - Rohit Bhargava
- Departments of Bioengineering, Electrical and Computer Engineering, Mechanical Science and Engineering, Chemical and Biomolecular Engineering and Chemistry, Beckman Institute for Advanced Science and Technology, Cancer Center at Illinois, University of Illinois at Urbana-Champaign, Urbana, IL, 61801, USA
| | - Anurag S Rathore
- Yardi School of Artificial Intelligence, Indian Institute of Technology Delhi, Hauz Khas, New Delhi, 110016, India.
- Department of Chemical Engineering, Indian Institute of Technology Delhi, Hauz Khas, New Delhi, India.
| |
Collapse
|
41
|
Valverde S, Coll L, Valencia L, Clèrigues A, Oliver A, Vilanova JC, Ramió-Torrentà L, Rovira À, Lladó X. Assessing the Accuracy and Reproducibility of PARIETAL: A Deep Learning Brain Extraction Algorithm. J Magn Reson Imaging 2024; 59:1991-2000. [PMID: 34137113 DOI: 10.1002/jmri.27776] [Citation(s) in RCA: 7] [Impact Index Per Article: 7.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/10/2020] [Revised: 05/31/2021] [Accepted: 06/01/2021] [Indexed: 01/18/2023] Open
Abstract
BACKGROUND Manual brain extraction from magnetic resonance (MR) images is time-consuming and prone to intra- and inter-rater variability. Several automated approaches have been developed to alleviate these constraints, including deep learning pipelines. However, these methods tend to reduce their performance in unseen magnetic resonance imaging (MRI) scanner vendors and different imaging protocols. PURPOSE To present and evaluate for clinical use PARIETAL, a pre-trained deep learning brain extraction method. We compare its reproducibility in a scan/rescan analysis and its robustness among scanners of different manufacturers. STUDY TYPE Retrospective. POPULATION Twenty-one subjects (12 women) with age range 22-48 years acquired using three different MRI scanner machines including scan/rescan in each of them. FIELD STRENGTH/SEQUENCE T1-weighted images acquired in a 3-T Siemens with magnetization prepared rapid gradient-echo sequence and two 1.5 T scanners, Philips and GE, with spin-echo and spoiled gradient-recalled (SPGR) sequences, respectively. ASSESSMENT Analysis of the intracranial cavity volumes obtained for each subject on the three different scanners and the scan/rescan acquisitions. STATISTICAL TESTS Parametric permutation tests of the differences in volumes to rank and statistically evaluate the performance of PARIETAL compared to state-of-the-art methods. RESULTS The mean absolute intracranial volume differences obtained by PARIETAL in the scan/rescan analysis were 1.88 mL, 3.91 mL, and 4.71 mL for Siemens, GE, and Philips scanners, respectively. PARIETAL was the best-ranked method on Siemens and GE scanners, while decreasing to Rank 2 on the Philips images. Intracranial differences for the same subject between scanners were 5.46 mL, 27.16 mL, and 30.44 mL for GE/Philips, Siemens/Philips, and Siemens/GE comparison, respectively. The permutation tests revealed that PARIETAL was always in Rank 1, obtaining the most similar volumetric results between scanners. DATA CONCLUSION PARIETAL accurately segments the brain and it generalizes to images acquired at different sites without the need of training or fine-tuning it again. PARIETAL is publicly available. LEVEL OF EVIDENCE 2 TECHNICAL EFFICACY STAGE: 2.
Collapse
Affiliation(s)
- Sergi Valverde
- Research Institute of Computer Vision and Robotics, University of Girona, Girona, Spain
| | - Llucia Coll
- Research Institute of Computer Vision and Robotics, University of Girona, Girona, Spain
| | - Liliana Valencia
- Research Institute of Computer Vision and Robotics, University of Girona, Girona, Spain
| | - Albert Clèrigues
- Research Institute of Computer Vision and Robotics, University of Girona, Girona, Spain
| | - Arnau Oliver
- Research Institute of Computer Vision and Robotics, University of Girona, Girona, Spain
- REEM, Red Española de Esclerosis Múltiple
| | | | - Lluís Ramió-Torrentà
- REEM, Red Española de Esclerosis Múltiple
- Multiple Sclerosis and Neuroimmunology Unit, Neurology Department, Dr. Josep Trueta University Hospital, Institut d'Investigació Biomèdica, Girona, Spain
- Medical Sciences Department, University of Girona, Girona, Spain
| | - Àlex Rovira
- Magnetic Resonance Unit, Department of Radiology, Vall d'Hebron University Hospital, Barcelona, Spain
| | - Xavier Lladó
- Research Institute of Computer Vision and Robotics, University of Girona, Girona, Spain
- REEM, Red Española de Esclerosis Múltiple
| |
Collapse
|
42
|
Seoni S, Shahini A, Meiburger KM, Marzola F, Rotunno G, Acharya UR, Molinari F, Salvi M. All you need is data preparation: A systematic review of image harmonization techniques in Multi-center/device studies for medical support systems. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 250:108200. [PMID: 38677080 DOI: 10.1016/j.cmpb.2024.108200] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/27/2024] [Revised: 04/20/2024] [Accepted: 04/22/2024] [Indexed: 04/29/2024]
Abstract
BACKGROUND AND OBJECTIVES Artificial intelligence (AI) models trained on multi-centric and multi-device studies can provide more robust insights and research findings compared to single-center studies. However, variability in acquisition protocols and equipment can introduce inconsistencies that hamper the effective pooling of multi-source datasets. This systematic review evaluates strategies for image harmonization, which standardizes appearances to enable reliable AI analysis of multi-source medical imaging. METHODS A literature search using PRISMA guidelines was conducted to identify relevant papers published between 2013 and 2023 analyzing multi-centric and multi-device medical imaging studies that utilized image harmonization approaches. RESULTS Common image harmonization techniques included grayscale normalization (improving classification accuracy by up to 24.42 %), resampling (increasing the percentage of robust radiomics features from 59.5 % to 89.25 %), and color normalization (enhancing AUC by up to 0.25 in external test sets). Initially, mathematical and statistical methods dominated, but machine and deep learning adoption has risen recently. Color imaging modalities like digital pathology and dermatology have remained prominent application areas, though harmonization efforts have expanded to diverse fields including radiology, nuclear medicine, and ultrasound imaging. In all the modalities covered by this review, image harmonization improved AI performance, with increasing of up to 24.42 % in classification accuracy and 47 % in segmentation Dice scores. CONCLUSIONS Continued progress in image harmonization represents a promising strategy for advancing healthcare by enabling large-scale, reliable analysis of integrated multi-source datasets using AI. Standardizing imaging data across clinical settings can help realize personalized, evidence-based care supported by data-driven technologies while mitigating biases associated with specific populations or acquisition protocols.
Collapse
Affiliation(s)
- Silvia Seoni
- Biolab, PolitoBIOMedLab, Department of Electronics and Telecommunications, Politecnico di Torino, Turin, Italy
| | - Alen Shahini
- Biolab, PolitoBIOMedLab, Department of Electronics and Telecommunications, Politecnico di Torino, Turin, Italy
| | - Kristen M Meiburger
- Biolab, PolitoBIOMedLab, Department of Electronics and Telecommunications, Politecnico di Torino, Turin, Italy
| | - Francesco Marzola
- Biolab, PolitoBIOMedLab, Department of Electronics and Telecommunications, Politecnico di Torino, Turin, Italy
| | - Giulia Rotunno
- Biolab, PolitoBIOMedLab, Department of Electronics and Telecommunications, Politecnico di Torino, Turin, Italy
| | - U Rajendra Acharya
- School of Mathematics, Physics and Computing, University of Southern Queensland, Springfield, Australia; Centre for Health Research, University of Southern Queensland, Australia
| | - Filippo Molinari
- Biolab, PolitoBIOMedLab, Department of Electronics and Telecommunications, Politecnico di Torino, Turin, Italy
| | - Massimo Salvi
- Biolab, PolitoBIOMedLab, Department of Electronics and Telecommunications, Politecnico di Torino, Turin, Italy.
| |
Collapse
|
43
|
Batool A, Byun YC. Brain tumor detection with integrating traditional and computational intelligence approaches across diverse imaging modalities - Challenges and future directions. Comput Biol Med 2024; 175:108412. [PMID: 38691914 DOI: 10.1016/j.compbiomed.2024.108412] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/19/2023] [Revised: 03/18/2024] [Accepted: 04/02/2024] [Indexed: 05/03/2024]
Abstract
Brain tumor segmentation and classification play a crucial role in the diagnosis and treatment planning of brain tumors. Accurate and efficient methods for identifying tumor regions and classifying different tumor types are essential for guiding medical interventions. This study comprehensively reviews brain tumor segmentation and classification techniques, exploring various approaches based on image processing, machine learning, and deep learning. Furthermore, our study aims to review existing methodologies, discuss their advantages and limitations, and highlight recent advancements in this field. The impact of existing segmentation and classification techniques for automated brain tumor detection is also critically examined using various open-source datasets of Magnetic Resonance Images (MRI) of different modalities. Moreover, our proposed study highlights the challenges related to segmentation and classification techniques and datasets having various MRI modalities to enable researchers to develop innovative and robust solutions for automated brain tumor detection. The results of this study contribute to the development of automated and robust solutions for analyzing brain tumors, ultimately aiding medical professionals in making informed decisions and providing better patient care.
Collapse
Affiliation(s)
- Amreen Batool
- Department of Electronic Engineering, Institute of Information Science & Technology, Jeju National University, Jeju, 63243, South Korea
| | - Yung-Cheol Byun
- Department of Computer Engineering, Major of Electronic Engineering, Jeju National University, Institute of Information Science Technology, Jeju, 63243, South Korea.
| |
Collapse
|
44
|
Mulqueeney JM, Searle-Barnes A, Brombacher A, Sweeney M, Goswami A, Ezard THG. How many specimens make a sufficient training set for automated three-dimensional feature extraction? ROYAL SOCIETY OPEN SCIENCE 2024; 11:rsos.240113. [PMID: 39100182 PMCID: PMC11296157 DOI: 10.1098/rsos.240113] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 01/18/2024] [Revised: 04/24/2024] [Accepted: 04/25/2024] [Indexed: 08/06/2024]
Abstract
Deep learning has emerged as a robust tool for automating feature extraction from three-dimensional images, offering an efficient alternative to labour-intensive and potentially biased manual image segmentation methods. However, there has been limited exploration into the optimal training set sizes, including assessing whether artficial expansion by data augmentation can achieve consistent results in less time and how consistent these benefits are across different types of traits. In this study, we manually segmented 50 planktonic foraminifera specimens from the genus Menardella to determine the minimum number of training images required to produce accurate volumetric and shape data from internal and external structures. The results reveal unsurprisingly that deep learning models improve with a larger number of training images with eight specimens being required to achieve 95% accuracy. Furthermore, data augmentation can enhance network accuracy by up to 8.0%. Notably, predicting both volumetric and shape measurements for the internal structure poses a greater challenge compared with the external structure, owing to low contrast differences between different materials and increased geometric complexity. These results provide novel insight into optimal training set sizes for precise image segmentation of diverse traits and highlight the potential of data augmentation for enhancing multivariate feature extraction from three-dimensional images.
Collapse
Affiliation(s)
- James M. Mulqueeney
- School of Ocean & Earth Science, National Oceanography Centre Southampton, University of Southampton Waterfront Campus, Southampton, UK
- Department of Life Sciences, Natural History Museum, London, UK
| | - Alex Searle-Barnes
- School of Ocean & Earth Science, National Oceanography Centre Southampton, University of Southampton Waterfront Campus, Southampton, UK
| | - Anieke Brombacher
- School of Ocean & Earth Science, National Oceanography Centre Southampton, University of Southampton Waterfront Campus, Southampton, UK
| | - Marisa Sweeney
- School of Ocean & Earth Science, National Oceanography Centre Southampton, University of Southampton Waterfront Campus, Southampton, UK
| | - Anjali Goswami
- Department of Life Sciences, Natural History Museum, London, UK
| | - Thomas H. G. Ezard
- School of Ocean & Earth Science, National Oceanography Centre Southampton, University of Southampton Waterfront Campus, Southampton, UK
| |
Collapse
|
45
|
Brüggemann D, Cener D, Kuzo N, Anwer S, Kebernik J, Eberhard M, Alkadhi H, Tanner FC, Konukoglu E. Predicting mortality after transcatheter aortic valve replacement using preprocedural CT. Sci Rep 2024; 14:12526. [PMID: 38822074 PMCID: PMC11143216 DOI: 10.1038/s41598-024-63022-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/14/2023] [Accepted: 05/23/2024] [Indexed: 06/02/2024] Open
Abstract
Transcatheter aortic valve replacement (TAVR) is a widely used intervention for patients with severe aortic stenosis. Identifying high-risk patients is crucial due to potential postprocedural complications. Currently, this involves manual clinical assessment and time-consuming radiological assessment of preprocedural computed tomography (CT) images by an expert radiologist. In this study, we introduce a probabilistic model that predicts post-TAVR mortality automatically using unprocessed, preprocedural CT and 25 baseline patient characteristics. The model utilizes CT volumes by automatically localizing and extracting a region of interest around the aortic root and ascending aorta. It then extracts task-specific features with a 3D deep neural network and integrates them with patient characteristics to perform outcome prediction. As missing measurements or even missing CT images are common in TAVR planning, the proposed model is designed with a probabilistic structure to allow for marginalization over such missing information. Our model demonstrates an AUROC of 0.725 for predicting all-cause mortality during postprocedure follow-up on a cohort of 1449 TAVR patients. This performance is on par with what can be achieved with lengthy radiological assessments performed by experts. Thus, these findings underscore the potential of the proposed model in automatically analyzing CT volumes and integrating them with patient characteristics for predicting mortality after TAVR.
Collapse
Affiliation(s)
- David Brüggemann
- Computer Vision Laboratory, ETH Zurich, 8092, Zurich, Switzerland
| | - Denis Cener
- Computer Vision Laboratory, ETH Zurich, 8092, Zurich, Switzerland
| | - Nazar Kuzo
- Department of Cardiology, University Heart Center, University Hospital Zurich, 8091, Zurich, Switzerland
| | - Shehab Anwer
- Department of Cardiology, University Heart Center, University Hospital Zurich, 8091, Zurich, Switzerland
| | - Julia Kebernik
- Institute for Diagnostic and Interventional Radiology, University Hospital Zurich, 8091, Zurich, Switzerland
| | - Matthias Eberhard
- Institute for Diagnostic and Interventional Radiology, University Hospital Zurich, 8091, Zurich, Switzerland
| | - Hatem Alkadhi
- Institute for Diagnostic and Interventional Radiology, University Hospital Zurich, 8091, Zurich, Switzerland
| | - Felix C Tanner
- Department of Cardiology, University Heart Center, University Hospital Zurich, 8091, Zurich, Switzerland
| | - Ender Konukoglu
- Computer Vision Laboratory, ETH Zurich, 8092, Zurich, Switzerland.
| |
Collapse
|
46
|
Usha MP, Kannan G, Ramamoorthy M. Multimodal Brain Tumor Classification Using Convolutional Tumnet Architecture. Behav Neurol 2024; 2024:4678554. [PMID: 38882177 PMCID: PMC11178426 DOI: 10.1155/2024/4678554] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2022] [Revised: 12/22/2023] [Accepted: 01/10/2024] [Indexed: 06/18/2024] Open
Abstract
The most common and aggressive tumor is brain malignancy, which has a short life span in the fourth grade of the disease. As a result, the medical plan may be a crucial step toward improving the well-being of a patient. Both diagnosis and therapy are part of the medical plan. Brain tumors are commonly imaged with magnetic resonance imaging (MRI), positron emission tomography (PET), and computed tomography (CT). In this paper, multimodal fused imaging with classification and segmentation for brain tumors was proposed using the deep learning method. The MRI and CT brain tumor images of the same slices (308 slices of meningioma and sarcoma) are combined using three different types of pixel-level fusion methods. The presence/absence of a tumor is classified using the proposed Tumnet technique, and the tumor area is found accordingly. In the other case, Tumnet is also applied for single-modal MRI/CT (561 image slices) for classification. The proposed Tumnet was modeled with 5 convolutional layers, 3 pooling layers with ReLU activation function, and 3 fully connected layers. The first-order statistical fusion metrics for an average method of MRI-CT images are obtained as SSIM tissue at 83%, SSIM bone at 84%, accuracy at 90%, sensitivity at 96%, and specificity at 95%, and the second-order statistical fusion metrics are obtained as the standard deviation of fused images at 79% and entropy at 0.99. The entropy value confirms the presence of additional features in the fused image. The proposed Tumnet yields a sensitivity of 96%, an accuracy of 98%, a specificity of 99%, normalized values of the mean of 0.75, a standard deviation of 0.4, a variance of 0.16, and an entropy of 0.90.
Collapse
Affiliation(s)
- M Padma Usha
- Department of Electronics and Communication Engineering B.S. Abdur Rahman Crescent Institute of Science and Technology, Vandalur, Chennai, India
| | - G Kannan
- Department of Electronics and Communication Engineering B.S. Abdur Rahman Crescent Institute of Science and Technology, Vandalur, Chennai, India
| | - M Ramamoorthy
- Department of Artificial Intelligence and Machine Learning Saveetha School of Engineering SIMATS, Chennai, 600124, India
| |
Collapse
|
47
|
Chen B, Sun Q, Han Y, Liu B, Zhang J, Zhang Q. Adaptive cascaded transformer U-Net for MRI brain tumor segmentation. Phys Med Biol 2024; 69:115036. [PMID: 38636503 DOI: 10.1088/1361-6560/ad4081] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2023] [Accepted: 04/18/2024] [Indexed: 04/20/2024]
Abstract
Objective.Brain tumor segmentation on magnetic resonance imaging (MRI) plays an important role in assisting the diagnosis and treatment of cancer patients. Recently, cascaded U-Net models have achieved excellent performance via conducting coarse-to-fine segmentation of MRI brain tumors. However, they are still restricted by obvious global and local differences among various brain tumors, which are difficult to solve with conventional convolutions.Approach.To address the issue, this study proposes a novel Adaptive Cascaded Transformer U-Net (ACTransU-Net) for MRI brain tumor segmentation, which simultaneously integrates Transformer and dynamic convolution into a single cascaded U-Net architecture to adaptively capture global information and local details of brain tumors. ACTransU-Net first cascades two 3D U-Nets into a two-stage network to segment brain tumors from coarse to fine. Subsequently, it integrates omni-dimensional dynamic convolution modules into the second-stage shallow encoder and decoder, thereby enhancing the local detail representation of various brain tumors through dynamically adjusting convolution kernel parameters. Moreover, 3D Swin-Transformer modules are introduced into the second-stage deep encoder and decoder to capture image long-range dependencies, which helps adapt the global representation of brain tumors.Main results.Extensive experiment results evaluated on the public BraTS 2020 and BraTS 2021 brain tumor data sets demonstrate the effectiveness of ACTransU-Net, with average DSC of 84.96% and 91.37%, and HD95 of 10.81 and 7.31 mm, proving competitiveness with the state-of-the-art methods.Significance.The proposed method focuses on adaptively capturing both global information and local details of brain tumors, aiding physicians in their accurate diagnosis. In addition, it has the potential to extend ACTransU-Net for segmenting other types of lesions. The source code is available at:https://github.com/chenbn266/ACTransUnet.
Collapse
Affiliation(s)
- Bonian Chen
- School of Computer Science and Engineering, Dalian Minzu University, Dalian, 116600, People's Republic of China
| | - Qiule Sun
- Faculty of Electronic and Information Engineering, Dalian University of Technology, Dalian, 116024, People's Republic of China
| | - Yutong Han
- School of Computer Science and Engineering, Dalian Minzu University, Dalian, 116600, People's Republic of China
| | - Bin Liu
- International School of Information Science and Engineering (DUT-RUISE), Dalian University of Technology, Dalian, 116620, People's Republic of China
| | - Jianxin Zhang
- School of Computer Science and Engineering, Dalian Minzu University, Dalian, 116600, People's Republic of China
| | - Qiang Zhang
- Faculty of Electronic and Information Engineering, Dalian University of Technology, Dalian, 116024, People's Republic of China
| |
Collapse
|
48
|
Lin CL, Wu MH, Ho YH, Lin FY, Lu YH, Hsueh YY, Chen CC. Multispectral Imaging-Based System for Detecting Tissue Oxygen Saturation With Wound Segmentation for Monitoring Wound Healing. IEEE JOURNAL OF TRANSLATIONAL ENGINEERING IN HEALTH AND MEDICINE 2024; 12:468-479. [PMID: 38899145 PMCID: PMC11186648 DOI: 10.1109/jtehm.2024.3399232] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/15/2024] [Revised: 04/13/2024] [Accepted: 05/07/2024] [Indexed: 06/21/2024]
Abstract
OBJECTIVE Blood circulation is an important indicator of wound healing. In this study, a tissue oxygen saturation detecting (TOSD) system that is based on multispectral imaging (MSI) is proposed to quantify the degree of tissue oxygen saturation (StO2) in cutaneous tissues. METHODS A wound segmentation algorithm is used to segment automatically wound and skin areas, eliminating the need for manual labeling and applying adaptive tissue optics. Animal experiments were conducted on six mice in which they were observed seven times, once every two days. The TOSD system illuminated cutaneous tissues with two wavelengths of light - red ([Formula: see text] nm) and near-infrared ([Formula: see text] nm), and StO2 levels were calculated using images that were captured using a monochrome camera. The wound segmentation algorithm using ResNet34-based U-Net was integrated with computer vision techniques to improve its performance. RESULTS Animal experiments revealed that the wound segmentation algorithm achieved a Dice score of 93.49%. The StO2 levels that were determined using the TOSD system varied significantly among the phases of wound healing. Changes in StO2 levels were detected before laser speckle contrast imaging (LSCI) detected changes in blood flux. Moreover, statistical features that were extracted from the TOSD system and LSCI were utilized in principal component analysis (PCA) to visualize different wound healing phases. The average silhouette coefficients of the TOSD system with segmentation (ResNet34-based U-Net) and LSCI were 0.2890 and 0.0194, respectively. CONCLUSION By detecting the StO2 levels of cutaneous tissues using the TOSD system with segmentation, the phases of wound healing were accurately distinguished. This method can support medical personnel in conducting precise wound assessments. Clinical and Translational Impact Statement-This study supports efforts in monitoring StO2 levels, wound segmentation, and wound healing phase classification to improve the efficiency and accuracy of preclinical research in the field.
Collapse
Affiliation(s)
- Chih-Lung Lin
- Department of Electrical EngineeringNational Cheng Kung UniversityTainan70101Taiwan
| | - Meng-Hsuan Wu
- Department of Electrical EngineeringNational Cheng Kung UniversityTainan70101Taiwan
| | - Yuan-Hao Ho
- Department of Electrical EngineeringNational Cheng Kung UniversityTainan70101Taiwan
| | - Fang-Yi Lin
- Department of Electrical EngineeringNational Cheng Kung UniversityTainan70101Taiwan
| | - Yu-Hsien Lu
- Department of Electrical EngineeringNational Cheng Kung UniversityTainan70101Taiwan
| | - Yuan-Yu Hsueh
- Division of Plastic and Reconstructive SurgeryNational Cheng Kung University HospitalTainan70428Taiwan
- Department of SurgeryNational Cheng Kung University HospitalTainan70428Taiwan
| | - Chia-Chen Chen
- Department of Electrical EngineeringNational Cheng Kung UniversityTainan70101Taiwan
| |
Collapse
|
49
|
Khodadadi Shoushtari F, Dehkordi ANV, Sina S. Quantitative and Visual Analysis of Data Augmentation and Hyperparameter Optimization in Deep Learning-Based Segmentation of Low-Grade Glioma Tumors Using Grad-CAM. Ann Biomed Eng 2024; 52:1359-1377. [PMID: 38409433 DOI: 10.1007/s10439-024-03461-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2023] [Accepted: 01/29/2024] [Indexed: 02/28/2024]
Abstract
This study executes a quantitative and visual investigation on the effectiveness of data augmentation and hyperparameter optimization on the accuracy of deep learning-based segmentation of LGG tumors. The study employed the MobileNetV2 and ResNet backbones with atrous convolution in DeepLabV3+ structure. The Grad-CAM tool was also used to interpret the effect of augmentation and network optimization on segmentation performance. A wide investigation was performed to optimize the network hyperparameters. In addition, the study examined 35 different models to evaluate different data augmentation techniques. The results of the study indicated that incorporating data augmentation techniques and optimization can improve the performance of segmenting brain LGG tumors up to 10%. Our extensive investigation of the data augmentation techniques indicated that enlargement of data from 90° and 225° rotated data,up to down and left to right flipping are the most effective techniques. MobilenetV2 as the backbone,"Focal Loss" as the loss function and "Adam" as the optimizer showed the superior results. The optimal model (DLG-Net) achieved an overall accuracy of 96.1% with a loss value of 0.006. Specifically, the segmentation performance for Whole Tumor (WT), Tumor Core (TC), and Enhanced Tumor (ET) reached a Dice Similarity Coefficient (DSC) of 89.4%, 70.1%, and 49.9%, respectively. Simultaneous visual and quantitative assessment of data augmentation and network optimization can lead to an optimal model with a reasonable performance in segmenting the LGG tumors.
Collapse
Affiliation(s)
| | - Azimeh N V Dehkordi
- Department of Physics, Faculty of Computer Engineering, Najafabad Branch, Islamic Azad University, Najafabad, Iran.
- Najafabad Branch, Islamic Azad University, Najafabad, 8514143131, Iran.
| | - Sedigheh Sina
- Nuclear Engineering Department, Shiraz University, Shiraz, Iran
- Radiation Research Center, Shiraz University, Shiraz, Iran
| |
Collapse
|
50
|
Turk O, Ozhan D, Acar E, Akinci TC, Yilmaz M. Automatic detection of brain tumors with the aid of ensemble deep learning architectures and class activation map indicators by employing magnetic resonance images. Z Med Phys 2024; 34:278-290. [PMID: 36593139 PMCID: PMC11156777 DOI: 10.1016/j.zemedi.2022.11.010] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/15/2022] [Accepted: 11/25/2022] [Indexed: 01/01/2023]
Abstract
Today, as in every life-threatening disease, early diagnosis of brain tumors plays a life-saving role. The brain tumor is formed by the transformation of brain cells from their normal structures into abnormal cell structures. These formed abnormal cells begin to form in masses in the brain regions. Nowadays, many different techniques are employed to detect these tumor masses, and the most common of these techniques is Magnetic Resonance Imaging (MRI). In this study, it is aimed to automatically detect brain tumors with the help of ensemble deep learning architectures (ResNet50, VGG19, InceptionV3 and MobileNet) and Class Activation Maps (CAMs) indicators by employing MRI images. The proposed system was implemented in three stages. In the first stage, it was determined whether there was a tumor in the MR images (Binary Approach). In the second stage, different tumor types (Normal, Glioma Tumor, Meningioma Tumor, Pituitary Tumor) were detected from MR images (Multi-class Approach). In the last stage, CAMs of each tumor group were created as an alternative tool to facilitate the work of specialists in tumor detection. The results showed that the overall accuracy of the binary approach was calculated as 100% on the ResNet50, InceptionV3 and MobileNet architectures, and 99.71% on the VGG19 architecture. Moreover, the accuracy values of 96.45% with ResNet50, 93.40% with VGG19, 85.03% with InceptionV3 and 89.34% with MobileNet architectures were obtained in the multi-class approach.
Collapse
Affiliation(s)
- Omer Turk
- Department of Computer Programming, Vocational School, Mardin Artuklu University, 47500 Mardin, Turkey.
| | - Davut Ozhan
- Department of Electronics, Vocational School, Mardin Artuklu University, 47500 Mardin, Turkey.
| | - Emrullah Acar
- Department of Electrical-Electronics Engineering, Architecture and Engineering Faculty, Batman University, Batman, Turkey.
| | - Tahir Cetin Akinci
- WCGEC, University of California Riverside, Riverside, CA, USA; Department of Electrical Engineering, Istanbul Technical University, Istanbul, Turkey.
| | - Musa Yilmaz
- Department of Electrical-Electronics Engineering, Architecture and Engineering Faculty, Batman University, Batman, Turkey.
| |
Collapse
|