1
|
Ding X, Qian K, Zhang Q, Jiang X, Dong L. Dual-channel compression mapping network with fused attention mechanism for medical image segmentation. Sci Rep 2025; 15:8906. [PMID: 40087522 PMCID: PMC11909205 DOI: 10.1038/s41598-025-93494-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/04/2024] [Accepted: 03/07/2025] [Indexed: 03/17/2025] Open
Abstract
Accurate image segmentation is the key to quantitative analysis and recognition of pathological tissues in medical imaging technology, which can provide important technical support for medical diagnosis and treatment. However, the task of lesion segmentation is particularly challenging due to the difficulty in identifying edges, the complexity of different tissues, and the variability in their shapes. To address these challenges, we propose a dual-channel compression mapping network (DCM-Net) with fused attention mechanism for medical image segmentation. Firstly, a dual-channel compression mapping module is added to U-Net's standard convolution blocks to capture inter-channel information. Secondly, we replace the traditional skip path with a fusion attention mechanism that can better present context information in high-level features. Finally, the combination of squeeze-and-excitation module and residual connection in the decoder part can improve the adaptive ability of the network. Through extensive experiments on various medical image datasets, DCM-Net has demonstrated superior performance compared to other models. For instance, on the ISIC database, our network achieved an Accuracy of 91.42%, True Positive Rate (TPR) of 88.93%, Dice of 86.09%, and Jaccard of 76.02%. Additionally, on the pituitary adenoma dataset from Quzhou People's Hospital, DCM-Net reached an Accuracy of 97.07%, TPR of 93.09%, Dice of 92.29%, and Jaccard of 87.73%. These results demonstrate the effectiveness of DCM-Net in providing accurate and reliable segmentation, and it shows valuable potential in the field of medical imaging technology.
Collapse
Affiliation(s)
- Xiaokang Ding
- College of Mechanical Engineering, Quzhou University, Quzhou, 324000, China
| | - Ke'er Qian
- College of Mechanical Engineering, Quzhou University, Quzhou, 324000, China.
| | - Qile Zhang
- Department of Rehabilitation, The Quzhou Affiliated Hospital of Wenzhou Medical University, Quzhou People's Hospital, Quzhou, 324000, China.
| | - Xiaoliang Jiang
- College of Mechanical Engineering, Quzhou University, Quzhou, 324000, China
| | - Ling Dong
- College of Mechanical Engineering, Quzhou University, Quzhou, 324000, China
| |
Collapse
|
2
|
Lee S, Jung JY, Mahatthanatrakul A, Kim JS. Artificial Intelligence in Spinal Imaging and Patient Care: A Review of Recent Advances. Neurospine 2024; 21:474-486. [PMID: 38955525 PMCID: PMC11224760 DOI: 10.14245/ns.2448388.194] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2024] [Revised: 05/14/2024] [Accepted: 05/23/2024] [Indexed: 07/04/2024] Open
Abstract
Artificial intelligence (AI) is transforming spinal imaging and patient care through automated analysis and enhanced decision-making. This review presents a clinical task-based evaluation, highlighting the specific impact of AI techniques on different aspects of spinal imaging and patient care. We first discuss how AI can potentially improve image quality through techniques like denoising or artifact reduction. We then explore how AI enables efficient quantification of anatomical measurements, spinal curvature parameters, vertebral segmentation, and disc grading. This facilitates objective, accurate interpretation and diagnosis. AI models now reliably detect key spinal pathologies, achieving expert-level performance in tasks like identifying fractures, stenosis, infections, and tumors. Beyond diagnosis, AI also assists surgical planning via synthetic computed tomography generation, augmented reality systems, and robotic guidance. Furthermore, AI image analysis combined with clinical data enables personalized predictions to guide treatment decisions, such as forecasting spine surgery outcomes. However, challenges still need to be addressed in implementing AI clinically, including model interpretability, generalizability, and data limitations. Multicenter collaboration using large, diverse datasets is critical to advance the field further. While adoption barriers persist, AI presents a transformative opportunity to revolutionize spinal imaging workflows, empowering clinicians to translate data into actionable insights for improved patient care.
Collapse
Affiliation(s)
- Sungwon Lee
- Department of Radiology, Seoul St. Mary’s Hospital, College of Medicine, The Catholic University of Korea, Seoul, Korea
- Visual Analysis and Learning for Improved Diagnostics (VALID) Lab, Seoul St. Mary’s Hospital, College of Medicine, The Catholic University of Korea, Seoul, Korea
| | - Joon-Yong Jung
- Department of Radiology, Seoul St. Mary’s Hospital, College of Medicine, The Catholic University of Korea, Seoul, Korea
- Visual Analysis and Learning for Improved Diagnostics (VALID) Lab, Seoul St. Mary’s Hospital, College of Medicine, The Catholic University of Korea, Seoul, Korea
| | - Akaworn Mahatthanatrakul
- Department of Orthopaedics, Faculty of Medicine, Naresuan University Hospital, Phitsanulok, Thailand
| | - Jin-Sung Kim
- Spine Center, Department of Neurosurgery, Seoul St. Mary’s Hospital, College of Medicine, The Catholic University of Korea, Seoul, Korea
| |
Collapse
|
3
|
Saeed MU, Dikaios N, Dastgir A, Ali G, Hamid M, Hajjej F. An Automated Deep Learning Approach for Spine Segmentation and Vertebrae Recognition Using Computed Tomography Images. Diagnostics (Basel) 2023; 13:2658. [PMID: 37627917 PMCID: PMC10453471 DOI: 10.3390/diagnostics13162658] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2023] [Revised: 04/09/2023] [Accepted: 04/20/2023] [Indexed: 08/27/2023] Open
Abstract
Spine image analysis is based on the accurate segmentation and vertebrae recognition of the spine. Several deep learning models have been proposed for spine segmentation and vertebrae recognition, but they are very computationally demanding. In this research, a novel deep learning model is introduced for spine segmentation and vertebrae recognition using CT images. The proposed model works in two steps: (1) A cascaded hierarchical atrous spatial pyramid pooling residual attention U-Net (CHASPPRAU-Net), which is a modified version of U-Net, is used for the segmentation of the spine. Cascaded spatial pyramid pooling layers, along with residual blocks, are used for feature extraction, while the attention module is used for focusing on regions of interest. (2) A 3D mobile residual U-Net (MRU-Net) is used for vertebrae recognition. MobileNetv2 includes residual and attention modules to accurately extract features from the axial, sagittal, and coronal views of 3D spine images. The features from these three views are concatenated to form a 3D feature map. After that, a 3D deep learning model is used for vertebrae recognition. The VerSe 20 and VerSe 19 datasets were used to validate the proposed model. The model achieved more accurate results in spine segmentation and vertebrae recognition than the state-of-the-art methods.
Collapse
Affiliation(s)
- Muhammad Usman Saeed
- Department of Computer Science, University of Okara, Okara 56310, Pakistan; (A.D.); (G.A.)
| | - Nikolaos Dikaios
- Mathematics Research Centre, Academy of Athens, 10679 Athens, Greece
| | - Aqsa Dastgir
- Department of Computer Science, University of Okara, Okara 56310, Pakistan; (A.D.); (G.A.)
| | - Ghulam Ali
- Department of Computer Science, University of Okara, Okara 56310, Pakistan; (A.D.); (G.A.)
| | - Muhammad Hamid
- Department of Computer Science, Government College Women University, Sialkot 51310, Pakistan;
| | - Fahima Hajjej
- Department of Information Systems, College of Computer and Information Science, Princess Nourah bint Abdulrahman University, Riyadh 11671, Saudi Arabia;
| |
Collapse
|
4
|
Mohanty R, Allabun S, Solanki SS, Pani SK, Alqahtani MS, Abbas M, Soufiene BO. NAMSTCD: A Novel Augmented Model for Spinal Cord Segmentation and Tumor Classification Using Deep Nets. Diagnostics (Basel) 2023; 13:diagnostics13081417. [PMID: 37189520 DOI: 10.3390/diagnostics13081417] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/05/2023] [Revised: 04/06/2023] [Accepted: 04/08/2023] [Indexed: 05/17/2023] Open
Abstract
Spinal cord segmentation is the process of identifying and delineating the boundaries of the spinal cord in medical images such as magnetic resonance imaging (MRI) or computed tomography (CT) scans. This process is important for many medical applications, including the diagnosis, treatment planning, and monitoring of spinal cord injuries and diseases. The segmentation process involves using image processing techniques to identify the spinal cord in the medical image and differentiate it from other structures, such as the vertebrae, cerebrospinal fluid, and tumors. There are several approaches to spinal cord segmentation, including manual segmentation by a trained expert, semi-automated segmentation using software tools that require some user input, and fully automated segmentation using deep learning algorithms. Researchers have proposed a wide range of system models for segmentation and tumor classification in spinal cord scans, but the majority of these models are designed for a specific segment of the spine. As a result, their performance is limited when applied to the entire lead, limiting their deployment scalability. This paper proposes a novel augmented model for spinal cord segmentation and tumor classification using deep nets to overcome this limitation. The model initially segments all five spinal cord regions and stores them as separate datasets. These datasets are manually tagged with cancer status and stage based on observations from multiple radiologist experts. Multiple Mask Regional Convolutional Neural Networks (MRCNNs) were trained on various datasets for region segmentation. The results of these segmentations were combined using a combination of VGGNet 19, YoLo V2, ResNet 101, and GoogLeNet models. These models were selected via performance validation on each segment. It was observed that VGGNet-19 was capable of classifying the thoracic and cervical regions, while YoLo V2 was able to efficiently classify the lumbar region, ResNet 101 exhibited better accuracy for sacral-region classification, and GoogLeNet was able to classify the coccygeal region with high performance accuracy. Due to use of specialized CNN models for different spinal cord segments, the proposed model was able to achieve a 14.5% better segmentation efficiency, 98.9% tumor classification accuracy, and a 15.6% higher speed performance when averaged over the entire dataset and compared with various state-of-the art models. This performance was observed to be better, due to which it can be used for various clinical deployments. Moreover, this performance was observed to be consistent across multiple tumor types and spinal cord regions, which makes the model highly scalable for a wide variety of spinal cord tumor classification scenarios.
Collapse
Affiliation(s)
- Ricky Mohanty
- School of Information System, ASBM University, Bhubaneswar 754012, Odisha, India
| | - Sarah Allabun
- Department of Medical Education, College of Medicine, Princess Nourah bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
| | - Sandeep Singh Solanki
- Department of Electronics and Communication Engineering, Birla Institute of Technology, Mesra 835215, Jharkhand, India
| | | | - Mohammed S Alqahtani
- Radiological Sciences Department, College of Applied Medical Sciences, King Khalid University, Abha 61421, Saudi Arabia
- BioImaging Unit, Space Research Centre, University of Leicester, Michael Atiyah Building, Leicester LE1 7RH, UK
| | - Mohamed Abbas
- Electrical Engineering Department, College of Engineering, King Khalid University, Abha 61421, Saudi Arabia
| | - Ben Othman Soufiene
- PRINCE Laboratory Research, ISITcom, Hammam Sousse, University of Sousse, Sousse 4000, Tunisia
| |
Collapse
|
5
|
Diniz JOB, Dias Júnior DA, da Cruz LB, da Silva GLF, Ferreira JL, Pontes DBQ, Silva AC, de Paiva AC, Gattas M. Heart segmentation in planning CT using 2.5D U-Net++ with attention gate. COMPUTER METHODS IN BIOMECHANICS AND BIOMEDICAL ENGINEERING: IMAGING & VISUALIZATION 2022. [DOI: 10.1080/21681163.2022.2043779] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/06/2023]
Affiliation(s)
- J. O. B. Diniz
- Laboratory Innovation Factory, Federal Institute of Maranhão (IFMA), Grajaú - MA, Brazil
- Applied Computing Group, Federal University of Maranhão (UFMA), São Luís - MA, Brazil
| | - D. A. Dias Júnior
- Applied Computing Group, Federal University of Maranhão (UFMA), São Luís - MA, Brazil
| | - L. B. da Cruz
- Applied Computing Group, Federal University of Maranhão (UFMA), São Luís - MA, Brazil
| | - G. L. F. da Silva
- Applied Computing Group, Federal University of Maranhão (UFMA), São Luís - MA, Brazil
- Dom Bosco Higher Education Unit (UNDB), São Luís - MA, Brazil
| | - J. L. Ferreira
- Applied Computing Group, Federal University of Maranhão (UFMA), São Luís - MA, Brazil
| | - D. B. Q. Pontes
- Applied Computing Group, Federal University of Maranhão (UFMA), São Luís - MA, Brazil
| | - A. C. Silva
- Applied Computing Group, Federal University of Maranhão (UFMA), São Luís - MA, Brazil
| | - A. C. de Paiva
- Applied Computing Group, Federal University of Maranhão (UFMA), São Luís - MA, Brazil
| | - M. Gattas
- Pontifical Catholic University of Rio de Janeiro (PUC-Rio), Rio de Janeiro, RJ, Brazil
| |
Collapse
|