1
|
Qiu J, Karageorgos GM, Peng X, Ghose S, Yang Z, Dentinger A, Xu Z, Jo J, Ragupathi S, Xu G, Abdulaziz N, Gandikota G, Wang X, Mills D. SwinDAF3D: Pyramid Swin Transformers with Deep Attentive Features for Automated Finger Joint Segmentation in 3D Ultrasound Images for Rheumatoid Arthritis Assessment. Bioengineering (Basel) 2025; 12:390. [PMID: 40281750 PMCID: PMC12025309 DOI: 10.3390/bioengineering12040390] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/07/2025] [Revised: 04/01/2025] [Accepted: 04/03/2025] [Indexed: 04/29/2025] Open
Abstract
Rheumatoid arthritis (RA) is a chronic autoimmune disease that can cause severe joint damage and functional impairment. Ultrasound imaging has shown promise in providing real-time assessment of synovium inflammation associated with the early stages of RA. Accurate segmentation of the synovium region and quantification of inflammation-specific imaging biomarkers are crucial for assessing and grading RA. However, automatic segmentation of the synovium in 3D ultrasound is challenging due to ambiguous boundaries, variability in synovium shape, and inhomogeneous intensity distribution. In this work, we introduce a novel network architecture, Swin Transformers with Deep Attentive Features for 3D segmentation (SwinDAF3D), which integrates Swin Transformers into a Deep Attentive Features framework. The developed architecture leverages the hierarchical structure and shifted windows of Swin Transformers to capture rich, multi-scale and attentive contextual information, improving the modeling of long-range dependencies and spatial hierarchies in 3D ultrasound images. In a six-fold cross-validation study with 3D ultrasound images of RA patients' finger joints (n = 72), our SwinDAF3D model achieved the highest performance with a Dice Score (DSC) of 0.838 ± 0.013, an Intersection over Union (IoU) of 0.719 ± 0.019, and Surface Dice Score (SDSC) of 0.852 ± 0.020, compared to 3D UNet (DSC: 0.742 ± 0.025; IoU: 0.589 ± 0.031; SDSC: 0.661 ± 0.029), DAF3D (DSC: 0.813 ± 0.017; IoU: 0.689 ± 0.022; SDSC: 0.817 ± 0.013), Swin UNETR (DSC: 0.808 ± 0.025; IoU: 0.678 ± 0.032; SDSC: 0.822 ± 0.039), UNETR++ (DSC: 0.810 ± 0.014; IoU: 0.684 ± 0.018; SDSC: 0.829 ± 0.027) and TransUNet (DSC: 0.818 ± 0.013; IoU: 0.692 ± 0.017; SDSC: 0.815 ± 0.016) models. This ablation study demonstrates the effectiveness of combining a Swin Transformers feature pyramid with a deep attention mechanism, improving the segmentation accuracy of the synovium in 3D ultrasound. This advancement shows great promise in enabling more efficient and standardized RA screening using ultrasound imaging.
Collapse
Affiliation(s)
- Jianwei Qiu
- GE HealthCare Technology & Innovation Center, Niskayuna, NY 12309, USA; (G.M.K.); (S.G.); (A.D.); (D.M.)
| | - Grigorios M. Karageorgos
- GE HealthCare Technology & Innovation Center, Niskayuna, NY 12309, USA; (G.M.K.); (S.G.); (A.D.); (D.M.)
| | - Xiaorui Peng
- Department of Biomedical Engineering, University of Michigan, Ann Arbor, MI 48109, USA; (X.P.); (Z.X.); (J.J.); (S.R.); (G.X.); (X.W.)
| | - Soumya Ghose
- GE HealthCare Technology & Innovation Center, Niskayuna, NY 12309, USA; (G.M.K.); (S.G.); (A.D.); (D.M.)
| | | | - Aaron Dentinger
- GE HealthCare Technology & Innovation Center, Niskayuna, NY 12309, USA; (G.M.K.); (S.G.); (A.D.); (D.M.)
| | - Zhanpeng Xu
- Department of Biomedical Engineering, University of Michigan, Ann Arbor, MI 48109, USA; (X.P.); (Z.X.); (J.J.); (S.R.); (G.X.); (X.W.)
| | - Janggun Jo
- Department of Biomedical Engineering, University of Michigan, Ann Arbor, MI 48109, USA; (X.P.); (Z.X.); (J.J.); (S.R.); (G.X.); (X.W.)
| | - Siddarth Ragupathi
- Department of Biomedical Engineering, University of Michigan, Ann Arbor, MI 48109, USA; (X.P.); (Z.X.); (J.J.); (S.R.); (G.X.); (X.W.)
| | - Guan Xu
- Department of Biomedical Engineering, University of Michigan, Ann Arbor, MI 48109, USA; (X.P.); (Z.X.); (J.J.); (S.R.); (G.X.); (X.W.)
| | - Nada Abdulaziz
- Division of Rheumatology, Department of Internal Medicine, University of Michigan, Ann Arbor, MI 48109, USA;
| | - Girish Gandikota
- Department of Radiology, University of North Carolina at Chapel Hill, Chapel Hill, NC 27599, USA;
| | - Xueding Wang
- Department of Biomedical Engineering, University of Michigan, Ann Arbor, MI 48109, USA; (X.P.); (Z.X.); (J.J.); (S.R.); (G.X.); (X.W.)
| | - David Mills
- GE HealthCare Technology & Innovation Center, Niskayuna, NY 12309, USA; (G.M.K.); (S.G.); (A.D.); (D.M.)
| |
Collapse
|
2
|
Elhadidy MS, Elgohr AT, El-Geneedy M, Akram S, Kasem HM. Comparative analysis for accurate multi-classification of brain tumor based on significant deep learning models. Comput Biol Med 2025; 188:109872. [PMID: 39970824 DOI: 10.1016/j.compbiomed.2025.109872] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2024] [Revised: 02/11/2025] [Accepted: 02/12/2025] [Indexed: 02/21/2025]
Abstract
Brain tumours are a significant health concern, often resulting in severe cognitive and physiological impairments. Accurate detection and classification of brain tumours, including glioma, meningioma, and pituitary tumours, are crucial for effective treatment. In this study, we present a comprehensive approach for brain tumor classification using MRI scans and deep learning models, specifically focusing on the use of Convolutional Neural Networks (CNN), Swin Transformer, and EfficientNet. MRI scans from four categories, including healthy brains, underwent pre-processing using normalisation, resizing, and data augmentation to mitigate problems associated with variability in image quality and tumor manifestation. Every deep learning model was trained on the pre-processed dataset, and their performance was assessed using accuracy, sensitivity, and specificity measures. The findings demonstrate that the Swin Transformer and EfficientNet models achieved superior classification testing accuracy, which are 98.08 % and 98.72 % respectively, surpassing conventional CNNs, which achieve 95.16 % testing accuracy. EfficientNet exhibited an optimal combination between computational economy and classification performance, making it an exemplary choice for resource-limited settings. Our results underscore the capability of sophisticated deep learning architectures to enhance diagnostic precision in brain tumor classification tasks.
Collapse
Affiliation(s)
- Mohamed S Elhadidy
- Department of Mechatronics Engineering, Faculty of Engineering, Horus University, New Damietta, 34517, Egypt.
| | - Abdelrahman T Elgohr
- Department of Mechatronics Engineering, Faculty of Engineering, Horus University, New Damietta, 34517, Egypt.
| | - Marwa El-Geneedy
- Department of Mechatronics Engineering, Faculty of Engineering, Horus University, New Damietta, 34517, Egypt.
| | - Shimaa Akram
- Communications and Electronics Engineering Dept., Faculty of Engineering, Horus University Egypt, New Damietta, Egypt.
| | - Hossam M Kasem
- Communications and Electronics Engineering Dept., Faculty of Engineering, Horus University Egypt, New Damietta, Egypt; Department of Electronics and Communications, Faculty of Engineering, Tanta University, Egypt; Department of Computer Science Engineering, Egypt - Japan University of Science and Technology (E-JUST), Borg Elarab, Alexandria, Egypt.
| |
Collapse
|
3
|
Wen K, Chen Y, Zhu Z, Yang J, Bao J, Fu D, Hu Z, Peng X, Jiao M. A novel real-time crayfish weight grading method based on improved Swin Transformer. J Food Sci 2025; 90:e70008. [PMID: 39902912 DOI: 10.1111/1750-3841.70008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2024] [Revised: 11/21/2024] [Accepted: 12/22/2024] [Indexed: 02/06/2025]
Abstract
This study proposed a novel detection method for crayfish weight classification based on an improved Swin-Transformer model. The model demonstrated a Mean Intersection over Union (MIOU) of 90.36% on the crayfish dataset, outperforming the IC-Net, DeepLabV3, and U-Net models by 17.44%, 5.55%, and 1.01%, respectively. Furthermore, the segmentation accuracy of the Swin-Transformer model reached 99.0%, surpassing the aforementioned models by 1.25%, 1.73%, and 0.46%, respectively. To facilitate weight prediction of crayfish from segmented images, this study also investigated the correlation between the projected area and the weight of each crayfish part, and developed a multiple regression model with a correlation coefficient of 0.983 by comparing the total projected area and the relationship between the projected area and the actual weight of each crayfish part. To validate this model, a test set of 40 samples was employed, with the average prediction accuracy reaching 98.34% based on 10 representative data points. Finally, grading experiments were carried out on the crayfish weight grading system, and the experimental results showed that the grading accuracy could reach more than 86.5%, confirming the system's feasibility. The detection method not only predicts the weight based on the area but also incorporates the proportional relationship of the area of each part to improve the accuracy of the prediction further. This innovation makes up for the limitations of traditional inspection methods and shows higher potential for application. This study has important applications in industrial automation, especially for real-time high-precision weight grading in the aquatic processing industry, which can improve production efficiency and optimize quality control.
Collapse
Affiliation(s)
- Ke Wen
- School of Mechanical Engineering, Wuhan Polytechnic University, Wuhan, China
| | - Yan Chen
- School of Mechanical Engineering, Wuhan Polytechnic University, Wuhan, China
| | - Zhengwei Zhu
- School of Mechanical Engineering, Wuhan Polytechnic University, Wuhan, China
| | - Jinzhou Yang
- School of Mechanical Engineering, Wuhan Polytechnic University, Wuhan, China
| | - Jinjin Bao
- School of Mechanical Engineering, Wuhan Polytechnic University, Wuhan, China
| | - Dandan Fu
- School of Mechanical Engineering, Wuhan Polytechnic University, Wuhan, China
| | - Zhigang Hu
- School of Mechanical Engineering, Wuhan Polytechnic University, Wuhan, China
| | - Xianhui Peng
- School of Mechanical Engineering, Wuhan Polytechnic University, Wuhan, China
| | - Ming Jiao
- School of Mechanical Engineering, Wuhan Polytechnic University, Wuhan, China
| |
Collapse
|
4
|
Zheng J, Wang L, Gui J, Yussuf AH. Study on lung CT image segmentation algorithm based on threshold-gradient combination and improved convex hull method. Sci Rep 2024; 14:17731. [PMID: 39085327 PMCID: PMC11291637 DOI: 10.1038/s41598-024-68409-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/20/2023] [Accepted: 07/23/2024] [Indexed: 08/02/2024] Open
Abstract
Lung images often have the characteristics of strong noise, uneven grayscale distribution, and complex pathological structures, which makes lung image segmentation a challenging task. To solve this problems, this paper proposes an initial lung mask extraction algorithm that combines threshold and gradient. The gradient used in the algorithm is obtained by the time series feature extraction method based on differential memory (TFDM), which is obtained by the grayscale threshold and image grayscale features. At the same time, we also proposed a lung contour repair algorithm based on the improved convex hull method to solve the contour loss caused by solid nodules and other lesions. Experimental results show that on the COVID-19 CT segmentation dataset, the advanced lung segmentation algorithm proposed in this article achieves better segmentation results and greatly improves the consistency and accuracy of lung segmentation. Our method can obtain more lung information, resulting in ideal segmentation effects with improved accuracy and robustness.
Collapse
Affiliation(s)
- Junbao Zheng
- School of Computer Science and Technology (School of Artificial Intelligence), Zhejiang Sci-tech University, Hangzhou, 310018, Zhejiang, People's Republic of China
| | - Lixian Wang
- School of Information Science and Engineering, Zhejiang Sci-tech University, Hangzhou, 310018, Zhejiang, People's Republic of China
| | - Jiangsheng Gui
- School of Computer Science and Technology (School of Artificial Intelligence), Zhejiang Sci-tech University, Hangzhou, 310018, Zhejiang, People's Republic of China.
| | - Abdulla Hamad Yussuf
- School of Computer Science and Technology (School of Artificial Intelligence), Zhejiang Sci-tech University, Hangzhou, 310018, Zhejiang, People's Republic of China
| |
Collapse
|
5
|
Ma J, Yuan G, Guo C, Gang X, Zheng M. SW-UNet: a U-Net fusing sliding window transformer block with CNN for segmentation of lung nodules. Front Med (Lausanne) 2023; 10:1273441. [PMID: 37841008 PMCID: PMC10569032 DOI: 10.3389/fmed.2023.1273441] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/06/2023] [Accepted: 09/12/2023] [Indexed: 10/17/2023] Open
Abstract
Medical images are information carriers that visually reflect and record the anatomical structure of the human body, and play an important role in clinical diagnosis, teaching and research, etc. Modern medicine has become increasingly inseparable from the intelligent processing of medical images. In recent years, there have been more and more attempts to apply deep learning theory to medical image segmentation tasks, and it is imperative to explore a simple and efficient deep learning algorithm for medical image segmentation. In this paper, we investigate the segmentation of lung nodule images. We address the above-mentioned problems of medical image segmentation algorithms and conduct research on medical image fusion algorithms based on a hybrid channel-space attention mechanism and medical image segmentation algorithms with a hybrid architecture of Convolutional Neural Networks (CNN) and Visual Transformer. To the problem that medical image segmentation algorithms are difficult to capture long-range feature dependencies, this paper proposes a medical image segmentation model SW-UNet based on a hybrid CNN and Vision Transformer (ViT) framework. Self-attention mechanism and sliding window design of Visual Transformer are used to capture global feature associations and break the perceptual field limitation of convolutional operations due to inductive bias. At the same time, a widened self-attentive vector is used to streamline the number of modules and compress the model size so as to fit the characteristics of a small amount of medical data, which makes the model easy to be overfitted. Experiments on the LUNA16 lung nodule image dataset validate the algorithm and show that the proposed network can achieve efficient medical image segmentation on a lightweight scale. In addition, to validate the migratability of the model, we performed additional validation on other tumor datasets with desirable results. Our research addresses the crucial need for improved medical image segmentation algorithms. By introducing the SW-UNet model, which combines CNN and ViT, we successfully capture long-range feature dependencies and break the perceptual field limitations of traditional convolutional operations. This approach not only enhances the efficiency of medical image segmentation but also maintains model scalability and adaptability to small medical datasets. The positive outcomes on various tumor datasets emphasize the potential migratability and broad applicability of our proposed model in the field of medical image analysis.
Collapse
Affiliation(s)
- Jiajun Ma
- Shenhua Hollysys Information Technology Co., Ltd., Beijing, China
| | - Gang Yuan
- The First Affiliated Hospital of Dalian Medical University, Dalian, China
| | - Chenhua Guo
- School of Software, North University of China, Taiyuan, China
| | | | - Minting Zheng
- The First Affiliated Hospital of Dalian Medical University, Dalian, China
| |
Collapse
|
6
|
Krinski BA, Ruiz DV, Laroca R, Todt E. DACov: a deeper analysis of data augmentation on the computed tomography segmentation problem. COMPUTER METHODS IN BIOMECHANICS AND BIOMEDICAL ENGINEERING: IMAGING & VISUALIZATION 2023. [DOI: 10.1080/21681163.2023.2183807] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/09/2023]
Affiliation(s)
- Bruno A. Krinski
- Department of Informatics, Federal University of Paraná, Curitiba, Brazil
| | - Daniel V. Ruiz
- Department of Informatics, Federal University of Paraná, Curitiba, Brazil
| | - Rayson Laroca
- Department of Informatics, Federal University of Paraná, Curitiba, Brazil
| | - Eduardo Todt
- Department of Informatics, Federal University of Paraná, Curitiba, Brazil
| |
Collapse
|
7
|
Sun W, Pang Y, Zhang G. CCT: Lightweight compact convolutional transformer for lung disease CT image classification. Front Physiol 2022; 13:1066999. [DOI: 10.3389/fphys.2022.1066999] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2022] [Accepted: 10/25/2022] [Indexed: 11/06/2022] Open
Abstract
Computed tomography (CT) imaging results are an important criterion for the diagnosis of lung disease. CT images can clearly show the characteristics of lung lesions. Early and accurate detection of lung diseases helps clinicians to improve patient care effectively. Therefore, in this study, we used a lightweight compact convolutional transformer (CCT) to build a prediction model for lung disease classification using chest CT images. We added a position offset term and changed the attention mechanism of the transformer encoder to an axial attention mechanism module. As a result, the classification performance of the model was improved in terms of height and width. We show that the model effectively classifies COVID-19, community pneumonia, and normal conditions on the CC-CCII dataset. The proposed model outperforms other comparable models in the test set, achieving an accuracy of 98.5% and a sensitivity of 98.6%. The results show that our method achieves a larger field of perception on CT images, which positively affects the classification of CT images. Thus, the method can provide adequate assistance to clinicians.
Collapse
|