1
|
Al-Absi AA, Fu R, Ebrahim N, Al-Absi MA, Kang DK. Brain Tumour Segmentation and Grading Using Local and Global Context-Aggregated Attention Network Architecture. Bioengineering (Basel) 2025; 12:552. [PMID: 40428171 PMCID: PMC12108882 DOI: 10.3390/bioengineering12050552] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2025] [Revised: 05/07/2025] [Accepted: 05/12/2025] [Indexed: 05/29/2025] Open
Abstract
Brain tumours (BTs) are among the most dangerous and life-threatening cancers in humans of all ages, and the early detection of BTs can make a huge difference to their treatment. However, grade recognition is a challenging issue for radiologists involved in automated diagnosis and healthcare monitoring. Recent research has been motivated by the search for deep learning-based mechanisms for segmentation and grading to assist radiologists in diagnostic analysis. Segmentation refers to the identification and delineation of tumour regions in medical images, while classification classifies based on tumour characteristics, such as the size, location and enhancement pattern. The main aim of this research is to design and develop an intelligent model that can detect and grade tumours more effectively. This research develops an aggregated architecture called LGCNet, which combines a local context attention network and a global context attention network. LGCNet makes use of information extracted through the task, dimension and scale. Specifically, a global context attention network is developed for capturing multiple-scale features, and a local context attention network is designed for specific tasks. Thereafter, both networks are aggregated, and the learning network is designed to balance all the tasks by combining the loss functions of the classification and segmentation. The main advantage of LGCNet is its dedicated network for a specific task. The proposed model is evaluated by considering the BraTS2019 dataset with different metrics, such as the Dice score, sensitivity, specificity and Hausdorff score. Comparative analysis with the existing model shows marginal improvement and provides scope for further research into BT segmentation and classification. The scope of this study focuses on the BraTS2019 dataset, with future work aiming to extend the applicability of the model to different clinical and imaging environments.
Collapse
Affiliation(s)
- Ahmed Abdulhakim Al-Absi
- Department of Smart Computing, Kyungdong University, 46 4-gil, Bongpo, Gosung 24764, Republic of Korea; (A.A.A.-A.); (M.A.A.-A.)
| | - Rui Fu
- College of Language Intelligence, Language & Brain Research Center, Sichuan International Studies University, Chongqing 400031, China;
| | - Nadhem Ebrahim
- Department of Computer Science, College of Engineering and Polymer Science, University of Akron Ohio, Akron, OH 44325, USA;
| | - Mohammed Abdulhakim Al-Absi
- Department of Smart Computing, Kyungdong University, 46 4-gil, Bongpo, Gosung 24764, Republic of Korea; (A.A.A.-A.); (M.A.A.-A.)
| | - Dae-Ki Kang
- Department of Computer & Information Engineering, Dongseo University, 47 Jurye-ro, Sasang-gu, Busan 47011, Republic of Korea
| |
Collapse
|
2
|
Chang YJ, Cho J, Shon B, Choi KY, Jeong S, Ryu JY. A novel clinical investigation using deep learning and human-in-the-loop approach in orbital volume measurement. J Craniomaxillofac Surg 2025; 53:498-506. [PMID: 39875226 DOI: 10.1016/j.jcms.2025.01.007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/11/2024] [Revised: 10/29/2024] [Accepted: 01/11/2025] [Indexed: 01/30/2025] Open
Abstract
Orbital volume assessment is crucial for surgical planning. Traditional methods lack efficiency and accuracy. Recent studies explore AI-driven techniques, but research on their clinical effectiveness is limited. This study included 349 patients aged 19 years and above, who underwent three-dimensional facial computed tomography (3DCT) without orbital trauma or congenital anomalies. To construct an AI training dataset, manual segmentation was performed on 178 patients' 3DCT using 3D Slicer. The remaining data of 171 patients underwent human-in-the-loop method, resulting in a dataset of 349 annotated samples. Comparative analysis of Dice coefficients and execution speeds was performed between manual and semi-automated segmentations. Comparing AI-assisted semi-automated segmentation with manual segmentation, all six annotators demonstrated lower average inference times without a significant difference in Dice coefficients (90.31% vs. 88.72%). For 178 patients' 3DCT, a high average Dice coefficient of 89.9% was observed, and a 38.42-ms inference time was recorded. For the full dataset, the AI model achieved a high average Dice coefficient of 94.1% and a fast average inference time of 32.55 ms per axial slice. This study demonstrates the potential of AI for maintaining high accuracy and time-efficiency in orbital region segmentation, with wide clinical applications.
Collapse
Affiliation(s)
- Yong June Chang
- Department of Plastic and Reconstructive Surgery, School of Medicine, Kyungpook National University, Daegu, Republic of Korea.
| | - Jungrae Cho
- Research Center for Artificial Intelligence in Medicine, Kyungpook National University Hospital, Daegu, Republic of Korea.
| | - Byungeun Shon
- Research Center for Artificial Intelligence in Medicine, Kyungpook National University Hospital, Daegu, Republic of Korea; Department of Medical Informatics, School of Medicine, Kyungpook National University, Daegu, Republic of Korea
| | - Kang Young Choi
- Department of Plastic and Reconstructive Surgery, School of Medicine, Kyungpook National University, Daegu, Republic of Korea
| | - Sungmoon Jeong
- Research Center for Artificial Intelligence in Medicine, Kyungpook National University Hospital, Daegu, Republic of Korea; Department of Medical Informatics, School of Medicine, Kyungpook National University, Daegu, Republic of Korea.
| | - Jeong Yeop Ryu
- Department of Plastic and Reconstructive Surgery, School of Medicine, Kyungpook National University, Daegu, Republic of Korea.
| |
Collapse
|
3
|
Lou X, Zhu J, Yang J, Zhu Y, Shu H, Li B. Enhanced Cross-stage-attention U-Net for esophageal target volume segmentation. BMC Med Imaging 2024; 24:339. [PMID: 39696039 DOI: 10.1186/s12880-024-01515-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/06/2024] [Accepted: 11/26/2024] [Indexed: 12/20/2024] Open
Abstract
PURPOSE The segmentation of target volume and organs at risk (OAR) was a significant part of radiotherapy. Specifically, determining the location and scale of the esophagus in simulated computed tomography images was difficult and time-consuming primarily due to its complex structure and low contrast with the surrounding tissues. In this study, an Enhanced Cross-stage-attention U-Net was proposed to solve the segmentation problem for the esophageal gross tumor volume (GTV) and clinical tumor volume (CTV) in CT images. METHODS First, a module based on principal component analysis theory was constructed to pre-extract the features of the input image. Then, a cross-stage based feature fusion model was designed to replace the skip concatenation of original UNet, which was composed of Wide Range Attention unit, Small-kernel Local Attention unit, and Inverted Bottleneck unit. WRA was employed to capture global attention, whose large convolution kernel was further decomposed to simplify the calculation. SLA was used to complement the local attention to WRA. IBN was structed to fuse the extracted features, where a global frequency response layer was built to redistribute the frequency response of the fused feature maps. RESULTS The proposed method was compared with relevant published esophageal segmentation methods. The prediction of the proposed network was MSD = 2.83(1.62, 4.76)mm, HD = 11.79 ± 6.02 mm, DC = 72.45 ± 19.18% in GTV; MSD = 5.26(2.18, 8.82)mm, HD = 16.22 ± 10.01 mm, DC = 71.06 ± 17.72% in CTV. CONCLUSION The reconstruction of the skip concatenation in UNet showed an improvement of performance for esophageal segmentation. The results showed the proposed network had better effect on esophageal GTV and CTV segmentation.
Collapse
Affiliation(s)
- Xiao Lou
- Laboratory of Image Science and Technology, Key Laboratory of New Generation Artificial Intelligence Technology and Its Interdisciplinary Applications, Ministry of Education, Southeast University, Sipailou 2, Nanjing, P.R. China
- Department of Radiotherapy, Lishui People's Hospital, No. 1188, Liyang Street, Lishui, P.R. China
| | - Juan Zhu
- Department of Respiratory Medicine, The People's Hospital of Zhangqiuqu Area, No. 1920, Huiquan Street, Jinan, P.R. China
| | - Jian Yang
- Department of Clinical Laboratory, The People's Hospital of Zhangqiuqu Area, No. 1920, Huiquan Street, Jinan, P.R. China
| | - Youzhe Zhu
- Laboratory of Image Science and Technology, Key Laboratory of New Generation Artificial Intelligence Technology and Its Interdisciplinary Applications, Ministry of Education, Southeast University, Sipailou 2, Nanjing, P.R. China.
- Department of Radiotherapy, Lishui People's Hospital, No. 1188, Liyang Street, Lishui, P.R. China.
| | - Huazhong Shu
- Laboratory of Image Science and Technology, Key Laboratory of New Generation Artificial Intelligence Technology and Its Interdisciplinary Applications, Ministry of Education, Southeast University, Sipailou 2, Nanjing, P.R. China.
| | - Baosheng Li
- Laboratory of Image Science and Technology, Key Laboratory of New Generation Artificial Intelligence Technology and Its Interdisciplinary Applications, Ministry of Education, Southeast University, Sipailou 2, Nanjing, P.R. China.
- Shandong First Medical University and Shandong Academy of Medical Sciences, Shandong Cancer Hospital and Institute, No. 440, Jiyan Street, Jinan, P.R. China.
| |
Collapse
|
4
|
Zheng K, Pan J, Jia Z, Xiao S, Tao W, Zhang D, Li Q, Pan L. A method of nucleus image segmentation and counting based on TC-UNet++ and distance watershed. Med Eng Phys 2024; 133:104244. [PMID: 39557502 DOI: 10.1016/j.medengphy.2024.104244] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2024] [Revised: 09/08/2024] [Accepted: 10/14/2024] [Indexed: 11/20/2024]
Abstract
Nucleus segmentation and counting play a crucial role in many cell analysis applications. However, the dense distribution and blurry boundaries of nucleus make nucleus segmentation tasks challenging. This paper proposes a novel segmentation and counting method. Firstly, TC-UNet++ is proposed to achieve a global segmentation. Then, the distance watershed method is used to finish local segmentation, which separate the adhesion and overlap part of the image. Finally, counting method is performed to obtain information on the counting number, area and center of mass of nucleus. TC-UNet++ achieved a Dice coefficient of 89.95% for cell instance segmentation on the Data Science Bowl dataset, surpassing the original U-Net++ by 0.23%. It also showed a 5.09% improvement in counting results compared to other methods. On the ALL-IDB dataset, TC-UNet++ reached a Dice coefficient of 83.97%, a 7.93% increase over the original U-Net++. Additionally, its counting results improved by 16.82% compared to other approaches. These results indicate that our method has a more complete and reasonable nucleus segmentation and counting scheme compared to other methods.
Collapse
Affiliation(s)
- Kaifeng Zheng
- School of Mechanical Engineering, Jiangsu University of Technology, Changzhou, People's Republic of China.
| | - Jie Pan
- Department of Laboratory, Hua Dong Sanatorium, Wuxi, People's Republic of China.
| | - Ziyan Jia
- School of Electrical and Information Engineering, Jiangsu University of Technology, Changzhou, People's Republic of China.
| | - Shuyan Xiao
- School of Electrical and Information Engineering, Jiangsu University of Technology, Changzhou, People's Republic of China.
| | - Weige Tao
- School of Electrical and Information Engineering, Jiangsu University of Technology, Changzhou, People's Republic of China.
| | - Dachuan Zhang
- The First People's Hospital of Changzhou, Changzhou, People's Republic of China.
| | - Qing Li
- The First People's Hospital of Changzhou, Changzhou, People's Republic of China.
| | - Lingjiao Pan
- School of Electrical and Information Engineering, Jiangsu University of Technology, Changzhou, People's Republic of China.
| |
Collapse
|
5
|
Rana K, Garg D, Yong LSS, Macri C, Tong JY, Patel S, Slattery J, Chan WO, Davis G, Selva D. Extraocular muscle enlargement in dysthyroid optic neuropathy. CANADIAN JOURNAL OF OPHTHALMOLOGY 2024; 59:e542-e546. [PMID: 38114063 DOI: 10.1016/j.jcjo.2023.11.015] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/11/2023] [Revised: 10/27/2023] [Accepted: 11/22/2023] [Indexed: 12/21/2023]
Abstract
OBJECTIVE To investigate extraocular muscle volumes in thyroid eye disease (TED) patients with and without dysthyroid optic neuropathy (DON). DESIGN Retrospective cohort study. PARTICIPANTS TED patients who had computed tomography of the orbits. METHODS The extraocular muscles were manually segmented in consecutive axial and coronal slices, and the volume was calculated by summing the areas in each slice and multiplying by the slice thickness. Data were collected on patient demographics, disease presentation, thyroid function tests, and antibody levels. RESULTS Imaging from 200 orbits was evaluated. The medial rectus, lateral rectus, superior muscle group, inferior rectus, and superior oblique volumes were significantly greater in orbits with DON compared with TED orbits without DON (p < 0.01 for all). There was no significant difference in the inferior oblique muscle volume (p = 0.19). Increase in volume of the superior oblique muscle showed the highest odds for DON. Each 100 mm3 increase in superior oblique, lateral rectus, inferior rectus, medial rectus, and superior muscle group volume was associated with 1.58, 1.25, 1.20, 1.16, and 1.14 times increased odds of DON. CONCLUSION All extraocular muscle volumes except for the inferior oblique were significantly greater in DON patients. Superior oblique enlargement was associated with the highest odds of DON, suggesting superior oblique enlargement to be a novel marker of DON.
Collapse
Affiliation(s)
- Khizar Rana
- Discipline of Ophthalmology and Visual Sciences, University of Adelaide, Adelaide, South Australia; Department of Ophthalmology, Royal Adelaide Hospital, Adelaide, South Australia.
| | - Devanshu Garg
- Discipline of Ophthalmology and Visual Sciences, University of Adelaide, Adelaide, South Australia; Department of Ophthalmology, Royal Adelaide Hospital, Adelaide, South Australia
| | - Lee Shien S Yong
- Discipline of Ophthalmology and Visual Sciences, University of Adelaide, Adelaide, South Australia; Department of Ophthalmology, Royal Adelaide Hospital, Adelaide, South Australia
| | - Carmelo Macri
- Discipline of Ophthalmology and Visual Sciences, University of Adelaide, Adelaide, South Australia; Department of Ophthalmology, Royal Adelaide Hospital, Adelaide, South Australia
| | - Jessica Y Tong
- Discipline of Ophthalmology and Visual Sciences, University of Adelaide, Adelaide, South Australia; Department of Ophthalmology, Royal Adelaide Hospital, Adelaide, South Australia
| | - Sandy Patel
- Department of Medical Imaging, Royal Adelaide Hospital, Adelaide, South Australia
| | - James Slattery
- Discipline of Ophthalmology and Visual Sciences, University of Adelaide, Adelaide, South Australia; Department of Ophthalmology, Royal Adelaide Hospital, Adelaide, South Australia
| | - Weng Onn Chan
- Discipline of Ophthalmology and Visual Sciences, University of Adelaide, Adelaide, South Australia; Department of Ophthalmology, Royal Adelaide Hospital, Adelaide, South Australia
| | - Garry Davis
- Discipline of Ophthalmology and Visual Sciences, University of Adelaide, Adelaide, South Australia; Department of Ophthalmology, Royal Adelaide Hospital, Adelaide, South Australia
| | - Dinesh Selva
- Discipline of Ophthalmology and Visual Sciences, University of Adelaide, Adelaide, South Australia; Department of Ophthalmology, Royal Adelaide Hospital, Adelaide, South Australia
| |
Collapse
|
6
|
Chen Y, Hu X, Zhu Y, Liu X, Yi B. Real-time non-invasive hemoglobin prediction using deep learning-enabled smartphone imaging. BMC Med Inform Decis Mak 2024; 24:187. [PMID: 38951831 PMCID: PMC11218390 DOI: 10.1186/s12911-024-02585-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/26/2024] [Accepted: 06/24/2024] [Indexed: 07/03/2024] Open
Abstract
BACKGROUND Accurate measurement of hemoglobin concentration is essential for various medical scenarios, including preoperative evaluations and determining blood loss. Traditional invasive methods are inconvenient and not suitable for rapid, point-of-care testing. Moreover, current models, due to their complex parameters, are not well-suited for mobile medical settings, which limits the ability to conduct frequent and rapid testing. This study aims to introduce a novel, compact, and efficient system that leverages deep learning and smartphone technology to accurately estimate hemoglobin levels, thereby facilitating rapid and accessible medical assessments. METHODS The study employed a smartphone application to capture images of the eye, which were subsequently analyzed by a deep neural network trained on data from invasive blood test data. Specifically, the EGE-Unet model was utilized for eyelid segmentation, while the DHA(C3AE) model was employed for hemoglobin level prediction. The performance of the EGE-Unet was evaluated using statistical metrics including mean intersection over union (MIOU), F1 Score, accuracy, specificity, and sensitivity. The DHA(C3AE) model's performance was assessed using mean absolute error (MAE), mean-square error (MSE), root mean square error (RMSE), and R^2. RESULTS The EGE-Unet model demonstrated robust performance in eyelid segmentation, achieving an MIOU of 0.78, an F1 Score of 0.87, an accuracy of 0.97, a specificity of 0.98, and a sensitivity of 0.86. The DHA(C3AE) model for hemoglobin level prediction yielded promising outcomes with an MAE of 1.34, an MSE of 2.85, an RMSE of 1.69, and an R^2 of 0.34. The overall size of the model is modest at 1.08 M, with a computational complexity of 0.12 FLOPs (G). CONCLUSIONS This system presents a groundbreaking approach that eliminates the need for supplementary devices, providing a cost-effective, swift, and accurate method for healthcare professionals to enhance treatment planning and improve patient care in perioperative environments. The proposed system has the potential to enable frequent and rapid testing of hemoglobin levels, which can be particularly beneficial in mobile medical settings. TRIAL REGISTRATION The clinical trial was registered on the Chinese Clinical Trial Registry (No. ChiCTR2100044138) on 20/02/2021.
Collapse
Affiliation(s)
- Yuwen Chen
- Chongqing Institute of Green and Intelligent Technology, Chinese Academy of Sciences, Chongqing, 400714, China
| | - Xiaoyan Hu
- Department of Anaesthesiology, Southwest Hospital, Third Military Medical University, Chongqing, 400038, China
| | - Yiziting Zhu
- Department of Anaesthesiology, Southwest Hospital, Third Military Medical University, Chongqing, 400038, China
| | - Xiang Liu
- Department of Anaesthesiology, Southwest Hospital, Third Military Medical University, Chongqing, 400038, China
| | - Bin Yi
- Department of Anaesthesiology, Southwest Hospital, Third Military Medical University, Chongqing, 400038, China.
| |
Collapse
|
7
|
Huang Z, Wang L, Xu L. DRA-Net: Medical image segmentation based on adaptive feature extraction and region-level information fusion. Sci Rep 2024; 14:9714. [PMID: 38678063 PMCID: PMC11584768 DOI: 10.1038/s41598-024-60475-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/20/2024] [Accepted: 04/23/2024] [Indexed: 04/29/2024] Open
Abstract
Medical image segmentation is a key task in computer aided diagnosis. In recent years, convolutional neural network (CNN) has made some achievements in medical image segmentation. However, the convolution operation can only extract features in a fixed size region at a time, which leads to the loss of some key features. The recently popular Transformer has global modeling capabilities, but it does not pay enough attention to local information and cannot accurately segment the edge details of the target area. Given these issues, we proposed dynamic regional attention network (DRA-Net). Different from the above methods, it first measures the similarity of features and concentrates attention on different dynamic regions. In this way, the network can adaptively select different modeling scopes for feature extraction, reducing information loss. Then, regional feature interaction is carried out to better learn local edge details. At the same time, we also design ordered shift multilayer perceptron (MLP) blocks to enhance communication within different regions, further enhancing the network's ability to learn local edge details. After several experiments, the results indicate that our network produces more accurate segmentation performance compared to other CNN and Transformer based networks.
Collapse
Affiliation(s)
- Zhongmiao Huang
- School of Computer Science and Technology, Xinjiang University, Urumqi, 830046, China
| | - Liejun Wang
- School of Computer Science and Technology, Xinjiang University, Urumqi, 830046, China.
| | - Lianghui Xu
- School of Computer Science and Technology, Xinjiang University, Urumqi, 830046, China
| |
Collapse
|
8
|
Zhou Z, Zheng Y, Zhou X, Yu J, Rong S. Self-supervised pre-training for joint optic disc and cup segmentation via attention-aware network. BMC Ophthalmol 2024; 24:98. [PMID: 38438876 PMCID: PMC10910696 DOI: 10.1186/s12886-024-03376-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/18/2023] [Accepted: 02/28/2024] [Indexed: 03/06/2024] Open
Abstract
Image segmentation is a fundamental task in deep learning, which is able to analyse the essence of the images for further development. However, for the supervised learning segmentation method, collecting pixel-level labels is very time-consuming and labour-intensive. In the medical image processing area for optic disc and cup segmentation, we consider there are two challenging problems that remain unsolved. One is how to design an efficient network to capture the global field of the medical image and execute fast in real applications. The other is how to train the deep segmentation network using a few training data due to some medical privacy issues. In this paper, to conquer such issues, we first design a novel attention-aware segmentation model equipped with the multi-scale attention module in the pyramid structure-like encoder-decoder network, which can efficiently learn the global semantics and the long-range dependencies of the input images. Furthermore, we also inject the prior knowledge that the optic cup lies inside the optic disc by a novel loss function. Then, we propose a self-supervised contrastive learning method for optic disc and cup segmentation. The unsupervised feature representation is learned by matching an encoded query to a dictionary of encoded keys using a contrastive technique. Finetuning the pre-trained model using the proposed loss function can help achieve good performance for the task. To validate the effectiveness of the proposed method, extensive systemic evaluations on different public challenging optic disc and cup benchmarks, including DRISHTI-GS and REFUGE datasets demonstrate the superiority of the proposed method, which can achieve new state-of-the-art performance approaching 0.9801 and 0.9087 F1 score respectively while gaining 0.9657 D C disc and 0.8976 D C cup . The code will be made publicly available.
Collapse
Affiliation(s)
- Zhiwang Zhou
- Institute for Advanced Study, Nanchang University, Nanchang, 330031, China.
| | - Yuanchang Zheng
- Institute for Advanced Study, Nanchang University, Nanchang, 330031, China
- Institute of Science and Technology, Waseda University, Tokyo, 63-8001, Japan
| | - Xiaoyu Zhou
- School of Transportation Engineering, Tongji University, Shanghai, 200000, China
| | - Jie Yu
- School of Electrical Automation and Information Engineering, Tianjin University, Tianjin, 300000, China
| | - Shangjie Rong
- School of Mathematical Sciences, Xiamen University, Xiamen, 361000, China
| |
Collapse
|
9
|
Li H, Liu M, Fan J, Liu Q. Biomedical image segmentation algorithm based on dense atrous convolution. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2024; 21:4351-4369. [PMID: 38549331 DOI: 10.3934/mbe.2024192] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/02/2024]
Abstract
Biomedical images have complex tissue structures, and there are great differences between images of the same part of different individuals. Although deep learning methods have made some progress in automatic segmentation of biomedical images, the segmentation accuracy is relatively low for biomedical images with significant changes in segmentation targets, and there are also problems of missegmentation and missed segmentation. To address these challenges, we proposed a biomedical image segmentation method based on dense atrous convolution. First, we added a dense atrous convolution module (DAC) between the encoding and decoding paths of the U-Net network. This module was based on the inception structure and atrous convolution design, which can effectively capture multi-scale features of images. Second, we introduced a dense residual pooling module to detect multi-scale features in images by connecting residual pooling blocks of different sizes. Finally, in the decoding part of the network, we adopted an attention mechanism to suppress background interference by enhancing the weight of the target area. These modules work together to improve the accuracy and robustness of biomedical image segmentation. The experimental results showed that compared to mainstream segmentation networks, our segmentation model exhibited stronger segmentation ability when processing biomedical images with multiple-shaped targets. At the same time, this model can significantly reduce the phenomenon of missed segmentation and missegmentation, improve segmentation accuracy, and make the segmentation results closer to the real situation.
Collapse
Affiliation(s)
- Hong'an Li
- College of Computer Science and Technology, Xi'an University of Science and Technology, Xi'an, 710054, China
- State Key Laboratory of Virtual Reality Technology and Systems, Beihang University, Beijing, 100191, China
| | - Man Liu
- College of Computer Science and Technology, Xi'an University of Science and Technology, Xi'an, 710054, China
| | - Jiangwen Fan
- College of Computer Science and Technology, Xi'an University of Science and Technology, Xi'an, 710054, China
| | - Qingfang Liu
- Information Center, Shijiazhuang Posts and Telecommunications Technical College, Shijiazhuang 050021, China
| |
Collapse
|
10
|
Li Y, Yan B, Hou J, Bai B, Huang X, Xu C, Fang L. UNet based on dynamic convolution decomposition and triplet attention. Sci Rep 2024; 14:271. [PMID: 38168684 PMCID: PMC10761743 DOI: 10.1038/s41598-023-50989-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/17/2023] [Accepted: 12/28/2023] [Indexed: 01/05/2024] Open
Abstract
The robustness and generalization of medical image segmentation models are being challenged by the differences between different disease types, different image types, and different cases.Deep learning based semantic segmentation methods have been providing state-of-the-art performance in the last few years. One deep learning technique, U-Net, has become the most popular architecture in the medical imaging segmentation. Despite outstanding overall performance in segmenting medical images, it still has the problems of limited feature expression ability and inaccurate segmentation. To this end, we propose a DTA-UNet based on Dynamic Convolution Decomposition (DCD) and Triple Attention (TA). Firstly, the model with Attention U-Net as the baseline network uses DCD to replace all the conventional convolution in the encoding-decoding process to enhance its feature extraction capability. Secondly, we combine TA with Attention Gate (AG) to be used for skip connection in order to highlight lesion regions by removing redundant information in both spatial and channel dimensions. The proposed model are tested on the two public datasets and actual clinical dataset such as the public COVID-SemiSeg dataset, the ISIC 2018 dataset, and the cooperative hospital stroke segmentation dataset. Ablation experiments on the clinical stroke segmentation dataset show the effectiveness of DCD and TA with only a 0.7628 M increase in the number of parameters compared to the baseline model. The proposed DTA-UNet is further evaluated on the three datasets of different types of images to verify its universality. Extensive experimental results show superior performance on different segmentation metrics compared to eight state-of-art methods.The GitHub URL of our code is https://github.com/shuaihou1234/DTA-UNet .
Collapse
Affiliation(s)
- Yang Li
- Academy for Advanced Interdisciplinary Studies, Northeast Normal University, Changchun, 130024, Jilin, China
- Shanghai Zhangjiang Institute of Mathematics, Shanghai, 201203, China
| | - Bobo Yan
- School of Computer Science and Engineering, Changchun University of Technology, Changchun, 130012, Jilin, China
- Pazhou Lab, Guangzhou, China
| | - Jianxin Hou
- School of Computer Science and Engineering, Changchun University of Technology, Changchun, 130012, Jilin, China
| | - Bingyang Bai
- School of Computer Science and Engineering, Changchun University of Technology, Changchun, 130012, Jilin, China
| | - Xiaoyu Huang
- School of Computer Science and Engineering, Changchun University of Technology, Changchun, 130012, Jilin, China
| | - Canfei Xu
- The Third Hospital of Jilin University, Changchun, 130117, Jilin, China
| | - Limei Fang
- Encephalopathy Center, The Third Affiliated Hospital of Changchun University of Chinese Medicine, Changchun, 130117, Jilin, China.
| |
Collapse
|
11
|
Zhu H, He H, Zhou H. IMoVR-Net: A robust interpretable network for multi-ocular lesion recognition from TAO facial images. Comput Biol Med 2024; 168:107771. [PMID: 38070200 DOI: 10.1016/j.compbiomed.2023.107771] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2023] [Revised: 11/06/2023] [Accepted: 11/26/2023] [Indexed: 01/10/2024]
Abstract
Thyroid-associated ophthalmopathy (TAO) is an organ-specific autoimmune disease that seriously affects patient's life and health. However, the early diagnosis of TAO is highly dependent on the physician's subjective experience. Moreover, the currently proposed deep learning networks for eye diseases do not provide robust interpretability concerning feature learning paradigm, model structure, and the number of neurons. But the mentioned components are very important for model interpretability and are key factors that severely affect the transparency of the model. Therefore, a robust interpretable multi-orientation visual recognition network (IMoVR-Net) for TAO multi-ocular lesion recognition is proposed in this paper. Firstly, a multi-orientation visual cascaded encoder composed of the DenseGabor module and the dilated Gabor convolution group is proposed to achieve the fine extraction of multi-directional TAO lesion features by using a novel feature learning paradigm called alternating filtering. Besides, combining information theory and topology tool, an optimization rule based on topological energy entropy is proposed to provide a solid interpretable theory for determining the model structure. Finally, a clustering correlation analysis method is developed to accomplish the determination of the number of convolutional hidden neurons, providing robust interpretability for the selection of the number of neurons. Compared to other advanced models, the IMoVR-Net achieved state-of-the-art performance on different TAO ocular datasets with an average accuracy, sensitivity, precision, and F1 score of 0.878, 87.27 %, 0.875, and 87.39 %, respectively. The IMoVR-Net has good clinical application prospects due to its strong recognition ability and robust interpretability in feature extraction paradigm, model structure, and number of convolutional neurons.
Collapse
Affiliation(s)
- Haipeng Zhu
- School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai, China
| | - Hong He
- School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai, China.
| | - Huifang Zhou
- Ninth People's Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China; Shanghai Key Laboratory of Orbital Diseases and Ocular Oncology, Shanghai, China
| |
Collapse
|
12
|
Park S, No C, Kim S, Han K, Jung JM, Kwon KY, Lee M. A multimodal screening system for elderly neurological diseases based on deep learning. Sci Rep 2023; 13:21013. [PMID: 38030653 PMCID: PMC10687257 DOI: 10.1038/s41598-023-48071-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/01/2023] [Accepted: 11/22/2023] [Indexed: 12/01/2023] Open
Abstract
In this paper, we propose a deep-learning-based algorithm for screening neurological diseases. We proposed various examination protocols for screening neurological diseases and collected data by video-recording persons performing these protocols. We converted video data into human landmarks that capture action information with a much smaller data dimension. We also used voice data which are also effective indicators of neurological disorders. We designed a subnetwork for each protocol to extract features from landmarks or voice and a feature aggregator that combines all the information extracted from the protocols to make a final decision. Multitask learning was applied to screen two neurological diseases. To capture meaningful information about these human landmarks and voices, we applied various pre-trained models to extract preliminary features. The spatiotemporal characteristics of landmarks are extracted using a pre-trained graph neural network, and voice features are extracted using a pre-trained time-delay neural network. These extracted high-level features are then passed onto the subnetworks and an additional feature aggregator that are simultaneously trained. We also used various data augmentation techniques to overcome the shortage of data. Using a frame-length staticizer that considers the characteristics of the data, we can capture momentary tremors without wasting information. Finally, we examine the effectiveness of different protocols and different modalities (different body parts and voice) through extensive experiments. The proposed method achieves AUC scores of 0.802 for stroke and 0.780 for Parkinson's disease, which is effective for a screening system.
Collapse
Affiliation(s)
- Sangyoung Park
- Department of Electrical and Electronic Engineering, Hanyang University ERICA, Ansan, 15588, South Korea
| | - Changho No
- Department of Electrical and Electronic Engineering, Hanyang University ERICA, Ansan, 15588, South Korea
| | - Sora Kim
- Department of Electrical and Electronic Engineering, Hanyang University ERICA, Ansan, 15588, South Korea
| | - Kyoungmin Han
- Department of Electrical and Electronic Engineering, Hanyang University ERICA, Ansan, 15588, South Korea
| | - Jin-Man Jung
- Department of Neurology, Korea University Ansan Hospital, Ansan, 15355, South Korea
| | - Kyum-Yil Kwon
- Department of Neurology, Soonchunhyang University Seoul Hospital, Seoul, 04401, South Korea
| | - Minsik Lee
- Department of Electrical and Electronic Engineering, Hanyang University ERICA, Ansan, 15588, South Korea.
| |
Collapse
|
13
|
Yao N, Li L, Gao Z, Zhao C, Li Y, Han C, Nan J, Zhu Z, Xiao Y, Zhu F, Zhao M, Zhou W. Deep learning-based diagnosis of disease activity in patients with Graves' orbitopathy using orbital SPECT/CT. Eur J Nucl Med Mol Imaging 2023; 50:3666-3674. [PMID: 37395800 PMCID: PMC10547836 DOI: 10.1007/s00259-023-06312-2] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/02/2023] [Accepted: 06/17/2023] [Indexed: 07/04/2023]
Abstract
PURPOSE Orbital [99mTc]TcDTPA orbital single-photon emission computed tomography (SPECT)/CT is an important method for assessing inflammatory activity in patients with Graves' orbitopathy (GO). However, interpreting the results requires substantial physician workload. We aim to propose an automated method called GO-Net to detect inflammatory activity in patients with GO. MATERIALS AND METHODS GO-Net had two stages: (1) a semantic V-Net segmentation network (SV-Net) that extracts extraocular muscles (EOMs) in orbital CT images and (2) a convolutional neural network (CNN) that uses SPECT/CT images and the segmentation results to classify inflammatory activity. A total of 956 eyes from 478 patients with GO (active: 475; inactive: 481) at Xiangya Hospital of Central South University were investigated. For the segmentation task, five-fold cross-validation with 194 eyes was used for training and internal validation. For the classification task, 80% of the eye data were used for training and internal fivefold cross-validation, and the remaining 20% of the eye data were used for testing. The EOM regions of interest (ROIs) were manually drawn by two readers and reviewed by an experienced physician as ground truth for segmentation GO activity was diagnosed according to clinical activity scores (CASs) and the SPECT/CT images. Furthermore, results are interpreted and visualized using gradient-weighted class activation mapping (Grad-CAM). RESULTS The GO-Net model combining CT, SPECT, and EOM masks achieved a sensitivity of 84.63%, a specificity of 83.87%, and an area under the receiver operating curve (AUC) of 0.89 (p < 0.01) on the test set for distinguishing active and inactive GO. Compared with the CT-only model, the GO-Net model showed superior diagnostic performance. Moreover, Grad-CAM demonstrated that the GO-Net model placed focus on the GO-active regions. For EOM segmentation, our segmentation model achieved a mean intersection over union (IOU) of 0.82. CONCLUSION The proposed Go-Net model accurately detected GO activity and has great potential in the diagnosis of GO.
Collapse
Affiliation(s)
- Ni Yao
- School of Computer and Communication Engineering, Zhengzhou University of Light Industry, Zhengzhou, 450002, Henan, China
| | - Longxi Li
- School of Computer and Communication Engineering, Zhengzhou University of Light Industry, Zhengzhou, 450002, Henan, China
| | - Zhengyuan Gao
- School of Biomedical Engineering, Shenzhen University Medical School, Shenzhen University, Shenzhen, 518060, China
| | - Chen Zhao
- Department of Applied Computing, Michigan Technological University, Houghton, MI, USA
| | - Yanting Li
- School of Computer and Communication Engineering, Zhengzhou University of Light Industry, Zhengzhou, 450002, Henan, China
| | - Chuang Han
- School of Computer and Communication Engineering, Zhengzhou University of Light Industry, Zhengzhou, 450002, Henan, China
| | - Jiaofen Nan
- School of Computer and Communication Engineering, Zhengzhou University of Light Industry, Zhengzhou, 450002, Henan, China
| | - Zelin Zhu
- School of Computer and Communication Engineering, Zhengzhou University of Light Industry, Zhengzhou, 450002, Henan, China
| | - Yi Xiao
- Department of Nuclear Medicine, Xiangya Hospital, Central South University, Changsha, China
| | - Fubao Zhu
- School of Computer and Communication Engineering, Zhengzhou University of Light Industry, Zhengzhou, 450002, Henan, China
| | - Min Zhao
- Department of Nuclear Medicine, The Third Xiangya Hospital, Central South University, No. 138, Tongzipo Road, Changsha, 410013, Hunan Province, China.
| | - Weihua Zhou
- Department of Applied Computing, Michigan Technological University, Houghton, MI, USA
- Center for Biocomputing and Digital Health, Institute of Computing and Cybersystems, and Health Research Institute, Michigan Technological University, Houghton, MI, USA
| |
Collapse
|
14
|
Krishnan SD, Pelusi D, Daniel A, Suresh V, Balusamy B. Improved graph neural network-based green anaconda optimization for segmenting and classifying the lung cancer. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2023; 20:17138-17157. [PMID: 37920050 DOI: 10.3934/mbe.2023764] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/04/2023]
Abstract
Normal lung cells incur genetic damage over time, which causes unchecked cell growth and ultimately leads to lung cancer. Nearly 85% of lung cancer cases are caused by smoking, but there exists factual evidence that beta-carotene supplements and arsenic in water may raise the risk of developing the illness. Asbestos, polycyclic aromatic hydrocarbons, arsenic, radon gas, nickel, chromium and hereditary factors represent various lung cancer-causing agents. Therefore, deep learning approaches are employed to quicken the crucial procedure of diagnosing lung cancer. The effectiveness of these methods has increased when used to examine cancer histopathology slides. Initially, the data is gathered from the standard benchmark dataset. Further, the pre-processing of the collected images is accomplished using the Gabor filter method. The segmentation of these pre-processed images is done through the modified expectation maximization (MEM) algorithm method. Next, using the histogram of oriented gradient (HOG) scheme, the features are extracted from these segmented images. Finally, the classification of lung cancer is performed by the improved graph neural network (IGNN), where the parameter optimization of graph neural network (GNN) is done by the green anaconda optimization (GAO) algorithm in order to derive the accuracy maximization as the major objective function. This IGNN classifies lung cancer into normal, adeno carcinoma and squamous cell carcinoma as the final output. On comparison with existing methods with respect to distinct performance measures, the simulation findings reveal the betterment of the introduced method.
Collapse
Affiliation(s)
- S Dinesh Krishnan
- Assistant professor, B V Raju Institute of Technology, Narsapur, Telangana, India
| | - Danilo Pelusi
- Department of Communication Sciences, University of Teramo, Italy
| | - A Daniel
- Associate Professor, Amity University, Gwalior, Madhya Pradesh, India
| | - V Suresh
- Assistant professor, Dr. N. G. P Institute of Technology, Coimbatore, India
| | | |
Collapse
|
15
|
Qureshi A, Lim S, Suh SY, Mutawak B, Chitnis PV, Demer JL, Wei Q. Deep-Learning-Based Segmentation of Extraocular Muscles from Magnetic Resonance Images. Bioengineering (Basel) 2023; 10:699. [PMID: 37370630 PMCID: PMC10295225 DOI: 10.3390/bioengineering10060699] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/30/2023] [Revised: 05/31/2023] [Accepted: 06/02/2023] [Indexed: 06/29/2023] Open
Abstract
In this study, we investigated the performance of four deep learning frameworks of U-Net, U-NeXt, DeepLabV3+, and ConResNet in multi-class pixel-based segmentation of the extraocular muscles (EOMs) from coronal MRI. Performances of the four models were evaluated and compared with the standard F-measure-based metrics of intersection over union (IoU) and Dice, where the U-Net achieved the highest overall IoU and Dice scores of 0.77 and 0.85, respectively. Centroid distance offset between identified and ground truth EOM centroids was measured where U-Net and DeepLabV3+ achieved low offsets (p > 0.05) of 0.33 mm and 0.35 mm, respectively. Our results also demonstrated that segmentation accuracy varies in spatially different image planes. This study systematically compared factors that impact the variability of segmentation and morphometric accuracy of the deep learning models when applied to segmenting EOMs from MRI.
Collapse
Affiliation(s)
- Amad Qureshi
- Department of Bioengineering, George Mason University, Fairfax, VA 22030, USA; (A.Q.)
| | - Seongjin Lim
- Department of Ophthalmology, Neurology and Bioengineering, Jules Stein Eye Institute, University of California, Los Angeles, CA 90095, USA; (S.L.)
| | - Soh Youn Suh
- Department of Ophthalmology, Neurology and Bioengineering, Jules Stein Eye Institute, University of California, Los Angeles, CA 90095, USA; (S.L.)
| | - Bassam Mutawak
- Department of Bioengineering, George Mason University, Fairfax, VA 22030, USA; (A.Q.)
| | - Parag V. Chitnis
- Department of Bioengineering, George Mason University, Fairfax, VA 22030, USA; (A.Q.)
| | - Joseph L. Demer
- Department of Ophthalmology, Neurology and Bioengineering, Jules Stein Eye Institute, University of California, Los Angeles, CA 90095, USA; (S.L.)
| | - Qi Wei
- Department of Bioengineering, George Mason University, Fairfax, VA 22030, USA; (A.Q.)
| |
Collapse
|
16
|
Jin R, Cai Y, Zhang S, Yang T, Feng H, Jiang H, Zhang X, Hu Y, Liu J. Computational approaches for the reconstruction of optic nerve fibers along the visual pathway from medical images: a comprehensive review. Front Neurosci 2023; 17:1191999. [PMID: 37304011 PMCID: PMC10250625 DOI: 10.3389/fnins.2023.1191999] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2023] [Accepted: 05/09/2023] [Indexed: 06/13/2023] Open
Abstract
Optic never fibers in the visual pathway play significant roles in vision formation. Damages of optic nerve fibers are biomarkers for the diagnosis of various ophthalmological and neurological diseases; also, there is a need to prevent the optic nerve fibers from getting damaged in neurosurgery and radiation therapy. Reconstruction of optic nerve fibers from medical images can facilitate all these clinical applications. Although many computational methods are developed for the reconstruction of optic nerve fibers, a comprehensive review of these methods is still lacking. This paper described both the two strategies for optic nerve fiber reconstruction applied in existing studies, i.e., image segmentation and fiber tracking. In comparison to image segmentation, fiber tracking can delineate more detailed structures of optic nerve fibers. For each strategy, both conventional and AI-based approaches were introduced, and the latter usually demonstrates better performance than the former. From the review, we concluded that AI-based methods are the trend for optic nerve fiber reconstruction and some new techniques like generative AI can help address the current challenges in optic nerve fiber reconstruction.
Collapse
Affiliation(s)
- Richu Jin
- Research Institute of Trustworthy Autonomous Systems, Southern University of Science and Technology, Shenzhen, China
- Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China
| | - Yongning Cai
- Research Institute of Trustworthy Autonomous Systems, Southern University of Science and Technology, Shenzhen, China
| | - Shiyang Zhang
- Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China
| | - Ting Yang
- Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China
| | - Haibo Feng
- Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China
| | - Hongyang Jiang
- Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China
| | - Xiaoqing Zhang
- Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China
| | - Yan Hu
- Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China
| | - Jiang Liu
- Research Institute of Trustworthy Autonomous Systems, Southern University of Science and Technology, Shenzhen, China
- Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China
- Guangdong Provincial Key Laboratory of Brain-inspired Intelligent Computation, Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China
| |
Collapse
|
17
|
Chen Y, Zhang X, Li D, Park H, Li X, Liu P, Jin J, Shen Y. Automatic segmentation of thyroid with the assistance of the devised boundary improvement based on multicomponent small dataset. APPL INTELL 2023; 53:1-16. [PMID: 37363389 PMCID: PMC10015528 DOI: 10.1007/s10489-023-04540-5] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 02/23/2023] [Indexed: 03/17/2023]
Abstract
Deep learning has been widely considered in medical image segmentation. However, the difficulty of acquiring medical images and labels can affect the accuracy of the segmentation results for deep learning methods. In this paper, an automatic segmentation method is proposed by devising a multicomponent neighborhood extreme learning machine to improve the boundary attention region of the preliminary segmentation results. The neighborhood features are acquired by training U-Nets with the multicomponent small dataset, which consists of original thyroid ultrasound images, Sobel edge images and superpixel images. Afterward, the neighborhood features are selected by min-redundancy and max-relevance filter in the designed extreme learning machine, and the selected features are used to train the extreme learning machine to obtain supplementary segmentation results. Finally, the accuracy of the segmentation results is improved by adjusting the boundary attention region of the preliminary segmentation results with the supplementary segmentation results. This method combines the advantages of deep learning and traditional machine learning, boosting the accuracy of thyroid segmentation accuracy with a small dataset in a multigroup test.
Collapse
Affiliation(s)
- Yifei Chen
- Control Science and Engineering, Harbin Institute of Technology, Harbin, 150001 China
- Electrical Engineering, Korea Advanced Institute of Science and Technology, Daejeon, 34141 Korea
| | - Xin Zhang
- Control Science and Engineering, Harbin Institute of Technology, Harbin, 150001 China
| | - Dandan Li
- Control Science and Engineering, Harbin Institute of Technology, Harbin, 150001 China
| | - HyunWook Park
- Electrical Engineering, Korea Advanced Institute of Science and Technology, Daejeon, 34141 Korea
| | - Xinran Li
- Mathematics, Harbin Institute of Technology, Harbin, 150001 China
| | - Peng Liu
- Heilongjiang Provincial Key Laboratory of Trace Elements and Human Health, Harbin Medical University, Harbin, 150081 China
| | - Jing Jin
- Control Science and Engineering, Harbin Institute of Technology, Harbin, 150001 China
| | - Yi Shen
- Control Science and Engineering, Harbin Institute of Technology, Harbin, 150001 China
| |
Collapse
|
18
|
Shanker RRBJ, Zhang MH, Ginat DT. Semantic Segmentation of Extraocular Muscles on Computed Tomography Images Using Convolutional Neural Networks. Diagnostics (Basel) 2022; 12:1553. [PMID: 35885459 PMCID: PMC9325103 DOI: 10.3390/diagnostics12071553] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/16/2022] [Revised: 06/23/2022] [Accepted: 06/24/2022] [Indexed: 11/22/2022] Open
Abstract
Computed tomography (CT) imaging of the orbit with measurement of extraocular muscle size can be useful for diagnosing and monitoring conditions that affect extraocular muscles. However, the manual measurement of extraocular muscle size can be time-consuming and tedious. The purpose of this study is to evaluate the effectiveness of deep learning algorithms in segmenting extraocular muscles and measuring muscle sizes from CT images. Consecutive CT scans of orbits from 210 patients between 1 January 2010 and 31 December 2019 were used. Extraocular muscles were manually annotated in the studies, which were then used to train the deep learning algorithms. The proposed U-net algorithm can segment extraocular muscles on coronal slices of 32 test samples with an average dice score of 0.92. The thickness and area measurements from predicted segmentations had a mean absolute error (MAE) of 0.35 mm and 3.87 mm2, respectively, with a corresponding mean absolute percentage error (MAPE) of 7 and 9%, respectively. On qualitative analysis of 32 test samples, 30 predicted segmentations from the U-net algorithm were accepted while 2 were rejected. Based on the results from quantitative and qualitative evaluation, this study demonstrates that CNN-based deep learning algorithms are effective at segmenting extraocular muscles and measuring muscles sizes.
Collapse
Affiliation(s)
| | - Michael H. Zhang
- Department of Radiology, University of Chicago, Chicago, IL 60615, USA; (R.R.B.J.S.); (M.H.Z.)
| | - Daniel T. Ginat
- Department of Radiology, Section of Neuroradiology, University of Chicago, Chicago, IL 60615, USA
| |
Collapse
|
19
|
Bi R, Ji C, Yang Z, Qiao M, Lv P, Wang H. Residual based attention-Unet combing DAC and RMP modules for automatic liver tumor segmentation in CT. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2022; 19:4703-4718. [PMID: 35430836 DOI: 10.3934/mbe.2022219] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Purpose: Due to the complex distribution of liver tumors in the abdomen, the accuracy of liver tumor segmentation cannot meet the needs of clinical assistance yet. This paper aims to propose a new end-to-end network to improve the segmentation accuracy of liver tumors from CT. Method: We proposed a hybrid network, leveraging the residual block, the context encoder (CE), and the Attention-Unet, called ResCEAttUnet. The CE comprises a dense atrous convolution (DAC) module and a residual multi-kernel pooling (RMP) module. The DAC module ensures the network derives high-level semantic information and minimizes detailed information loss. The RMP module improves the ability of the network to extract multi-scale features. Moreover, a hybrid loss function based on cross-entropy and Tversky loss function is employed to distribute the weights of the two-loss parts through training iterations. Results: We evaluated the proposed method in LiTS17 and 3DIRCADb databases. It significantly improved the segmentation accuracy compared to state-of-the-art methods. Conclusions: Experimental results demonstrate the satisfying effects of the proposed method through both quantitative and qualitative analyses, thus proving a promising tool in liver tumor segmentation.
Collapse
Affiliation(s)
- Rongrong Bi
- Department of Software Engineering, Harbin University of Science and Technology, Rongcheng 264300, China
| | - Chunlei Ji
- Department of Software Engineering, Harbin University of Science and Technology, Rongcheng 264300, China
| | - Zhipeng Yang
- School of Automation, Harbin University of Science and Technology, Harbin 150080, China
| | - Meixia Qiao
- Department of Software Engineering, Harbin University of Science and Technology, Rongcheng 264300, China
| | - Peiqing Lv
- School of Automation, Harbin University of Science and Technology, Harbin 150080, China
| | - Haiying Wang
- School of Automation, Harbin University of Science and Technology, Harbin 150080, China
| |
Collapse
|
20
|
Nerve optic segmentation in CT images using a deep learning model and a texture descriptor. COMPLEX INTELL SYST 2022. [DOI: 10.1007/s40747-022-00694-w] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/17/2022]
Abstract
AbstractThe increased intracranial pressure (ICP) can be described as an increase in pressure around the brain and can lead to serious health problems. The assessment of ultrasound images is commonly conducted by skilled experts which is a time-consuming approach, but advanced computer-aided diagnosis (CAD) systems can assist the physician to decrease the time of ICP diagnosis. The accurate detection of the nerve optic regions, with drawing a precise slope line behind the eyeball and calculating the diameter of nerve optic, are the main aims of this research. First, the Fuzzy C-mean (FCM) clustering is employed for segmenting the input CT screening images into the different parts. Second, a histogram equalization approach is used for region-based image quality enhancement. Then, the Local Directional Number method (LDN) is used for representing some key information in a new image. Finally, a cascade Convolutional Neural Network (CNN) is employed for nerve optic segmentation by two distinct input images. Comprehensive experiments on the CT screening dataset [The Cancer Imaging Archive (TCIA)] consisting of 1600 images show the competitive results of inaccurate extraction of the brain features. Also, the indexes such as Dice, Specificity, and Precision for the proposed approach are reported 87.7%, 91.3%, and 90.1%, respectively. The final classification results show that the proposed approach effectively and accurately detects the nerve optic and its diameter in comparison with the other methods. Therefore, this method can be used for early diagnose of ICP and preventing the occurrence of serious health problems in patients.
Collapse
|