1
|
Zhu M, Yang Q, Gao Z, Yuan Y, Liu J. FedBM: Stealing knowledge from pre-trained language models for heterogeneous federated learning. Med Image Anal 2025; 102:103524. [PMID: 40073584 DOI: 10.1016/j.media.2025.103524] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2024] [Revised: 01/21/2025] [Accepted: 02/22/2025] [Indexed: 03/14/2025]
Abstract
Federated learning (FL) has shown great potential in medical image computing since it provides a decentralized learning paradigm that allows multiple clients to train a model collaboratively without privacy leakage. However, current studies have shown that data heterogeneity incurs local learning bias in classifiers and feature extractors of client models during local training, leading to the performance degradation of a federation system. To address these issues, we propose a novel framework called Federated Bias eliMinating (FedBM) to get rid of local learning bias in heterogeneous federated learning (FL), which mainly consists of two modules, i.e., Linguistic Knowledge-based Classifier Construction (LKCC) and Concept-guided Global Distribution Estimation (CGDE). Specifically, LKCC exploits class concepts, prompts and pre-trained language models (PLMs) to obtain concept embeddings. These embeddings are used to estimate the latent concept distribution of each class in the linguistic space. Based on the theoretical derivation, we can rely on these distributions to pre-construct a high-quality classifier for clients to achieve classification optimization, which is frozen to avoid classifier bias during local training. CGDE samples probabilistic concept embeddings from the latent concept distributions to learn a conditional generator to capture the input space of the global model. Three regularization terms are introduced to improve the quality and utility of the generator. The generator is shared by all clients and produces pseudo data to calibrate updates of local feature extractors. Extensive comparison experiments and ablation studies on public datasets demonstrate the superior performance of FedBM over state-of-the-arts and confirm the effectiveness of each module, respectively. The code is available at https://github.com/CUHK-AIM-Group/FedBM.
Collapse
Affiliation(s)
- Meilu Zhu
- Department of Mechanical Engineering, City University of Hong Kong, Hong Kong Special Administrative Region of China
| | - Qiushi Yang
- Department of Electrical Engineering, City University of Hong Kong, Hong Kong Special Administrative Region of China
| | - Zhifan Gao
- School of Biomedical Engineering, Sun Yat-sen University, China
| | - Yixuan Yuan
- Department of Electronic Engineering, Chinese University of Hong Kong, Hong Kong Special Administrative Region of China.
| | - Jun Liu
- Department of Mechanical Engineering, City University of Hong Kong, Hong Kong Special Administrative Region of China; Department of Data and Systems Engineering, The University of Hong Kong, Hong Kong Special Administrative Region of China.
| |
Collapse
|
2
|
Li G, Li K, Zhang G, Pan K, Ding Y, Wang Z, Fu C, Zhu Z. A landslide area segmentation method based on an improved UNet. Sci Rep 2025; 15:11852. [PMID: 40195381 PMCID: PMC11976986 DOI: 10.1038/s41598-025-94039-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/06/2024] [Accepted: 03/11/2025] [Indexed: 04/09/2025] Open
Abstract
As remote sensing technology matures, landslide target segmentation has become increasingly important in disaster prevention, control, and urban construction, playing a crucial role in disaster loss assessment and post-disaster rescue. Therefore, this paper proposes an improved UNet-based landslide segmentation algorithm. Firstly, the feature extraction structure of the model was redesigned by integrating dilated convolution and EMA attention mechanism to enhance the model's ability to extract image features. Additionally, this study introduces the Pag module to replace the original skip connection method, thereby enhancing information fusion between feature maps, reducing pixel information loss, and further improving the model's overall performance. Experimental results show that compared to the original model, our model improves mIoU, Precision, Recall, and F1-score by approximately 2.4%, 2.4%, 3.2%, and 2.8%, respectively. This study not only provides an effective method for landslide segmentation tasks but also offers new perspectives for further research in related fields.
Collapse
Affiliation(s)
- Guangchen Li
- Shandong Jiaotong University, Haitang Road 5001, Jinan, 250357, China
| | - Kefeng Li
- Shandong Jiaotong University, Haitang Road 5001, Jinan, 250357, China
| | - Guangyuan Zhang
- Shandong Jiaotong University, Haitang Road 5001, Jinan, 250357, China.
| | - Ke Pan
- Shandong Jiaotong University, Haitang Road 5001, Jinan, 250357, China
| | - Yuxuan Ding
- Shandong Jiaotong University, Haitang Road 5001, Jinan, 250357, China
| | - Zhenfei Wang
- Shandong Zhengyuan Yeda Environmental Technology Co., Ltd, Jinan, 250101, China
| | - Chen Fu
- Shandong Jiaotong University, Haitang Road 5001, Jinan, 250357, China
| | - Zhenfang Zhu
- Shandong Jiaotong University, Haitang Road 5001, Jinan, 250357, China
| |
Collapse
|
3
|
Bhati A, Jain S, Gour N, Khanna P, Ojha A, Werghi N. Dynamic Statistical Attention-based lightweight model for Retinal Vessel Segmentation: DyStA-RetNet. Comput Biol Med 2025; 186:109592. [PMID: 39731923 DOI: 10.1016/j.compbiomed.2024.109592] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/09/2024] [Revised: 12/12/2024] [Accepted: 12/16/2024] [Indexed: 12/30/2024]
Abstract
BACKGROUND AND OBJECTIVE Accurate extraction of retinal vascular components is vital in diagnosing and treating retinal diseases. Achieving precise segmentation of retinal blood vessels is challenging due to their complex structure and overlapping vessels with other anatomical features. Existing deep neural networks often suffer from false positives at vessel branches or missing fragile vessel patterns. Also, deployment of the existing models in resource-constrained environments is challenging due to their computational complexity. An attention-based and computationally efficient architecture is proposed in this work to bridge this gap while enabling improved segmentation of retinal vascular structures. METHODS The proposed dynamic statistical attention-based lightweight model for retinal vessel segmentation (DyStA-RetNet) employs a shallow CNN-based encoder-decoder architecture. One branch of the decoder utilizes a partial decoder connecting encoder layers with decoder layers to allow the transfer of high-level semantic information, whereas the other branch helps to incorporate low-level information. The multi-scale dynamic attention block empowers the network to accurately identify different-sized tree-shaped vessel patterns during the reconstruction phase in the decoder. The statistical spatial attention block improves the feature learning capability. By effectively integrating low-level and high-level semantic information, DYStA-RetNet significantly improves the performance of vessel segmentation. RESULTS Experiments performed on four benchmark datasets (DRIVE, STARE, CHASEDB, and HRF) exhibit the adaptability of DYStA-RetNet for clinical applications with a significantly smaller number of trainable parameters (37.19K) and GFLOPS (0.75), and superior segmentation performance. CONCLUSION The proposed lightweight CNN-based DYStA-RetNet efficiently extracts complex retinal vascular components from fundus images. It is computationally efficient and deployable in resource-constrained environments.
Collapse
Affiliation(s)
- Amit Bhati
- PDPM Indian Institute of Information Technology, Design and Manufacturing, Jabalpur 482005, India
| | - Samir Jain
- PDPM Indian Institute of Information Technology, Design and Manufacturing, Jabalpur 482005, India
| | - Neha Gour
- Khalifa University, Abu Dhabi, United Arab Emirates
| | - Pritee Khanna
- PDPM Indian Institute of Information Technology, Design and Manufacturing, Jabalpur 482005, India.
| | - Aparajita Ojha
- PDPM Indian Institute of Information Technology, Design and Manufacturing, Jabalpur 482005, India
| | | |
Collapse
|
4
|
Kande GB, Nalluri MR, Manikandan R, Cho J, Veerappampalayam Easwaramoorthy S. Multi scale multi attention network for blood vessel segmentation in fundus images. Sci Rep 2025; 15:3438. [PMID: 39870673 PMCID: PMC11772654 DOI: 10.1038/s41598-024-84255-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2024] [Accepted: 12/20/2024] [Indexed: 01/29/2025] Open
Abstract
Precise segmentation of retinal vasculature is crucial for the early detection, diagnosis, and treatment of vision-threatening ailments. However, this task is challenging due to limited contextual information, variations in vessel thicknesses, the complexity of vessel structures, and the potential for confusion with lesions. In this paper, we introduce a novel approach, the MSMA Net model, which overcomes these challenges by replacing traditional convolution blocks and skip connections with an improved multi-scale squeeze and excitation block (MSSE Block) and Bottleneck residual paths (B-Res paths) with spatial attention blocks (SAB). Our experimental findings on publicly available datasets of fundus images, specifically DRIVE, STARE, CHASE_DB1, HRF and DR HAGIS consistently demonstrate that our approach outperforms other segmentation techniques, achieving higher accuracy, sensitivity, Dice score, and area under the receiver operator characteristic (AUC) in the segmentation of blood vessels with different thicknesses, even in situations involving diverse contextual information, the presence of coexisting lesions, and intricate vessel morphologies.
Collapse
Affiliation(s)
- Giri Babu Kande
- Vasireddy Venkatadri Institute of Technology, Nambur, 522508, India
| | - Madhusudana Rao Nalluri
- School of Computing, Amrita Vishwa Vidyapeetham, Amaravati, 522503, India.
- Department of Computer Science & Engineering, Faculty of Science and Technology (IcfaiTech), ICFAI Foundation for Higher Education, Hyderabad, India.
| | - R Manikandan
- School of Computing, SASTRA Deemed University, Thanjavur, 613401, India
| | - Jaehyuk Cho
- Department of Software Engineering & Division of Electronics and Information Engineering, Jeonbuk National University, Jeonju-si, 54896, Republic of Korea.
| | | |
Collapse
|
5
|
Chetoui M, Akhloufi MA. A Novel Ensemble Meta-Model for Enhanced Retinal Blood Vessel Segmentation Using Deep Learning Architectures. Biomedicines 2025; 13:141. [PMID: 39857725 PMCID: PMC11760907 DOI: 10.3390/biomedicines13010141] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2024] [Revised: 01/05/2025] [Accepted: 01/07/2025] [Indexed: 01/27/2025] Open
Abstract
Background: Retinal blood vessel segmentation plays an important role in diagnosing retinal diseases such as diabetic retinopathy, glaucoma, and hypertensive retinopathy. Accurate segmentation of blood vessels in retinal images presents a challenging task due to noise, low contrast, and the complex morphology of blood vessel structures. Methods: In this study, we propose a novel ensemble learning framework combining four deep learning architectures: U-Net, ResNet50, U-Net with a ResNet50 backbone, and U-Net with a transformer block. Each architecture is customized to enhance feature extraction and segmentation performance. The models are trained on the DRIVE and STARE datasets to improve the degree of generalization and evaluated using the performance metric accuracy, F1-Score, sensitivity, specificity, and AUC. Results: The ensemble meta-model integrates predictions from these architectures using a stacking approach, achieving state-of-the-art performance with an accuracy of 0.9778, an AUC of 0.9912, and an F1-Score of 0.8231. These results demonstrate the performance of the proposed technique in identifying thin retinal blood vessels. Conclusions: A comparative analysis using qualitative and quantitative results with individual models highlights the robustness of the ensemble framework, especially under conditions of noise and poor visibility.
Collapse
Affiliation(s)
| | - Moulay A. Akhloufi
- Perception, Robotics, and Intelligent Machines Lab (PRIME), Department of Computer Science, Université de Moncton, Moncton, NB E1A 3E9, Canada;
| |
Collapse
|
6
|
Deng H, Wang Y, Cheng V, He Y, Wen Z, Chen S, Guo S, Zhou P, Wang Y. Research trends and hotspots in fundus image segmentation from 2007 to 2023: A bibliometric analysis. Heliyon 2024; 10:e39329. [PMID: 39524903 PMCID: PMC11544040 DOI: 10.1016/j.heliyon.2024.e39329] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2024] [Revised: 09/23/2024] [Accepted: 10/11/2024] [Indexed: 11/16/2024] Open
Abstract
Objective To understand the current status, research hotspots, and trends of automatic segmentation of fundus lesion images worldwide, providing a reference for subsequent related studies. Methods The electronic database Web of Science Core Collection was searched for research in the field of automatic segmentation of fundus lesion images from 2007 to 2023. Visualization maps of countries, authors, institutions, journals, references, and keywords were generated and analyzed using the CiteSpace and VOSviewer software. Results After deduplication, 707 publications were sorted out, showing an overall increasing trend in publication volume. The countries with the highest publication counts were China, followed by India, the USA, the UK, Spain, Pakistan, and Singapore. A high degree of collaboration was observed among authors, and they cooperated widely. The keywords included "diabetic retinopathy," "deep learning," "vessel segmentation," "retinal images," "optic disc localization," and so forth, with keyword bursts starting in 2018 for "retinal images," "machine learning," "biomedical imaging," "deep learning," "convolutional neural networks," and "transfer learning." The most prolific author was U Rajendra Acharya from the University of Southern Queensland, and the journal with the most publications was Computer Methods and Programs in Biomedicine. Conclusions Compared with manual segmentation of fundus lesion images, the use of deep learning models for segmentation is more efficient and accurate, which is crucial for patients with eye diseases. Although the number of related publications globally is relatively small, a growing trend is still witnessed, with broad connections between countries and authors, mainly concentrated in East Asia and Europe. Research institutions in this field are limited, and hence, the research on diabetic retinopathy and retinal vessel segmentation should be strengthened to promote the development of this area.
Collapse
Affiliation(s)
- Hairui Deng
- School of Nursing, Southwest Medical University, Luzhou, 646000, China
- Wound Healing Basic Research and Clinical Application Key Laboratory of Luzhou, School of Nursing, Southwest Medical University, Luzhou, 646000, China
| | - Yiren Wang
- School of Nursing, Southwest Medical University, Luzhou, 646000, China
- Wound Healing Basic Research and Clinical Application Key Laboratory of Luzhou, School of Nursing, Southwest Medical University, Luzhou, 646000, China
| | - Venhui Cheng
- Department of Ophthalmology, The Affiliated Hospital of Southwest Medical University, Luzhou, 646000, China
| | - Yongcheng He
- Department of Pharmacy, Sichuan Agricultural University, Chengdu, 610000, China
| | - Zhongjian Wen
- School of Nursing, Southwest Medical University, Luzhou, 646000, China
- Wound Healing Basic Research and Clinical Application Key Laboratory of Luzhou, School of Nursing, Southwest Medical University, Luzhou, 646000, China
| | - Shouying Chen
- School of Nursing, Southwest Medical University, Luzhou, 646000, China
- Wound Healing Basic Research and Clinical Application Key Laboratory of Luzhou, School of Nursing, Southwest Medical University, Luzhou, 646000, China
| | - Shengmin Guo
- Department of Nursing, The Affiliated Hospital of Southwest Medical University, Luzhou, 646000, China
| | - Ping Zhou
- Wound Healing Basic Research and Clinical Application Key Laboratory of Luzhou, School of Nursing, Southwest Medical University, Luzhou, 646000, China
- Department of Radiology, The Affiliated Hospital of Southwest Medical University, Luzhou, 646000, China
| | - Yi Wang
- Department of Publicity, The Affiliated Hospital of Southwest Medical University, Luzhou, 646000, China
| |
Collapse
|
7
|
Seong D, Lee E, Kim Y, Yae CG, Choi J, Kim HK, Jeon M, Kim J. Deep learning based highly accurate transplanted bioengineered corneal equivalent thickness measurement using optical coherence tomography. NPJ Digit Med 2024; 7:308. [PMID: 39501083 PMCID: PMC11538249 DOI: 10.1038/s41746-024-01305-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/21/2024] [Accepted: 10/15/2024] [Indexed: 11/08/2024] Open
Abstract
Corneal transplantation is the primary treatment for irreversible corneal diseases, but due to limited donor availability, bioengineered corneal equivalents are being developed as a solution, with biocompatibility, structural integrity, and physical function considered key factors. Since conventional evaluation methods may not fully capture the complex properties of the cornea, there is a need for advanced imaging and assessment techniques. In this study, we proposed a deep learning-based automatic segmentation method for transplanted bioengineered corneal equivalents using optical coherence tomography to achieve a highly accurate evaluation of graft integrity and biocompatibility. Our method provides quantitative individual thickness values, detailed maps, and volume measurements of the bioengineered corneal equivalents, and has been validated through 14 days of monitoring. Based on the results, it is expected to have high clinical utility as a quantitative assessment method for human keratoplasties, including automatic opacity area segmentation and implanted graft part extraction, beyond animal studies.
Collapse
Affiliation(s)
- Daewoon Seong
- School of Electronic and Electrical Engineering, College of IT engineering, Kyungpook National University, Daegu, Republic of Korea
| | - Euimin Lee
- School of Electronic and Electrical Engineering, College of IT engineering, Kyungpook National University, Daegu, Republic of Korea
| | - Yoonseok Kim
- School of Electronic and Electrical Engineering, College of IT engineering, Kyungpook National University, Daegu, Republic of Korea
| | - Che Gyem Yae
- Bio-Medical Institute, Kyungpook National University Hospital, Daegu, Korea
- Department of Ophthalmology, School of Medicine, Kyungpook National University, Daegu, Korea
| | - JeongMun Choi
- Bio-Medical Institute, Kyungpook National University Hospital, Daegu, Korea
- Department of Ophthalmology, School of Medicine, Kyungpook National University, Daegu, Korea
| | - Hong Kyun Kim
- Bio-Medical Institute, Kyungpook National University Hospital, Daegu, Korea.
- Department of Ophthalmology, School of Medicine, Kyungpook National University, Daegu, Korea.
| | - Mansik Jeon
- School of Electronic and Electrical Engineering, College of IT engineering, Kyungpook National University, Daegu, Republic of Korea.
| | - Jeehyun Kim
- School of Electronic and Electrical Engineering, College of IT engineering, Kyungpook National University, Daegu, Republic of Korea
| |
Collapse
|
8
|
Lv N, Xu L, Chen Y, Sun W, Tian J, Zhang S. TCDDU-Net: combining transformer and convolutional dual-path decoding U-Net for retinal vessel segmentation. Sci Rep 2024; 14:25978. [PMID: 39472606 PMCID: PMC11522399 DOI: 10.1038/s41598-024-77464-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/06/2024] [Accepted: 10/22/2024] [Indexed: 11/02/2024] Open
Abstract
Accurate segmentation of retinal blood vessels is crucial for enhancing diagnostic efficiency and preventing disease progression. However, the small size and complex structure of retinal blood vessels, coupled with low contrast in corresponding fundus images, pose significant challenges for this task. We propose a novel approach for retinal vessel segmentation, which combines the transformer and convolutional dual-path decoding U-Net (TCDDU-Net). We propose the selective dense connection swin transformer block, which converts the input feature map into patches, introduces MLPs to generate probabilities, and performs selective fusion at different stages. This structure forms a dense connection framework, enabling the capture of long-distance dependencies and effective fusion of features across different stages. The subsequent stage involves the design of the background decoder, which utilizes deformable convolution to learn the background information of retinal vessels by treating them as segmentation objects. This is then combined with the foreground decoder to form a dual-path decoding U-Net. Finally, the foreground segmentation results and the processed background segmentation results are fused to obtain the final retinal vessel segmentation map. To evaluate the effectiveness of our method, we performed experiments on the DRIVE, STARE, and CHASE datasets for retinal vessel segmentation. Experimental results show that the segmentation accuracies of our algorithms are 96.98, 97.40, and 97.23, and the AUC metrics are 98.68, 98.56, and 98.50, respectively.In addition, we evaluated our methods using F1 score, specificity, and sensitivity metrics. Through a comparative analysis, we found that our proposed TCDDU-Net method effectively improves retinal vessel segmentation performance and achieves impressive results on multiple datasets compared to existing methods.
Collapse
Affiliation(s)
- Nianzu Lv
- College of Information Engineering, Xinjiang Institute of Technology, No.1 Xuefu West Road, Aksu, 843100, Xinjiang, China
| | - Li Xu
- College of Information Engineering, Xinjiang Institute of Technology, No.1 Xuefu West Road, Aksu, 843100, Xinjiang, China.
| | - Yuling Chen
- School of Information Engineering, Mianyang Teachers' College, No. 166 Mianxing West Road, High Tech Zone, Mianyang, 621000, Sichuan, China
| | - Wei Sun
- CISDI Engineering Co., LTD, Chongqing, 401120, China
| | - Jiya Tian
- College of Information Engineering, Xinjiang Institute of Technology, No.1 Xuefu West Road, Aksu, 843100, Xinjiang, China
| | - Shuping Zhang
- College of Information Engineering, Xinjiang Institute of Technology, No.1 Xuefu West Road, Aksu, 843100, Xinjiang, China
| |
Collapse
|
9
|
Liu Q, Zhou F, Shen J, Xu J, Wan C, Xu X, Yan Z, Yao J. A fundus vessel segmentation method based on double skip connections combined with deep supervision. Front Cell Dev Biol 2024; 12:1477819. [PMID: 39430046 PMCID: PMC11487527 DOI: 10.3389/fcell.2024.1477819] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/08/2024] [Accepted: 09/20/2024] [Indexed: 10/22/2024] Open
Abstract
Background Fundus vessel segmentation is vital for diagnosing ophthalmic diseases like central serous chorioretinopathy (CSC), diabetic retinopathy, and glaucoma. Accurate segmentation provides crucial vessel morphology details, aiding the early detection and intervention of ophthalmic diseases. However, current algorithms struggle with fine vessel segmentation and maintaining sensitivity in complex regions. Challenges also stem from imaging variability and poor generalization across multimodal datasets, highlighting the need for more advanced algorithms in clinical practice. Methods This paper aims to explore a new vessel segmentation method to alleviate the above problems. We propose a fundus vessel segmentation model based on a combination of double skip connections, deep supervision, and TransUNet, namely DS2TUNet. Initially, the original fundus images are improved through grayscale conversion, normalization, histogram equalization, gamma correction, and other preprocessing techniques. Subsequently, by utilizing the U-Net architecture, the preprocessed fundus images are segmented to obtain the final vessel information. Specifically, the encoder firstly incorporates the ResNetV1 downsampling, dilated convolution downsampling, and Transformer to capture both local and global features, which upgrades its vessel feature extraction ability. Then, the decoder introduces the double skip connections to facilitate upsampling and refine segmentation outcomes. Finally, the deep supervision module introduces multiple upsampling vessel features from the decoder into the loss function, so that the model can learn vessel feature representations more effectively and alleviate gradient vanishing during the training phase. Results Extensive experiments on publicly available multimodal fundus datasets such as DRIVE, CHASE_DB1, and ROSE-1 demonstrate that the DS2TUNet model attains F1-scores of 0.8195, 0.8362, and 0.8425, with Accuracy of 0.9664, 0.9741, and 0.9557, Sensitivity of 0.8071, 0.8101, and 0.8586, and Specificity of 0.9823, 0.9869, and 0.9713, respectively. Additionally, the model also exhibits excellent test performance on the clinical fundus dataset CSC, with F1-score of 0.7757, Accuracy of 0.9688, Sensitivity of 0.8141, and Specificity of 0.9801 based on the weight trained on the CHASE_DB1 dataset. These results comprehensively validate that the proposed method obtains good performance in fundus vessel segmentation, thereby aiding clinicians in the further diagnosis and treatment of fundus diseases in terms of effectiveness and feasibility.
Collapse
Affiliation(s)
- Qingyou Liu
- College of Mechanical and Electrical Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing, China
| | - Fen Zhou
- The Affiliated Eye Hospital of Nanjing Medical University, Nanjing, China
| | - Jianxin Shen
- College of Mechanical and Electrical Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing, China
| | - Jianguo Xu
- College of Mechanical and Electrical Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing, China
| | - Cheng Wan
- College of Electronic and Information Engineering, Nanjing University of Aeronautics and Astronautics, Nanjing, China
| | - Xiangzhong Xu
- The Affiliated Eye Hospital of Nanjing Medical University, Nanjing, China
| | - Zhipeng Yan
- The Affiliated Eye Hospital of Nanjing Medical University, Nanjing, China
| | - Jin Yao
- The Affiliated Eye Hospital of Nanjing Medical University, Nanjing, China
| |
Collapse
|
10
|
Zhang Y, Chung ACS. Retinal Vessel Segmentation by a Transformer-U-Net Hybrid Model With Dual-Path Decoder. IEEE J Biomed Health Inform 2024; 28:5347-5359. [PMID: 38669172 DOI: 10.1109/jbhi.2024.3394151] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/28/2024]
Abstract
This paper introduces an effective and efficient framework for retinal vessel segmentation. First, we design a Transformer-CNN hybrid model in which a Transformer module is inserted inside the U-Net to capture long-range interactions. Second, we design a dual-path decoder in the U-Net framework, which contains two decoding paths for multi-task outputs. Specifically, we train the extra decoder to predict vessel skeletons as an auxiliary task which helps the model learn balanced features. The proposed framework, named as TSNet, not only achieves good performances in a fully supervised learning manner but also enables a rough skeleton annotation process. The annotators only need to roughly delineate vessel skeletons instead of giving precise pixel-wise vessel annotations. To learn with rough skeleton annotations plus a few precise vessel annotations, we propose a skeleton semi-supervised learning scheme. We adopt a mean teacher model to produce pseudo vessel annotations and conduct annotation correction for roughly labeled skeletons annotations. This learning scheme can achieve promising performance with fewer annotation efforts. We have evaluated TSNet through extensive experiments on five benchmarking datasets. Experimental results show that TSNet yields state-of-the-art performances on retinal vessel segmentation and provides an efficient training scheme in practice.
Collapse
|
11
|
Khan MZ, Gajendran MK. Generative Neural Framework for Micro-Vessels Classification. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2024; 2024:1-4. [PMID: 40039239 DOI: 10.1109/embc53108.2024.10782802] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/06/2025]
Abstract
The morphological abnormalities in the retinal blood vessel have a close association with cerebrovascular, cardio-vascular, and systemic diseases. It makes the retinal artery/vein (A/V) classification salient for clinical decision-making. The existing methods find it challenging to correctly classify A/V with non-uniform brightness and vessel thickness, especially at the bifurcation and endpoints. To avoid these problems and increase precision, AV-Net is proposed. It uses the context information and performs data fusion to improve A/V classification. Specifically, the AV-Net offers a module that fuses local and global vessel information for creating a weight map to constrain the A/V features. It helps suppress the background-prone features and improve region extraction at the bifurcation and endpoints. In addition, to improve model robustness, the AV-Net uses a multiscale-feature module that captures coarse and fine details.
Collapse
|
12
|
Liu X, Tan H, Wang W, Chen Z. Deep learning based retinal vessel segmentation and hypertensive retinopathy quantification using heterogeneous features cross-attention neural network. Front Med (Lausanne) 2024; 11:1377479. [PMID: 38841586 PMCID: PMC11150614 DOI: 10.3389/fmed.2024.1377479] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/27/2024] [Accepted: 05/09/2024] [Indexed: 06/07/2024] Open
Abstract
Retinal vessels play a pivotal role as biomarkers in the detection of retinal diseases, including hypertensive retinopathy. The manual identification of these retinal vessels is both resource-intensive and time-consuming. The fidelity of vessel segmentation in automated methods directly depends on the fundus images' quality. In instances of sub-optimal image quality, applying deep learning-based methodologies emerges as a more effective approach for precise segmentation. We propose a heterogeneous neural network combining the benefit of local semantic information extraction of convolutional neural network and long-range spatial features mining of transformer network structures. Such cross-attention network structure boosts the model's ability to tackle vessel structures in the retinal images. Experiments on four publicly available datasets demonstrate our model's superior performance on vessel segmentation and the big potential of hypertensive retinopathy quantification.
Collapse
Affiliation(s)
- Xinghui Liu
- School of Clinical Medicine, Guizhou Medical University, Guiyang, China
- Department of Cardiovascular Medicine, Guizhou Provincial People's Hospital, Guiyang, China
| | - Hongwen Tan
- Department of Cardiovascular Medicine, Guizhou Provincial People's Hospital, Guiyang, China
| | - Wu Wang
- Electrical Engineering College, Guizhou University, Guiyang, China
| | - Zhangrong Chen
- School of Clinical Medicine, Guizhou Medical University, Guiyang, China
- Department of Cardiovascular Medicine, The Affiliated Hospital of Guizhou Medical University, Guiyang, China
| |
Collapse
|
13
|
Huang Z, Zhao Y, Yu Z, Qin P, Han X, Wang M, Liu M, Gregersen H. BiU-net: A dual-branch structure based on two-stage fusion strategy for biomedical image segmentation. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 252:108235. [PMID: 38776830 DOI: 10.1016/j.cmpb.2024.108235] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/12/2023] [Revised: 04/28/2024] [Accepted: 05/17/2024] [Indexed: 05/25/2024]
Abstract
BACKGROUND AND OBJECTIVE Computer-based biomedical image segmentation plays a crucial role in planning of assisted diagnostics and therapy. However, due to the variable size and irregular shape of the segmentation target, it is still a challenge to construct an effective medical image segmentation structure. Recently, hybrid architectures based on convolutional neural networks (CNNs) and transformers were proposed. However, most current backbones directly replace one or all convolutional layers with transformer blocks, regardless of the semantic gap between features. Thus, how to sufficiently and effectively eliminate the semantic gap as well as combine the global and local information is a critical challenge. METHODS To address the challenge, we propose a novel structure, called BiU-Net, which integrates CNNs and transformers with a two-stage fusion strategy. In the first fusion stage, called Single-Scale Fusion (SSF) stage, the encoding layers of the CNNs and transformers are coupled, with both having the same feature map size. The SSF stage aims to reconstruct local features based on CNNs and long-range information based on transformers in each encoding block. In the second stage, Multi-Scale Fusion (MSF), BiU-Net interacts with multi-scale features from various encoding layers to eliminate the semantic gap between deep and shallow layers. Furthermore, a Context-Aware Block (CAB) is embedded in the bottleneck to reinforce multi-scale features in the decoder. RESULTS Experiments on four public datasets were conducted. On the BUSI dataset, our BiU-Net achieved 85.50 % on Dice coefficient (Dice), 76.73 % on intersection over union (IoU), and 97.23 % on accuracy (ACC). Compared to the state-of-the-art method, BiU-Net improves Dice by 1.17 %. For the Monuseg dataset, the proposed method attained the highest scores, reaching 80.27 % and 67.22 % for Dice and IoU. The BiU-Net achieves 95.33 % and 81.22 % Dice on the PH2 and DRIVE datasets. CONCLUSIONS The results of our experiments showed that BiU-Net transcends existing state-of-the-art methods on four publicly available biomedical datasets. Due to the powerful multi-scale feature extraction ability, our proposed BiU-Net is a versatile medical image segmentation framework for various types of medical images. The source code is released on (https://github.com/ZYLandy/BiU-Net).
Collapse
Affiliation(s)
- Zhiyong Huang
- School of Microelectronics and Communication Engineering, Chongqing University, Chongqing 400044, China.
| | - Yunlan Zhao
- School of Microelectronics and Communication Engineering, Chongqing University, Chongqing 400044, China
| | - Zhi Yu
- School of Microelectronics and Communication Engineering, Chongqing University, Chongqing 400044, China
| | - Pinzhong Qin
- School of Microelectronics and Communication Engineering, Chongqing University, Chongqing 400044, China
| | - Xiao Han
- School of Microelectronics and Communication Engineering, Chongqing University, Chongqing 400044, China
| | - Mengyao Wang
- School of Microelectronics and Communication Engineering, Chongqing University, Chongqing 400044, China
| | - Man Liu
- School of Microelectronics and Communication Engineering, Chongqing University, Chongqing 400044, China
| | - Hans Gregersen
- California Medical Innovations Institute, San Diego 92121, California
| |
Collapse
|
14
|
Yang Y, Yue S, Quan H. CS-UNet: Cross-scale U-Net with Semantic-position dependencies for retinal vessel segmentation. NETWORK (BRISTOL, ENGLAND) 2024; 35:134-153. [PMID: 38050997 DOI: 10.1080/0954898x.2023.2288858] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/21/2022] [Accepted: 11/23/2023] [Indexed: 12/07/2023]
Abstract
Accurate retinal vessel segmentation is the prerequisite for early recognition and treatment of retina-related diseases. However, segmenting retinal vessels is still challenging due to the intricate vessel tree in fundus images, which has a significant number of tiny vessels, low contrast, and lesion interference. For this task, the u-shaped architecture (U-Net) has become the de-facto standard and has achieved considerable success. However, U-Net is a pure convolutional network, which usually shows limitations in global modelling. In this paper, we propose a novel Cross-scale U-Net with Semantic-position Dependencies (CS-UNet) for retinal vessel segmentation. In particular, we first designed a Semantic-position Dependencies Aggregator (SPDA) and incorporate it into each layer of the encoder to better focus on global contextual information by integrating the relationship of semantic and position. To endow the model with the capability of cross-scale interaction, the Cross-scale Relation Refine Module (CSRR) is designed to dynamically select the information associated with the vessels, which helps guide the up-sampling operation. Finally, we have evaluated CS-UNet on three public datasets: DRIVE, CHASE_DB1, and STARE. Compared to most existing state-of-the-art methods, CS-UNet demonstrated better performance.
Collapse
Affiliation(s)
- Ying Yang
- College of Information Engineering and Automation, Kunming University of Science and Technology, Kunming, Yunnan, China
| | - Shengbin Yue
- College of Information Engineering and Automation, Kunming University of Science and Technology, Kunming, Yunnan, China
- Yunnan Provincial Key Laboratory of Artificial Intelligence, Kunming University of Science and Technology, Kunming, Yunnan, China
| | - Haiyan Quan
- College of Information Engineering and Automation, Kunming University of Science and Technology, Kunming, Yunnan, China
| |
Collapse
|
15
|
Poddar R, Shukla V, Alam Z, Mohan M. Automatic segmentation of layers in chorio-retinal complex using Graph-based method for ultra-speed 1.7 MHz wide field swept source FDML optical coherence tomography. Med Biol Eng Comput 2024; 62:1375-1393. [PMID: 38191981 DOI: 10.1007/s11517-023-03007-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/23/2023] [Accepted: 12/20/2023] [Indexed: 01/10/2024]
Abstract
The posterior segment of the human eye complex contains two discrete microstructure and vasculature network systems, namely, the retina and choroid. We present a single segmentation framework technique for segmenting the entire layers present in the chorio-retinal complex of the human eye using optical coherence tomography (OCT) images. This automatic program is based on the graph theory method. This single program is capable of segmenting seven layers of the retina and choroid scleral interface. The graph theory was utilized to find the probability matrix and subsequent boundaries of different layers. The program was also implemented to segment angiographic maps of different chorio-retinal layers using "segmentation matrices." The method was tested and successfully validated on OCT images from six normal human eyes as well as eyes with non-exudative age-related macular degeneration (AMD). The thickness of microstructure and microvasculature for different layers located in the chorio-retinal segment of the eye was also generated and compared. A decent efficiency in terms of processing time, sensitivity, and accuracy was observed compared to the manual segmentation and other existing methods. The proposed method automatically segments whole OCT images of chorio-retinal complex with augmented probability maps generation in OCT volume dataset. We have also evaluated the segmentation results using quantitative metrics such as Dice coefficient and Hausdorff distance This method realizes a mean descent Dice similarity coefficient (DSC) value of 0.82 (range, 0.816-0.864) for RPE and CSI layer.
Collapse
Affiliation(s)
- Raju Poddar
- Biophotonics Lab, Department of Bioengineering & Biotechnology, Birla Institute of Technology-Mesra, Ranchi, JH, 835 215, India.
| | - Vinita Shukla
- Biophotonics Lab, Department of Bioengineering & Biotechnology, Birla Institute of Technology-Mesra, Ranchi, JH, 835 215, India
| | - Zoya Alam
- Biophotonics Lab, Department of Bioengineering & Biotechnology, Birla Institute of Technology-Mesra, Ranchi, JH, 835 215, India
| | - Muktesh Mohan
- Biophotonics Lab, Department of Bioengineering & Biotechnology, Birla Institute of Technology-Mesra, Ranchi, JH, 835 215, India
| |
Collapse
|
16
|
Driban M, Yan A, Selvam A, Ong J, Vupparaboina KK, Chhablani J. Artificial intelligence in chorioretinal pathology through fundoscopy: a comprehensive review. Int J Retina Vitreous 2024; 10:36. [PMID: 38654344 PMCID: PMC11036694 DOI: 10.1186/s40942-024-00554-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2024] [Accepted: 04/02/2024] [Indexed: 04/25/2024] Open
Abstract
BACKGROUND Applications for artificial intelligence (AI) in ophthalmology are continually evolving. Fundoscopy is one of the oldest ocular imaging techniques but remains a mainstay in posterior segment imaging due to its prevalence, ease of use, and ongoing technological advancement. AI has been leveraged for fundoscopy to accomplish core tasks including segmentation, classification, and prediction. MAIN BODY In this article we provide a review of AI in fundoscopy applied to representative chorioretinal pathologies, including diabetic retinopathy and age-related macular degeneration, among others. We conclude with a discussion of future directions and current limitations. SHORT CONCLUSION As AI evolves, it will become increasingly essential for the modern ophthalmologist to understand its applications and limitations to improve patient outcomes and continue to innovate.
Collapse
Affiliation(s)
- Matthew Driban
- Department of Ophthalmology, University of Pittsburgh School of Medicine, Pittsburgh, PA, USA
| | - Audrey Yan
- Department of Medicine, West Virginia School of Osteopathic Medicine, Lewisburg, WV, USA
| | - Amrish Selvam
- Department of Ophthalmology, University of Pittsburgh School of Medicine, Pittsburgh, PA, USA
| | - Joshua Ong
- Michigan Medicine, University of Michigan, Ann Arbor, USA
| | | | - Jay Chhablani
- Department of Ophthalmology, University of Pittsburgh School of Medicine, Pittsburgh, PA, USA.
| |
Collapse
|
17
|
Guo H, Meng J, Zhao Y, Zhang H, Dai C. High-precision retinal blood vessel segmentation based on a multi-stage and dual-channel deep learning network. Phys Med Biol 2024; 69:045007. [PMID: 38198716 DOI: 10.1088/1361-6560/ad1cf6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/09/2023] [Accepted: 01/10/2024] [Indexed: 01/12/2024]
Abstract
Objective.The high-precision segmentation of retinal vessels in fundus images is important for the early diagnosis of ophthalmic diseases. However, the extraction for microvessels is challenging due to their characteristics of low contrast and high structural complexity. Although some works have been developed to improve the segmentation ability in thin vessels, they have only been successful in recognizing small vessels with relatively high contrast.Approach.Therefore, we develop a deep learning (DL) framework with a multi-stage and dual-channel network model (MSDC_NET) to further improve the thin-vessel segmentation with low contrast. Specifically, an adaptive image enhancement strategy combining multiple preprocessing and the DL method is firstly proposed to elevate the contrast of thin vessels; then, a two-channel model with multi-scale perception is developed to implement whole- and thin-vessel segmentation; and finally, a series of post-processing operations are designed to extract more small vessels in the predicted maps from thin-vessel channels.Main results.Experiments on DRIVE, STARE and CHASE_DB1 demonstrate the superiorities of the proposed MSDC_NET in extracting more thin vessels in fundus images, and quantitative evaluations on several parameters based on the advanced ground truth further verify the advantages of our proposed DL model. Compared with the previous multi-branch method, the specificity and F1score are improved by about 2.18%, 0.68%, 1.73% and 2.91%, 0.24%, 8.38% on the three datasets, respectively.Significance.This work may provide richer information to ophthalmologists for the diagnosis and treatment of vascular-related ophthalmic diseases.
Collapse
Affiliation(s)
- Hui Guo
- School of Computer, Qufu Normal University, 276826 Rizhao, People's Republic of China
| | - Jing Meng
- School of Computer, Qufu Normal University, 276826 Rizhao, People's Republic of China
| | - Yongfu Zhao
- School of Computer, Qufu Normal University, 276826 Rizhao, People's Republic of China
| | - Hongdong Zhang
- School of Computer, Qufu Normal University, 276826 Rizhao, People's Republic of China
| | - Cuixia Dai
- College of Science, Shanghai Institute of Technology, 201418 Shanghai, People's Republic of China
| |
Collapse
|
18
|
Zheng C, Li H, Ge Y, He Y, Yi Y, Zhu M, Sun H, Kong J. Retinal vessel segmentation based on multi-scale feature and style transfer. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2024; 21:49-74. [PMID: 38303413 DOI: 10.3934/mbe.2024003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/03/2024]
Abstract
Retinal vessel segmentation is very important for diagnosing and treating certain eye diseases. Recently, many deep learning-based retinal vessel segmentation methods have been proposed; however, there are still many shortcomings (e.g., they cannot obtain satisfactory results when dealing with cross-domain data or segmenting small blood vessels). To alleviate these problems and avoid overly complex models, we propose a novel network based on a multi-scale feature and style transfer (MSFST-NET) for retinal vessel segmentation. Specifically, we first construct a lightweight segmentation module named MSF-Net, which introduces the selective kernel (SK) module to increase the multi-scale feature extraction ability of the model to achieve improved small blood vessel segmentation. Then, to alleviate the problem of model performance degradation when segmenting cross-domain datasets, we propose a style transfer module and a pseudo-label learning strategy. The style transfer module is used to reduce the style difference between the source domain image and the target domain image to improve the segmentation performance for the target domain image. The pseudo-label learning strategy is designed to be combined with the style transfer module to further boost the generalization ability of the model. Moreover, we trained and tested our proposed MSFST-NET in experiments on the DRIVE and CHASE_DB1 datasets. The experimental results demonstrate that MSFST-NET can effectively improve the generalization ability of the model on cross-domain datasets and achieve improved retinal vessel segmentation results than other state-of-the-art methods.
Collapse
Affiliation(s)
- Caixia Zheng
- Jilin Animation Institute, Changchun 130013, China
- College of Information Science and Technology, Northeast Normal University, Changchun 130117, China
| | - Huican Li
- College of Information Science and Technology, Northeast Normal University, Changchun 130117, China
| | - Yingying Ge
- Jilin Animation Institute, Changchun 130013, China
| | - Yanlin He
- College of Information Science and Technology, Northeast Normal University, Changchun 130117, China
| | - Yugen Yi
- School of Software, Jiangxi Normal University, Nanchang 330022, China
| | - Meili Zhu
- Jilin Animation Institute, Changchun 130013, China
| | - Hui Sun
- School of Science and Technology, Changchun Humanities and Sciences College, Changchun 130117, China
| | - Jun Kong
- College of Information Science and Technology, Northeast Normal University, Changchun 130117, China
| |
Collapse
|
19
|
Ma Z, Li X. An improved supervised and attention mechanism-based U-Net algorithm for retinal vessel segmentation. Comput Biol Med 2024; 168:107770. [PMID: 38056215 DOI: 10.1016/j.compbiomed.2023.107770] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2023] [Revised: 11/08/2023] [Accepted: 11/26/2023] [Indexed: 12/08/2023]
Abstract
The segmentation results of retinal blood vessels are crucial for automatically diagnosing ophthalmic diseases such as diabetic retinopathy, hypertension, cardiovascular and cerebrovascular diseases. To improve the accuracy of vessel segmentation and better extract information about small vessels and edges, we introduce the U-Net algorithm with a supervised attention mechanism for retinal vessel segmentation. We achieve this by introducing a decoder fusion module (DFM) in the encoding part, effectively combining different convolutional blocks to extract features comprehensively. Additionally, in the decoding part of U-Net, we propose the context squeeze and excitation (CSE) decoding module to enhance important contextual feature information and the detection of tiny blood vessels. For the final output, we introduce the supervised fusion mechanism (SFM), which combines multiple branches from shallow to deep layers, effectively fusing multi-scale features and capturing information from different levels, fully integrating low-level and high-level features to improve segmentation performance. Our experimental results on the public datasets of DRIVE, STARE, and CHASED_B1 demonstrate the excellent performance of our proposed network.
Collapse
Affiliation(s)
- Zhendi Ma
- School of Computer Science and Technology, Zhejiang Normal University, Jinhua 321004, China
| | - Xiaobo Li
- School of Computer Science and Technology, Zhejiang Normal University, Jinhua 321004, China.
| |
Collapse
|
20
|
Li C, Li Z, Liu W. TDCAU-Net: retinal vessel segmentation using transformer dilated convolutional attention-based U-Net method. Phys Med Biol 2023; 69:015003. [PMID: 38052089 DOI: 10.1088/1361-6560/ad1273] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2023] [Accepted: 12/05/2023] [Indexed: 12/07/2023]
Abstract
Retinal vessel segmentation plays a vital role in the medical field, facilitating the identification of numerous chronic conditions based on retinal vessel images. These conditions include diabetic retinopathy, hypertensive retinopathy, glaucoma, and others. Although the U-Net model has shown promising results in retinal vessel segmentation, it tends to struggle with fine branching and dense vessel segmentation. To further enhance the precision of retinal vessel segmentation, we propose a novel approach called transformer dilated convolution attention U-Net (TDCAU-Net), which builds upon the U-Net architecture with improved Transformer-based dilated convolution attention mechanisms. The proposed model retains the three-layer architecture of the U-Net network. The Transformer component enables the learning of contextual information for each pixel in the image, while the dilated convolution attention prevents information loss. The algorithm efficiently addresses several challenges to optimize blood vessel detection. The process starts with five-step preprocessing of the images, followed by chunking them into segments. Subsequently, the retinal images are fed into the modified U-Net network introduced in this paper for segmentation. The study employs eye fundus images from the DRIVE and CHASEDB1 databases for both training and testing purposes. Evaluation metrics are utilized to compare the algorithm's results with state-of-the-art methods. The experimental analysis on both databases demonstrates that the algorithm achieves high values of sensitivity, specificity, accuracy, and AUC. Specifically, for the first database, the achieved values are 0.8187, 0.9756, 0.9556, and 0.9795, respectively. For the second database, the corresponding values are 0.8243, 0.9836, 0.9738, and 0.9878, respectively. These results demonstrate that the proposed approach outperforms state-of-the-art methods, achieving higher performance on both datasets. The TDCAU-Net model presented in this study exhibits substantial capabilities in accurately segmenting fine branching and dense vessels. The segmentation performance of the network surpasses that of the U-Net algorithm and several mainstream methods.
Collapse
Affiliation(s)
- Chunyang Li
- School of Electronics and Information Engineering, University of Science and Technology Liaoning, Anshan, People's Republic of China
| | - Zhigang Li
- School of Electronics and Information Engineering, University of Science and Technology Liaoning, Anshan, People's Republic of China
| | - Weikang Liu
- School of Electronics and Information Engineering, University of Science and Technology Liaoning, Anshan, People's Republic of China
| |
Collapse
|
21
|
Zhu YF, Xu X, Zhang XD, Jiang MS. CCS-UNet: a cross-channel spatial attention model for accurate retinal vessel segmentation. BIOMEDICAL OPTICS EXPRESS 2023; 14:4739-4758. [PMID: 37791275 PMCID: PMC10545190 DOI: 10.1364/boe.495766] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/18/2023] [Revised: 07/14/2023] [Accepted: 08/09/2023] [Indexed: 10/05/2023]
Abstract
Precise segmentation of retinal vessels plays an important role in computer-assisted diagnosis. Deep learning models have been applied to retinal vessel segmentation, but the efficacy is limited by the significant scale variation of vascular structures and the intricate background of retinal images. This paper supposes a cross-channel spatial attention U-Net (CCS-UNet) for accurate retinal vessel segmentation. In comparison to other models based on U-Net, our model employes a ResNeSt block for the encoder-decoder architecture. The block has a multi-branch structure that enables the model to extract more diverse vascular features. It facilitates weight distribution across channels through the incorporation of soft attention, which effectively aggregates contextual information in vascular images. Furthermore, we suppose an attention mechanism within the skip connection. This mechanism serves to enhance feature integration across various layers, thereby mitigating the degradation of effective information. It helps acquire cross-channel information and enhance the localization of regions of interest, ultimately leading to improved recognition of vascular structures. In addition, the feature fusion module (FFM) module is used to provide semantic information for a more refined vascular segmentation map. We evaluated CCS-UNet based on five benchmark retinal image datasets, DRIVE, CHASEDB1, STARE, IOSTAR and HRF. Our proposed method exhibits superior segmentation efficacy compared to other state-of-the-art techniques with a global accuracy of 0.9617/0.9806/0.9766/0.9786/0.9834 and AUC of 0.9863/0.9894/0.9938/0.9902/0.9855 on DRIVE, CHASEDB1, STARE, IOSTAR and HRF respectively. Ablation studies are also performed to evaluate the the relative contributions of different architectural components. Our proposed model is potential for diagnostic aid of retinal diseases.
Collapse
Affiliation(s)
| | | | - Xue-dian Zhang
- Shanghai Key Laboratory of Contemporary Optics System, College of Optical-Electrical and Computer Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China
| | - Min-shan Jiang
- Shanghai Key Laboratory of Contemporary Optics System, College of Optical-Electrical and Computer Engineering, University of Shanghai for Science and Technology, Shanghai 200093, China
| |
Collapse
|
22
|
Sun Q, Yang J, Ma S, Huang Y, Yuan Y, Hou Y. 3D vessel extraction using a scale-adaptive hybrid parametric tracker. Med Biol Eng Comput 2023; 61:2467-2480. [PMID: 37184591 DOI: 10.1007/s11517-023-02815-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/11/2022] [Accepted: 02/28/2023] [Indexed: 05/16/2023]
Abstract
3D vessel extraction has great significance in the diagnosis of vascular diseases. However, accurate extraction of vessels from computed tomography angiography (CTA) data is challenging. For one thing, vessels in different body parts have a wide range of scales and large curvatures; for another, the intensity distributions of vessels in different CTA data vary considerably. Besides, surrounding interfering tissue, like bones or veins with similar intensity, also seriously affects vessel extraction. Considering all the above imaging and structural features of vessels, we propose a new scale-adaptive hybrid parametric tracker (SAHPT) to extract arbitrary vessels of different body parts. First, a geometry-intensity parametric model is constructed to calculate the geometry-intensity response. While geometry parameters are calculated to adapt to the variation in scale, intensity parameters can also be estimated to meet non-uniform intensity distributions. Then, a gradient parametric model is proposed to calculate the gradient response based on a multiscale symmetric normalized gradient filter which can effectively separate the target vessel from surrounding interfering tissue. Last, a hybrid parametric model that combines the geometry-intensity and gradient parametric models is constructed to evaluate how well it fits a local image patch. In the extraction process, a multipath spherical sampling strategy is used to solve the problem of anatomical complexity. We have conducted many quantitative experiments using the synthetic and clinical CTA data, asserting its superior performance compared to traditional or deep learning-based baselines.
Collapse
Affiliation(s)
- Qi Sun
- Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, Liaoning, China
- School of Computer Science and Engineering, Northeastern University, Shenyang, Liaoning, China
| | - Jinzhu Yang
- Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, Liaoning, China.
- School of Computer Science and Engineering, Northeastern University, Shenyang, Liaoning, China.
| | - Shuang Ma
- Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, Liaoning, China
- School of Computer Science and Engineering, Northeastern University, Shenyang, Liaoning, China
| | - Yan Huang
- Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, Liaoning, China
- School of Computer Science and Engineering, Northeastern University, Shenyang, Liaoning, China
| | - Yuliang Yuan
- Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Northeastern University, Shenyang, Liaoning, China
- School of Computer Science and Engineering, Northeastern University, Shenyang, Liaoning, China
| | - Yang Hou
- Department of Radiology, ShengJing Hospital of China Medical University, Shenyang, Liaoning, China
| |
Collapse
|
23
|
Li Y, Zhang Y, Liu JY, Wang K, Zhang K, Zhang GS, Liao XF, Yang G. Global Transformer and Dual Local Attention Network via Deep-Shallow Hierarchical Feature Fusion for Retinal Vessel Segmentation. IEEE TRANSACTIONS ON CYBERNETICS 2023; 53:5826-5839. [PMID: 35984806 DOI: 10.1109/tcyb.2022.3194099] [Citation(s) in RCA: 12] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Clinically, retinal vessel segmentation is a significant step in the diagnosis of fundus diseases. However, recent methods generally neglect the difference of semantic information between deep and shallow features, which fail to capture the global and local characterizations in fundus images simultaneously, resulting in the limited segmentation performance for fine vessels. In this article, a global transformer (GT) and dual local attention (DLA) network via deep-shallow hierarchical feature fusion (GT-DLA-dsHFF) are investigated to solve the above limitations. First, the GT is developed to integrate the global information in the retinal image, which effectively captures the long-distance dependence between pixels, alleviating the discontinuity of blood vessels in the segmentation results. Second, DLA, which is constructed using dilated convolutions with varied dilation rates, unsupervised edge detection, and squeeze-excitation block, is proposed to extract local vessel information, consolidating the edge details in the segmentation result. Finally, a novel deep-shallow hierarchical feature fusion (dsHFF) algorithm is studied to fuse the features in different scales in the deep learning framework, respectively, which can mitigate the attenuation of valid information in the process of feature fusion. We verified the GT-DLA-dsHFF on four typical fundus image datasets. The experimental results demonstrate our GT-DLA-dsHFF achieves superior performance against the current methods and detailed discussions verify the efficacy of the proposed three modules. Segmentation results of diseased images show the robustness of our proposed GT-DLA-dsHFF. Implementation codes will be available on https://github.com/YangLibuaa/GT-DLA-dsHFF.
Collapse
|
24
|
Gao Y, Ma C, Guo L, Zhang X, Ji X. MIL-CT: Multiple Instance Learning via a Cross-Scale Transformer for Enhanced Arterial Light Reflex Detection. Bioengineering (Basel) 2023; 10:971. [PMID: 37627856 PMCID: PMC10451897 DOI: 10.3390/bioengineering10080971] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/06/2023] [Revised: 08/03/2023] [Accepted: 08/04/2023] [Indexed: 08/27/2023] Open
Abstract
One of the early manifestations of systemic atherosclerosis, which leads to blood circulation issues, is the enhanced arterial light reflex (EALR). Fundus images are commonly used for regular screening purposes to intervene and assess the severity of systemic atherosclerosis in a timely manner. However, there is a lack of automated methods that can meet the demands of large-scale population screening. Therefore, this study introduces a novel cross-scale transformer-based multi-instance learning method, named MIL-CT, for the detection of early arterial lesions (e.g., EALR) in fundus images. MIL-CT utilizes the cross-scale vision transformer to extract retinal features in a multi-granularity perceptual domain. It incorporates a multi-head cross-scale attention fusion module to enhance global perceptual capability and feature representation. By integrating information from different scales and minimizing information loss, the method significantly improves the performance of the EALR detection task. Furthermore, a multi-instance learning module is implemented to enable the model to better comprehend local details and features in fundus images, facilitating the classification of patch tokens related to retinal lesions. To effectively learn the features associated with retinal lesions, we utilize weights pre-trained on a large fundus image Kaggle dataset. Our validation and comparison experiments conducted on our collected EALR dataset demonstrate the effectiveness of the MIL-CT method in reducing generalization errors while maintaining efficient attention to retinal vascular details. Moreover, the method surpasses existing models in EALR detection, achieving an accuracy, precision, sensitivity, specificity, and F1 score of 97.62%, 97.63%, 97.05%, 96.48%, and 97.62%, respectively. These results exhibit the significant enhancement in diagnostic accuracy of fundus images brought about by the MIL-CT method. Thus, it holds potential for various applications, particularly in the early screening of cardiovascular diseases such as hypertension and atherosclerosis.
Collapse
Affiliation(s)
- Yuan Gao
- Department of Biomedical Engineering, School of Biological Science and Medical Engineering, Beihang University, Beijing 100191, China
- Department of Ophthalmology, Xuanwu Hospital, Capital Medical University, Beijing 100053, China
| | - Chenbin Ma
- Department of Biomedical Engineering, School of Biological Science and Medical Engineering, Beihang University, Beijing 100191, China
- Shen Yuan Honors College, Beihang University, Beijing 100191, China
| | - Lishuang Guo
- Department of Biomedical Engineering, School of Biological Science and Medical Engineering, Beihang University, Beijing 100191, China
| | - Xuxiang Zhang
- Department of Ophthalmology, Beijing Tiantan Hospital, Capital Medical University, Beijing 100050, China
| | - Xunming Ji
- Department of Biomedical Engineering, School of Biological Science and Medical Engineering, Beihang University, Beijing 100191, China
| |
Collapse
|
25
|
Shi T, Ding X, Zhou W, Pan F, Yan Z, Bai X, Yang X. Affinity Feature Strengthening for Accurate, Complete and Robust Vessel Segmentation. IEEE J Biomed Health Inform 2023; 27:4006-4017. [PMID: 37163397 DOI: 10.1109/jbhi.2023.3274789] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/12/2023]
Abstract
Vessel segmentation is crucial in many medical image applications, such as detecting coronary stenoses, retinal vessel diseases and brain aneurysms. However, achieving high pixel-wise accuracy, complete topology structure and robustness to various contrast variations are critical and challenging, and most existing methods focus only on achieving one or two of these aspects. In this paper, we present a novel approach, the affinity feature strengthening network (AFN), which jointly models geometry and refines pixel-wise segmentation features using a contrast-insensitive, multiscale affinity approach. Specifically, we compute a multiscale affinity field for each pixel, capturing its semantic relationships with neighboring pixels in the predicted mask image. This field represents the local geometry of vessel segments of different sizes, allowing us to learn spatial- and scale-aware adaptive weights to strengthen vessel features. We evaluate our AFN on four different types of vascular datasets: X-ray angiography coronary vessel dataset (XCAD), portal vein dataset (PV), digital subtraction angiography cerebrovascular vessel dataset (DSA) and retinal vessel dataset (DRIVE). Extensive experimental results demonstrate that our AFN outperforms the state-of-the-art methods in terms of both higher accuracy and topological metrics, while also being more robust to various contrast changes.
Collapse
|
26
|
Zhang H, Qiu Y, Song C, Li J. Landmark-Assisted Anatomy-Sensitive Retinal Vessel Segmentation Network. Diagnostics (Basel) 2023; 13:2260. [PMID: 37443654 DOI: 10.3390/diagnostics13132260] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/31/2023] [Revised: 06/28/2023] [Accepted: 06/30/2023] [Indexed: 07/15/2023] Open
Abstract
Automatic retinal vessel segmentation is important for assisting clinicians in diagnosing ophthalmic diseases. The existing deep learning methods remain constrained in instance connectivity and thin vessel detection. To this end, we propose a novel anatomy-sensitive retinal vessel segmentation framework to preserve instance connectivity and improve the segmentation accuracy of thin vessels. This framework uses TransUNet as its backbone and utilizes self-supervised extracted landmarks to guide network learning. TransUNet is designed to simultaneously benefit from the advantages of convolutional and multi-head attention mechanisms in extracting local features and modeling global dependencies. In particular, we introduce contrastive learning-based self-supervised extraction anatomical landmarks to guide the model to focus on learning the morphological information of retinal vessels. We evaluated the proposed method on three public datasets: DRIVE, CHASE-DB1, and STARE. Our method demonstrates promising results on the DRIVE and CHASE-DB1 datasets, outperforming state-of-the-art methods by improving the F1 scores by 0.36% and 0.31%, respectively. On the STARE dataset, our method achieves results close to the best-performing methods. Visualizations of the results highlight the potential of our method in maintaining topological continuity and identifying thin blood vessels. Furthermore, we conducted a series of ablation experiments to validate the effectiveness of each module in our model and considered the impact of image resolution on the results.
Collapse
Affiliation(s)
- Haifeng Zhang
- College of Information Science and Engineering, Northeastern University, Shenyang 110819, China
| | - Yunlong Qiu
- College of Information Science and Engineering, Northeastern University, Shenyang 110819, China
| | - Chonghui Song
- College of Information Science and Engineering, Northeastern University, Shenyang 110819, China
| | - Jiale Li
- College of Information Science and Engineering, Northeastern University, Shenyang 110819, China
| |
Collapse
|
27
|
Lin X, Yu L, Cheng KT, Yan Z. BATFormer: Towards Boundary-Aware Lightweight Transformer for Efficient Medical Image Segmentation. IEEE J Biomed Health Inform 2023; 27:3501-3512. [PMID: 37053058 DOI: 10.1109/jbhi.2023.3266977] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/14/2023]
Abstract
OBJECTIVE Transformers, born to remedy the inadequate receptive fields of CNNs, have drawn explosive attention recently. However, the daunting computational complexity of global representation learning, together with rigid window partitioning, hinders their deployment in medical image segmentation. This work aims to address the above two issues in transformers for better medical image segmentation. METHODS We propose a boundary-aware lightweight transformer (BATFormer) that can build cross-scale global interaction with lower computational complexity and generate windows flexibly under the guidance of entropy. Specifically, to fully explore the benefits of transformers in long-range dependency establishment, a cross-scale global transformer (CGT) module is introduced to jointly utilize multiple small-scale feature maps for richer global features with lower computational complexity. Given the importance of shape modeling in medical image segmentation, a boundary-aware local transformer (BLT) module is constructed. Different from rigid window partitioning in vanilla transformers which would produce boundary distortion, BLT adopts an adaptive window partitioning scheme under the guidance of entropy for both computational complexity reduction and shape preservation. RESULTS BATFormer achieves the best performance in Dice of 92.84 %, 91.97 %, 90.26 %, and 96.30 % for the average, right ventricle, myocardium, and left ventricle respectively on the ACDC dataset and the best performance in Dice, IoU, and ACC of 90.76 %, 84.64 %, and 96.76 % respectively on the ISIC 2018 dataset. More importantly, BATFormer requires the least amount of model parameters and the lowest computational complexity compared to the state-of-the-art approaches. CONCLUSION AND SIGNIFICANCE Our results demonstrate the necessity of developing customized transformers for efficient and better medical image segmentation. We believe the design of BATFormer is inspiring and extendable to other applications/frameworks.
Collapse
|
28
|
Tan Y, Zhao SX, Yang KF, Li YJ. A lightweight network guided with differential matched filtering for retinal vessel segmentation. Comput Biol Med 2023; 160:106924. [PMID: 37146492 DOI: 10.1016/j.compbiomed.2023.106924] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/18/2022] [Revised: 04/03/2023] [Accepted: 04/13/2023] [Indexed: 05/07/2023]
Abstract
The geometric morphology of retinal vessels reflects the state of cardiovascular health, and fundus images are important reference materials for ophthalmologists. Great progress has been made in automated vessel segmentation, but few studies have focused on thin vessel breakage and false-positives in areas with lesions or low contrast. In this work, we propose a new network, differential matched filtering guided attention UNet (DMF-AU), to address these issues, incorporating a differential matched filtering layer, feature anisotropic attention, and a multiscale consistency constrained backbone to perform thin vessel segmentation. The differential matched filtering is used for the early identification of locally linear vessels, and the resulting rough vessel map guides the backbone to learn vascular details. Feature anisotropic attention reinforces the vessel features of spatial linearity at each stage of the model. Multiscale constraints reduce the loss of vessel information while pooling within large receptive fields. In tests on multiple classical datasets, the proposed model performed well compared with other algorithms on several specially designed criteria for vessel segmentation. DMF-AU is a high-performance, lightweight vessel segmentation model. The source code is at https://github.com/tyb311/DMF-AU.
Collapse
Affiliation(s)
- Yubo Tan
- The MOE Key Laboratory for Neuroinformation, Radiation Oncology Key Laboratory of Sichuan Province, University of Electronic Science and Technology of China, China.
| | - Shi-Xuan Zhao
- The MOE Key Laboratory for Neuroinformation, Radiation Oncology Key Laboratory of Sichuan Province, University of Electronic Science and Technology of China, China.
| | - Kai-Fu Yang
- The MOE Key Laboratory for Neuroinformation, Radiation Oncology Key Laboratory of Sichuan Province, University of Electronic Science and Technology of China, China.
| | - Yong-Jie Li
- The MOE Key Laboratory for Neuroinformation, Radiation Oncology Key Laboratory of Sichuan Province, University of Electronic Science and Technology of China, China.
| |
Collapse
|
29
|
Zhang H, Zhong X, Li G, Liu W, Liu J, Ji D, Li X, Wu J. BCU-Net: Bridging ConvNeXt and U-Net for medical image segmentation. Comput Biol Med 2023; 159:106960. [PMID: 37099973 DOI: 10.1016/j.compbiomed.2023.106960] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/18/2022] [Revised: 04/12/2023] [Accepted: 04/17/2023] [Indexed: 04/28/2023]
Abstract
Medical image segmentation enables doctors to observe lesion regions better and make accurate diagnostic decisions. Single-branch models such as U-Net have achieved great progress in this field. However, the complementary local and global pathological semantics of heterogeneous neural networks have not yet been fully explored. The class-imbalance problem remains a serious issue. To alleviate these two problems, we propose a novel model called BCU-Net, which leverages the advantages of ConvNeXt in global interaction and U-Net in local processing. We propose a new multilabel recall loss (MRL) module to relieve the class imbalance problem and facilitate deep-level fusion of local and global pathological semantics between the two heterogeneous branches. Extensive experiments were conducted on six medical image datasets including retinal vessel and polyp images. The qualitative and quantitative results demonstrate the superiority and generalizability of BCU-Net. In particular, BCU-Net can handle diverse medical images with diverse resolutions. It has a flexible structure owing to its plug-and-play characteristics, which promotes its practicality.
Collapse
Affiliation(s)
- Hongbin Zhang
- School of Software, East China Jiaotong University, China.
| | - Xiang Zhong
- School of Software, East China Jiaotong University, China.
| | - Guangli Li
- School of Information Engineering, East China Jiaotong University, China.
| | - Wei Liu
- School of Software, East China Jiaotong University, China.
| | - Jiawei Liu
- School of Software, East China Jiaotong University, China.
| | - Donghong Ji
- School of Cyber Science and Engineering, Wuhan University, China.
| | - Xiong Li
- School of Software, East China Jiaotong University, China.
| | - Jianguo Wu
- The Second Affiliated Hospital of Nanchang University, China.
| |
Collapse
|
30
|
Qu Z, Zhuo L, Cao J, Li X, Yin H, Wang Z. TP-Net: Two-Path Network for Retinal Vessel Segmentation. IEEE J Biomed Health Inform 2023; 27:1979-1990. [PMID: 37021912 DOI: 10.1109/jbhi.2023.3237704] [Citation(s) in RCA: 9] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/19/2023]
Abstract
Refined and automatic retinal vessel segmentation is crucial for computer-aided early diagnosis of retinopathy. However, existing methods often suffer from mis-segmentation when dealing with thin and low-contrast vessels. In this paper, a two-path retinal vessel segmentation network is proposed, namely TP-Net, which consists of three core parts, i.e., main-path, sub-path, and multi-scale feature aggregation module (MFAM). Main-path is to detect the trunk area of the retinal vessels, and the sub-path to effectively capture edge information of the retinal vessels. The prediction results of the two paths are combined by MFAM, obtaining refined segmentation of retinal vessels. In the main-path, a three-layer lightweight backbone network is elaborately designed according to the characteristics of retinal vessels, and then a global feature selection mechanism (GFSM) is proposed, which can autonomously select features that are more important for the segmentation task from the features at different layers of the network, thereby, enhancing the segmentation capability for low-contrast vessels. In the sub-path, an edge feature extraction method and an edge loss function are proposed, which can enhance the ability of the network to capture edge information and reduce the mis-segmentation of thin vessels. Finally, MFAM is proposed to fuse the prediction results of main-path and sub-path, which can remove background noises while preserving edge details, and thus, obtaining refined segmentation of retinal vessels. The proposed TP-Net has been evaluated on three public retinal vessel datasets, namely DRIVE, STARE, and CHASE DB1. The experimental results show that the TP-Net achieved a superior performance and generalization ability with fewer model parameters compared with the state-of-the-art methods.
Collapse
|
31
|
Computational intelligence in eye disease diagnosis: a comparative study. Med Biol Eng Comput 2023; 61:593-615. [PMID: 36595155 DOI: 10.1007/s11517-022-02737-3] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/21/2021] [Accepted: 12/09/2022] [Indexed: 01/04/2023]
Abstract
In recent years, eye disorders are an important health issue among older people. Generally, individuals with eye diseases are unaware of the gradual growth of symptoms. Therefore, routine eye examinations are required for early diagnosis. Usually, eye disorders are identified by an ophthalmologist via a slit-lamp investigation. Slit-lamp interpretations are inadequate due to the differences in the analytical skills of the ophthalmologist, inconsistency in eye disorder analysis, and record maintenance issues. Therefore, digital images of an eye and computational intelligence (CI)-based approaches are preferred as assistive methods for eye disease diagnosis. A comparative study of CI-based decision support models for eye disorder diagnosis is presented in this paper. The CI-based decision support systems used for eye abnormalities diagnosis were grouped as anterior and retinal eye abnormalities diagnostic systems, and numerous algorithms used for diagnosing the eye abnormalities were also briefed. Various eye imaging modalities, pre-processing methods such as reflection removal, contrast enhancement, region of interest segmentation methods, and public eye image databases used for CI-based eye disease diagnosis system development were also discussed in this paper. In this comparative study, the reliability of various CI-based systems used for anterior eye and retinal disorder diagnosis was compared based on the precision, sensitivity, and specificity in eye disease diagnosis. The outcomes of the comparative analysis indicate that the CI-based anterior and retinal disease diagnosis systems attained significant prediction accuracy. Hence, these CI-based diagnosis systems can be used in clinics to reduce the burden on physicians, minimize fatigue-related misdetection, and take precise clinical decisions.
Collapse
|
32
|
GDF-Net: A multi-task symmetrical network for retinal vessel segmentation. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104426] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/02/2022]
|
33
|
Arsalan M, Khan TM, Naqvi SS, Nawaz M, Razzak I. Prompt Deep Light-Weight Vessel Segmentation Network (PLVS-Net). IEEE/ACM TRANSACTIONS ON COMPUTATIONAL BIOLOGY AND BIOINFORMATICS 2023; 20:1363-1371. [PMID: 36194721 DOI: 10.1109/tcbb.2022.3211936] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/04/2023]
Abstract
Achieving accurate retinal vessel segmentation is critical in the progression and diagnosis of vision-threatening diseases such as diabetic retinopathy and age-related macular degeneration. Existing vessel segmentation methods are based on encoder-decoder architectures, which frequently fail to take into account the retinal vessel structure's context in their analysis. As a result, such methods have difficulty bridging the semantic gap between encoder and decoder characteristics. This paper proposes a Prompt Deep Light-weight Vessel Segmentation Network (PLVS-Net) to address these issues by using prompt blocks. Each prompt block use combination of asymmetric kernel convolutions, depth-wise separable convolutions, and ordinary convolutions to extract useful features. This novel strategy improves the performance of the segmentation network while simultaneously decreasing the number of trainable parameters. Our method outperformed competing approaches in the literature on three benchmark datasets, including DRIVE, STARE, and CHASE.
Collapse
|
34
|
Wang J, Zhou L, Yuan Z, Wang H, Shi C. MIC-Net: multi-scale integrated context network for automatic retinal vessel segmentation in fundus image. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2023; 20:6912-6931. [PMID: 37161134 DOI: 10.3934/mbe.2023298] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/11/2023]
Abstract
PURPOSE Accurate retinal vessel segmentation is of great value in the auxiliary screening of various diseases. However, due to the low contrast between the ends of the branches of the fundus blood vessels and the background, and the variable morphology of the optic disc and cup in the retinal image, the task of high-precision retinal blood vessel segmentation still faces difficulties. METHOD This paper proposes a multi-scale integrated context network, MIC-Net, which fully fuses the encoder-decoder features, and extracts multi-scale information. First, a hybrid stride sampling (HSS) block was designed in the encoder to minimize the loss of helpful information caused by the downsampling operation. Second, a dense hybrid dilated convolution (DHDC) was employed in the connection layer. On the premise of preserving feature resolution, it can perceive richer contextual information. Third, a squeeze-and-excitation with residual connections (SERC) was introduced in the decoder to adjust the channel attention adaptively. Finally, we utilized a multi-layer feature fusion mechanism in the skip connection part, which enables the network to consider both low-level details and high-level semantic information. RESULTS We evaluated the proposed method on three public datasets DRIVE, STARE and CHASE. In the experimental results, the Area under the receiver operating characteristic (ROC) and the accuracy rate (Acc) achieved high performances of 98.62%/97.02%, 98.60%/97.76% and 98.73%/97.38%, respectively. CONCLUSIONS Experimental results show that the proposed method can obtain comparable segmentation performance compared with the state-of-the-art (SOTA) methods. Specifically, the proposed method can effectively reduce the small blood vessel segmentation error, thus proving it a promising tool for auxiliary diagnosis of ophthalmic diseases.
Collapse
Affiliation(s)
- Jinke Wang
- Department of Software Engineering, Harbin University of Science and Technology, Rongcheng 264300, China
- School of Automation, Harbin University of Science and Technology, Harbin 150080, China
| | - Lubiao Zhou
- School of Automation, Harbin University of Science and Technology, Harbin 150080, China
| | - Zhongzheng Yuan
- Department of Software Engineering, Harbin University of Science and Technology, Rongcheng 264300, China
| | - Haiying Wang
- School of Automation, Harbin University of Science and Technology, Harbin 150080, China
| | - Changfa Shi
- Mobile E-business Collaborative Innovation Center of Hunan Province, Hunan University of Technology and Business, Changsha 410205, China
| |
Collapse
|
35
|
Retinal blood vessel segmentation by using the MS-LSDNet network and geometric skeleton reconnection method. Comput Biol Med 2023; 153:106416. [PMID: 36586230 DOI: 10.1016/j.compbiomed.2022.106416] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2022] [Revised: 11/21/2022] [Accepted: 12/04/2022] [Indexed: 12/29/2022]
Abstract
Automatic retinal blood vessel segmentation is a key link in the diagnosis of ophthalmic diseases. Recent deep learning methods have achieved high accuracy in vessel segmentation but still face challenges in maintaining vascular structural connectivity. Therefore, this paper proposes a novel retinal blood vessel segmentation strategy that includes three stages: vessel structure detection, vessel branch extraction and broken vessel segment reconnection. First, we propose a multiscale linear structure detection network (MS-LSDNet), which improves the detection ability of fine blood vessels by learning the types of rich hierarchical features. In addition, to maintain the connectivity of the vascular structure in the process of binarization of the vascular probability map, an adaptive hysteresis threshold method for vascular extraction is proposed. Finally, we propose a vascular tree structure reconstruction algorithm based on a geometric skeleton to connect the broken vessel segments. Experimental results on three publicly available datasets show that compared with current state-of-the-art algorithms, our strategy effectively maintains the connectivity of retinal vascular tree structure.
Collapse
|
36
|
Imran SMA, Saleem MW, Hameed MT, Hussain A, Naqvi RA, Lee SW. Feature preserving mesh network for semantic segmentation of retinal vasculature to support ophthalmic disease analysis. Front Med (Lausanne) 2023; 9:1040562. [PMID: 36714120 PMCID: PMC9880050 DOI: 10.3389/fmed.2022.1040562] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/09/2022] [Accepted: 12/20/2022] [Indexed: 01/14/2023] Open
Abstract
Introduction Ophthalmic diseases are approaching an alarming count across the globe. Typically, ophthalmologists depend on manual methods for the analysis of different ophthalmic diseases such as glaucoma, Sickle cell retinopathy (SCR), diabetic retinopathy, and hypertensive retinopathy. All these manual assessments are not reliable, time-consuming, tedious, and prone to error. Therefore, automatic methods are desirable to replace conventional approaches. The accuracy of this segmentation of these vessels using automated approaches directly depends on the quality of fundus images. Retinal vessels are assumed as a potential biomarker for the diagnosis of many ophthalmic diseases. Mostly newly developed ophthalmic diseases contain minor changes in vasculature which is a critical job for the early detection and analysis of disease. Method Several artificial intelligence-based methods suggested intelligent solutions for automated retinal vessel detection. However, existing methods exhibited significant limitations in segmentation performance, complexity, and computational efficiency. Specifically, most of the existing methods failed in detecting small vessels owing to vanishing gradient problems. To overcome the stated problems, an intelligence-based automated shallow network with high performance and low cost is designed named Feature Preserving Mesh Network (FPM-Net) for the accurate segmentation of retinal vessels. FPM-Net employs a feature-preserving block that preserves the spatial features and helps in maintaining a better segmentation performance. Similarly, FPM-Net architecture uses a series of feature concatenation that also boosts the overall segmentation performance. Finally, preserved features, low-level input image information, and up-sampled spatial features are aggregated at the final concatenation stage for improved pixel prediction accuracy. The technique is reliable since it performs better on the DRIVE database, CHASE-DB1 database, and STARE dataset. Results and discussion Experimental outcomes confirm that FPM-Net outperforms state-of-the-art techniques with superior computational efficiency. In addition, presented results are achieved without using any preprocessing or postprocessing scheme. Our proposed method FPM-Net gives improvement results which can be observed with DRIVE datasets, it gives Se, Sp, and Acc as 0.8285, 0.98270, 0.92920, for CHASE-DB1 dataset 0.8219, 0.9840, 0.9728 and STARE datasets it produces 0.8618, 0.9819 and 0.9727 respectively. Which is a remarkable difference and enhancement as compared to the conventional methods using only 2.45 million trainable parameters.
Collapse
Affiliation(s)
| | | | | | - Abida Hussain
- Faculty of CS and IT, Superior University, Lahore, Pakistan
| | - Rizwan Ali Naqvi
- Department of Unmanned Vehicle Engineering, Sejong University, Seoul, Republic of Korea,*Correspondence: Rizwan Ali Naqvi ✉
| | - Seung Won Lee
- School of Medicine, Sungkyunkwan University, Suwon, Republic of Korea,Seung Won Lee ✉
| |
Collapse
|
37
|
Li L, Verma M, Wang B, Nakashima Y, Nagahara H, Kawasaki R. Automated grading system of retinal arterio-venous crossing patterns: A deep learning approach replicating ophthalmologist's diagnostic process of arteriolosclerosis. PLOS DIGITAL HEALTH 2023; 2:e0000174. [PMID: 36812612 PMCID: PMC9931248 DOI: 10.1371/journal.pdig.0000174] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 06/20/2022] [Accepted: 11/29/2022] [Indexed: 01/13/2023]
Abstract
The morphological feature of retinal arterio-venous crossing patterns is a valuable source of cardiovascular risk stratification as it directly captures vascular health. Although Scheie's classification, which was proposed in 1953, has been used to grade the severity of arteriolosclerosis as diagnostic criteria, it is not widely used in clinical settings as mastering this grading is challenging as it requires vast experience. In this paper, we propose a deep learning approach to replicate a diagnostic process of ophthalmologists while providing a checkpoint to secure explainability to understand the grading process. The proposed pipeline is three-fold to replicate a diagnostic process of ophthalmologists. First, we adopt segmentation and classification models to automatically obtain vessels in a retinal image with the corresponding artery/vein labels and find candidate arterio-venous crossing points. Second, we use a classification model to validate the true crossing point. At last, the grade of severity for the vessel crossings is classified. To better address the problem of label ambiguity and imbalanced label distribution, we propose a new model, named multi-diagnosis team network (MDTNet), in which the sub-models with different structures or different loss functions provide different decisions. MDTNet unifies these diverse theories to give the final decision with high accuracy. Our automated grading pipeline was able to validate crossing points with precision and recall of 96.3% and 96.3%, respectively. Among correctly detected crossing points, the kappa value for the agreement between the grading by a retina specialist and the estimated score was 0.85, with an accuracy of 0.92. The numerical results demonstrate that our method can achieve a good performance in both arterio-venous crossing validation and severity grading tasks following the diagnostic process of ophthalmologists. By the proposed models, we could build a pipeline reproducing ophthalmologists' diagnostic process without requiring subjective feature extractions. The code is available (https://github.com/conscienceli/MDTNet).
Collapse
Affiliation(s)
- Liangzhi Li
- Institute for Datability Science (IDS), Osaka University, Osaka, Japan
| | - Manisha Verma
- Institute for Datability Science (IDS), Osaka University, Osaka, Japan
| | - Bowen Wang
- Institute for Datability Science (IDS), Osaka University, Osaka, Japan
| | - Yuta Nakashima
- Institute for Datability Science (IDS), Osaka University, Osaka, Japan
| | - Hajime Nagahara
- Institute for Datability Science (IDS), Osaka University, Osaka, Japan
| | - Ryo Kawasaki
- Graduate School of Medicine, Osaka University, Osaka, Japan
- * E-mail:
| |
Collapse
|
38
|
Liu Y, Shen J, Yang L, Bian G, Yu H. ResDO-UNet: A deep residual network for accurate retinal vessel segmentation from fundus images. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104087] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
39
|
Xu GX, Ren CX. SPNet: A novel deep neural network for retinal vessel segmentation based on shared decoder and pyramid-like loss. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.12.039] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/24/2022]
|
40
|
Improving vessel connectivity in retinal vessel segmentation via adversarial learning. Knowl Based Syst 2022. [DOI: 10.1016/j.knosys.2022.110243] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/29/2022]
|
41
|
Kumar KS, Singh NP. An efficient registration-based approach for retinal blood vessel segmentation using generalized Pareto and fatigue pdf. Med Eng Phys 2022; 110:103936. [PMID: 36529622 DOI: 10.1016/j.medengphy.2022.103936] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/05/2022] [Revised: 11/05/2022] [Accepted: 12/01/2022] [Indexed: 12/12/2022]
Abstract
Segmentation of Retinal Blood Vessel (RBV) extraction in the retina images and Registration of segmented RBV structure is implemented to identify changes in vessel structure by ophthalmologists in diagnosis of various illnesses like Glaucoma, Diabetes, and Hypertension's. The Retinal Blood Vessel provides blood to the inner retinal neurons, RBV are located mainly in internal retina but it may partly in the ganglion cell layer, following network failure haven't been identified with past methods. Classifications of accurate RBV and Registration of segmented blood vessels are challenging tasks in the low intensity background of Retinal Image. So, we projected a novel approach of segmentation of RBV extraction used matched filter of Generalized Pareto Probability Distribution Function (pdf) and Registration approach on feature-based segmented retinal blood vessel of Binary Robust Invariant Scalable Key point (BRISK). The BRISK provides the predefined sampling pattern as compared to Pdf. The BRISK feature is implemented for attention point recognition & matching approach for change in vessel structure. The proposed approaches contain 3 levels: pre-processing, matched filter-based Generalized Pareto pdf as a source along with the novel approach of fatigue pdf as a target, and BRISK framework is used for Registration on segmented retinal images of supply & intention images. This implemented system's performance is estimated in experimental analysis by the Average accuracy, Normalized Cross-Correlation (NCC), and computation time process of the segmented retinal source and target image. The NCC is main element to give more statistical information about retinal image segmentation. The proposed approach of Generalized Pareto value pdf has Average Accuracy of 95.21%, NCC of both image pairs is 93%, and Average accuracy of Registration of segmented source images and the target image is 98.51% respectively. The proposed approach of average computational time taken is around 1.4 s, which has been identified on boundary condition of Pdf function.
Collapse
Affiliation(s)
- K Susheel Kumar
- GITAM University, Bengaluru, 561203, India; National Institute of Technology Hamirpur, Himachal Pradesh 177005, India.
| | | |
Collapse
|
42
|
Sun Q, Dai M, Lan Z, Cai F, Wei L, Yang C, Chen R. UCR-Net: U-shaped context residual network for medical image segmentation. Comput Biol Med 2022; 151:106203. [PMID: 36306581 DOI: 10.1016/j.compbiomed.2022.106203] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/11/2022] [Revised: 09/04/2022] [Accepted: 10/09/2022] [Indexed: 12/27/2022]
Abstract
Medical image segmentation prerequisite for numerous clinical needs is a critical step in biomedical image analysis. The U-Net framework is one of the most popular deep networks in this field. However, U-Net's successive pooling and downsampling operations result in some loss of spatial information. In this paper, we propose a U-shaped context residual network, called UCR-Net, to capture more context and high-level information for medical image segmentation. The proposed UCR-Net is an encoder-decoder framework comprising a feature encoder module and a feature decoder module. The feature decoder module contains four newly proposed context attention exploration(CAE) modules, a newly proposed global and spatial attention (GSA) module, and four decoder blocks. We use the proposed CAE module to capture more multi-scale context features from the encoder. The proposed GSA module further explores global context features and semantically enhanced deep-level features. The proposed UCR-Net can recover more high-level semantic features and fuse context attention information from CAE and global and spatial attention information from GSA module. Experiments on the retinal vessel, femoropopliteal artery stent, and polyp datasets demonstrate that the proposed UCR-Net performs favorably against the original U-Net and other advanced methods.
Collapse
Affiliation(s)
- Qi Sun
- Digital Fujian Research Institute of Big Data for Agriculture and Forestry, College of Computer and Information Sciences, Fujian Agriculture and Forestry University, Fuzhou 350002, China.
| | - Mengyun Dai
- Digital Fujian Research Institute of Big Data for Agriculture and Forestry, College of Computer and Information Sciences, Fujian Agriculture and Forestry University, Fuzhou 350002, China.
| | - Ziyang Lan
- Digital Fujian Research Institute of Big Data for Agriculture and Forestry, College of Computer and Information Sciences, Fujian Agriculture and Forestry University, Fuzhou 350002, China.
| | - Fanggang Cai
- Department of vascular surgery, the First Affiliated Hospital, Fujian Medical University, Fuzhou 350108, China.
| | - Lifang Wei
- Digital Fujian Research Institute of Big Data for Agriculture and Forestry, College of Computer and Information Sciences, Fujian Agriculture and Forestry University, Fuzhou 350002, China.
| | - Changcai Yang
- Digital Fujian Research Institute of Big Data for Agriculture and Forestry, College of Computer and Information Sciences, Fujian Agriculture and Forestry University, Fuzhou 350002, China.
| | - Riqing Chen
- Digital Fujian Research Institute of Big Data for Agriculture and Forestry, College of Computer and Information Sciences, Fujian Agriculture and Forestry University, Fuzhou 350002, China.
| |
Collapse
|
43
|
Tang W, Deng H, Yin S. CPMF-Net: Multi-Feature Network Based on Collaborative Patches for Retinal Vessel Segmentation. SENSORS (BASEL, SWITZERLAND) 2022; 22:9210. [PMID: 36501911 PMCID: PMC9736046 DOI: 10.3390/s22239210] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/07/2022] [Revised: 11/23/2022] [Accepted: 11/24/2022] [Indexed: 06/17/2023]
Abstract
As an important basis of clinical diagnosis, the morphology of retinal vessels is very useful for the early diagnosis of some eye diseases. In recent years, with the rapid development of deep learning technology, automatic segmentation methods based on it have made considerable progresses in the field of retinal blood vessel segmentation. However, due to the complexity of vessel structure and the poor quality of some images, retinal vessel segmentation, especially the segmentation of Capillaries, is still a challenging task. In this work, we propose a new retinal blood vessel segmentation method, called multi-feature segmentation, based on collaborative patches. First, we design a new collaborative patch training method which effectively compensates for the pixel information loss in the patch extraction through information transmission between collaborative patches. Additionally, the collaborative patch training strategy can simultaneously have the characteristics of low occupancy, easy structure and high accuracy. Then, we design a multi-feature network to gather a variety of information features. The hierarchical network structure, together with the integration of the adaptive coordinate attention module and the gated self-attention module, enables these rich information features to be used for segmentation. Finally, we evaluate the proposed method on two public datasets, namely DRIVE and STARE, and compare the results of our method with those of other nine advanced methods. The results show that our method outperforms other existing methods.
Collapse
|
44
|
RADCU-Net: residual attention and dual-supervision cascaded U-Net for retinal blood vessel segmentation. INT J MACH LEARN CYB 2022. [DOI: 10.1007/s13042-022-01715-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
45
|
Zhong X, Zhang H, Li G, Ji D. Do you need sharpened details? Asking MMDC-Net: Multi-layer multi-scale dilated convolution network for retinal vessel segmentation. Comput Biol Med 2022; 150:106198. [PMID: 37859292 DOI: 10.1016/j.compbiomed.2022.106198] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/16/2022] [Revised: 09/19/2022] [Accepted: 10/09/2022] [Indexed: 11/24/2022]
Abstract
Convolutional neural networks (CNN), especially numerous U-shaped models, have achieved great progress in retinal vessel segmentation. However, a great quantity of global information in fundus images has not been fully explored. And the class imbalance problem of background and blood vessels is still serious. To alleviate these issues, we design a novel multi-layer multi-scale dilated convolution network (MMDC-Net) based on U-Net. We propose an MMDC module to capture sufficient global information under diverse receptive fields through a cascaded mode. Then, we place a new multi-layer fusion (MLF) module behind the decoder, which can not only fuse complementary features but filter noisy information. This enables MMDC-Net to capture the blood vessel details after continuous up-sampling. Finally, we employ a recall loss to resolve the class imbalance problem. Extensive experiments have been done on diverse fundus color image datasets, including STARE, CHASEDB1, DRIVE, and HRF. HRF has a large resolution of 3504 × 2336 whereas others have a small resolution of slightly more than 512 × 512. Qualitative and quantitative results verify the superiority of MMDC-Net. Notably, satisfactory accuracy and sensitivity are acquired by our model. Hence, some key blood vessel details are sharpened. In addition, a large number of further validations and discussions prove the effectiveness and generalization of the proposed MMDC-Net.
Collapse
Affiliation(s)
- Xiang Zhong
- School of Software, East China Jiaotong University, China
| | - Hongbin Zhang
- School of Software, East China Jiaotong University, China.
| | - Guangli Li
- School of Information Engineering, East China Jiaotong University, China
| | - Donghong Ji
- School of Cyber Science and Engineering, Wuhan University, China
| |
Collapse
|
46
|
Rodrigues EO, Rodrigues LO, Machado JHP, Casanova D, Teixeira M, Oliva JT, Bernardes G, Liatsis P. Local-Sensitive Connectivity Filter (LS-CF): A Post-Processing Unsupervised Improvement of the Frangi, Hessian and Vesselness Filters for Multimodal Vessel Segmentation. J Imaging 2022; 8:jimaging8100291. [PMID: 36286385 PMCID: PMC9604711 DOI: 10.3390/jimaging8100291] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/05/2022] [Revised: 09/13/2022] [Accepted: 09/16/2022] [Indexed: 11/07/2022] Open
Abstract
A retinal vessel analysis is a procedure that can be used as an assessment of risks to the eye. This work proposes an unsupervised multimodal approach that improves the response of the Frangi filter, enabling automatic vessel segmentation. We propose a filter that computes pixel-level vessel continuity while introducing a local tolerance heuristic to fill in vessel discontinuities produced by the Frangi response. This proposal, called the local-sensitive connectivity filter (LS-CF), is compared against a naive connectivity filter to the baseline thresholded Frangi filter response and to the naive connectivity filter response in combination with the morphological closing and to the current approaches in the literature. The proposal was able to achieve competitive results in a variety of multimodal datasets. It was robust enough to outperform all the state-of-the-art approaches in the literature for the OSIRIX angiographic dataset in terms of accuracy and 4 out of 5 works in the case of the IOSTAR dataset while also outperforming several works in the case of the DRIVE and STARE datasets and 6 out of 10 in the CHASE-DB dataset. For the CHASE-DB, it also outperformed all the state-of-the-art unsupervised methods.
Collapse
Affiliation(s)
- Erick O. Rodrigues
- Department of Academic Informatics (DAINF), Universidade Tecnologica Federal do Parana (UTFPR), Pato Branco 85503-390, PR, Brazil
- Correspondence:
| | - Lucas O. Rodrigues
- Graduate Program of Sciences Applied to Health Products, Universidade Federal Fluminense (UFF), Niteroi 24241-000, RJ, Brazil
| | - João H. P. Machado
- Department of Academic Informatics (DAINF), Universidade Tecnologica Federal do Parana (UTFPR), Pato Branco 85503-390, PR, Brazil
| | - Dalcimar Casanova
- Department of Academic Informatics (DAINF), Universidade Tecnologica Federal do Parana (UTFPR), Pato Branco 85503-390, PR, Brazil
| | - Marcelo Teixeira
- Department of Academic Informatics (DAINF), Universidade Tecnologica Federal do Parana (UTFPR), Pato Branco 85503-390, PR, Brazil
| | - Jeferson T. Oliva
- Department of Academic Informatics (DAINF), Universidade Tecnologica Federal do Parana (UTFPR), Pato Branco 85503-390, PR, Brazil
| | - Giovani Bernardes
- Institute of Technological Sciences (ICT), Universidade Federal de Itajuba (UNIFEI), Itabira 35903-087, MG, Brazil
| | - Panos Liatsis
- Department of Electrical Engineering and Computer Science, Khalifa University of Science and Technology, Abu Dhabi P.O. Box 127788, United Arab Emirates
| |
Collapse
|
47
|
MHA-Net: A Multibranch Hybrid Attention Network for Medical Image Segmentation. COMPUTATIONAL AND MATHEMATICAL METHODS IN MEDICINE 2022; 2022:8375981. [PMID: 36245836 PMCID: PMC9560845 DOI: 10.1155/2022/8375981] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/15/2022] [Revised: 09/11/2022] [Accepted: 09/13/2022] [Indexed: 12/02/2022]
Abstract
The robust segmentation of organs from the medical image is the key technique in medical image analysis for disease diagnosis. U-Net is a robust structure for medical image segmentation. However, U-Net adopts consecutive downsampling encoders to capture multiscale features, resulting in the loss of contextual information and insufficient recovery of high-level semantic features. In this paper, we present a new multibranch hybrid attention network (MHA-Net) to capture more contextual information and high-level semantic features. The main idea of our proposed MHA-Net is to use the multibranch hybrid attention feature decoder to recover more high-level semantic features. The lightweight pyramid split attention (PSA) module is used to connect the encoder and decoder subnetwork to obtain a richer multiscale feature map. We compare the proposed MHA-Net to state-of-art approaches on the DRIVE dataset, the fluoroscopic roentgenographic stereophotogrammetric analysis X-ray dataset, and the polyp dataset. The experimental results on different modal images reveal that our proposed MHA-Net provides better segmentation results than other segmentation approaches.
Collapse
|
48
|
A Hybrid Fusion Method Combining Spatial Image Filtering with Parallel Channel Network for Retinal Vessel Segmentation. ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING 2022. [DOI: 10.1007/s13369-022-07311-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
49
|
Abdominal vessel segmentation using vessel model embedded fuzzy C-means and similarity from CT angiography. Med Biol Eng Comput 2022; 60:3325-3340. [DOI: 10.1007/s11517-022-02644-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/18/2021] [Accepted: 08/09/2022] [Indexed: 11/25/2022]
|
50
|
Hu X, Wang L, Li Y. HT-Net: A Hybrid Transformer Network for Fundus Vessel Segmentation. SENSORS (BASEL, SWITZERLAND) 2022; 22:6782. [PMID: 36146132 PMCID: PMC9504252 DOI: 10.3390/s22186782] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/12/2022] [Revised: 09/06/2022] [Accepted: 09/06/2022] [Indexed: 06/16/2023]
Abstract
Doctors usually diagnose a disease by evaluating the pattern of abnormal blood vessels in the fundus. At present, the segmentation of fundus blood vessels based on deep learning has achieved great success, but it still faces the problems of low accuracy and capillary rupture. A good vessel segmentation method can guide the early diagnosis of eye diseases, so we propose a novel hybrid Transformer network (HT-Net) for fundus imaging analysis. HT-Net can improve the vessel segmentation quality by capturing detailed local information and implementing long-range information interactions, and it mainly consists of the following blocks. The feature fusion block (FFB) is embedded in the shallow levels, and FFB enriches the feature space. In addition, the feature refinement block (FRB) is added to the shallow position of the network, which solves the problem of vessel scale change by fusing multi-scale feature information to improve the accuracy of segmentation. Finally, HT-Net's bottom-level position can capture remote dependencies by combining the Transformer and CNN. We prove the performance of HT-Net on the DRIVE, CHASE_DB1, and STARE datasets. The experiment shows that FFB and FRB can effectively improve the quality of microvessel segmentation by extracting multi-scale information. Embedding efficient self-attention mechanisms in the network can effectively improve the vessel segmentation accuracy. The HT-Net exceeds most existing methods, indicating that it can perform the task of vessel segmentation competently.
Collapse
|