1
|
Freire TP, Braz Júnior G, de Almeida JDS, Rodrigues Junior JRD. Cup and Disc Segmentation in Smartphone Handheld Ophthalmoscope Images with a Composite Backbone and Double Decoder Architecture. Vision (Basel) 2025; 9:32. [PMID: 40265400 PMCID: PMC12015843 DOI: 10.3390/vision9020032] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/27/2025] [Revised: 04/01/2025] [Accepted: 04/09/2025] [Indexed: 04/24/2025] Open
Abstract
Glaucoma is a visual disease that affects millions of people, and early diagnosis can prevent total blindness. One way to diagnose the disease is through fundus image examination, which analyzes the optic disc and cup structures. However, screening programs in primary care are costly and unfeasible. Neural network models have been used to segment optic nerve structures, assisting physicians in this task and reducing fatigue. This work presents a methodology to enhance morphological biomarkers of the optic disc and cup in images obtained by a smartphone coupled to an ophthalmoscope through a deep neural network, which combines two backbones and a dual decoder approach to improve the segmentation of these structures, as well as a new way to combine the loss weights in the training process. The models obtained were numerically evaluated through Dice and IoU measures. The dice values obtained in the experiments reached a Dice of 95.92% and 85.30% for the optical disc and cup and an IoU of 92.22% and 75.68% for the optical disc and cup, respectively, in the BrG dataset. These findings indicate promising architectures in the fundus image segmentation task.
Collapse
Affiliation(s)
- Thiago Paiva Freire
- UFMA/Computer Science Department, Universidade Federal do Maranhão, Campus do Bacanga, São Luís 65085-580, Brazil;
| | - Geraldo Braz Júnior
- UFMA/Computer Science Department, Universidade Federal do Maranhão, Campus do Bacanga, São Luís 65085-580, Brazil;
| | | | | |
Collapse
|
2
|
Banerjee R, Mujib R, Sanyal P, Chakraborti T, Saha SK. Pan-Ret: a semi-supervised framework for scalable detection of pan-retinal diseases. Med Biol Eng Comput 2025; 63:959-974. [PMID: 39672991 DOI: 10.1007/s11517-024-03250-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/09/2024] [Accepted: 11/19/2024] [Indexed: 12/15/2024]
Abstract
It has been shown in recent years that a range of optical diseases have early manifestation in retinal fundus images. It is becoming increasingly important to separate the regions of interest (RoI) upfront in the automated classification pipeline in order to ensure the alignment of the disease diagnosis with clinically relevant visual features. In this work, we introduce Pan-Ret, a semi-supervised framework which starts with locating the abnormalities in the biomedically relevant RoIs of a retinal image in an "annotation-agnostic" fashion. It does so by leveraging an anomaly detection setup using parallel autoencoders that are trained only on healthy population initially. Then, the anomalous images are separated based on the RoIs using a fully interpretable classifier like support vector machine (SVM). Experimental results show that the proposed approach yields an overall F1-score of 0.95 and 0.96 in detecting abnormalities on two different public datasets covering a diverse range of retinal diseases including diabetic retinopathy, hypertensive retinopathy, glaucoma, age-related macular degeneration, and several more in a staged manner. Thus, the work presents a milestone towards a pan-retinal disease diagnostic pipeline that can not only cater to the current set of disease classes, but has the capacity of adding further classes down the line. This is due to an anomaly detection style one-class learning setup of the deep autoencoder piece of the proposed pipeline, thus improving the generalizability of this approach compared to usual fully supervised competitors. This is also expected to increase the practical translational potential of Pan-Ret in a real-life scalable clinical setting.
Collapse
|
3
|
Wang X, Gong D, Chen Y, Zong Z, Li M, Fan K, Jia L, Cao Q, Liu Q, Yang Q. Hybrid CNN-Mamba model for multi-scale fundus image enhancement. BIOMEDICAL OPTICS EXPRESS 2025; 16:1104-1117. [PMID: 40109520 PMCID: PMC11919352 DOI: 10.1364/boe.542471] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/18/2024] [Revised: 11/26/2024] [Accepted: 02/06/2025] [Indexed: 03/22/2025]
Abstract
This study proposes a multi-scale fundus image enhancement approach that combines CNN with Mamba, demonstrating clear superiority across multiple benchmarks. The model consistently achieves top performance on public datasets, with the lowest FID and KID scores, and the highest PSNR and SSIM values, particularly excelling at larger image resolutions. Notably, its performance improves as the image size increases, with several metrics reaching optimal values at 1024 × 1024 resolution. Scale generalizability further highlights the model's exceptional structural preservation capability. Additionally, its high VSD and IOU scores in segmentation tasks further validate its practical effectiveness, making it a valuable tool for enhancing fundus images and improving diagnostic accuracy.
Collapse
Affiliation(s)
- Xiaopeng Wang
- Academy of Artificial Intelligence, Beijing Institute of Petrochemical Technology, Beijing 102617, China
| | - Di Gong
- China-Japan Friendship Hospital, Beijing 100029, China
| | - Yi Chen
- China-Japan Friendship Hospital, Beijing 100029, China
| | - Zheng Zong
- Beijing Information Science and Technology University, Beijing 100096, China
| | - Meng Li
- Tianjin University of Technology, Tianjin 300384, China
| | - Kun Fan
- Taihe Intelligent Technology Group Co., Ltd., Anhui 230000, China
| | - Lina Jia
- Beijing Etop Smartinfo Technology Co., Ltd., Beijing 102617, China
| | - Qiyuan Cao
- Academy of Artificial Intelligence, Beijing Institute of Petrochemical Technology, Beijing 102617, China
| | - Qiang Liu
- Academy of Artificial Intelligence, Beijing Institute of Petrochemical Technology, Beijing 102617, China
| | - Qiang Yang
- Academy of Artificial Intelligence, Beijing Institute of Petrochemical Technology, Beijing 102617, China
| |
Collapse
|
4
|
Abdalla MMI, Mohanraj J. Revolutionizing diabetic retinopathy screening and management: The role of artificial intelligence and machine learning. World J Clin Cases 2025; 13:101306. [PMID: 39959767 PMCID: PMC11606367 DOI: 10.12998/wjcc.v13.i5.101306] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 09/10/2024] [Revised: 10/09/2024] [Accepted: 11/05/2024] [Indexed: 11/18/2024] Open
Abstract
Diabetic retinopathy (DR) remains a leading cause of vision impairment and blindness among individuals with diabetes, necessitating innovative approaches to screening and management. This editorial explores the transformative potential of artificial intelligence (AI) and machine learning (ML) in revolutionizing DR care. AI and ML technologies have demonstrated remarkable advancements in enhancing the accuracy, efficiency, and accessibility of DR screening, helping to overcome barriers to early detection. These technologies leverage vast datasets to identify patterns and predict disease progression with unprecedented precision, enabling clinicians to make more informed decisions. Furthermore, AI-driven solutions hold promise in personalizing management strategies for DR, incorporating predictive analytics to tailor interventions and optimize treatment pathways. By automating routine tasks, AI can reduce the burden on healthcare providers, allowing for a more focused allocation of resources towards complex patient care. This review aims to evaluate the current advancements and applications of AI and ML in DR screening, and to discuss the potential of these technologies in developing personalized management strategies, ultimately aiming to improve patient outcomes and reduce the global burden of DR. The integration of AI and ML in DR care represents a paradigm shift, offering a glimpse into the future of ophthalmic healthcare.
Collapse
Affiliation(s)
- Mona Mohamed Ibrahim Abdalla
- Department of Human Biology, School of Medicine, International Medical University, Bukit Jalil 57000, Kuala Lumpur, Malaysia
| | - Jaiprakash Mohanraj
- Department of Human Biology, School of Medicine, International Medical University, Bukit Jalil 57000, Kuala Lumpur, Malaysia
| |
Collapse
|
5
|
Din S, Shoaib M, Serpedin E. CXR-Seg: A Novel Deep Learning Network for Lung Segmentation from Chest X-Ray Images. Bioengineering (Basel) 2025; 12:167. [PMID: 40001687 PMCID: PMC11851456 DOI: 10.3390/bioengineering12020167] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/04/2025] [Revised: 01/26/2025] [Accepted: 02/07/2025] [Indexed: 02/27/2025] Open
Abstract
Over the past decade, deep learning techniques, particularly neural networks, have become essential in medical imaging for tasks like image detection, classification, and segmentation. These methods have greatly enhanced diagnostic accuracy, enabling quicker identification and more effective treatments. In chest X-ray analysis, however, challenges remain in accurately segmenting and classifying organs such as the lungs, heart, diaphragm, sternum, and clavicles, as well as detecting abnormalities in the thoracic cavity. Despite progress, these issues highlight the need for improved approaches to overcome segmentation difficulties and enhance diagnostic reliability. In this context, we propose a novel architecture named CXR-Seg, tailored for semantic segmentation of lungs from chest X-ray images. The proposed network mainly consists of four components, including a pre-trained EfficientNet as an encoder to extract feature encodings, a spatial enhancement module embedded in the skip connection to promote the adjacent feature fusion, a transformer attention module at the bottleneck layer, and a multi-scale feature fusion block at the decoder. The performance of the proposed CRX-Seg was evaluated on four publicly available datasets (MC, Darwin, and Shenzhen for chest X-rays, and TCIA for brain flair segmentation from MRI images). The proposed method achieved a Jaccard index, Dice coefficient, accuracy, sensitivity, and specificity of 95.63%, 97.76%, 98.77%, 98.00%, and 99.05%on MC; 91.66%, 95.62%, 96.35%, 95.53%, and 96.94% on V7 Darwin COVID-19; and 92.97%, 96.32%, 96.69%, 96.01%, and 97.40% on the Shenzhen Tuberculosis CXR Dataset, respectively. Conclusively, the proposed network offers improved performance in comparison with state-of-the-art methods, and better generalization for the semantic segmentation of lungs from chest X-ray images.
Collapse
Affiliation(s)
- Sadia Din
- Electrical and Computer Engineering Program, Texas A&M University, Doha 23874, Qatar
| | - Muhammad Shoaib
- Department of Electrical and Computer Engineering, Abbottabad Campus, COMSATS University Islamabad, Abbottabad 22060, Pakistan
| | - Erchin Serpedin
- Department of Electrical and Computer Engineering, Texas A&M University, College Station, TX 77840, USA
| |
Collapse
|
6
|
Muthukumar KA, Nandi D, Ranjan P, Ramachandran K, Pj S, Ghosh A, M A, Radhakrishnan A, Dhandapani VE, Janardhanan R. Integrating electrocardiogram and fundus images for early detection of cardiovascular diseases. Sci Rep 2025; 15:4390. [PMID: 39910082 PMCID: PMC11799439 DOI: 10.1038/s41598-025-87634-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/26/2024] [Accepted: 01/21/2025] [Indexed: 02/07/2025] Open
Abstract
Cardiovascular diseases (CVD) are a predominant health concern globally, emphasizing the need for advanced diagnostic techniques. In our research, we present an avant-garde methodology that synergistically integrates ECG readings and retinal fundus images to facilitate the early disease tagging as well as triaging of the CVDs in the order of disease priority. Recognizing the intricate vascular network of the retina as a reflection of the cardiovascular system, alongwith the dynamic cardiac insights from ECG, we sought to provide a holistic diagnostic perspective. Initially, a Fast Fourier Transform (FFT) was applied to both the ECG and fundus images, transforming the data into the frequency domain. Subsequently, the Earth Mover's Distance (EMD) was computed for the frequency-domain features of both modalities. These EMD values were then concatenated, forming a comprehensive feature set that was fed into a Neural Network classifier. This approach, leveraging the FFT's spectral insights and EMD's capability to capture nuanced data differences, offers a robust representation for CVD classification. Preliminary tests yielded a commendable accuracy of 84%, underscoring the potential of this combined diagnostic strategy. As we continue our research, we anticipate refining and validating the model further to enhance its clinical applicability in resource limited healthcare ecosystems prevalent across the Indian sub-continent and also the world at large.
Collapse
Affiliation(s)
- K A Muthukumar
- University of Petroleum and Energy Studies, Dehradun, Uttarakhand, India.
| | - Dhruva Nandi
- Faculty of Medicine and Health Sciences , SRM Medical College Hospital and Research Centre, SRM IST, Kattankulathur, Chengalpattu, Tamil Nadu, India
| | - Priya Ranjan
- University of Petroleum and Energy Studies, Dehradun, Uttarakhand, India
| | - Krithika Ramachandran
- Centre for High Impact Neuroscience and Translational Applications, TCG Crest, Kolkata, West Bengal, India
| | - Shiny Pj
- Faculty of Medicine and Health Sciences , SRM Medical College Hospital and Research Centre, SRM IST, Kattankulathur, Chengalpattu, Tamil Nadu, India
| | - Anirban Ghosh
- Department of Electronics and Communication, SRM University AP, Neerukonda, Andhra Pradesh, India
| | - Ashwini M
- Ashwini Eye Care, Chennai, Tamil Nadu, India
| | - Aiswaryah Radhakrishnan
- Faculty of Medicine and Health Sciences , SRM Medical College Hospital and Research Centre, SRM IST, Kattankulathur, Chengalpattu, Tamil Nadu, India
| | | | - Rajiv Janardhanan
- Faculty of Medicine and Health Sciences , SRM Medical College Hospital and Research Centre, SRM IST, Kattankulathur, Chengalpattu, Tamil Nadu, India.
| |
Collapse
|
7
|
Deng T, Huang Y, Yang C. Parallel Multi-Path Network for Ocular Disease Detection Inspired by Visual Cognition Mechanism. IEEE J Biomed Health Inform 2025; 29:345-357. [PMID: 39348249 DOI: 10.1109/jbhi.2024.3471807] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/02/2024]
Abstract
Various ocular diseases such as cataracts, glaucoma, and diabetic retinopathy have become several major factors causing non-congenital visual impairment, which seriously threatens people's vision health. The shortage of ophthalmic medical resources has brought huge obstacles to large-scale ocular disease screening. Therefore, it is necessary to use computer-aided diagnosis (CAD) technology to achieve large-scale screening and diagnosis of ocular diseases. In this work, inspired by the human visual cognition mechanism, we propose a parallel multi-path network for multiple ocular diseases detection, called PMP-OD, which integrates the detection of multiple common ocular diseases, including cataracts, glaucoma, diabetic retinopathy, and pathological myopia. The bottom-up features of the fundus image are extracted by a common convolutional module, the Low-level Feature Extraction module, which simulates the non-selective pathway. Simultaneously, the top-down vessel and other lesion features are extracted by the High-level Feature Extraction module that simulates the selective pathway. The retinal vessel and lesion features can be regarded as task-driven high-level semantic information in the physician's disease diagnosis process. Then, the features are fused by a feature fusion module based on the attention mechanism. Finally, the disease classifier gives prediction results according to the integrated multi-features. The experimental results indicate that our PMP-OD model outperforms other state-of-the-art (SOTA) models on an ocular disease dataset reconstructed from ODIR-5K, APTOS-2019, ORIGA-light, and Kaggle.
Collapse
|
8
|
Farooq H, Zafar Z, Saadat A, Khan TM, Iqbal S, Razzak I. LSSF-Net: Lightweight segmentation with self-awareness, spatial attention, and focal modulation. Artif Intell Med 2024; 158:103012. [PMID: 39547075 DOI: 10.1016/j.artmed.2024.103012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/22/2024] [Revised: 09/02/2024] [Accepted: 11/02/2024] [Indexed: 11/17/2024]
Abstract
Accurate segmentation of skin lesions within dermoscopic images plays a crucial role in the timely identification of skin cancer for computer-aided diagnosis on mobile platforms. However, varying shapes of the lesions, lack of defined edges, and the presence of obstructions such as hair strands and marker colours make this challenge more complex. Additionally, skin lesions often exhibit subtle variations in texture and colour that are difficult to differentiate from surrounding healthy skin, necessitating models that can capture both fine-grained details and broader contextual information. Currently, melanoma segmentation models are commonly based on fully connected networks and U-Nets. However, these models often struggle with capturing the complex and varied characteristics of skin lesions, such as the presence of indistinct boundaries and diverse lesion appearances, which can lead to suboptimal segmentation performance. To address these challenges, we propose a novel lightweight network specifically designed for skin lesion segmentation utilising mobile devices, featuring a minimal number of learnable parameters (only 0.8 million). This network comprises an encoder-decoder architecture that incorporates conformer-based focal modulation attention, self-aware local and global spatial attention, and split channel-shuffle. The efficacy of our model has been evaluated on four well-established benchmark datasets for skin lesion segmentation: ISIC 2016, ISIC 2017, ISIC 2018, and PH2. Empirical findings substantiate its state-of-the-art performance, notably reflected in a high Jaccard index.
Collapse
Affiliation(s)
- Hamza Farooq
- School of Electrical Engineering and Computer Science (SEECS), National University of Sciences & Technology (NUST), Islamabad 44000, Pakistan
| | - Zuhair Zafar
- School of Electrical Engineering and Computer Science (SEECS), National University of Sciences & Technology (NUST), Islamabad 44000, Pakistan
| | - Ahsan Saadat
- School of Electrical Engineering and Computer Science (SEECS), National University of Sciences & Technology (NUST), Islamabad 44000, Pakistan; Sydney City Campus, Western Sydney University, Sydney, Australia
| | - Tariq M Khan
- School of Computer Science and Engineering, UNSW, Sydney, Australia
| | - Shahzaib Iqbal
- Department of Computing, Abasyn University Islamabad Campus, Pakistan.
| | - Imran Razzak
- School of Computer Science and Engineering, UNSW, Sydney, Australia
| |
Collapse
|
9
|
Giangaspero S, Nicchiotti G, Venier P, Genilloud L, Pirrami L. AI-Driven Electrical Fast Transient Suppression for Enhanced Electromagnetic Interference Immunity in Inductive Smart Proximity Sensors. SENSORS (BASEL, SWITZERLAND) 2024; 24:7372. [PMID: 39599148 PMCID: PMC11598082 DOI: 10.3390/s24227372] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/12/2024] [Revised: 11/14/2024] [Accepted: 11/15/2024] [Indexed: 11/29/2024]
Abstract
Inductive proximity sensors are relevant in position-sensing applications in many industries but, in order to be used in harsh industrial environments, they need to be immune to electromagnetic interference (EMI). The use of conventional filters to mitigate these perturbations often compromises signal bandwidth, ranging from 100 Hz to 1.6 kHz. We have exploited recent advances in the field of artificial intelligence (AI) to study the ability of neural networks (NNs) to automatically filter out EMI features. This study offers an analysis and comparison of possible NN models (a 1D convolutional NN, a recurrent NN, and a hybrid convolutional and recurrent approach) for denoising EMI-perturbed signals and proposes a final model, which is based on gated recurrent unit (GRU) layers. This network is compressed and optimised to meet memory requirements, so that in future developments it could be implemented in application-specific integrated circuits (ASICs) for inductive sensors. The final RNN manages to reduce noise by 70% (MSEred) while occupying 2 KB of memory.
Collapse
Affiliation(s)
- Silvia Giangaspero
- iSIS Institute, HEIA-FR, HES-SO University of Applied Sciences and Arts Western Switzerland, 1700 Fribourg, Switzerland
| | - Gianluca Nicchiotti
- iSIS Institute, HEIA-FR, HES-SO University of Applied Sciences and Arts Western Switzerland, 1700 Fribourg, Switzerland
| | - Philippe Venier
- Contrinex SA, Route du Pâqui 3, 1720 Corminboeuf, Switzerland; (P.V.); (L.G.)
| | - Laurent Genilloud
- Contrinex SA, Route du Pâqui 3, 1720 Corminboeuf, Switzerland; (P.V.); (L.G.)
| | - Lorenzo Pirrami
- iSIS Institute, HEIA-FR, HES-SO University of Applied Sciences and Arts Western Switzerland, 1700 Fribourg, Switzerland
| |
Collapse
|
10
|
Jia Y, Chen G, Chi H. Retinal fundus image super-resolution based on generative adversarial network guided with vascular structure prior. Sci Rep 2024; 14:22786. [PMID: 39354105 PMCID: PMC11445418 DOI: 10.1038/s41598-024-74186-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/26/2024] [Accepted: 09/24/2024] [Indexed: 10/03/2024] Open
Abstract
Many ophthalmic and systemic diseases can be screened by analyzing retinal fundus images. The clarity and resolution of retinal fundus images directly determine the effectiveness of clinical diagnosis. Deep learning methods based on generative adversarial networks are used in various research fields due to their powerful generative capabilities, especially image super-resolution. Although Real-ESRGAN is a recently proposed method that excels in processing real-world degraded images, it suffers from structural distortions when super-resolving retinal fundus images are rich in structural information. To address this shortcoming, we first process the input image using a pre-trained U-Net model to obtain a structural segmentation map of the retinal vessels and use the segmentation map as the structural prior. The spatial feature transform layer is then used to better integrate the structural prior into the generation process of the generator. In addition, we introduce channel and spatial attention modules into the skip connections of the discriminator to emphasize meaningful features and accordingly enhance the discriminative power of the discriminator. Based on the original loss functions, we introduce the L1 loss function to measure the pixel-level differences between the segmentation maps of retinal vascular structures in the high-resolution images and the super-resolution images to further constrain the super-resolution images. Simulation results on retinal image datasets show that our improved algorithm results have a better visual performance by suppressing structural distortions in the super-resolution images.
Collapse
Affiliation(s)
- Yanfei Jia
- School of Electrical and Information Engineering, Beihua University, Jilin, 132013, China.
| | - Guangda Chen
- School of Electrical and Information Engineering, Beihua University, Jilin, 132013, China
| | - Haotian Chi
- College of electronic science and engineering, Jilin University, Changchun, 130015, China
| |
Collapse
|
11
|
Plećaš D, Gotovac Đogaš V, Polašek O, Škunca Herman J. Determinants of Human Asymmetry: Does Asymmetrical Retinal Vasculature Predict Asymmetry Elsewhere in the Body? Life (Basel) 2024; 14:929. [PMID: 39202672 PMCID: PMC11355915 DOI: 10.3390/life14080929] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/15/2024] [Revised: 07/21/2024] [Accepted: 07/22/2024] [Indexed: 09/03/2024] Open
Abstract
The aim of this study was to explore retinal vasculature asymmetry (ReVA) patterns in subjects from the islands of Vis and Korcula and the city of Split, Croatia. Asymmetry estimates were based on topographic image analysis of non-mydriatic retinal fundus photographs and compared with nine ophthalmic measurements, three Doppler-based pressure indices and eight frequencies of audiometry. ReVA was also correlated to the genomic runs of homozygosity (ROHs) and used in a Cox regression survival model, where we adjusted for the effects of sex, age and comorbidity. In 1873 subjects, ReVA estimates were significantly correlated with most ophthalmic asymmetry measures, less strongly with the ankle-brachial pressure index and only modestly with higher-amplitude audiometry asymmetries (lowest p = 0.020). ReVA was significantly correlated with the number of ROHs (r = 0.229, p < 0.001) but less strongly with the ROH length (r = 0.101, p < 0.001). The overlap of asymmetries was low, with only 107 subjects (5.7% of the total sample) who had two or more instances in which they were among the top 10%. Multiple asymmetries did not affect survival (HR = 0.74, 95% confidence intervals 0.45-1.22). Retinal vasculature asymmetry is a poor predictor of asymmetry elsewhere in the body. Despite its existence and apparent association with comorbidities, the observed extent of retinal vasculature asymmetry did not affect the lifespan in this population.
Collapse
Affiliation(s)
- Doris Plećaš
- University of Split School of Medicine, 21000 Split, Croatia;
| | | | - Ozren Polašek
- University of Split School of Medicine, 21000 Split, Croatia;
- Croatian Science Foundation, 10000 Zagreb, Croatia
| | | |
Collapse
|
12
|
Matloob Abbasi M, Iqbal S, Aurangzeb K, Alhussein M, Khan TM. LMBiS-Net: A lightweight bidirectional skip connection based multipath CNN for retinal blood vessel segmentation. Sci Rep 2024; 14:15219. [PMID: 38956117 PMCID: PMC11219784 DOI: 10.1038/s41598-024-63496-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2023] [Accepted: 05/29/2024] [Indexed: 07/04/2024] Open
Abstract
Blinding eye diseases are often related to changes in retinal structure, which can be detected by analysing retinal blood vessels in fundus images. However, existing techniques struggle to accurately segment these delicate vessels. Although deep learning has shown promise in medical image segmentation, its reliance on specific operations can limit its ability to capture crucial details such as the edges of the vessel. This paper introduces LMBiS-Net, a lightweight convolutional neural network designed for the segmentation of retinal vessels. LMBiS-Net achieves exceptional performance with a remarkably low number of learnable parameters (only 0.172 million). The network used multipath feature extraction blocks and incorporates bidirectional skip connections for the information flow between the encoder and decoder. In addition, we have optimised the efficiency of the model by carefully selecting the number of filters to avoid filter overlap. This optimisation significantly reduces training time and improves computational efficiency. To assess LMBiS-Net's robustness and ability to generalise to unseen data, we conducted comprehensive evaluations on four publicly available datasets: DRIVE, STARE, CHASE_DB1, and HRF The proposed LMBiS-Net achieves significant performance metrics in various datasets. It obtains sensitivity values of 83.60%, 84.37%, 86.05%, and 83.48%, specificity values of 98.83%, 98.77%, 98.96%, and 98.77%, accuracy (acc) scores of 97.08%, 97.69%, 97.75%, and 96.90%, and AUC values of 98.80%, 98.82%, 98.71%, and 88.77% on the DRIVE, STARE, CHEASE_DB, and HRF datasets, respectively. In addition, it records F1 scores of 83.43%, 84.44%, 83.54%, and 78.73% on the same datasets. Our evaluations demonstrate that LMBiS-Net achieves high segmentation accuracy (acc) while exhibiting both robustness and generalisability across various retinal image datasets. This combination of qualities makes LMBiS-Net a promising tool for various clinical applications.
Collapse
Affiliation(s)
- Mufassir Matloob Abbasi
- Department of Electrical Engineering, Abasyn University Islamabad Campus (AUIC), Islamabad, 44000, Pakistan
| | - Shahzaib Iqbal
- Department of Electrical Engineering, Abasyn University Islamabad Campus (AUIC), Islamabad, 44000, Pakistan.
| | - Khursheed Aurangzeb
- Department of Computer Engineering, College of Computer and Information Sciences, King Saud University, Riyadh, P. O. Box 51178, 11543, Saudi Arabia
| | - Musaed Alhussein
- Department of Computer Engineering, College of Computer and Information Sciences, King Saud University, Riyadh, P. O. Box 51178, 11543, Saudi Arabia
| | - Tariq M Khan
- School of Computer Science and Engineering, University of New South Wales, Sydney, NSW, Australia
| |
Collapse
|
13
|
Iqbal S, Khan TM, Naqvi SS, Naveed A, Usman M, Khan HA, Razzak I. LDMRes-Net: A Lightweight Neural Network for Efficient Medical Image Segmentation on IoT and Edge Devices. IEEE J Biomed Health Inform 2024; 28:3860-3871. [PMID: 37938951 DOI: 10.1109/jbhi.2023.3331278] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2023]
Abstract
In this study, we propose LDMRes-Net, a lightweight dual-multiscale residual block-based convolutional neural network tailored for medical image segmentation on IoT and edge platforms. Conventional U-Net-based models face challenges in meeting the speed and efficiency demands of real-time clinical applications, such as disease monitoring, radiation therapy, and image-guided surgery. In this study, we present the Lightweight Dual Multiscale Residual Block-based Convolutional Neural Network (LDMRes-Net), which is specifically designed to overcome these difficulties. LDMRes-Net overcomes these limitations with its remarkably low number of learnable parameters (0.072 M), making it highly suitable for resource-constrained devices. The model's key innovation lies in its dual multiscale residual block architecture, which enables the extraction of refined features on multiple scales, enhancing overall segmentation performance. To further optimize efficiency, the number of filters is carefully selected to prevent overlap, reduce training time, and improve computational efficiency. The study includes comprehensive evaluations, focusing on the segmentation of the retinal image of vessels and hard exudates crucial for the diagnosis and treatment of ophthalmology. The results demonstrate the robustness, generalizability, and high segmentation accuracy of LDMRes-Net, positioning it as an efficient tool for accurate and rapid medical image segmentation in diverse clinical applications, particularly on IoT and edge platforms. Such advances hold significant promise for improving healthcare outcomes and enabling real-time medical image analysis in resource-limited settings.
Collapse
|
14
|
Zhang S, Webers CAB, Berendschot TTJM. Computational single fundus image restoration techniques: a review. FRONTIERS IN OPHTHALMOLOGY 2024; 4:1332197. [PMID: 38984141 PMCID: PMC11199880 DOI: 10.3389/fopht.2024.1332197] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/02/2023] [Accepted: 04/19/2024] [Indexed: 07/11/2024]
Abstract
Fundus cameras are widely used by ophthalmologists for monitoring and diagnosing retinal pathologies. Unfortunately, no optical system is perfect, and the visibility of retinal images can be greatly degraded due to the presence of problematic illumination, intraocular scattering, or blurriness caused by sudden movements. To improve image quality, different retinal image restoration/enhancement techniques have been developed, which play an important role in improving the performance of various clinical and computer-assisted applications. This paper gives a comprehensive review of these restoration/enhancement techniques, discusses their underlying mathematical models, and shows how they may be effectively applied in real-life practice to increase the visual quality of retinal images for potential clinical applications including diagnosis and retinal structure recognition. All three main topics of retinal image restoration/enhancement techniques, i.e., illumination correction, dehazing, and deblurring, are addressed. Finally, some considerations about challenges and the future scope of retinal image restoration/enhancement techniques will be discussed.
Collapse
Affiliation(s)
- Shuhe Zhang
- University Eye Clinic Maastricht, Maastricht University Medical Center, Maastricht, Netherlands
| | - Carroll A B Webers
- University Eye Clinic Maastricht, Maastricht University Medical Center, Maastricht, Netherlands
| | - Tos T J M Berendschot
- University Eye Clinic Maastricht, Maastricht University Medical Center, Maastricht, Netherlands
| |
Collapse
|
15
|
Chen Q, Peng J, Zhao S, Liu W. Automatic artery/vein classification methods for retinal blood vessel: A review. Comput Med Imaging Graph 2024; 113:102355. [PMID: 38377630 DOI: 10.1016/j.compmedimag.2024.102355] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/26/2023] [Revised: 02/06/2024] [Accepted: 02/06/2024] [Indexed: 02/22/2024]
Abstract
Automatic retinal arteriovenous classification can assist ophthalmologists in disease early diagnosis. Deep learning-based methods and topological graph-based methods have become the main solutions for retinal arteriovenous classification in recent years. This paper reviews the automatic retinal arteriovenous classification methods from 2003 to 2022. Firstly, we compare different methods and provide comparison tables of the summary results. Secondly, we complete the classification of the public arteriovenous classification datasets and provide the annotation development tables of different datasets. Finally, we sort out the challenges of evaluation methods and provide a comprehensive evaluation system. Quantitative and qualitative analysis shows the changes in research hotspots over time, Quantitative and qualitative analyses reveal the evolution of research hotspots over time, highlighting the significance of exploring the integration of deep learning with topological information in future research.
Collapse
Affiliation(s)
- Qihan Chen
- School of Intelligent Systems Engineering, Shenzhen Campus of Sun Yat-sen University, Shenzhen 518107, China
| | - Jianqing Peng
- School of Intelligent Systems Engineering, Shenzhen Campus of Sun Yat-sen University, Shenzhen 518107, China; Guangdong Provincial Key Laboratory of Fire Science and Technology, Guangzhou 510006, China.
| | - Shen Zhao
- School of Intelligent Systems Engineering, Shenzhen Campus of Sun Yat-sen University, Shenzhen 518107, China.
| | - Wanquan Liu
- School of Intelligent Systems Engineering, Shenzhen Campus of Sun Yat-sen University, Shenzhen 518107, China
| |
Collapse
|
16
|
Zang F, Ma H. CRA-Net: Transformer guided category-relation attention network for diabetic retinopathy grading. Comput Biol Med 2024; 170:107993. [PMID: 38277925 DOI: 10.1016/j.compbiomed.2024.107993] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/24/2023] [Revised: 12/30/2023] [Accepted: 01/13/2024] [Indexed: 01/28/2024]
Abstract
Automated grading of diabetic retinopathy (DR) is an important means for assisting clinical diagnosis and preventing further retinal damage. However, imbalances and similarities between categories in the DR dataset make it highly challenging to accurately grade the severity of the condition. Furthermore, DR images encompass various lesions, and the pathological relationship information among these lesions can be easily overlooked. For instance, under different severity levels, the varying contributions of different lesions to accurate model grading differ significantly. To address the aforementioned issues, we design a transformer guided category-relation attention network (CRA-Net). Specifically, we propose a novel category attention block that enhances feature information within the class from the perspective of DR image categories, thereby alleviating class imbalance problems. Additionally, we design a lesion relation attention block that captures relationships between lesions by incorporating attention mechanisms in two primary aspects: capsule attention models the relative importance of different lesions, allowing the model to focus on more "informative" ones. Spatial attention captures the global position relationship between lesion features under transformer guidance, facilitating more accurate localization of lesions. Experimental and ablation studies on two datasets DDR and APTOS 2019 demonstrate the effectiveness of CRA-Net and obtain competitive performance.
Collapse
Affiliation(s)
- Feng Zang
- School of Electronic Engineering, Heilongjiang University, Harbin 150080, China.
| | - Hui Ma
- School of Electronic Engineering, Heilongjiang University, Harbin 150080, China.
| |
Collapse
|
17
|
Guo H, Meng J, Zhao Y, Zhang H, Dai C. High-precision retinal blood vessel segmentation based on a multi-stage and dual-channel deep learning network. Phys Med Biol 2024; 69:045007. [PMID: 38198716 DOI: 10.1088/1361-6560/ad1cf6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/09/2023] [Accepted: 01/10/2024] [Indexed: 01/12/2024]
Abstract
Objective.The high-precision segmentation of retinal vessels in fundus images is important for the early diagnosis of ophthalmic diseases. However, the extraction for microvessels is challenging due to their characteristics of low contrast and high structural complexity. Although some works have been developed to improve the segmentation ability in thin vessels, they have only been successful in recognizing small vessels with relatively high contrast.Approach.Therefore, we develop a deep learning (DL) framework with a multi-stage and dual-channel network model (MSDC_NET) to further improve the thin-vessel segmentation with low contrast. Specifically, an adaptive image enhancement strategy combining multiple preprocessing and the DL method is firstly proposed to elevate the contrast of thin vessels; then, a two-channel model with multi-scale perception is developed to implement whole- and thin-vessel segmentation; and finally, a series of post-processing operations are designed to extract more small vessels in the predicted maps from thin-vessel channels.Main results.Experiments on DRIVE, STARE and CHASE_DB1 demonstrate the superiorities of the proposed MSDC_NET in extracting more thin vessels in fundus images, and quantitative evaluations on several parameters based on the advanced ground truth further verify the advantages of our proposed DL model. Compared with the previous multi-branch method, the specificity and F1score are improved by about 2.18%, 0.68%, 1.73% and 2.91%, 0.24%, 8.38% on the three datasets, respectively.Significance.This work may provide richer information to ophthalmologists for the diagnosis and treatment of vascular-related ophthalmic diseases.
Collapse
Affiliation(s)
- Hui Guo
- School of Computer, Qufu Normal University, 276826 Rizhao, People's Republic of China
| | - Jing Meng
- School of Computer, Qufu Normal University, 276826 Rizhao, People's Republic of China
| | - Yongfu Zhao
- School of Computer, Qufu Normal University, 276826 Rizhao, People's Republic of China
| | - Hongdong Zhang
- School of Computer, Qufu Normal University, 276826 Rizhao, People's Republic of China
| | - Cuixia Dai
- College of Science, Shanghai Institute of Technology, 201418 Shanghai, People's Republic of China
| |
Collapse
|
18
|
Zhao L, Chi H, Zhong T, Jia Y. Perception-oriented generative adversarial network for retinal fundus image super-resolution. Comput Biol Med 2024; 168:107708. [PMID: 37995535 DOI: 10.1016/j.compbiomed.2023.107708] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/02/2023] [Revised: 10/07/2023] [Accepted: 11/15/2023] [Indexed: 11/25/2023]
Abstract
Retinal fundus imaging is a crucial diagnostic tool in ophthalmology, enabling the early detection and monitoring of various ocular diseases. However, capturing high-resolution fundus images often presents challenges due to factors such as defocusing and diffraction in the digital imaging process, limited shutter speed, sensor unit density, and random noise in the image sensor or during image transmission. Super-resolution techniques offer a promising solution to overcome these limitations and enhance the visual details in retinal fundus images. Since the retina has rich texture details, the super-resolution images often introduce artifacts into texture details and lose some fine retinal vessel structures. To improve the perceptual quality of the retinal fundus image, a generative adversarial network that consists of a generator and a discriminator is proposed. The proposed generator mainly comprises 23 multi-scale feature extraction blocks, an image segmentation network, and 23 residual-in-residual dense blocks. These components are employed to extract features at different scales, acquire the retinal vessel grayscale image, and extract retinal vascular features, respectively. The generator has two branches that are mainly responsible for extracting global features and vascular features, respectively. The extracted features from the two branches are fused to better restore the super-resolution image. The proposed generator can restore more details and more accurate fine vessel structures in retinal images. The improved discriminator is proposed by introducing our designed attention modules to help the generator yield clearer super-resolution images. Additionally, an artifact loss function is also introduced to enhance the generative adversarial network, enabling more accurate measurement of the disparity between the high-resolution image and the restored image. Experimental results show that the generated images obtained by our proposed method have a better perceptual quality than the state-of-the-art image super-resolution methods.
Collapse
Affiliation(s)
- Liquan Zhao
- Key Laboratory of Modern Power System Simulation and Control & Renewable Energy Technology, Ministry of Education, Northeast Electric Power University, Jilin, China
| | - Haotian Chi
- Key Laboratory of Modern Power System Simulation and Control & Renewable Energy Technology, Ministry of Education, Northeast Electric Power University, Jilin, China
| | - Tie Zhong
- Key Laboratory of Modern Power System Simulation and Control & Renewable Energy Technology, Ministry of Education, Northeast Electric Power University, Jilin, China.
| | - Yanfei Jia
- College of Electric Power Engineering, Beihua University, Jilin, China
| |
Collapse
|
19
|
Manan MA, Jinchao F, Khan TM, Yaqub M, Ahmed S, Chuhan IS. Semantic segmentation of retinal exudates using a residual encoder-decoder architecture in diabetic retinopathy. Microsc Res Tech 2023; 86:1443-1460. [PMID: 37194727 DOI: 10.1002/jemt.24345] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/11/2023] [Revised: 04/21/2023] [Accepted: 05/04/2023] [Indexed: 05/18/2023]
Abstract
Exudates are a common sign of diabetic retinopathy, which is a disease that affects the blood vessels in the retina. Early detection of exudates is critical to avoiding vision problems through continuous screening and treatment. In traditional clinical practice, the involved lesions are manually detected using photographs of the fundus. However, this task is cumbersome and time-consuming and requires intense effort due to the small size of the lesion and the low contrast of the images. Thus, computer-assisted diagnosis of retinal disease based on the detection of red lesions has been actively explored recently. In this paper, we present a comparison of deep convolutional neural network (CNN) architectures and propose a residual CNN with residual skip connections to reduce the parameter for the semantic segmentation of exudates in retinal images. A suitable image augmentation technique is used to improve the performance of network architecture. The proposed network can robustly segment exudates with high accuracy, which makes it suitable for diabetic retinopathy screening. A comparative performance analysis of three benchmark databases: E-ophtha, DIARETDB1, and Hamilton Ophthalmology Institute's Macular Edema, is presented. The proposed method achieves a precision of 0.95, 0.92, 0.97, accuracy of 0.98, 0.98, 0.98, sensitivity of 0.97, 0.95, 0.95, specificity of 0.99, 0.99, 0.99, and area under the curve of 0.97, 0.94, and 0.96, respectively. RESEARCH HIGHLIGHTS: The research focuses on the detection and segmentation of exudates in diabetic retinopathy, a disease affecting the retina. Early detection of exudates is important to avoid vision problems and requires continuous screening and treatment. Currently, manual detection is time-consuming and requires intense effort. The authors compare qualitative results of the state-of-the-art convolutional neural network (CNN) architectures and propose a computer-assisted diagnosis approach based on deep learning, using a residual CNN with residual skip connections to reduce parameters. The proposed method is evaluated on three benchmark databases and demonstrates high accuracy and suitability for diabetic retinopathy screening.
Collapse
Affiliation(s)
- Malik Abdul Manan
- Beijing Key Laboratory of Computational Intelligence and Intelligent System, Faculty of Information Technology, Beijing University of Technology, Beijing, China
| | - Feng Jinchao
- Beijing Key Laboratory of Computational Intelligence and Intelligent System, Faculty of Information Technology, Beijing University of Technology, Beijing, China
| | - Tariq M Khan
- School of IT, Deakin University, Waurn Ponds, Australia
| | - Muhammad Yaqub
- Beijing Key Laboratory of Computational Intelligence and Intelligent System, Faculty of Information Technology, Beijing University of Technology, Beijing, China
| | - Shahzad Ahmed
- Beijing Key Laboratory of Computational Intelligence and Intelligent System, Faculty of Information Technology, Beijing University of Technology, Beijing, China
| | - Imran Shabir Chuhan
- Interdisciplinary Research Institute, Faculty of Science, Beijing University of Technology, Beijing, China
| |
Collapse
|
20
|
Canales-Fiscal MR, Tamez-Peña JG. Hybrid morphological-convolutional neural networks for computer-aided diagnosis. Front Artif Intell 2023; 6:1253183. [PMID: 37795497 PMCID: PMC10546173 DOI: 10.3389/frai.2023.1253183] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/08/2023] [Accepted: 08/30/2023] [Indexed: 10/06/2023] Open
Abstract
Training deep Convolutional Neural Networks (CNNs) presents challenges in terms of memory requirements and computational resources, often resulting in issues such as model overfitting and lack of generalization. These challenges can only be mitigated by using an excessive number of training images. However, medical image datasets commonly suffer from data scarcity due to the complexities involved in their acquisition, preparation, and curation. To address this issue, we propose a compact and hybrid machine learning architecture based on the Morphological and Convolutional Neural Network (MCNN), followed by a Random Forest classifier. Unlike deep CNN architectures, the MCNN was specifically designed to achieve effective performance with medical image datasets limited to a few hundred samples. It incorporates various morphological operations into a single layer and uses independent neural networks to extract information from each signal channel. The final classification is obtained by utilizing a Random Forest classifier on the outputs of the last neural network layer. We compare the classification performance of our proposed method with three popular deep CNN architectures (ResNet-18, ShuffleNet-V2, and MobileNet-V2) using two training approaches: full training and transfer learning. The evaluation was conducted on two distinct medical image datasets: the ISIC dataset for melanoma classification and the ORIGA dataset for glaucoma classification. Results demonstrate that the MCNN method exhibits reliable performance in melanoma classification, achieving an AUC of 0.94 (95% CI: 0.91 to 0.97), outperforming the popular CNN architectures. For the glaucoma dataset, the MCNN achieved an AUC of 0.65 (95% CI: 0.53 to 0.74), which was similar to the performance of the popular CNN architectures. This study contributes to the understanding of mathematical morphology in shallow neural networks for medical image classification and highlights the potential of hybrid architectures in effectively learning from medical image datasets that are limited by a small number of case samples.
Collapse
|
21
|
Yi Y, Jiang Y, Zhou B, Zhang N, Dai J, Huang X, Zeng Q, Zhou W. C2FTFNet: Coarse-to-fine transformer network for joint optic disc and cup segmentation. Comput Biol Med 2023; 164:107215. [PMID: 37481947 DOI: 10.1016/j.compbiomed.2023.107215] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/27/2023] [Revised: 06/07/2023] [Accepted: 06/25/2023] [Indexed: 07/25/2023]
Abstract
Glaucoma is a leading cause of worldwide blindness and visual impairment, making early screening and diagnosis is crucial to prevent vision loss. Cup-to-Disk Ratio (CDR) evaluation serves as a widely applied approach for effective glaucoma screening. At present, deep learning methods have exhibited outstanding performance in optic disk (OD) and optic cup (OC) segmentation and maturely deployed in CAD system. However, owning to the complexity of clinical data, these techniques could be constrained. Therefore, an original Coarse-to-Fine Transformer Network (C2FTFNet) is designed to segment OD and OC jointly , which is composed of two stages. In the coarse stage, to eliminate the effects of irrelevant organization on the segmented OC and OD regions, we employ U-Net and Circular Hough Transform (CHT) to segment the Region of Interest (ROI) of OD. Meanwhile, a TransUnet3+ model is designed in the fine segmentation stage to extract the OC and OD regions more accurately from ROI. In this model, to alleviate the limitation of the receptive field caused by traditional convolutional methods, a Transformer module is introduced into the backbone to capture long-distance dependent features for retaining more global information. Then, a Multi-Scale Dense Skip Connection (MSDC) module is proposed to fuse the low-level and high-level features from different layers for reducing the semantic gap among different level features. Comprehensive experiments conducted on DRIONS-DB, Drishti-GS, and REFUGE datasets validate the superior effectiveness of the proposed C2FTFNet compared to existing state-of-the-art approaches.
Collapse
Affiliation(s)
- Yugen Yi
- School of Software, Jiangxi Normal University, Nanchang, 330022, China; Jiangxi Provincial Engineering Research Center of Blockchain Data Security and Governance, Nanchang, 330022, China
| | - Yan Jiang
- School of Software, Jiangxi Normal University, Nanchang, 330022, China
| | - Bin Zhou
- School of Software, Jiangxi Normal University, Nanchang, 330022, China
| | - Ningyi Zhang
- School of Software, Jiangxi Normal University, Nanchang, 330022, China
| | - Jiangyan Dai
- School of Computer Engineering, Weifang University, 261061, China.
| | - Xin Huang
- School of Software, Jiangxi Normal University, Nanchang, 330022, China
| | - Qinqin Zeng
- Department of Ophthalmology, The Second Affiliated Hospital of Nanchang University, Nanchang, 330006, China
| | - Wei Zhou
- College of Computer Science, Shenyang Aerospace University, Shenyang, 110136, China.
| |
Collapse
|
22
|
Khan TM, Naqvi SS, Robles-Kelly A, Razzak I. Retinal vessel segmentation via a Multi-resolution Contextual Network and adversarial learning. Neural Netw 2023; 165:310-320. [PMID: 37327578 DOI: 10.1016/j.neunet.2023.05.029] [Citation(s) in RCA: 6] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/27/2022] [Revised: 04/24/2023] [Accepted: 05/17/2023] [Indexed: 06/18/2023]
Abstract
Timely and affordable computer-aided diagnosis of retinal diseases is pivotal in precluding blindness. Accurate retinal vessel segmentation plays an important role in disease progression and diagnosis of such vision-threatening diseases. To this end, we propose a Multi-resolution Contextual Network (MRC-Net) that addresses these issues by extracting multi-scale features to learn contextual dependencies between semantically different features and using bi-directional recurrent learning to model former-latter and latter-former dependencies. Another key idea is training in adversarial settings for foreground segmentation improvement through optimization of the region-based scores. This novel strategy boosts the performance of the segmentation network in terms of the Dice score (and correspondingly Jaccard index) while keeping the number of trainable parameters comparatively low. We have evaluated our method on three benchmark datasets, including DRIVE, STARE, and CHASE, demonstrating its superior performance as compared with competitive approaches elsewhere in the literature.
Collapse
Affiliation(s)
- Tariq M Khan
- School of Computer Science and Engineering, University of New South Wales, Sydney, NSW, Australia.
| | - Syed S Naqvi
- Department of Electrical and Computer Engineering, COMSATS University Islamabad, Pakistan
| | - Antonio Robles-Kelly
- School of Information Technology, Faculty of Science Engineering & Built Environment, Deakin University, Locked Bag 20000, Geelong, Australia; Defence Science and Technology Group, 5111, Edinburgh, SA, Australia
| | - Imran Razzak
- School of Computer Science and Engineering, University of New South Wales, Sydney, NSW, Australia
| |
Collapse
|
23
|
Islam MT, Khan HA, Naveed K, Nauman A, Gulfam SM, Kim SW. LUVS-Net: A Lightweight U-Net Vessel Segmentor for Retinal Vasculature Detection in Fundus Images. ELECTRONICS 2023; 12:1786. [DOI: 10.3390/electronics12081786] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/01/2023]
Abstract
This paper presents LUVS-Net, which is a lightweight convolutional network for retinal vessel segmentation in fundus images that is designed for resource-constrained devices that are typically unable to meet the computational requirements of large neural networks. The computational challenges arise due to low-quality retinal images, wide variance in image acquisition conditions and disparities in intensity. Consequently, the training of existing segmentation methods requires a multitude of trainable parameters for the training of networks, resulting in computational complexity. The proposed Lightweight U-Net for Vessel Segmentation Network (LUVS-Net) can achieve high segmentation performance with only a few trainable parameters. This network uses an encoder–decoder framework in which edge data are transposed from the first layers of the encoder to the last layer of the decoder, massively improving the convergence latency. Additionally, LUVS-Net’s design allows for a dual-stream information flow both inside as well as outside of the encoder–decoder pair. The network width is enhanced using group convolutions, which allow the network to learn a larger number of low- and intermediate-level features. Spatial information loss is minimized using skip connections, and class imbalances are mitigated using dice loss for pixel-wise classification. The performance of the proposed network is evaluated on the publicly available retinal blood vessel datasets DRIVE, CHASE_DB1 and STARE. LUVS-Net proves to be quite competitive, outperforming alternative state-of-the-art segmentation methods and achieving comparable accuracy using trainable parameters that are reduced by two to three orders of magnitude compared with those of comparative state-of-the-art methods.
Collapse
Affiliation(s)
- Muhammad Talha Islam
- Department of Computer Science, COMSATS University Islamabad (CUI), Islamabad 45550, Pakistan
| | - Haroon Ahmed Khan
- Department of Electrical and Computer Engineering, COMSATS University Islamabad (CUI), Islamabad 45550, Pakistan
| | - Khuram Naveed
- Department of Electrical and Computer Engineering, COMSATS University Islamabad (CUI), Islamabad 45550, Pakistan
- Department of Electrical and Computer Engineering, Aarhus University, 8000 Aarhus, Denmark
| | - Ali Nauman
- Department of Information and Communication Engineering, Yeungnam University, Gyeongsan-si 38541, Republic of Korea
| | - Sardar Muhammad Gulfam
- Department of Electrical and Computer Engineering, Abbottabad Campus, COMSATS University Islamabad (CUI), Abbottabad 22060, Pakistan
| | - Sung Won Kim
- Department of Information and Communication Engineering, Yeungnam University, Gyeongsan-si 38541, Republic of Korea
| |
Collapse
|