1
|
Liu S, Zhang Z, Gu Y, Hao J, Liu Y, Fu H, Guo X, Song H, Zhang S, Zhao Y. Beyond the eye: A relational model for early dementia detection using retinal OCTA images. Med Image Anal 2025; 102:103513. [PMID: 40022853 DOI: 10.1016/j.media.2025.103513] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/13/2025] [Accepted: 02/15/2025] [Indexed: 03/04/2025]
Abstract
Early detection of dementia, such as Alzheimer's disease (AD) or mild cognitive impairment (MCI), is essential to enable timely intervention and potential treatment. Accurate detection of AD/MCI is challenging due to the high complexity, cost, and often invasive nature of current diagnostic techniques, which limit their suitability for large-scale population screening. Given the shared embryological origins and physiological characteristics of the retina and brain, retinal imaging is emerging as a potentially rapid and cost-effective alternative for the identification of individuals with or at high risk of AD. In this paper, we present a novel PolarNet+ that uses retinal optical coherence tomography angiography (OCTA) to discriminate early-onset AD (EOAD) and MCI subjects from controls. Our method first maps OCTA images from Cartesian coordinates to polar coordinates, allowing approximate sub-region calculation to implement the clinician-friendly early treatment of diabetic retinopathy study (ETDRS) grid analysis. We then introduce a multi-view module to serialize and analyze the images along three dimensions for comprehensive, clinically useful information extraction. Finally, we abstract the sequence embedding into a graph, transforming the detection task into a general graph classification problem. A regional relationship module is applied after the multi-view module to explore the relationship between the sub-regions. Such regional relationship analyses validate known eye-brain links and reveal new discriminative patterns. The proposed model is trained, tested, and validated on four retinal OCTA datasets, including 1,671 participants with AD, MCI, and healthy controls. Experimental results demonstrate the performance of our model in detecting AD and MCI with an AUC of 88.69% and 88.02%, respectively. Our results provide evidence that retinal OCTA imaging, coupled with artificial intelligence, may serve as a rapid and non-invasive approach for large-scale screening of AD and MCI. The code is available at https://github.com/iMED-Lab/PolarNet-Plus-PyTorch, and the dataset is also available upon request.
Collapse
Affiliation(s)
- Shouyue Liu
- Ningbo Institute of Materials Technology and Engineering, Chinese Academy of Sciences, Ningbo, 315000, China
| | - Ziyi Zhang
- Department of Neurology, West China Hospital, Sichuan University, Chengdu, 610000, China
| | - Yuanyuan Gu
- Ningbo Institute of Materials Technology and Engineering, Chinese Academy of Sciences, Ningbo, 315000, China
| | - Jinkui Hao
- Ningbo Institute of Materials Technology and Engineering, Chinese Academy of Sciences, Ningbo, 315000, China
| | - Yonghuai Liu
- Department of Computer Science, Edge Hill University, Ormskirk, L39 4QP, United Kingdom
| | - Huazhu Fu
- Institute of High-Performance Computing, Agency for Science, Technology and Research, Singapore, 138632, Singapore
| | - Xinyu Guo
- Ningbo Institute of Materials Technology and Engineering, Chinese Academy of Sciences, Ningbo, 315000, China
| | - Hong Song
- School of Computer Science and Technology, Beijing Institute of Technology, Beijing, 100081, China.
| | - Shuting Zhang
- Department of Neurology, West China Hospital, Sichuan University, Chengdu, 610000, China.
| | - Yitian Zhao
- Ningbo Institute of Materials Technology and Engineering, Chinese Academy of Sciences, Ningbo, 315000, China.
| |
Collapse
|
2
|
Huang X, Yue C, Guo Y, Huang J, Jiang Z, Wang M, Xu Z, Zhang G, Liu J, Zhang T, Zheng Z, Zhang X, He H, Jiang S, Sun Y. Multidimensional Directionality-Enhanced Segmentation via large vision model. Med Image Anal 2025; 101:103395. [PMID: 39644753 DOI: 10.1016/j.media.2024.103395] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/09/2024] [Revised: 10/21/2024] [Accepted: 11/15/2024] [Indexed: 12/09/2024]
Abstract
Optical Coherence Tomography (OCT) facilitates a comprehensive examination of macular edema and associated lesions. Manual delineation of retinal fluid is labor-intensive and error-prone, necessitating an automated diagnostic and therapeutic planning mechanism. Conventional supervised learning models are hindered by dataset limitations, while Transformer-based large vision models exhibit challenges in medical image segmentation, particularly in detecting small, subtle lesions in OCT images. This paper introduces the Multidimensional Directionality-Enhanced Retinal Fluid Segmentation framework (MD-DERFS), which reduces the limitations inherent in conventional supervised models by adapting a transformer-based large vision model for macular edema segmentation. The proposed MD-DERFS introduces a Multi-Dimensional Feature Re-Encoder Unit (MFU) to augment the model's proficiency in recognizing specific textures and pathological features through directional prior extraction and an Edema Texture Mapping Unit (ETMU), a Cross-scale Directional Insight Network (CDIN) furnishes a holistic perspective spanning local to global details, mitigating the large vision model's deficiencies in capturing localized feature information. Additionally, the framework is augmented by a Harmonic Minutiae Segmentation Equilibrium loss (LHMSE) that can address the challenges of data imbalance and annotation scarcity in macular edema datasets. Empirical validation on the MacuScan-8k dataset shows that MD-DERFS surpasses existing segmentation methodologies, demonstrating its efficacy in adapting large vision models for boundary-sensitive medical imaging tasks. The code is publicly available at https://github.com/IMOP-lab/MD-DERFS-Pytorch.git.
Collapse
Affiliation(s)
- Xingru Huang
- Hangzhou Dianzi University, Hangzhou, China; School of Electronic Engineering and Computer Science, Queen Mary University, London, UK
| | | | - Yihao Guo
- Hangzhou Dianzi University, Hangzhou, China
| | - Jian Huang
- Hangzhou Dianzi University, Hangzhou, China
| | | | | | - Zhaoyang Xu
- Department of Paediatrics, University of Cambridge, Cambridge, UK
| | - Guangyuan Zhang
- College of Engineering, College of Engineering, Peking University, Beijing, China
| | - Jin Liu
- Hangzhou Dianzi University, Hangzhou, China; School of Remote Sensing and Information Engineering, Wuhan University, Wuhan, China.
| | | | | | - Xiaoshuai Zhang
- Faculty of Information Science and Engineering, Ocean University of China, Qingdao, China.
| | - Hong He
- Hangzhou Dianzi University, Hangzhou, China.
| | | | - Yaoqi Sun
- Hangzhou Dianzi University, Hangzhou, China.
| |
Collapse
|
3
|
Sangchocanonta S, Pooprasert P, Lerthirunvibul N, Patchimnan K, Phienphanich P, Munthuli A, Puangarom S, Itthipanichpong R, Ratanawongphaibul K, Chansangpetch S, Manassakorn A, Tantisevi V, Rojanapongpun P, Tantibundhit C. Optimizing deep learning models for glaucoma screening with vision transformers for resource efficiency and the pie augmentation method. PLoS One 2025; 20:e0314111. [PMID: 40117284 PMCID: PMC11927916 DOI: 10.1371/journal.pone.0314111] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/18/2023] [Accepted: 11/06/2024] [Indexed: 03/23/2025] Open
Abstract
Glaucoma is the leading cause of irreversible vision impairment, emphasizing the critical need for early detection. Typically, AI-based glaucoma screening relies on fundus imaging. To tackle the resource and time challenges in glaucoma screening with convolutional neural network (CNN), we chose the Data-efficient image Transformers (DeiT), a vision transformer, known for its reduced computational demands, with preprocessing time decreased by a factor of 10. Our approach utilized the meticulously annotated GlauCUTU-DATA dataset, curated by ophthalmologists through consensus, encompassing both unanimous agreement (3/3) and majority agreement (2/3) data. However, DeiT's performance was initially lower than CNN. Therefore, we introduced the "pie method," an augmentation method aligned with the ISNT rule. Along with employing polar transformation to improved cup region visibility and alignment with the vision transformer's input to elevated performance levels. The classification results demonstrated improvements comparable to CNN. Using the 3/3 data, excluding the superior and nasal regions, especially in glaucoma suspects, sensitivity increased by 40.18% from 47.06% to 88.24%. The average area under the curve (AUC) ± standard deviation (SD) for glaucoma, glaucoma suspects, and no glaucoma were 92.63 ± 4.39%, 92.35 ± 4.39%, and 92.32 ± 1.45%, respectively. With the 2/3 data, excluding the superior and temporal regions, sensitivity for diagnosing glaucoma increased by 11.36% from 47.73% to 59.09%. The average AUC ± SD for glaucoma, glaucoma suspects, and no glaucoma were 68.22 ± 4.45%, 68.23 ± 4.39%, and 73.09 ± 3.05%, respectively. For both datasets, the AUC values for glaucoma, glaucoma suspects, and no glaucoma were 84.53%, 84.54%, and 91.05%, respectively, which approach the performance of a CNN model that achieved 84.70%, 84.69%, and 93.19%, respectively. Moreover, the incorporation of attention maps from DeiT facilitated the precise localization of clinically significant areas, such as the disc rim and notching, thereby enhancing the overall effectiveness of glaucoma screening.
Collapse
Affiliation(s)
- Sirikorn Sangchocanonta
- Center of Excellence in Nexus for Advanced Intelligence in Law, Engineering, and Medicine (Nail'Em), Faculty of Engineering, Thammasat School of Engineering, Thammasat University, Pathum Thani, Thailand
| | - Pakinee Pooprasert
- Center of Excellence in Nexus for Advanced Intelligence in Law, Engineering, and Medicine (Nail'Em), Faculty of Engineering, Thammasat School of Engineering, Thammasat University, Pathum Thani, Thailand
| | - Nichapa Lerthirunvibul
- Center of Excellence in Nexus for Advanced Intelligence in Law, Engineering, and Medicine (Nail'Em), Faculty of Engineering, Thammasat School of Engineering, Thammasat University, Pathum Thani, Thailand
| | - Kanyarak Patchimnan
- Center of Excellence in Nexus for Advanced Intelligence in Law, Engineering, and Medicine (Nail'Em), Faculty of Engineering, Thammasat School of Engineering, Thammasat University, Pathum Thani, Thailand
| | - Phongphan Phienphanich
- Center of Excellence in Nexus for Advanced Intelligence in Law, Engineering, and Medicine (Nail'Em), Faculty of Engineering, Thammasat School of Engineering, Thammasat University, Pathum Thani, Thailand
| | - Adirek Munthuli
- Center of Excellence in Glaucoma, Department of Ophthalmology, Chulalongkorn University, Bangkok, Thailand
| | - Sujittra Puangarom
- Center of Excellence in Nexus for Advanced Intelligence in Law, Engineering, and Medicine (Nail'Em), Faculty of Engineering, Thammasat School of Engineering, Thammasat University, Pathum Thani, Thailand
| | - Rath Itthipanichpong
- Center of Excellence in Glaucoma, Department of Ophthalmology, Chulalongkorn University, Bangkok, Thailand
- King Chulalongkorn Memorial Hospital, Thai Red Cross Society, Bangkok, Thailand
| | - Kitiya Ratanawongphaibul
- Center of Excellence in Glaucoma, Department of Ophthalmology, Chulalongkorn University, Bangkok, Thailand
- King Chulalongkorn Memorial Hospital, Thai Red Cross Society, Bangkok, Thailand
| | - Sunee Chansangpetch
- Center of Excellence in Glaucoma, Department of Ophthalmology, Chulalongkorn University, Bangkok, Thailand
- King Chulalongkorn Memorial Hospital, Thai Red Cross Society, Bangkok, Thailand
| | - Anita Manassakorn
- Center of Excellence in Glaucoma, Department of Ophthalmology, Chulalongkorn University, Bangkok, Thailand
- King Chulalongkorn Memorial Hospital, Thai Red Cross Society, Bangkok, Thailand
| | - Visanee Tantisevi
- Center of Excellence in Glaucoma, Department of Ophthalmology, Chulalongkorn University, Bangkok, Thailand
- King Chulalongkorn Memorial Hospital, Thai Red Cross Society, Bangkok, Thailand
| | - Prin Rojanapongpun
- Center of Excellence in Glaucoma, Department of Ophthalmology, Chulalongkorn University, Bangkok, Thailand
- King Chulalongkorn Memorial Hospital, Thai Red Cross Society, Bangkok, Thailand
| | - Charturong Tantibundhit
- Center of Excellence in Nexus for Advanced Intelligence in Law, Engineering, and Medicine (Nail'Em), Faculty of Engineering, Thammasat School of Engineering, Thammasat University, Pathum Thani, Thailand
| |
Collapse
|
4
|
Guo X, Wen H, Hao H, Zhao Y, Meng Y, Liu J, Zheng Y, Chen W, Zhao Y. Randomness-Restricted Diffusion Model for Ocular Surface Structure Segmentation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2025; 44:1359-1372. [PMID: 39527437 DOI: 10.1109/tmi.2024.3494762] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2024]
Abstract
Ocular surface diseases affect a significant portion of the population worldwide. Accurate segmentation and quantification of different ocular surface structures are crucial for the understanding of these diseases and clinical decision-making. However, the automated segmentation of the ocular surface structure is relatively unexplored and faces several challenges. Ocular surface structure boundaries are often inconspicuous and obscured by glare from reflections. In addition, the segmentation of different ocular structures always requires training of multiple individual models. Thus, developing a one-model-fits-all segmentation approach is desirable. In this paper, we introduce a randomness-restricted diffusion model for multiple ocular surface structure segmentation. First, a time-controlled fusion-attention module (TFM) is proposed to dynamically adjust the information flow within the diffusion model, based on the temporal relationships between the network's input and time. TFM enables the network to effectively utilize image features to constrain the randomness of the generation process. We further propose a low-frequency consistency filter and a new loss to alleviate model uncertainty and error accumulation caused by the multi-step denoising process. Extensive experiments have shown that our approach can segment seven different ocular surface structures. Our method performs better than both dedicated ocular surface segmentation methods and general medical image segmentation methods. We further validated the proposed method over two clinical datasets, and the results demonstrated that it is beneficial to clinical applications, such as the meibomian gland dysfunction grading and aqueous deficient dry eye diagnosis.
Collapse
|
5
|
Ke X, Chen G, Liu H, Guo W. MEFA-Net: A mask enhanced feature aggregation network for polyp segmentation. Comput Biol Med 2025; 186:109601. [PMID: 39740513 DOI: 10.1016/j.compbiomed.2024.109601] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/27/2024] [Revised: 11/30/2024] [Accepted: 12/18/2024] [Indexed: 01/02/2025]
Abstract
Accurate polyp segmentation is crucial for early diagnosis and treatment of colorectal cancer. This is a challenging task for three main reasons: (i) the problem of model overfitting and weak generalization due to the multi-center distribution of data; (ii) the problem of interclass ambiguity caused by motion blur and overexposure to endoscopic light; and (iii) the problem of intraclass inconsistency caused by the variety of morphologies and sizes of the same type of polyps. To address these challenges, we propose a new high-precision polyp segmentation framework, MEFA-Net, which consists of three modules, including the plug-and-play Mask Enhancement Module (MEG), Separable Path Attention Enhancement Module (SPAE), and Dynamic Global Attention Pool Module (DGAP). Specifically, firstly, the MEG module regionally masks the high-energy regions of the environment and polyps through a mask, which guides the model to rely on only a small amount of information to distinguish between polyps and background features, avoiding the model from overfitting the environmental information, and improving the robustness of the model. At the same time, this module can effectively counteract the "dark corner phenomenon" in the dataset and further improve the generalization performance of the model. Next, the SPAE module can effectively alleviate the inter-class fuzzy problem by strengthening the feature expression. Then, the DGAP module solves the intra-class inconsistency problem by extracting the invariance of scale, shape and position. Finally, we propose a new evaluation metric, MultiColoScore, for comprehensively evaluating the segmentation performance of the model on five datasets with different domains. We evaluated the new method quantitatively and qualitatively on five datasets using four metrics. Experimental results show that MEFA-Net significantly improves the accuracy of polyp segmentation and outperforms current state-of-the-art algorithms. Code posted on https://github.com/847001315/MEFA-Net.
Collapse
Affiliation(s)
- Xiao Ke
- Fujian Provincial Key Laboratory of Networking Computing and Intelligent Information Processing, College of Computer and Data Science, Fuzhou University, Fuzhou 350116, China; Engineering Research Center of Big Data Intelligence, Ministry of Education, Fuzhou 350116, China
| | - Guanhong Chen
- Fujian Provincial Key Laboratory of Networking Computing and Intelligent Information Processing, College of Computer and Data Science, Fuzhou University, Fuzhou 350116, China; Engineering Research Center of Big Data Intelligence, Ministry of Education, Fuzhou 350116, China
| | - Hao Liu
- Fujian Provincial Key Laboratory of Networking Computing and Intelligent Information Processing, College of Computer and Data Science, Fuzhou University, Fuzhou 350116, China; Engineering Research Center of Big Data Intelligence, Ministry of Education, Fuzhou 350116, China
| | - Wenzhong Guo
- Fujian Provincial Key Laboratory of Networking Computing and Intelligent Information Processing, College of Computer and Data Science, Fuzhou University, Fuzhou 350116, China; Engineering Research Center of Big Data Intelligence, Ministry of Education, Fuzhou 350116, China.
| |
Collapse
|
6
|
Chowdhury A, Lodh A, Agarwal R, Garai R, Nandi A, Murmu N, Banerjee S, Nandi D. Rim learning framework based on TS-GAN: A new paradigm of automated glaucoma screening from fundus images. Comput Biol Med 2025; 187:109752. [PMID: 39904104 DOI: 10.1016/j.compbiomed.2025.109752] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2024] [Revised: 01/22/2025] [Accepted: 01/23/2025] [Indexed: 02/06/2025]
Abstract
Glaucoma detection from fundus images often relies on biomarkers such as the Cup-to-Disc Ratio (CDR) and Rim-to-Disc Ratio (RDR). However, precise segmentation of the optic cup and disc is challenging due to low-contrast boundaries and the interference of blood vessels and optic nerves. This article presents a novel automated framework for glaucoma detection that focuses on the rim structure as a biomarker, excluding the conventional reliance on CDR and RDR. The proposed framework introduces a Teacher-Student Generative Adversarial Network (TS-GAN) for precise segmentation of the optic cup and disc, along with a SqueezeNet for glaucoma detection. The Teacher model uses an attention-based CNN encoder-decoder, while the Student model incorporates Expectation Maximization to enhance segmentation performance. By combining implicit generative modeling and explicit probability density modeling, the TS-GAN effectively addresses the mode collapse problem seen in existing GANs. A rim generator processes the segmented cup and disc to produce the rim, which serves as input to SqueezeNet for glaucoma classification. The framework has been extensively tested on diverse fundus image datasets, including a private dataset, demonstrating superior segmentation and detection accuracy compared to state-of-the-art models. Results show its effectiveness in early glaucoma detection, offering higher accuracy and reliability. This innovative framework provides a robust tool for ophthalmologists, enabling efficient glaucoma management and reducing the risk of vision loss.
Collapse
Affiliation(s)
- Arindam Chowdhury
- Department of Computer Science and Engineering, National Institute of Technology, Durgapur 713209, West Bengal, India
| | - Ankit Lodh
- Department of Electronics and Communication Engineering, University Institute of Technology, Burdwan 713104, West Bengal, India
| | - Rohit Agarwal
- Department of Computer Science and Engineering, National Institute of Technology, Durgapur 713209, West Bengal, India
| | - Rahul Garai
- Department of Information Technology, University Institute of Technology, Burdwan 713104, West Bengal, India
| | - Ahana Nandi
- Department of Computer Science and Engineering, Government College of Engineering and Textile Technology, Serampore 712201, West Bengal, India
| | - Narayan Murmu
- Department of Computer Science and Engineering, National Institute of Technology, Durgapur 713209, West Bengal, India
| | - Sumit Banerjee
- Department of Ophthalmology, Megha Eye Centre, Burdwan, West Bengal, India
| | - Debashis Nandi
- Department of Computer Science and Engineering, National Institute of Technology, Durgapur 713209, West Bengal, India.
| |
Collapse
|
7
|
Xu J, Gao J, Jiang S, Wang C, Smedby O, Wu Y, Jiang X, Chen X. Automatic Segmentation of Bone Graft in Maxillary Sinus via Distance Constrained Network Guided by Prior Anatomical Knowledge. IEEE J Biomed Health Inform 2025; 29:1995-2005. [PMID: 40030351 DOI: 10.1109/jbhi.2024.3505262] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/08/2025]
Abstract
Maxillary Sinus Lifting is a crucial surgical procedure for addressing insufficient alveolar bone mass andsevere resorption in dental implant therapy. To accurately analyze the geometry changesof the bone graft (BG) in the maxillary sinus (MS), it is essential to perform quantitative analysis. However, automated BG segmentation remains a major challenge due to the complex local appearance, including blurred boundaries, lesion interference, implant and artifact interference, and BG exceeding the MS. Currently, there are few tools available that can efficiently and accurately segment BG from cone beam computed tomography (CBCT) image. In this paper, we propose a distance-constrained attention network guided by prior anatomical knowledge for the automatic segmentation of BG. First, a guidance strategy of preoperative prior anatomical knowledge is added to a deep neural network (DNN), which improves its ability to identify the dividing line between the MS and BG. Next, a coordinate attention gate is proposed, which utilizes the synergy of channel and position attention to highlight salient features from the skip connections. Additionally, the geodesic distance constraint is introduced into the DNN to form multi-task predictions, which reduces the deviation of the segmentation result. In the test experiment, the proposed DNN achieved a Dice similarity coefficient of 85.48 6.38%, an average surface distance error is 0.57 0.34mm, and a 95% Hausdorff distance of 2.64 2.09mm, which is superior to the comparison networks. It markedly improves the segmentation accuracy and efficiency of BG and has potential applications in analyzing its volume change and absorption rate in the future.
Collapse
|
8
|
Choi M, Jang JS. Heatmap-Based Active Shape Model for Landmark Detection in Lumbar X-ray Images. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025; 38:291-308. [PMID: 39103566 PMCID: PMC11811376 DOI: 10.1007/s10278-024-01210-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/01/2024] [Revised: 07/04/2024] [Accepted: 07/09/2024] [Indexed: 08/07/2024]
Abstract
Medical staff inspect lumbar X-ray images to diagnose lumbar spine diseases, and the analysis process is currently automated using deep-learning techniques. The detection of landmarks is necessary in the automatic process of localizing the position and identifying the morphological features of the vertebrae. However, detection errors may occur owing to the noise and ambiguity of images, as well as individual variations in the shape of the lumbar vertebrae. This study proposes a method to improve the robustness of landmark detection results. This method assumes that landmarks are detected by a convolutional neural network-based two-step model consisting of Pose-Net and M-Net. The model generates a heatmap response to indicate the probable landmark positions. The proposed method then corrects the landmark positions using the heatmap response and active shape model, which employs statistical information on the landmark distribution. Experiments were conducted using 3600 lumbar X-ray images, and the results showed that the landmark detection error was reduced by the proposed method. The average value of maximum errors decreased by 5.58% after applying the proposed method, which combines the outstanding image analysis capabilities of deep learning with statistical shape constraints on landmark distribution. The proposed method could also be easily integrated with other techniques to increase the robustness of landmark detection results such as CoordConv layers and non-directional part affinity field. This resulted in a further enhancement in the landmark detection performance. These advantages can improve the reliability of automatic systems used to inspect lumbar X-ray images. This will benefit both patients and medical staff by reducing medical expenses and increasing diagnostic efficiency.
Collapse
Affiliation(s)
- Minho Choi
- Digital Health Research Division, Korea Institute of Oriental Medicine, 1672 Yuseong-daero, Yuseong-gu, Daejeon, 34054, Republic of Korea
| | - Jun-Su Jang
- Digital Health Research Division, Korea Institute of Oriental Medicine, 1672 Yuseong-daero, Yuseong-gu, Daejeon, 34054, Republic of Korea.
| |
Collapse
|
9
|
Xiao Y, Shao Y, Chen Z, Zhang R, Ding X, Zhao J, Liu S, Fukuyama T, Zhao Y, Peng X, Tian G, Wen S, Zhou X. MIU-Net: Advanced multi-scale feature extraction and imbalance mitigation for optic disc segmentation. Neural Netw 2025; 182:106895. [PMID: 39549494 DOI: 10.1016/j.neunet.2024.106895] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/09/2024] [Revised: 10/07/2024] [Accepted: 11/05/2024] [Indexed: 11/18/2024]
Abstract
Pathological myopia is a severe eye condition that can cause serious complications like retinal detachment and macular degeneration, posing a threat to vision. Optic disc segmentation helps measure changes in the optic disc and observe the surrounding retina, aiding early detection of pathological myopia. However, these changes make segmentation difficult, resulting in accuracy levels that are not suitable for clinical use. To address this, we propose a new model called MIU-Net, which improves segmentation performance through several innovations. First, we introduce a multi-scale feature extraction (MFE) module to capture features at different scales, helping the model better identify optic disc boundaries in complex images. Second, we design a dual attention module that combines channel and spatial attention to focus on important features and improve feature use. To tackle the imbalance between optic disc and background pixels, we use focal loss to enhance the model's ability to detect minority optic disc pixels. We also apply data augmentation techniques to increase data diversity and address the lack of training data. Our model was tested on the iChallenge-PM and iChallenge-AMD datasets, showing clear improvements in accuracy and robustness compared to existing methods. The experimental results demonstrate the effectiveness and potential of our model in diagnosing pathological myopia and other medical image processing tasks.
Collapse
Affiliation(s)
- Yichen Xiao
- Eye Institute and Department of Ophthalmology, Eye and ENT Hospital, Fudan University, Shanghai, 200031, China; NHC Key Laboratory of Myopia (Fudan University), Key Laboratory of Myopia, Chinese Academy of Medical Sciences, Shanghai, 200031, China; Shanghai Research Center of Ophthalmology and Optometry, Shanghai, 200031, China; Shanghai Key Laboratory of Visual Impairment and Restoration, Shanghai, 200031, China.
| | - Yi Shao
- Eye Institute and Department of Ophthalmology, Eye and ENT Hospital, Fudan University, Shanghai, 200031, China; NHC Key Laboratory of Myopia (Fudan University), Key Laboratory of Myopia, Chinese Academy of Medical Sciences, Shanghai, 200031, China; Shanghai Research Center of Ophthalmology and Optometry, Shanghai, 200031, China; Shanghai Key Laboratory of Visual Impairment and Restoration, Shanghai, 200031, China.
| | - Zhi Chen
- Eye Institute and Department of Ophthalmology, Eye and ENT Hospital, Fudan University, Shanghai, 200031, China; NHC Key Laboratory of Myopia (Fudan University), Key Laboratory of Myopia, Chinese Academy of Medical Sciences, Shanghai, 200031, China; Shanghai Research Center of Ophthalmology and Optometry, Shanghai, 200031, China; Shanghai Key Laboratory of Visual Impairment and Restoration, Shanghai, 200031, China.
| | - Ruyi Zhang
- Eye Institute and Department of Ophthalmology, Eye and ENT Hospital, Fudan University, Shanghai, 200031, China.
| | - Xuan Ding
- Eye Institute and Department of Ophthalmology, Eye and ENT Hospital, Fudan University, Shanghai, 200031, China; NHC Key Laboratory of Myopia (Fudan University), Key Laboratory of Myopia, Chinese Academy of Medical Sciences, Shanghai, 200031, China; Shanghai Research Center of Ophthalmology and Optometry, Shanghai, 200031, China; Shanghai Key Laboratory of Visual Impairment and Restoration, Shanghai, 200031, China.
| | - Jing Zhao
- Eye Institute and Department of Ophthalmology, Eye and ENT Hospital, Fudan University, Shanghai, 200031, China; NHC Key Laboratory of Myopia (Fudan University), Key Laboratory of Myopia, Chinese Academy of Medical Sciences, Shanghai, 200031, China; Shanghai Research Center of Ophthalmology and Optometry, Shanghai, 200031, China; Shanghai Key Laboratory of Visual Impairment and Restoration, Shanghai, 200031, China.
| | - Shengtao Liu
- Eye Institute and Department of Ophthalmology, Eye and ENT Hospital, Fudan University, Shanghai, 200031, China; NHC Key Laboratory of Myopia (Fudan University), Key Laboratory of Myopia, Chinese Academy of Medical Sciences, Shanghai, 200031, China; Shanghai Research Center of Ophthalmology and Optometry, Shanghai, 200031, China; Shanghai Key Laboratory of Visual Impairment and Restoration, Shanghai, 200031, China.
| | - Teruko Fukuyama
- Eye Institute and Department of Ophthalmology, Eye and ENT Hospital, Fudan University, Shanghai, 200031, China; NHC Key Laboratory of Myopia (Fudan University), Key Laboratory of Myopia, Chinese Academy of Medical Sciences, Shanghai, 200031, China; Shanghai Research Center of Ophthalmology and Optometry, Shanghai, 200031, China; Shanghai Key Laboratory of Visual Impairment and Restoration, Shanghai, 200031, China.
| | - Yu Zhao
- Eye Institute and Department of Ophthalmology, Eye and ENT Hospital, Fudan University, Shanghai, 200031, China; NHC Key Laboratory of Myopia (Fudan University), Key Laboratory of Myopia, Chinese Academy of Medical Sciences, Shanghai, 200031, China; Shanghai Research Center of Ophthalmology and Optometry, Shanghai, 200031, China; Shanghai Key Laboratory of Visual Impairment and Restoration, Shanghai, 200031, China.
| | - Xiaoliao Peng
- Eye Institute and Department of Ophthalmology, Eye and ENT Hospital, Fudan University, Shanghai, 200031, China; NHC Key Laboratory of Myopia (Fudan University), Key Laboratory of Myopia, Chinese Academy of Medical Sciences, Shanghai, 200031, China; Shanghai Research Center of Ophthalmology and Optometry, Shanghai, 200031, China; Shanghai Key Laboratory of Visual Impairment and Restoration, Shanghai, 200031, China.
| | - Guangyang Tian
- Australian AI Institute, Faculty of Engineering and Information Technology, University of Technology Sydney, Sydney, NSW 2007, Australia.
| | - Shiping Wen
- Australian AI Institute, Faculty of Engineering and Information Technology, University of Technology Sydney, Sydney, NSW 2007, Australia.
| | - Xingtao Zhou
- Eye Institute and Department of Ophthalmology, Eye and ENT Hospital, Fudan University, Shanghai, 200031, China; NHC Key Laboratory of Myopia (Fudan University), Key Laboratory of Myopia, Chinese Academy of Medical Sciences, Shanghai, 200031, China; Shanghai Research Center of Ophthalmology and Optometry, Shanghai, 200031, China; Shanghai Key Laboratory of Visual Impairment and Restoration, Shanghai, 200031, China.
| |
Collapse
|
10
|
Kande GB, Nalluri MR, Manikandan R, Cho J, Veerappampalayam Easwaramoorthy S. Multi scale multi attention network for blood vessel segmentation in fundus images. Sci Rep 2025; 15:3438. [PMID: 39870673 PMCID: PMC11772654 DOI: 10.1038/s41598-024-84255-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2024] [Accepted: 12/20/2024] [Indexed: 01/29/2025] Open
Abstract
Precise segmentation of retinal vasculature is crucial for the early detection, diagnosis, and treatment of vision-threatening ailments. However, this task is challenging due to limited contextual information, variations in vessel thicknesses, the complexity of vessel structures, and the potential for confusion with lesions. In this paper, we introduce a novel approach, the MSMA Net model, which overcomes these challenges by replacing traditional convolution blocks and skip connections with an improved multi-scale squeeze and excitation block (MSSE Block) and Bottleneck residual paths (B-Res paths) with spatial attention blocks (SAB). Our experimental findings on publicly available datasets of fundus images, specifically DRIVE, STARE, CHASE_DB1, HRF and DR HAGIS consistently demonstrate that our approach outperforms other segmentation techniques, achieving higher accuracy, sensitivity, Dice score, and area under the receiver operator characteristic (AUC) in the segmentation of blood vessels with different thicknesses, even in situations involving diverse contextual information, the presence of coexisting lesions, and intricate vessel morphologies.
Collapse
Affiliation(s)
- Giri Babu Kande
- Vasireddy Venkatadri Institute of Technology, Nambur, 522508, India
| | - Madhusudana Rao Nalluri
- School of Computing, Amrita Vishwa Vidyapeetham, Amaravati, 522503, India.
- Department of Computer Science & Engineering, Faculty of Science and Technology (IcfaiTech), ICFAI Foundation for Higher Education, Hyderabad, India.
| | - R Manikandan
- School of Computing, SASTRA Deemed University, Thanjavur, 613401, India
| | - Jaehyuk Cho
- Department of Software Engineering & Division of Electronics and Information Engineering, Jeonbuk National University, Jeonju-si, 54896, Republic of Korea.
| | | |
Collapse
|
11
|
Zhou W, Yang X, Ji J, Yi Y. C 2 MAL: cascaded network-guided class-balanced multi-prototype auxiliary learning for source-free domain adaptive medical image segmentation. Med Biol Eng Comput 2025:10.1007/s11517-025-03287-0. [PMID: 39831950 DOI: 10.1007/s11517-025-03287-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/08/2024] [Accepted: 12/31/2024] [Indexed: 01/22/2025]
Abstract
Source-free domain adaptation (SFDA) has become crucial in medical image analysis, enabling the adaptation of source models across diverse datasets without labeled target domain images. Self-training, a popular SFDA approach, iteratively refines self-generated pseudo-labels using unlabeled target domain data to adapt a pre-trained model from the source domain. However, it often faces model instability due to incorrect pseudo-label accumulation and foreground-background class imbalance. This paper presents a pioneering SFDA framework, named cascaded network-guided class-balanced multi-prototype auxiliary learning (C2 MAL), to enhance model stability. Firstly, we introduce the cascaded translation-segmentation network (CTS-Net), which employs iterative learning between translation and segmentation networks to generate accurate pseudo-labels. The CTS-Net employs a translation network to synthesize target-like images from unreliable predictions of the initial target domain images. The synthesized results refine segmentation network training, ensuring semantic alignment and minimizing visual disparities. Subsequently, reliable pseudo-labels guide the class-balanced multi-prototype auxiliary learning network (CMAL-Net) for effective model adaptation. CMAL-Net incorporates a new multi-prototype auxiliary learning strategy with a memory network to complement source domain data. We propose a class-balanced calibration loss and multi-prototype-guided symmetry cross-entropy loss to tackle class imbalance issue and enhance model adaptability to the target domain. Extensive experiments on four benchmark fundus image datasets validate the superiority of C2 MAL over state-of-the-art methods, especially in scenarios with significant domain shifts. Our code is available at https://github.com/yxk-art/C2MAL .
Collapse
Affiliation(s)
- Wei Zhou
- College of Computer Science, Shenyang Aerospace University, Shenyang, 110136, China.
| | - Xuekun Yang
- College of Computer Science, Shenyang Aerospace University, Shenyang, 110136, China
| | - Jianhang Ji
- Faculty of Data Science, City University of Macau, Macau, China
| | - Yugen Yi
- School of Software, Jiangxi Normal University, Nanchang, 330022, China
| |
Collapse
|
12
|
Huang L, Gao X, Li Y, Lyu F, Gao Y, Bai Y, Ma M, Liu S, Chen J, Ren X, Shang S, Ding X. Enhancing stereotactic ablative boost radiotherapy dose prediction for bulky lung cancer: A multi-scale dilated network approach with scale-balanced structure loss. J Appl Clin Med Phys 2025; 26:e14546. [PMID: 39374302 PMCID: PMC11712318 DOI: 10.1002/acm2.14546] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2024] [Revised: 08/25/2024] [Accepted: 09/21/2024] [Indexed: 10/09/2024] Open
Abstract
PURPOSE Partial stereotactic ablative boost radiotherapy (P-SABR) effectively treats bulky lung cancer; however, the planning process for P-SABR requires repeated dose calculations. To improve planning efficiency, we proposed a novel deep learning method that utilizes limited data to accurately predict the three-dimensional (3D) dose distribution of the P-SABR plan for bulky lung cancer. METHODS We utilized data on 74 patients diagnosed with bulky lung cancer who received P-SABR treatment. The patient dataset was randomly divided into a training set (51 plans) with augmentation, validation set (7 plans), and testing set (16 plans). We devised a 3D multi-scale dilated network (MD-Net) and integrated a scale-balanced structure loss into the loss function. A comparative analysis with a classical network and other advanced networks with multi-scale analysis capabilities and other loss functions was conducted based on the dose distributions in terms of the axial view, average dose scores (ADSs), and average absolute differences of dosimetric indices (AADDIs). Finally, we analyzed the predicted dosimetric indices against the ground-truth values and compared the predicted dose-volume histogram (DVH) with the ground-truth DVH. RESULTS Our proposed dose prediction method for P-SABR plans for bulky lung cancer demonstrated strong performance, exhibiting a significant improvement in predicting multiple indicators of regions of interest (ROIs), particularly the gross target volume (GTV). Our network demonstrated increased accuracy in most dosimetric indices and dose scores in different ROIs. The proposed loss function significantly enhanced the predictive performance of the dosimetric indices. The predicted dosimetric indices and DVHs were equivalent to the ground-truth values. CONCLUSION Our study presents an effective model based on limited datasets, and it exhibits high accuracy in the dose prediction of P-SABR plans for bulky lung cancer. This method has potential as an automated tool for P-SABR planning and can help optimize treatments and improve planning efficiency.
Collapse
Affiliation(s)
- Lei Huang
- Department of Radiation OncologyPeking University First HospitalBeijingChina
| | - Xianshu Gao
- Department of Radiation OncologyPeking University First HospitalBeijingChina
| | - Yue Li
- Department of Medical Biochemistry and BiophysicsKarolinska InstitutetStockholmSweden
| | - Feng Lyu
- Department of Radiation OncologyPeking University First HospitalBeijingChina
| | - Yan Gao
- Department of Radiation OncologyPeking University First HospitalBeijingChina
| | - Yun Bai
- Department of Radiation OncologyPeking University First HospitalBeijingChina
| | - Mingwei Ma
- Department of Radiation OncologyPeking University First HospitalBeijingChina
| | - Siwei Liu
- Department of Radiation OncologyPeking University First HospitalBeijingChina
| | - Jiayan Chen
- Department of Radiation OncologyPeking University First HospitalBeijingChina
| | - Xueying Ren
- Department of Radiation OncologyPeking University First HospitalBeijingChina
| | - Shiyu Shang
- Department of Radiation OncologyPeking University First HospitalBeijingChina
- National Cancer Centre/National Clinical Research Centre for Cancer/Cancer HospitalChinese Academy of Medical Sciences and Peking Union Medical CollegeBeijingChina
| | - Xuanfeng Ding
- Department of Radiation OncologyWilliam Beaumont University Hospital, Cordell HealthRoyal OakMichiganUSA
| |
Collapse
|
13
|
Azad R, Aghdam EK, Rauland A, Jia Y, Avval AH, Bozorgpour A, Karimijafarbigloo S, Cohen JP, Adeli E, Merhof D. Medical Image Segmentation Review: The Success of U-Net. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2024; 46:10076-10095. [PMID: 39167505 DOI: 10.1109/tpami.2024.3435571] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/23/2024]
Abstract
Automatic medical image segmentation is a crucial topic in the medical domain and successively a critical counterpart in the computer-aided diagnosis paradigm. U-Net is the most widespread image segmentation architecture due to its flexibility, optimized modular design, and success in all medical image modalities. Over the years, the U-Net model has received tremendous attention from academic and industrial researchers who have extended it to address the scale and complexity created by medical tasks. These extensions are commonly related to enhancing the U-Net's backbone, bottleneck, or skip connections, or including representation learning, or combining it with a Transformer architecture, or even addressing probabilistic prediction of the segmentation map. Having a compendium of different previously proposed U-Net variants makes it easier for machine learning researchers to identify relevant research questions and understand the challenges of the biological tasks that challenge the model. In this work, we discuss the practical aspects of the U-Net model and organize each variant model into a taxonomy. Moreover, to measure the performance of these strategies in a clinical application, we propose fair evaluations of some unique and famous designs on well-known datasets. Furthermore, we provide a comprehensive implementation library with trained models. In addition, for ease of future studies, we created an online list of U-Net papers with their possible official implementation.
Collapse
|
14
|
Deng H, Wang Y, Cheng V, He Y, Wen Z, Chen S, Guo S, Zhou P, Wang Y. Research trends and hotspots in fundus image segmentation from 2007 to 2023: A bibliometric analysis. Heliyon 2024; 10:e39329. [PMID: 39524903 PMCID: PMC11544040 DOI: 10.1016/j.heliyon.2024.e39329] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/16/2024] [Revised: 09/23/2024] [Accepted: 10/11/2024] [Indexed: 11/16/2024] Open
Abstract
Objective To understand the current status, research hotspots, and trends of automatic segmentation of fundus lesion images worldwide, providing a reference for subsequent related studies. Methods The electronic database Web of Science Core Collection was searched for research in the field of automatic segmentation of fundus lesion images from 2007 to 2023. Visualization maps of countries, authors, institutions, journals, references, and keywords were generated and analyzed using the CiteSpace and VOSviewer software. Results After deduplication, 707 publications were sorted out, showing an overall increasing trend in publication volume. The countries with the highest publication counts were China, followed by India, the USA, the UK, Spain, Pakistan, and Singapore. A high degree of collaboration was observed among authors, and they cooperated widely. The keywords included "diabetic retinopathy," "deep learning," "vessel segmentation," "retinal images," "optic disc localization," and so forth, with keyword bursts starting in 2018 for "retinal images," "machine learning," "biomedical imaging," "deep learning," "convolutional neural networks," and "transfer learning." The most prolific author was U Rajendra Acharya from the University of Southern Queensland, and the journal with the most publications was Computer Methods and Programs in Biomedicine. Conclusions Compared with manual segmentation of fundus lesion images, the use of deep learning models for segmentation is more efficient and accurate, which is crucial for patients with eye diseases. Although the number of related publications globally is relatively small, a growing trend is still witnessed, with broad connections between countries and authors, mainly concentrated in East Asia and Europe. Research institutions in this field are limited, and hence, the research on diabetic retinopathy and retinal vessel segmentation should be strengthened to promote the development of this area.
Collapse
Affiliation(s)
- Hairui Deng
- School of Nursing, Southwest Medical University, Luzhou, 646000, China
- Wound Healing Basic Research and Clinical Application Key Laboratory of Luzhou, School of Nursing, Southwest Medical University, Luzhou, 646000, China
| | - Yiren Wang
- School of Nursing, Southwest Medical University, Luzhou, 646000, China
- Wound Healing Basic Research and Clinical Application Key Laboratory of Luzhou, School of Nursing, Southwest Medical University, Luzhou, 646000, China
| | - Venhui Cheng
- Department of Ophthalmology, The Affiliated Hospital of Southwest Medical University, Luzhou, 646000, China
| | - Yongcheng He
- Department of Pharmacy, Sichuan Agricultural University, Chengdu, 610000, China
| | - Zhongjian Wen
- School of Nursing, Southwest Medical University, Luzhou, 646000, China
- Wound Healing Basic Research and Clinical Application Key Laboratory of Luzhou, School of Nursing, Southwest Medical University, Luzhou, 646000, China
| | - Shouying Chen
- School of Nursing, Southwest Medical University, Luzhou, 646000, China
- Wound Healing Basic Research and Clinical Application Key Laboratory of Luzhou, School of Nursing, Southwest Medical University, Luzhou, 646000, China
| | - Shengmin Guo
- Department of Nursing, The Affiliated Hospital of Southwest Medical University, Luzhou, 646000, China
| | - Ping Zhou
- Wound Healing Basic Research and Clinical Application Key Laboratory of Luzhou, School of Nursing, Southwest Medical University, Luzhou, 646000, China
- Department of Radiology, The Affiliated Hospital of Southwest Medical University, Luzhou, 646000, China
| | - Yi Wang
- Department of Publicity, The Affiliated Hospital of Southwest Medical University, Luzhou, 646000, China
| |
Collapse
|
15
|
Liu K, Zhang J. Development of a Cost-Efficient and Glaucoma-Specialized OD/OC Segmentation Model for Varying Clinical Scenarios. SENSORS (BASEL, SWITZERLAND) 2024; 24:7255. [PMID: 39599032 PMCID: PMC11597940 DOI: 10.3390/s24227255] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/01/2024] [Revised: 10/31/2024] [Accepted: 11/11/2024] [Indexed: 11/29/2024]
Abstract
Most existing optic disc (OD) and cup (OC) segmentation models are biased to the dominant size and easy class (normal class), resulting in suboptimal performances on glaucoma-confirmed samples. Thus, these models are not optimal choices for assisting in tracking glaucoma progression and prognosis. Moreover, fully supervised models employing annotated glaucoma samples can achieve superior performances, although restricted by the high cost of collecting and annotating the glaucoma samples. Therefore, in this paper, we are dedicated to developing a glaucoma-specialized model by exploiting low-cost annotated normal fundus images, simultaneously adapting various common scenarios in clinical practice. We employ a contrastive learning and domain adaptation-based model by exploiting shared knowledge from normal samples. To capture glaucoma-related features, we utilize a Gram matrix to encode style information and the domain adaptation strategy to encode domain information, followed by narrowing the style and domain gaps between normal and glaucoma samples by contrastive and adversarial learning, respectively. To validate the efficacy of our proposed model, we conducted experiments utilizing two public datasets to mimic various common scenarios. The results demonstrate the superior performance of our proposed model across multi-scenarios, showcasing its proficiency in both the segmentation- and glaucoma-related metrics. In summary, our study illustrates a concerted effort to target confirmed glaucoma samples, mitigating the inherent bias issue in most existing models. Moreover, we propose an annotation-efficient strategy that exploits low-cost, normal-labeled fundus samples, mitigating the economic- and labor-related burdens by employing a fully supervised strategy. Simultaneously, our approach demonstrates its adaptability across various scenarios, highlighting its potential utility in both assisting in the monitoring of glaucoma progression and assessing glaucoma prognosis.
Collapse
Affiliation(s)
- Kai Liu
- School of Biological Science and Medical Engineering, Beihang University, Beijing 100083, China;
- Beijing Advanced Innovation Centre for Biomedical Engineering, Beihang University, Beijing 100083, China
- Department of Computer Science, City University of Hong Kong, Hong Kong 98121, China
| | - Jicong Zhang
- School of Biological Science and Medical Engineering, Beihang University, Beijing 100083, China;
- Beijing Advanced Innovation Centre for Biomedical Engineering, Beihang University, Beijing 100083, China
- Hefei Innovation Research Institute, Beihang University, Hefei 230012, China
| |
Collapse
|
16
|
Faisal A, Munilla J, Rahebi J. Detection of optic disc in human retinal images utilizing the Bitterling Fish Optimization (BFO) algorithm. Sci Rep 2024; 14:25824. [PMID: 39468169 PMCID: PMC11519936 DOI: 10.1038/s41598-024-76134-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/13/2024] [Accepted: 10/10/2024] [Indexed: 10/30/2024] Open
Abstract
Early detection and correct identification of the optic disc (OD) on scanned retinal images are significant for diagnosing and treating several ophthalmic conditions, including glaucoma and diabetic retinopathy. Conventional methods for detecting the OD often struggle with processing retinal images due to noise, changes in illumination, and complex overlapping images. This study presents the development of effective and accurate fixation of the optic disc using the Bitterling Fish Optimization (BFO) algorithm, which enhances the processes of OD imaging in speed and precision. The proposed method begins with image enhancement and noise suppression for preprocessing, followed by applying the BFO algorithm to locate and delineate the OD region. The performance evaluation of the algorithm was conducted within several public domain retinal images, including DRIVE, STARE, ORIGA, DRISHTI-GS, DiaRetDB0, and DiaRetDB1 datasets about some internal metrics: sensitivity (SE), specificity (SP), accuracy (ACC), DICE overlap coefficient, overlap and time of processing respectively. The technique based on BFO provided better results, with 99.33%, 99.94%, and 98.22% accuracy achieved for OD in DRIVE, DRISHTI-GS, and DiaRetDB 1, respectively. The approach also demonstrated high overlaps and good DICE results, with a DICE coefficient of 0.9501 for the DRISHTI-GS database. On average, the processing time per image was less than 2.5 s, proving the approach's efficiency in computations. The BFO approach has demonstrated its effectiveness and scalability in detecting optic discs in retinal images in an automated manner. It showed impressive performance levels in terms of computation time and accuracy and was variation resistant irrespective of the quality of the image and the pathology present on it. This method holds significant potential for clinical use, providing a meaningful way of diagnosing and managing ocular disease at an early stage.
Collapse
Affiliation(s)
- Azhar Faisal
- Department of Telecommunication Engineering, Malaga University, Malaga, Spain
| | - Jorge Munilla
- Department of Telecommunication Engineering, Malaga University, Malaga, Spain
| | - Javad Rahebi
- Department of Software Engineering, Istanbul Topkapi University, Istanbul, Turkey.
| |
Collapse
|
17
|
Almarri B, Naveen Kumar B, Aditya Pai H, Bhatia Khan S, Asiri F, Mahesh TR. Redefining retinal vessel segmentation: empowering advanced fundus image analysis with the potential of GANs. Front Med (Lausanne) 2024; 11:1470941. [PMID: 39497847 PMCID: PMC11532151 DOI: 10.3389/fmed.2024.1470941] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2024] [Accepted: 09/13/2024] [Indexed: 11/07/2024] Open
Abstract
Retinal vessel segmentation is a critical task in fundus image analysis, providing essential insights for diagnosing various retinal diseases. In recent years, deep learning (DL) techniques, particularly Generative Adversarial Networks (GANs), have garnered significant attention for their potential to enhance medical image analysis. This paper presents a novel approach for retinal vessel segmentation by harnessing the capabilities of GANs. Our method, termed GANVesselNet, employs a specialized GAN architecture tailored to the intricacies of retinal vessel structures. In GANVesselNet, a dual-path network architecture is employed, featuring an Auto Encoder-Decoder (AED) pathway and a UNet-inspired pathway. This unique combination enables the network to efficiently capture multi-scale contextual information, improving the accuracy of vessel segmentation. Through extensive experimentation on publicly available retinal datasets, including STARE and DRIVE, GANVesselNet demonstrates remarkable performance compared to traditional methods and state-of-the-art deep learning approaches. The proposed GANVesselNet exhibits superior sensitivity (0.8174), specificity (0.9862), and accuracy (0.9827) in segmenting retinal vessels on the STARE dataset, and achieves commendable results on the DRIVE dataset with sensitivity (0.7834), specificity (0.9846), and accuracy (0.9709). Notably, GANVesselNet achieves remarkable performance on previously unseen data, underscoring its potential for real-world clinical applications. Furthermore, we present qualitative visualizations of the generated vessel segmentations, illustrating the network's proficiency in accurately delineating retinal vessels. In summary, this paper introduces GANVesselNet, a novel and powerful approach for retinal vessel segmentation. By capitalizing on the advanced capabilities of GANs and incorporating a tailored network architecture, GANVesselNet offers a quantum leap in retinal vessel segmentation accuracy, opening new avenues for enhanced fundus image analysis and improved clinical decision-making.
Collapse
Affiliation(s)
- Badar Almarri
- Department of Computer Science, College of Computer Sciences and Information Technology, King Faisal University, Alhasa, Saudi Arabia
| | - Baskaran Naveen Kumar
- Department of Computer Science and Engineering, Faculty of Engineering and Technology, JAIN (Deemed-to-be University), Bengaluru, India
| | - Haradi Aditya Pai
- Department of Computer Science and Engineering, MIT School of Computing, MIT Art, Design and Technology University, Pune, India
| | - Surbhi Bhatia Khan
- School of Science, Engineering and Environment, University of Salford, Manchester, United Kingdom
- Adjunct Research Faculty at the Centre for Research Impact & Outcome, Chitkara University, Punjab, India
| | - Fatima Asiri
- College of Computer Science, Informatics and Computer Systems Department, King Khalid University, Abha, Saudi Arabia
| | - Thyluru Ramakrishna Mahesh
- Department of Computer Science and Engineering, Faculty of Engineering and Technology, JAIN (Deemed-to-be University), Bengaluru, India
| |
Collapse
|
18
|
Chen Y, Liu Z, Meng Y, Li J. Lightweight Optic Disc and Optic Cup Segmentation Based on MobileNetv3 Convolutional Neural Network. Biomimetics (Basel) 2024; 9:637. [PMID: 39451843 PMCID: PMC11506706 DOI: 10.3390/biomimetics9100637] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/30/2024] [Revised: 10/07/2024] [Accepted: 10/16/2024] [Indexed: 10/26/2024] Open
Abstract
Glaucoma represents a significant global contributor to blindness. Accurately segmenting the optic disc (OD) and optic cup (OC) to obtain precise CDR is essential for effective screening. However, existing convolutional neural network (CNN)-based segmentation techniques are often limited by high computational demands and long inference times. This paper proposes an efficient end-to-end method for OD and OC segmentation, utilizing the lightweight MobileNetv3 network as the core feature-extraction module. Our approach combines boundary branches with adversarial learning, to achieve multi-label segmentation of the OD and OC. We validated our proposed approach across three public available datasets: Drishti-GS, RIM-ONE-r3, and REFUGE. The outcomes reveal that the Dice coefficients for the segmentation of OD and OC within these datasets are 0.974/0.900, 0.966/0.875, and 0.962/0.880, respectively. Additionally, our method substantially lowers computational complexity and inference time, thereby enabling efficient and precise segmentation of the optic disc and optic cup.
Collapse
Affiliation(s)
- Yuanqiong Chen
- School of Computer Science and Engineering, Central South University, Changsha 410000, China;
- School of Computer Science and Engineering, Jishou University, Zhangjiajie 427000, China
| | - Zhijie Liu
- School of Computer Science and Engineering, Jishou University, Zhangjiajie 427000, China
| | - Yujia Meng
- School of Computer Science and Engineering, Jishou University, Zhangjiajie 427000, China
| | - Jianfeng Li
- School of Computer Science and Engineering, Jishou University, Zhangjiajie 427000, China
| |
Collapse
|
19
|
Abbas S, Qaisar A, Farooq MS, Saleem M, Ahmad M, Khan MA. Smart Vision Transparency: Efficient Ocular Disease Prediction Model Using Explainable Artificial Intelligence. SENSORS (BASEL, SWITZERLAND) 2024; 24:6618. [PMID: 39460097 PMCID: PMC11510864 DOI: 10.3390/s24206618] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/10/2024] [Revised: 09/26/2024] [Accepted: 10/12/2024] [Indexed: 10/28/2024]
Abstract
The early prediction of ocular disease is certainly an obligatory concern in the domain of ophthalmic medicine. Although modern scientific discoveries have shown the potential to treat eye diseases by using artificial intelligence (AI) and machine learning, explainable AI remains a crucial challenge confronting this area of research. Although some traditional methods put in significant effort, they cannot accurately predict the proper ocular diseases. However, incorporating AI into diagnosing eye diseases in healthcare complicates the situation as the decision-making process of AI demonstrates complexity, which is a significant concern, especially in major sectors like ocular disease prediction. The lack of transparency in the AI models may hinder the confidence and trust of the doctors and the patients, as well as their perception of the AI and its abilities. Accordingly, explainable AI is significant in ensuring trust in the technology, enhancing clinical decision-making ability, and deploying ocular disease detection. This research proposed an efficient transfer learning model for eye disease prediction to transform smart vision potential in the healthcare sector and meet conventional approaches' challenges while integrating explainable artificial intelligence (XAI). The integration of XAI in the proposed model ensures the transparency of the decision-making process through the comprehensive provision of rationale. This proposed model provides promising results with 95.74% accuracy and explains the transformative potential of XAI in advancing ocular healthcare. This significant milestone underscores the effectiveness of the proposed model in accurately determining various types of ocular disease. It is clearly shown that the proposed model is performing better than the previously published methods.
Collapse
Affiliation(s)
- Sagheer Abbas
- Department of Computer Science, Prince Mohammad Bin Fahd University, Dhahran 34754, Saudi Arabia
| | - Adnan Qaisar
- Department of Computer Science, Lahore Garrison University, Lahore 54000, Pakistan
| | - Muhammad Sajid Farooq
- Department of Computer Science, Lahore Garrison University, Lahore 54000, Pakistan
- Department of Cyber Security, NASTP Institute of Information Technology, Lahore 54000, Pakistan
| | - Muhammad Saleem
- School of Computer Science, Minhaj University Lahore, Lahore 54000, Pakistan
| | - Munir Ahmad
- College of Informatics, Korea University, Seoul 02841, Republic of Korea
- Department of Computer Science, National College of Business Administration and Economics, Lahore 54000, Pakistan
| | - Muhammad Adnan Khan
- Department of Software, Faculty of Artificial Intelligence and Software, Gachon University, Seongnam-si 13120, Gyeonggi-do, Republic of Korea
| |
Collapse
|
20
|
Lan X, Jin W. Multi-scale input layers and dense decoder aggregation network for COVID-19 lesion segmentation from CT scans. Sci Rep 2024; 14:23729. [PMID: 39390053 PMCID: PMC11467340 DOI: 10.1038/s41598-024-74701-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/02/2024] [Accepted: 09/27/2024] [Indexed: 10/12/2024] Open
Abstract
Accurate segmentation of COVID-19 lesions from medical images is essential for achieving precise diagnosis and developing effective treatment strategies. Unfortunately, this task presents significant challenges, owing to the complex and diverse characteristics of opaque areas, subtle differences between infected and healthy tissue, and the presence of noise in CT images. To address these difficulties, this paper designs a new deep-learning architecture (named MD-Net) based on multi-scale input layers and dense decoder aggregation network for COVID-19 lesion segmentation. In our framework, the U-shaped structure serves as the cornerstone to facilitate complex hierarchical representations essential for accurate segmentation. Then, by introducing the multi-scale input layers (MIL), the network can effectively analyze both fine-grained details and contextual information in the original image. Furthermore, we introduce an SE-Conv module in the encoder network, which can enhance the ability to identify relevant information while simultaneously suppressing the transmission of extraneous or non-lesion information. Additionally, we design a dense decoder aggregation (DDA) module to integrate feature distributions and important COVID-19 lesion information from adjacent encoder layers. Finally, we conducted a comprehensive quantitative analysis and comparison between two publicly available datasets, namely Vid-QU-EX and QaTa-COV19-v2, to assess the robustness and versatility of MD-Net in segmenting COVID-19 lesions. The experimental results show that the proposed MD-Net has superior performance compared to its competitors, and it exhibits higher scores on the Dice value, Matthews correlation coefficient (Mcc), and Jaccard index. In addition, we also conducted ablation studies on the Vid-QU-EX dataset to evaluate the contributions of each key component within the proposed architecture.
Collapse
Affiliation(s)
- Xiaoke Lan
- College of Internet of Things Technology, Hangzhou Polytechnic, Hangzhou, 311402, China.
| | - Wenbing Jin
- College of Internet of Things Technology, Hangzhou Polytechnic, Hangzhou, 311402, China
| |
Collapse
|
21
|
Jiang W, Li Y, Jia Y, Feng Y, Yi Z, Wang J, Chen M. Segmentation of coronary artery based on discriminative frequency learning and coronary-geometric refinement. Comput Biol Med 2024; 181:109045. [PMID: 39180858 DOI: 10.1016/j.compbiomed.2024.109045] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/29/2023] [Revised: 07/10/2024] [Accepted: 08/17/2024] [Indexed: 08/27/2024]
Abstract
Coronary artery segmentation is crucial for physicians to identify and locate plaques and stenosis using coronary computed tomography angiography (CCTA). However, the low contrast of CCTA images and the intricate structures of coronary arteries make this task challenging. To address these difficulties, we propose a novel model, the DFS-PDS network. This network comprises two subnetworks: a discriminative frequency segment subnetwork (DFS) and a position domain scales subnetwork (PDS). DFS introduced a gated mechanism within the feed-forward network, leveraging the Joint Photographic Experts Group (JPEG) compression algorithm, to discriminatively determine which low- and high-frequency information of the features should be preserved for latent image segmentation. The PDS aims to learn the shape prototype by predicting the radius. Additionally, our model has the consistent ability to guarantee region and boundary features through boundary consistency loss. During training, both subnetworks are optimized jointly, and in the testing stage, the coarse segmentation and radius prediction are produced. A coronary-geometric refinement method refines the segmentation masks by leveraging the shape prior to being reconstructed from the radius map, reducing the difficulty of segmenting coronary artery structures from complex surrounding structures. The DFS-PDS network is compared with state-of-the-art (SOTA) methods on two coronary artery datasets to evaluate its performance. The experimental results demonstrate that the DFS-PDS network performs better than the SOTA models, including Vnet, nnUnet, DDT, CS2-Net, Unetr, and CAS-Net, in terms of Dice or connectivity evaluation metrics.
Collapse
Affiliation(s)
- Weili Jiang
- Machine Intelligence Laboratory, College of Computer Science, Sichuan University, Chengdu, Sichuan, PR China
| | - Yiming Li
- Department of Cardiology, West China Hospital, Sichuan University, Chengdu, Sichuan, PR China
| | - Yuheng Jia
- Department of Cardiology, West China Hospital, Sichuan University, Chengdu, Sichuan, PR China
| | - Yuan Feng
- Department of Cardiology, West China Hospital, Sichuan University, Chengdu, Sichuan, PR China
| | - Zhang Yi
- Machine Intelligence Laboratory, College of Computer Science, Sichuan University, Chengdu, Sichuan, PR China
| | - Jianyong Wang
- Machine Intelligence Laboratory, College of Computer Science, Sichuan University, Chengdu, Sichuan, PR China.
| | - Mao Chen
- Department of Cardiology, West China Hospital, Sichuan University, Chengdu, Sichuan, PR China.
| |
Collapse
|
22
|
Huang J, Luo Y, Guo Y, Li W, Wang Z, Liu G, Yang G. Accurate segmentation of intracellular organelle networks using low-level features and topological self-similarity. Bioinformatics 2024; 40:btae559. [PMID: 39302662 PMCID: PMC11467052 DOI: 10.1093/bioinformatics/btae559] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/18/2024] [Revised: 08/12/2024] [Accepted: 09/19/2024] [Indexed: 09/22/2024] Open
Abstract
MOTIVATION Intracellular organelle networks (IONs) such as the endoplasmic reticulum (ER) network and the mitochondrial (MITO) network serve crucial physiological functions. The morphology of these networks plays a critical role in mediating their functions. Accurate image segmentation is required for analyzing the morphology and topology of these networks for applications such as molecular mechanism analysis and drug target screening. So far, however, progress has been hindered by their structural complexity and density. RESULTS In this study, we first establish a rigorous performance baseline for accurate segmentation of these organelle networks from fluorescence microscopy images by optimizing a baseline U-Net model. We then develop the multi-resolution encoder (MRE) and the hierarchical fusion loss (Lhf) based on two inductive components, namely low-level features and topological self-similarity, to assist the model in better adapting to the task of segmenting IONs. Empowered by MRE and Lhf, both U-Net and Pyramid Vision Transformer (PVT) outperform competing state-of-the-art models such as U-Net++, HR-Net, nnU-Net, and TransUNet on custom datasets of the ER network and the MITO network, as well as on public datasets of another biological network, the retinal blood vessel network. In addition, integrating MRE and Lhf with models such as HR-Net and TransUNet also enhances their segmentation performance. These experimental results confirm the generalization capability and potential of our approach. Furthermore, accurate segmentation of the ER network enables analysis that provides novel insights into its dynamic morphological and topological properties. AVAILABILITY AND IMPLEMENTATION Code and data are openly accessible at https://github.com/cbmi-group/MRE.
Collapse
Affiliation(s)
- Jiaxing Huang
- State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100049, China
| | - Yaoru Luo
- State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100049, China
| | - Yuanhao Guo
- State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100049, China
| | - Wenjing Li
- State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100049, China
| | - Zichen Wang
- State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100049, China
| | - Guole Liu
- State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100049, China
| | - Ge Yang
- State Key Laboratory of Multimodal Artificial Intelligence Systems, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100049, China
| |
Collapse
|
23
|
Zhou W, Ji J, Cui W, Wang Y, Yi Y. Unsupervised Domain Adaptation Fundus Image Segmentation via Multi-Scale Adaptive Adversarial Learning. IEEE J Biomed Health Inform 2024; 28:5792-5803. [PMID: 38090822 DOI: 10.1109/jbhi.2023.3342422] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/19/2023]
Abstract
Segmentation of the Optic Disc (OD) and Optic Cup (OC) is crucial for the early detection and treatment of glaucoma. Despite the strides made in deep neural networks, incorporating trained segmentation models for clinical application remains challenging due to domain shifts arising from disparities in fundus images across different healthcare institutions. To tackle this challenge, this study introduces an innovative unsupervised domain adaptation technique called Multi-scale Adaptive Adversarial Learning (MAAL), which consists of three key components. The Multi-scale Wasserstein Patch Discriminator (MWPD) module is designed to extract domain-specific features at multiple scales, enhancing domain classification performance and offering valuable guidance for the segmentation network. To further enhance model generalizability and explore domain-invariant features, we introduce the Adaptive Weighted Domain Constraint (AWDC) module. During training, this module dynamically assigns varying weights to different scales, allowing the model to adaptively focus on informative features. Furthermore, the Pixel-level Feature Enhancement (PFE) module enhances low-level features extracted at shallow network layers by incorporating refined high-level features. This integration ensures the preservation of domain-invariant information, effectively addressing domain variation and mitigating the loss of global features. Two publicly accessible fundus image databases are employed to demonstrate the effectiveness of our MAAL method in mitigating model degradation and improving segmentation performance. The achieved results outperform current state-of-the-art (SOTA) methods in both OD and OC segmentation.
Collapse
|
24
|
Chen Z, Pan Y, Ye Y, Wang Z, Xia Y. TriLA: Triple-Level Alignment Based Unsupervised Domain Adaptation for Joint Segmentation of Optic Disc and Optic Cup. IEEE J Biomed Health Inform 2024; 28:5497-5508. [PMID: 38805331 DOI: 10.1109/jbhi.2024.3406447] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/30/2024]
Abstract
Cross-domain joint segmentation of optic disc and optic cup on fundus images is essential, yet challenging, for effective glaucoma screening. Although many unsupervised domain adaptation (UDA) methods have been proposed, these methods can hardly achieve complete domain alignment, leading to suboptimal performance. In this paper, we propose a triple-level alignment (TriLA) model to address this issue by aligning the source and target domains at the input level, feature level, and output level simultaneously. At the input level, a learnable Fourier domain adaptation (LFDA) module is developed to learn the cut-off frequency adaptively for frequency-domain translation. At the feature level, we disentangle the style and content features and align them in the corresponding feature spaces using consistency constraints. At the output level, we design a segmentation consistency constraint to emphasize the segmentation consistency across domains. The proposed model is trained on the RIGA+ dataset and widely evaluated on six different UDA scenarios. Our comprehensive results not only demonstrate that the proposed TriLA substantially outperforms other state-of-the-art UDA methods in joint segmentation of optic disc and optic cup, but also suggest the effectiveness of the triple-level alignment strategy.
Collapse
|
25
|
Chaurasia AK, Greatbatch CJ, Han X, Gharahkhani P, Mackey DA, MacGregor S, Craig JE, Hewitt AW. Highly Accurate and Precise Automated Cup-to-Disc Ratio Quantification for Glaucoma Screening. OPHTHALMOLOGY SCIENCE 2024; 4:100540. [PMID: 39051045 PMCID: PMC11268341 DOI: 10.1016/j.xops.2024.100540] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 08/24/2023] [Revised: 03/26/2024] [Accepted: 04/22/2024] [Indexed: 07/27/2024]
Abstract
Objective An enlarged cup-to-disc ratio (CDR) is a hallmark of glaucomatous optic neuropathy. Manual assessment of the CDR may be less accurate and more time-consuming than automated methods. Here, we sought to develop and validate a deep learning-based algorithm to automatically determine the CDR from fundus images. Design Algorithm development for estimating CDR using fundus data from a population-based observational study. Participants A total of 181 768 fundus images from the United Kingdom Biobank (UKBB), Drishti_GS, and EyePACS. Methods FastAI and PyTorch libraries were used to train a convolutional neural network-based model on fundus images from the UKBB. Models were constructed to determine image gradability (classification analysis) as well as to estimate CDR (regression analysis). The best-performing model was then validated for use in glaucoma screening using a multiethnic dataset from EyePACS and Drishti_GS. Main Outcome Measures The area under the receiver operating characteristic curve and coefficient of determination. Results Our gradability model vgg19_batch normalization (bn) achieved an accuracy of 97.13% on a validation set of 16 045 images, with 99.26% precision and area under the receiver operating characteristic curve of 96.56%. Using regression analysis, our best-performing model (trained on the vgg19_bn architecture) attained a coefficient of determination of 0.8514 (95% confidence interval [CI]: 0.8459-0.8568), while the mean squared error was 0.0050 (95% CI: 0.0048-0.0051) and mean absolute error was 0.0551 (95% CI: 0.0543-0.0559) on a validation set of 12 183 images for determining CDR. The regression point was converted into classification metrics using a tolerance of 0.2 for 20 classes; the classification metrics achieved an accuracy of 99.20%. The EyePACS dataset (98 172 healthy, 3270 glaucoma) was then used to externally validate the model for glaucoma classification, with an accuracy, sensitivity, and specificity of 82.49%, 72.02%, and 82.83%, respectively. Conclusions Our models were precise in determining image gradability and estimating CDR. Although our artificial intelligence-derived CDR estimates achieve high accuracy, the CDR threshold for glaucoma screening will vary depending on other clinical parameters. Financial Disclosures Proprietary or commercial disclosure may be found in the Footnotes and Disclosures at the end of this article.
Collapse
Affiliation(s)
- Abadh K. Chaurasia
- Menzies Institute for Medical Research, University of Tasmania, Hobart, Australia
| | - Connor J. Greatbatch
- Menzies Institute for Medical Research, University of Tasmania, Hobart, Australia
| | - Xikun Han
- QIMR Berghofer Medical Research Institute, Brisbane, Australia
- School of Medicine, University of Queensland, Brisbane, Australia
| | - Puya Gharahkhani
- QIMR Berghofer Medical Research Institute, Brisbane, Australia
- School of Medicine, University of Queensland, Brisbane, Australia
- Faculty of Health, School of Biomedical Sciences, Queensland University of Technology, Brisbane, Queensland, Australia
| | - David A. Mackey
- Lions Eye Institute, Centre for Vision Sciences, University of Western Australia, Nedlands, Australia
| | - Stuart MacGregor
- QIMR Berghofer Medical Research Institute, Brisbane, Australia
- School of Medicine, University of Queensland, Brisbane, Australia
| | - Jamie E. Craig
- Department of Ophthalmology, Flinders University, Flinders Medical Centre, Bedford Park, Australia
| | - Alex W. Hewitt
- Menzies Institute for Medical Research, University of Tasmania, Hobart, Australia
- Centre for Eye Research Australia, University of Melbourne, Melbourne, Australia
| |
Collapse
|
26
|
Eswari MS, Balamurali S, Ramasamy LK. Hybrid convolutional neural network optimized with an artificial algae algorithm for glaucoma screening using fundus images. J Int Med Res 2024; 52:3000605241271766. [PMID: 39301801 PMCID: PMC11539265 DOI: 10.1177/03000605241271766] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/09/2023] [Accepted: 06/05/2024] [Indexed: 09/22/2024] Open
Abstract
OBJECTIVE We developed an optimized decision support system for retinal fundus image-based glaucoma screening. METHODS We combined computer vision algorithms with a convolutional network for fundus images and applied a faster region-based convolutional neural network (FRCNN) and artificial algae algorithm with support vector machine (AAASVM) classifiers. Optic boundary detection, optic cup, and optic disc segmentations were conducted using TernausNet. Glaucoma screening was performed using the optimized FRCNN. The Softmax layer was replaced with an SVM classifier layer and optimized with an AAA to attain enhanced accuracy. RESULTS Using three retinal fundus image datasets (G1020, digital retinal images vessel extraction, and high-resolution fundus), we obtained accuracy of 95.11%, 92.87%, and 93.7%, respectively. Framework accuracy was amplified with an adaptive gradient algorithm optimizer FRCNN (AFRCNN), which achieved average accuracy 94.06%, sensitivity 93.353%, and specificity 94.706%. AAASVM obtained average accuracy of 96.52%, which was 3% ahead of the FRCNN classifier. These classifiers had areas under the curve of 0.9, 0.85, and 0.87, respectively. CONCLUSION Based on statistical Friedman evaluation, AAASVM was the best glaucoma screening model. Segmented and classified images can be directed to the health care system to assess patients' progress. This computer-aided decision support system will be useful for optometrists.
Collapse
Affiliation(s)
- M. Shanmuga Eswari
- Department of Computer Applications, Kalasalingam Academy of Research and Education, Krishnankoil, Tamilnadu, India
| | - S. Balamurali
- Department of Computer Applications, Kalasalingam Academy of Research and Education, Krishnankoil, Tamilnadu, India
| | | |
Collapse
|
27
|
Qiu J, Liu W, Lin C, Li J, Yu H, Boumaraf S. Occlusion-aware deep convolutional neural network via homogeneous Tanh-transforms for face parsing. IMAGE AND VISION COMPUTING 2024; 148:105120. [DOI: 10.1016/j.imavis.2024.105120] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/06/2025]
|
28
|
Luengnaruemitchai G, Sangchocanonta S, Munthuli A, Phienphanich P, Puangarom S, Jariyakosol S, Hirunwiwatkul P, Tantibundhit C. Automated Alzheimer's, Mild Cognitive Impairment, and Normal Aging Screening using Polar Transformation of Optic Disc and Central Zone of Fundus Images. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2024; 2024:1-4. [PMID: 40038980 DOI: 10.1109/embc53108.2024.10782014] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/06/2025]
Abstract
Detecting Mild Cognitive Impairment (MCI) is crucial for mitigating the risk of Alzheimer's disease (AD), a leading global cause of death. However, the current gold standard for AD and MCI detection relies on specialized equipment often limited to large testing centers, particularly in low-resource settings like Thailand. Our previous work aimed to create a cost-effective MCI and AD screening method using fundus images but struggled to differentiate between AD and MCI. Henceforth, we developed the proposed methodology, utilizing DenseNet-121 on polar-transformed and zone-selected fundus images, which significantly enhances AD and MCI classification, achieving 83% accuracy, 90% sensitivity, 77% specificity, 87% precision, and an F-1 score of 88%. Moreover, the model's Grad-Cam++ heatmap highlights vasculature differences, particularly in tortuosity and thickness, between AD and MCI fundus images. Combined with our previous work, we created a fully automated pipeline model for MCI, AD, and Normal aging classification, which is inexpensive, fast, and non-invasive with an overall 3-class accuracy of 88%.
Collapse
|
29
|
Meng Y, Zhang Y, Xie J, Duan J, Joddrell M, Madhusudhan S, Peto T, Zhao Y, Zheng Y. Multi-granularity learning of explicit geometric constraint and contrast for label-efficient medical image segmentation and differentiable clinical function assessment. Med Image Anal 2024; 95:103183. [PMID: 38692098 DOI: 10.1016/j.media.2024.103183] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/04/2023] [Revised: 01/26/2024] [Accepted: 04/18/2024] [Indexed: 05/03/2024]
Abstract
Automated segmentation is a challenging task in medical image analysis that usually requires a large amount of manually labeled data. However, most current supervised learning based algorithms suffer from insufficient manual annotations, posing a significant difficulty for accurate and robust segmentation. In addition, most current semi-supervised methods lack explicit representations of geometric structure and semantic information, restricting segmentation accuracy. In this work, we propose a hybrid framework to learn polygon vertices, region masks, and their boundaries in a weakly/semi-supervised manner that significantly advances geometric and semantic representations. Firstly, we propose multi-granularity learning of explicit geometric structure constraints via polygon vertices (PolyV) and pixel-wise region (PixelR) segmentation masks in a semi-supervised manner. Secondly, we propose eliminating boundary ambiguity by using an explicit contrastive objective to learn a discriminative feature space of boundary contours at the pixel level with limited annotations. Thirdly, we exploit the task-specific clinical domain knowledge to differentiate the clinical function assessment end-to-end. The ground truth of clinical function assessment, on the other hand, can serve as auxiliary weak supervision for PolyV and PixelR learning. We evaluate the proposed framework on two tasks, including optic disc (OD) and cup (OC) segmentation along with vertical cup-to-disc ratio (vCDR) estimation in fundus images; left ventricle (LV) segmentation at end-diastolic and end-systolic frames along with ejection fraction (LVEF) estimation in two-dimensional echocardiography images. Experiments on nine large-scale datasets of the two tasks under different label settings demonstrate our model's superior performance on segmentation and clinical function assessment.
Collapse
Affiliation(s)
- Yanda Meng
- Department of Eye and Vision Sciences, University of Liverpool, Liverpool, United Kingdom
| | - Yuchen Zhang
- Center for Bioinformatics, Peking University, Beijing, China
| | - Jianyang Xie
- Department of Eye and Vision Sciences, University of Liverpool, Liverpool, United Kingdom
| | - Jinming Duan
- School of Computer Science, University of Birmingham, Birmingham, United Kingdom
| | - Martha Joddrell
- Liverpool Centre for Cardiovascular Science, University of Liverpool and Liverpool Heart & Chest Hospital, Liverpool, United Kingdom; Department of Cardiovascular and Metabolic Medicine, University of Liverpool, Liverpool, United Kingdom
| | - Savita Madhusudhan
- St Paul's Eye Unit, Liverpool University Hospitals NHS Foundation Trust, Liverpool, United Kingdom
| | - Tunde Peto
- School of Medicine, Dentistry and Biomedical Sciences, Queen's University Belfast, Belfast, United Kingdom
| | - Yitian Zhao
- Institute of Biomedical Engineering, Ningbo Institute of Materials Technology and Engineering, Chinese Academy of Science, Ningbo, China; Ningbo Eye Hospital, Ningbo, China.
| | - Yalin Zheng
- Department of Eye and Vision Sciences, University of Liverpool, Liverpool, United Kingdom; Liverpool Centre for Cardiovascular Science, University of Liverpool and Liverpool Heart & Chest Hospital, Liverpool, United Kingdom.
| |
Collapse
|
30
|
He H, Qiu J, Lin L, Cai Z, Cheng P, Tang X. JOINEDTrans: Prior guided multi-task transformer for joint optic disc/cup segmentation and fovea detection. Comput Biol Med 2024; 177:108613. [PMID: 38781644 DOI: 10.1016/j.compbiomed.2024.108613] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/30/2023] [Revised: 01/18/2024] [Accepted: 05/11/2024] [Indexed: 05/25/2024]
Abstract
Deep learning-based image segmentation and detection models have largely improved the efficiency of analyzing retinal landmarks such as optic disc (OD), optic cup (OC), and fovea. However, factors including ophthalmic disease-related lesions and low image quality issues may severely complicate automatic OD/OC segmentation and fovea detection. Most existing works treat the identification of each landmark as a single task, and take into account no prior information. To address these issues, we propose a prior guided multi-task transformer framework for joint OD/OC segmentation and fovea detection, named JOINEDTrans. JOINEDTrans effectively combines various spatial features of the fundus images, relieving the structural distortions induced by lesions and other imaging issues. It contains a segmentation branch and a detection branch. To be noted, we employ an encoder with prior-learning in a vessel segmentation task to effectively exploit the positional relationship among vessel, OD/OC, and fovea, successfully incorporating spatial prior into the proposed JOINEDTrans framework. There are a coarse stage and a fine stage in JOINEDTrans. In the coarse stage, OD/OC coarse segmentation and fovea heatmap localization are obtained through a joint segmentation and detection module. In the fine stage, we crop regions of interest for subsequent refinement and use predictions obtained in the coarse stage to provide additional information for better performance and faster convergence. Experimental results demonstrate that JOINEDTrans outperforms existing state-of-the-art methods on the publicly available GAMMA, REFUGE, and PALM fundus image datasets. We make our code available at https://github.com/HuaqingHe/JOINEDTrans.
Collapse
Affiliation(s)
- Huaqing He
- Department of Electronic and Electrical Engineering, Southern University of Science and Technology, Shenzhen, Guangdong, China; Jiaxing Research Institute, Southern University of Science and Technology, Jiaxing, Zhejiang, China.
| | - Jiaming Qiu
- Department of Electronic and Electrical Engineering, Southern University of Science and Technology, Shenzhen, Guangdong, China.
| | - Li Lin
- Department of Electronic and Electrical Engineering, Southern University of Science and Technology, Shenzhen, Guangdong, China; Jiaxing Research Institute, Southern University of Science and Technology, Jiaxing, Zhejiang, China; Department of Electrical and Electronic Engineering, The University of Hong Kong, Hong Kong, China.
| | - Zhiyuan Cai
- Department of Electronic and Electrical Engineering, Southern University of Science and Technology, Shenzhen, Guangdong, China; Department of Computer Science and Engineering, The Hong Kong University of Science and Technology, Hong Kong, China.
| | - Pujin Cheng
- Department of Electronic and Electrical Engineering, Southern University of Science and Technology, Shenzhen, Guangdong, China; Department of Electrical and Electronic Engineering, The University of Hong Kong, Hong Kong, China.
| | - Xiaoying Tang
- Department of Electronic and Electrical Engineering, Southern University of Science and Technology, Shenzhen, Guangdong, China; Jiaxing Research Institute, Southern University of Science and Technology, Jiaxing, Zhejiang, China.
| |
Collapse
|
31
|
Chen N, Lv X. Research on segmentation model of optic disc and optic cup in fundus. BMC Ophthalmol 2024; 24:273. [PMID: 38943095 PMCID: PMC11214242 DOI: 10.1186/s12886-024-03532-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/04/2024] [Accepted: 06/20/2024] [Indexed: 07/01/2024] Open
Abstract
BACKGROUND Glaucoma is a worldwide eye disease that can cause irreversible vision loss. Early detection of glaucoma is important to reduce vision loss, and retinal fundus image examination is one of the most commonly used solutions for glaucoma diagnosis due to its low cost. Clinically, the cup-disc ratio of fundus images is an important indicator for glaucoma diagnosis. In recent years, there have been an increasing number of algorithms for segmentation and recognition of the optic disc (OD) and optic cup (OC), but these algorithms generally have poor universality, segmentation performance, and segmentation accuracy. METHODS By improving the YOLOv8 algorithm for segmentation of OD and OC. Firstly, a set of algorithms was designed to adapt the REFUGE dataset's result images to the input format of the YOLOv8 algorithm. Secondly, in order to improve segmentation performance, the network structure of YOLOv8 was improved, including adding a ROI (Region of Interest) module, modifying the bounding box regression loss function from CIOU to Focal-EIoU. Finally, by training and testing the REFUGE dataset, the improved YOLOv8 algorithm was evaluated. RESULTS The experimental results show that the improved YOLOv8 algorithm achieves good segmentation performance on the REFUGE dataset. In the OD and OC segmentation tests, the F1 score is 0.999. CONCLUSIONS We improved the YOLOv8 algorithm and applied the improved model to the segmentation task of OD and OC in fundus images. The results show that our improved model is far superior to the mainstream U-Net model in terms of training speed, segmentation performance, and segmentation accuracy.
Collapse
Affiliation(s)
- Naigong Chen
- National Clinical Research Center for Ocular Diseases, Eye Hospital, Wenzhou Medical University, Wenzhou, 325000, China.
- State Key Laboratory of Ophthalmology, Optometry and Vision Science, Eye Hospital, Wenzhou Medical University, Wenzhou, 325027, China.
- National Engineering Research Center of Ophthalmology and Optometry, Eye Hospital, Wenzhou Medical University, Wenzhou, 325027, China.
| | - Xiujuan Lv
- National Clinical Research Center for Ocular Diseases, Eye Hospital, Wenzhou Medical University, Wenzhou, 325000, China
- State Key Laboratory of Ophthalmology, Optometry and Vision Science, Eye Hospital, Wenzhou Medical University, Wenzhou, 325027, China
- National Engineering Research Center of Ophthalmology and Optometry, Eye Hospital, Wenzhou Medical University, Wenzhou, 325027, China
| |
Collapse
|
32
|
Likassa HT, Chen DG, Chen K, Wang Y, Zhu W. Robust PCA with Lw,∗ and L2,1 Norms: A Novel Method for Low-Quality Retinal Image Enhancement. J Imaging 2024; 10:151. [PMID: 39057722 PMCID: PMC11277667 DOI: 10.3390/jimaging10070151] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/10/2024] [Revised: 06/12/2024] [Accepted: 06/17/2024] [Indexed: 07/28/2024] Open
Abstract
Nonmydriatic retinal fundus images often suffer from quality issues and artifacts due to ocular or systemic comorbidities, leading to potential inaccuracies in clinical diagnoses. In recent times, deep learning methods have been widely employed to improve retinal image quality. However, these methods often require large datasets and lack robustness in clinical settings. Conversely, the inherent stability and adaptability of traditional unsupervised learning methods, coupled with their reduced reliance on extensive data, render them more suitable for real-world clinical applications, particularly in the limited data context of high noise levels or a significant presence of artifacts. However, existing unsupervised learning methods encounter challenges such as sensitivity to noise and outliers, reliance on assumptions like cluster shapes, and difficulties with scalability and interpretability, particularly when utilized for retinal image enhancement. To tackle these challenges, we propose a novel robust PCA (RPCA) method with low-rank sparse decomposition that also integrates affine transformations τi, weighted nuclear norm, and the L2,1 norms, aiming to overcome existing method limitations and to achieve image quality improvement unseen by these methods. We employ the weighted nuclear norm (Lw,∗) to assign weights to singular values to each retinal images and utilize the L2,1 norm to eliminate correlated samples and outliers in the retinal images. Moreover, τi is employed to enhance retinal image alignment, making the new method more robust to variations, outliers, noise, and image blurring. The Alternating Direction Method of Multipliers (ADMM) method is used to optimally determine parameters, including τi, by solving an optimization problem. Each parameter is addressed separately, harnessing the benefits of ADMM. Our method introduces a novel parameter update approach and significantly improves retinal image quality, detecting cataracts, and diabetic retinopathy. Simulation results confirm our method's superiority over existing state-of-the-art methods across various datasets.
Collapse
Affiliation(s)
- Habte Tadesse Likassa
- Department of Biostatistics, College of Health Solutions, Arizona State University, Phoenix, AZ 85004, USA
| | - Ding-Geng Chen
- Department of Biostatistics, College of Health Solutions, Arizona State University, Phoenix, AZ 85004, USA
- Department of Statistics, University of Pretoria, Pretoria 0028, South Africa
| | - Kewei Chen
- Department of Biostatistics, College of Health Solutions, Arizona State University, Phoenix, AZ 85004, USA
| | - Yalin Wang
- Computer Science and Engineering, School of Computing and Augmented Intelligence, Arizona State University, Phoenix, AZ 85287-8809, USA
| | - Wenhui Zhu
- Computer Science and Engineering, School of Computing and Augmented Intelligence, Arizona State University, Phoenix, AZ 85287-8809, USA
| |
Collapse
|
33
|
He Y, Kong J, Li J, Zheng C. Entropy and distance-guided super self-ensembling for optic disc and cup segmentation. BIOMEDICAL OPTICS EXPRESS 2024; 15:3975-3992. [PMID: 38867792 PMCID: PMC11166439 DOI: 10.1364/boe.521778] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 02/21/2024] [Revised: 04/14/2024] [Accepted: 05/06/2024] [Indexed: 06/14/2024]
Abstract
Segmenting the optic disc (OD) and optic cup (OC) is crucial to accurately detect changes in glaucoma progression in the elderly. Recently, various convolutional neural networks have emerged to deal with OD and OC segmentation. Due to the domain shift problem, achieving high-accuracy segmentation of OD and OC from different domain datasets remains highly challenging. Unsupervised domain adaptation has taken extensive focus as a way to address this problem. In this work, we propose a novel unsupervised domain adaptation method, called entropy and distance-guided super self-ensembling (EDSS), to enhance the segmentation performance of OD and OC. EDSS is comprised of two self-ensembling models, and the Gaussian noise is added to the weights of the whole network. Firstly, we design a super self-ensembling (SSE) framework, which can combine two self-ensembling to learn more discriminative information about images. Secondly, we propose a novel exponential moving average with Gaussian noise (G-EMA) to enhance the robustness of the self-ensembling framework. Thirdly, we propose an effective multi-information fusion strategy (MFS) to guide and improve the domain adaptation process. We evaluate the proposed EDSS on two public fundus image datasets RIGA+ and REFUGE. Large amounts of experimental results demonstrate that the proposed EDSS outperforms state-of-the-art segmentation methods with unsupervised domain adaptation, e.g., the Dicemean score on three test sub-datasets of RIGA+ are 0.8442, 0.8772 and 0.9006, respectively, and the Dicemean score on the REFUGE dataset is 0.9154.
Collapse
Affiliation(s)
- Yanlin He
- College of Information Sciences and Technology, Northeast Normal University, Changchun 130117, China
| | - Jun Kong
- College of Information Sciences and Technology, Northeast Normal University, Changchun 130117, China
| | - Juan Li
- Jilin Engineering Normal University, Changchun 130052, China
- Business School, Northeast Normal University, Changchun 130117, China
| | - Caixia Zheng
- College of Information Sciences and Technology, Northeast Normal University, Changchun 130117, China
- Key Laboratory of Applied Statistics of MOE, Northeast Normal University, Changchun 130024, China
| |
Collapse
|
34
|
Fu H, Zhang J, Li B, Chen L, Zou J, Zhang Z, Zou H. Abdominal multi-organ segmentation in Multi-sequence MRIs based on visual attention guided network and knowledge distillation. Phys Med 2024; 122:103385. [PMID: 38810392 DOI: 10.1016/j.ejmp.2024.103385] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 04/03/2023] [Revised: 03/17/2024] [Accepted: 05/21/2024] [Indexed: 05/31/2024] Open
Abstract
PURPOSE The segmentation of abdominal organs in magnetic resonance imaging (MRI) plays a pivotal role in various therapeutic applications. Nevertheless, the application of deep-learning methods to abdominal organ segmentation encounters numerous challenges, especially in addressing blurred boundaries and regions characterized by low-contrast. METHODS In this study, a multi-scale visual attention-guided network (VAG-Net) was proposed for abdominal multi-organ segmentation based on unpaired multi-sequence MRI. A new visual attention-guided (VAG) mechanism was designed to enhance the extraction of contextual information, particularly at the edge of organs. Furthermore, a new loss function inspired by knowledge distillation was introduced to minimize the semantic disparity between different MRI sequences. RESULTS The proposed method was evaluated on the CHAOS 2019 Challenge dataset and compared with six state-of-the-art methods. The results demonstrated that our model outperformed these methods, achieving DSC values of 91.83 ± 0.24% and 94.09 ± 0.66% for abdominal multi-organ segmentation in T1-DUAL and T2-SPIR modality, respectively. CONCLUSION The experimental results show that our proposed method has superior performance in abdominal multi-organ segmentation, especially in the case of small organs such as the kidneys.
Collapse
Affiliation(s)
- Hao Fu
- Department of Automation, School of Information Science and Engineering, East China University of Science and Technology, Shanghai 200237, China
| | - Jian Zhang
- Department of Automation, School of Information Science and Engineering, East China University of Science and Technology, Shanghai 200237, China
| | - Bin Li
- Jiangnan University Medical Center, Wuxi No. 2 People's Hospital, Wu Xi, Jiangsu 214000, China
| | - Lanlan Chen
- Department of Automation, School of Information Science and Engineering, East China University of Science and Technology, Shanghai 200237, China
| | - Junzhong Zou
- Department of Automation, School of Information Science and Engineering, East China University of Science and Technology, Shanghai 200237, China.
| | - ZhuiYang Zhang
- Jiangnan University Medical Center, Wuxi No. 2 People's Hospital, Wu Xi, Jiangsu 214000, China.
| | - Hao Zou
- Center for Intelligent Medical Imaging and Health, Research Institute of Tsinghua University in Shenzhen, Shenzhen 518000, China.
| |
Collapse
|
35
|
AlJabri M, Alghamdi M, Collado-Mesa F, Abdel-Mottaleb M. Recurrent attention U-Net for segmentation and quantification of breast arterial calcifications on synthesized 2D mammograms. PeerJ Comput Sci 2024; 10:e2076. [PMID: 38855260 PMCID: PMC11157579 DOI: 10.7717/peerj-cs.2076] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/19/2024] [Accepted: 04/30/2024] [Indexed: 06/11/2024]
Abstract
Breast arterial calcifications (BAC) are a type of calcification commonly observed on mammograms and are generally considered benign and not associated with breast cancer. However, there is accumulating observational evidence of an association between BAC and cardiovascular disease, the leading cause of death in women. We present a deep learning method that could assist radiologists in detecting and quantifying BAC in synthesized 2D mammograms. We present a recurrent attention U-Net model consisting of encoder and decoder modules that include multiple blocks that each use a recurrent mechanism, a recurrent mechanism, and an attention module between them. The model also includes a skip connection between the encoder and the decoder, similar to a U-shaped network. The attention module was used to enhance the capture of long-range dependencies and enable the network to effectively classify BAC from the background, whereas the recurrent blocks ensured better feature representation. The model was evaluated using a dataset containing 2,000 synthesized 2D mammogram images. We obtained 99.8861% overall accuracy, 69.6107% sensitivity, 66.5758% F-1 score, and 59.5498% Jaccard coefficient, respectively. The presented model achieved promising performance compared with related models.
Collapse
Affiliation(s)
- Manar AlJabri
- Department of Computer Science and Artificial Intelligence, Umm Al-Qura University, Makkah, Makkah, Saudi Arabia
- King Abdul Aziz University, Jeddah, Makkah, Saudi Arabia
| | - Manal Alghamdi
- Department of Computer Science and Artificial Intelligence, Umm Al-Qura University, Makkah, Makkah, Saudi Arabia
| | - Fernando Collado-Mesa
- Department of Radiology, Miller School of Medicine, University of Miami, Miami, Florida, United States
| | - Mohamed Abdel-Mottaleb
- Department of Electrical and Computer Engineering, University of Miami, Miami, Florida, United States
| |
Collapse
|
36
|
Gibbon S, Muniz-Terrera G, Yii FSL, Hamid C, Cox S, Maccormick IJC, Tatham AJ, Ritchie C, Trucco E, Dhillon B, MacGillivray TJ. PallorMetrics: Software for Automatically Quantifying Optic Disc Pallor in Fundus Photographs, and Associations With Peripapillary RNFL Thickness. Transl Vis Sci Technol 2024; 13:20. [PMID: 38780955 PMCID: PMC11127490 DOI: 10.1167/tvst.13.5.20] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/08/2023] [Accepted: 04/10/2024] [Indexed: 05/25/2024] Open
Abstract
Purpose We sough to develop an automatic method of quantifying optic disc pallor in fundus photographs and determine associations with peripapillary retinal nerve fiber layer (pRNFL) thickness. Methods We used deep learning to segment the optic disc, fovea, and vessels in fundus photographs, and measured pallor. We assessed the relationship between pallor and pRNFL thickness derived from optical coherence tomography scans in 118 participants. Separately, we used images diagnosed by clinical inspection as pale (n = 45) and assessed how measurements compared with healthy controls (n = 46). We also developed automatic rejection thresholds and tested the software for robustness to camera type, image format, and resolution. Results We developed software that automatically quantified disc pallor across several zones in fundus photographs. Pallor was associated with pRNFL thickness globally (β = -9.81; standard error [SE] = 3.16; P < 0.05), in the temporal inferior zone (β = -29.78; SE = 8.32; P < 0.01), with the nasal/temporal ratio (β = 0.88; SE = 0.34; P < 0.05), and in the whole disc (β = -8.22; SE = 2.92; P < 0.05). Furthermore, pallor was significantly higher in the patient group. Last, we demonstrate the analysis to be robust to camera type, image format, and resolution. Conclusions We developed software that automatically locates and quantifies disc pallor in fundus photographs and found associations between pallor measurements and pRNFL thickness. Translational Relevance We think our method will be useful for the identification, monitoring, and progression of diseases characterized by disc pallor and optic atrophy, including glaucoma, compression, and potentially in neurodegenerative disorders.
Collapse
Affiliation(s)
- Samuel Gibbon
- Centre for Clinical Brain Sciences, Edinburgh, UK
- Robert O Curle Ophthalmology Suite, Institute for Regeneration and Repair, University of Edinburgh, UK, Edinburgh, UK
| | | | - Fabian S. L. Yii
- Centre for Clinical Brain Sciences, Edinburgh, UK
- Robert O Curle Ophthalmology Suite, Institute for Regeneration and Repair, University of Edinburgh, UK, Edinburgh, UK
| | | | - Simon Cox
- Lothian Birth Cohorts, Department of Psychology, University of Edinburgh, Edinburgh, UK
| | - Ian J. C. Maccormick
- Centre for Inflammation Research, University of Edinburgh, Edinburgh, UK
- Institute for Adaptive and Neural Computation, University of Edinburgh, Edinburgh, UK
| | - Andrew J. Tatham
- Centre for Clinical Brain Sciences, Edinburgh, UK
- Princess Alexandra Eye Pavilion, Chalmers Street, Edinburgh, UK
| | - Craig Ritchie
- Centre for Clinical Brain Sciences, Edinburgh, UK
- Centre for Dementia Prevention, University of Edinburgh, Edinburgh, UK
| | - Emanuele Trucco
- VAMPIRE Project, Computing (SSEN), University of Dundee, Dundee, UK
| | - Baljean Dhillon
- Centre for Clinical Brain Sciences, Edinburgh, UK
- Princess Alexandra Eye Pavilion, Chalmers Street, Edinburgh, UK
| | - Thomas J. MacGillivray
- Centre for Clinical Brain Sciences, Edinburgh, UK
- Robert O Curle Ophthalmology Suite, Institute for Regeneration and Repair, University of Edinburgh, UK, Edinburgh, UK
- VAMPIRE Project, Edinburgh Clinical Research facility, University of Edinburgh, Edinburgh, UK
| |
Collapse
|
37
|
Shi K, Zhang X, Wang X, Xu J, Mu B, Yan J, Wang F, Ding Y, Wang Z. ICF-PR-Net: a deep phase retrieval neural network for X-ray phase contrast imaging of inertial confinement fusion capsules. OPTICS EXPRESS 2024; 32:14356-14376. [PMID: 38859383 DOI: 10.1364/oe.518249] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 01/09/2024] [Accepted: 03/25/2024] [Indexed: 06/12/2024]
Abstract
X-ray phase contrast imaging (XPCI) has demonstrated capability to characterize inertial confinement fusion (ICF) capsules, and phase retrieval can reconstruct phase information from intensity images. This study introduces ICF-PR-Net, a novel deep learning-based phase retrieval method for ICF-XPCI. We numerically constructed datasets based on ICF capsule shape features, and proposed an object-image loss function to add image formation physics to network training. ICF-PR-Net outperformed traditional methods as it exhibited satisfactory robustness against strong noise and nonuniform background and was well-suited for ICF-XPCI's constrained experimental conditions and single exposure limit. Numerical and experimental results showed that ICF-PR-Net accurately retrieved the phase and absorption while maintaining retrieval quality in different situations. Overall, the ICF-PR-Net enables the diagnosis of the inner interface and electron density of capsules to address ignition-preventing problems, such as hydrodynamic instability growth.
Collapse
|
38
|
Yap BP, Kelvin LZ, Toh EQ, Low KY, Rani SK, Goh EJH, Hui VYC, Ng BK, Lim TH. Generalizability of Deep Neural Networks for Vertical Cup-to-Disc Ratio Estimation in Ultra-Widefield and Smartphone-Based Fundus Images. Transl Vis Sci Technol 2024; 13:6. [PMID: 38568608 PMCID: PMC10996969 DOI: 10.1167/tvst.13.4.6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/12/2023] [Accepted: 02/19/2024] [Indexed: 04/05/2024] Open
Abstract
Purpose To develop and validate a deep learning system (DLS) for estimation of vertical cup-to-disc ratio (vCDR) in ultra-widefield (UWF) and smartphone-based fundus images. Methods A DLS consisting of two sequential convolutional neural networks (CNNs) to delineate optic disc (OD) and optic cup (OC) boundaries was developed using 800 standard fundus images from the public REFUGE data set. The CNNs were tested on 400 test images from the REFUGE data set and 296 UWF and 300 smartphone-based images from a teleophthalmology clinic. vCDRs derived from the delineated OD/OC boundaries were compared with optometrists' annotations using mean absolute error (MAE). Subgroup analysis was conducted to study the impact of peripapillary atrophy (PPA), and correlation study was performed to investigate potential correlations between sectoral CDR (sCDR) and retinal nerve fiber layer (RNFL) thickness. Results The system achieved MAEs of 0.040 (95% CI, 0.037-0.043) in the REFUGE test images, 0.068 (95% CI, 0.061-0.075) in the UWF images, and 0.084 (95% CI, 0.075-0.092) in the smartphone-based images. There was no statistical significance in differences between PPA and non-PPA images. Weak correlation (r = -0.4046, P < 0.05) between sCDR and RNFL thickness was found only in the superior sector. Conclusions We developed a deep learning system that estimates vCDR from standard, UWF, and smartphone-based images. We also described anatomic peripapillary adversarial lesion and its potential impact on OD/OC delineation. Translational Relevance Artificial intelligence can estimate vCDR from different types of fundus images and may be used as a general and interpretable screening tool to improve community reach for diagnosis and management of glaucoma.
Collapse
Affiliation(s)
- Boon Peng Yap
- School of Electrical and Electronic Engineering, Nanyang Technological University, Singapore, Singapore
| | - Li Zhenghao Kelvin
- Department of Ophthalmology, Tan Tock Seng Hospital, Singapore, Singapore
- Lee Kong Chian School of Medicine, Nanyang Technological University, Singapore, Singapore
- National Healthcare Group Eye Institute, Singapore, Singapore
| | - En Qi Toh
- Lee Kong Chian School of Medicine, Nanyang Technological University, Singapore, Singapore
| | - Kok Yao Low
- Department of Ophthalmology, Tan Tock Seng Hospital, Singapore, Singapore
- National Healthcare Group Eye Institute, Singapore, Singapore
| | - Sumaya Khan Rani
- Department of Ophthalmology, Tan Tock Seng Hospital, Singapore, Singapore
- National Healthcare Group Eye Institute, Singapore, Singapore
| | - Eunice Jin Hui Goh
- Department of Ophthalmology, Tan Tock Seng Hospital, Singapore, Singapore
- National Healthcare Group Eye Institute, Singapore, Singapore
| | - Vivien Yip Cherng Hui
- Department of Ophthalmology, Tan Tock Seng Hospital, Singapore, Singapore
- National Healthcare Group Eye Institute, Singapore, Singapore
| | - Beng Koon Ng
- School of Electrical and Electronic Engineering, Nanyang Technological University, Singapore, Singapore
| | - Tock Han Lim
- Department of Ophthalmology, Tan Tock Seng Hospital, Singapore, Singapore
- Lee Kong Chian School of Medicine, Nanyang Technological University, Singapore, Singapore
- National Healthcare Group Eye Institute, Singapore, Singapore
| |
Collapse
|
39
|
Chen Y, Bai Y, Zhang Y. Optic disc and cup segmentation for glaucoma detection using Attention U-Net incorporating residual mechanism. PeerJ Comput Sci 2024; 10:e1941. [PMID: 38660163 PMCID: PMC11042003 DOI: 10.7717/peerj-cs.1941] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/20/2023] [Accepted: 02/26/2024] [Indexed: 04/26/2024]
Abstract
Glaucoma is a common eye disease that can cause blindness. Accurate detection of the optic disc and cup disc is crucial for glaucoma diagnosis. Algorithm models based on artificial intelligence can assist doctors in improving detection performance. In this article, U-Net is used as the backbone network, and the attention and residual modules are integrated to construct an end-to-end convolutional neural network model for optic disc and cup disc segmentation. The U-Net backbone is used to infer the basic position information of optic disc and cup disc, the attention module enhances the model's ability to represent and extract features of optic disc and cup disc, and the residual module alleviates gradient disappearance or explosion that may occur during feature representation of the neural network. The proposed model is trained and tested on the DRISHTI-GS1 dataset. Results show that compared with the original U-Net method, our model can more effectively separate optic disc and cup disc in terms of overlap error, sensitivity, and specificity.
Collapse
Affiliation(s)
- Yuanyuan Chen
- School of Information Technology, Luoyang Normal University, Luoyang, China
| | - Yongpeng Bai
- School of Information Technology, Luoyang Normal University, Luoyang, China
| | - Yifan Zhang
- School of Information Technology, Luoyang Normal University, Luoyang, China
| |
Collapse
|
40
|
Zhou Z, Zheng Y, Zhou X, Yu J, Rong S. Self-supervised pre-training for joint optic disc and cup segmentation via attention-aware network. BMC Ophthalmol 2024; 24:98. [PMID: 38438876 PMCID: PMC10910696 DOI: 10.1186/s12886-024-03376-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/18/2023] [Accepted: 02/28/2024] [Indexed: 03/06/2024] Open
Abstract
Image segmentation is a fundamental task in deep learning, which is able to analyse the essence of the images for further development. However, for the supervised learning segmentation method, collecting pixel-level labels is very time-consuming and labour-intensive. In the medical image processing area for optic disc and cup segmentation, we consider there are two challenging problems that remain unsolved. One is how to design an efficient network to capture the global field of the medical image and execute fast in real applications. The other is how to train the deep segmentation network using a few training data due to some medical privacy issues. In this paper, to conquer such issues, we first design a novel attention-aware segmentation model equipped with the multi-scale attention module in the pyramid structure-like encoder-decoder network, which can efficiently learn the global semantics and the long-range dependencies of the input images. Furthermore, we also inject the prior knowledge that the optic cup lies inside the optic disc by a novel loss function. Then, we propose a self-supervised contrastive learning method for optic disc and cup segmentation. The unsupervised feature representation is learned by matching an encoded query to a dictionary of encoded keys using a contrastive technique. Finetuning the pre-trained model using the proposed loss function can help achieve good performance for the task. To validate the effectiveness of the proposed method, extensive systemic evaluations on different public challenging optic disc and cup benchmarks, including DRISHTI-GS and REFUGE datasets demonstrate the superiority of the proposed method, which can achieve new state-of-the-art performance approaching 0.9801 and 0.9087 F1 score respectively while gaining 0.9657 D C disc and 0.8976 D C cup . The code will be made publicly available.
Collapse
Affiliation(s)
- Zhiwang Zhou
- Institute for Advanced Study, Nanchang University, Nanchang, 330031, China.
| | - Yuanchang Zheng
- Institute for Advanced Study, Nanchang University, Nanchang, 330031, China
- Institute of Science and Technology, Waseda University, Tokyo, 63-8001, Japan
| | - Xiaoyu Zhou
- School of Transportation Engineering, Tongji University, Shanghai, 200000, China
| | - Jie Yu
- School of Electrical Automation and Information Engineering, Tianjin University, Tianjin, 300000, China
| | - Shangjie Rong
- School of Mathematical Sciences, Xiamen University, Xiamen, 361000, China
| |
Collapse
|
41
|
Song Y, Zhang W, Zhang Y. A novel lightweight deep learning approach for simultaneous optic cup and optic disc segmentation in glaucoma detection. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2024; 21:5092-5117. [PMID: 38872528 DOI: 10.3934/mbe.2024225] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2024]
Abstract
Glaucoma is a chronic neurodegenerative disease that can result in irreversible vision loss if not treated in its early stages. The cup-to-disc ratio is a key criterion for glaucoma screening and diagnosis, and it is determined by dividing the area of the optic cup (OC) by that of the optic disc (OD) in fundus images. Consequently, the automatic and accurate segmentation of the OC and OD is a pivotal step in glaucoma detection. In recent years, numerous methods have resulted in great success on this task. However, most existing methods either have unsatisfactory segmentation accuracy or high time costs. In this paper, we propose a lightweight deep-learning architecture for the simultaneous segmentation of the OC and OD, where we have adopted fuzzy learning and a multi-layer perceptron to simplify the learning complexity and improve segmentation accuracy. Experimental results demonstrate the superiority of our proposed method as compared to most state-of-the-art approaches in terms of both training time and segmentation accuracy.
Collapse
Affiliation(s)
- Yantao Song
- Institute of Big Data Science and Industry, Shanxi University, Taiyuan 030006, China
- School of Computer and Information Technology, Shanxi University, Taiyuan 030006, China
| | - Wenjie Zhang
- Institute of Big Data Science and Industry, Shanxi University, Taiyuan 030006, China
- School of Computer and Information Technology, Shanxi University, Taiyuan 030006, China
| | - Yue Zhang
- School of Computer and Information Technology, Shanxi University, Taiyuan 030006, China
| |
Collapse
|
42
|
Ling Y, Wang Y, Liu Q, Yu J, Xu L, Zhang X, Liang P, Kong D. EPolar-UNet: An edge-attending polar UNet for automatic medical image segmentation with small datasets. Med Phys 2024; 51:1702-1713. [PMID: 38299370 DOI: 10.1002/mp.16957] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/09/2023] [Revised: 12/29/2023] [Accepted: 01/14/2024] [Indexed: 02/02/2024] Open
Abstract
BACKGROUND Medical image segmentation is one of the most key steps in computer-aided clinical diagnosis, geometric characterization, measurement, image registration, and so forth. Convolutional neural networks especially UNet and its variants have been successfully used in many medical image segmentation tasks. However, the results are limited by the deficiency in extracting high resolution edge information because of the design of the skip connections in UNet and the need for large available datasets. PURPOSE In this paper, we proposed an edge-attending polar UNet (EPolar-UNet), which was trained on the polar coordinate system instead of classic Cartesian coordinate system with an edge-attending construction in skip connection path. METHODS EPolar-UNet extracted the location information from an eight-stacked hourglass network as the pole for polar transformation and extracted the boundary cues from an edge-attending UNet, which consisted of a deconvolution layer and a subtraction operation. RESULTS We evaluated the performance of EPolar-UNet across three imaging modalities for different segmentation tasks: CVC-ClinicDB dataset for polyp, ISIC-2018 dataset for skin lesion, and our private ultrasound dataset for liver tumor segmentation. Our proposed model outperformed state-of-the-art models on all three datasets and needed only 30%-60% of training data compared with the benchmark UNet model to achieve similar performances for medical image segmentation tasks. CONCLUSIONS We proposed an end-to-end EPolar-UNet for automatic medical image segmentation and showed good performance on small datasets, which was critical in the field of medical image segmentation.
Collapse
Affiliation(s)
- Yating Ling
- School of Mathematical Sciences, Zhejiang University, Hangzhou, China
| | - Yuling Wang
- Department of Interventional Ultrasound, The Fifth Medical Center of Chinese PLA General Hospital, Beijing, China
| | - Qian Liu
- School of Mathematical Sciences, Zhejiang University, Hangzhou, China
| | - Jie Yu
- Department of Interventional Ultrasound, The Fifth Medical Center of Chinese PLA General Hospital, Beijing, China
| | - Lei Xu
- Zhejiang Qiushi Institute for Mathematical Medicine, Hangzhou, China
| | - Xiaoqian Zhang
- School of Mathematical Sciences, Zhejiang University, Hangzhou, China
| | - Ping Liang
- Department of Interventional Ultrasound, The Fifth Medical Center of Chinese PLA General Hospital, Beijing, China
| | - Dexing Kong
- School of Mathematical Sciences, Zhejiang University, Hangzhou, China
| |
Collapse
|
43
|
Chen S, Xie F, Chen S, Liu S, Li H, Gong Q, Ruan G, Liu L, Chen H. TdDS-UNet: top-down deeply supervised U-Net for the delineation of 3D colorectal cancer. Phys Med Biol 2024; 69:055018. [PMID: 38306960 DOI: 10.1088/1361-6560/ad25c5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/25/2023] [Accepted: 02/01/2024] [Indexed: 02/04/2024]
Abstract
Automatically delineating colorectal cancers with fuzzy boundaries from 3D images is a challenging task, but the problem of fuzzy boundary delineation in existing deep learning-based methods have not been investigated in depth. Here, an encoder-decoder-based U-shaped network (U-Net) based on top-down deep supervision (TdDS) was designed to accurately and automatically delineate the fuzzy boundaries of colorectal cancer. TdDS refines the semantic targets of the upper and lower stages by mapping ground truths that are more consistent with the stage properties than upsampling deep supervision. This stage-specific approach can guide the model to learn a coarse-to-fine delineation process and improve the delineation accuracy of fuzzy boundaries by gradually shrinking the boundaries. Experimental results showed that TdDS is more customizable and plays a role similar to the attentional mechanism, and it can further improve the capability of the model to delineate colorectal cancer contours. A total of 103, 12, and 29 3D pelvic magnetic resonance imaging volumes were used for training, validation, and testing, respectively. The comparative results indicate that the proposed method exhibits the best comprehensive performance, with a dice similarity coefficient (DSC) of 0.805 ± 0.053 and a hausdorff distance (HD) of 9.28 ± 5.14 voxels. In the delineation performance analysis section also showed that 44.49% of the delineation results are satisfactory and do not require revisions. This study can provide new technical support for the delineation of 3D colorectal cancer. Our method is open source, and the code is available athttps://github.com/odindis/TdDS/tree/main.
Collapse
Affiliation(s)
- Shuchao Chen
- School of Life & Environmental Science, Guilin University of Electronic Technology, Guilin 541004, People's Republic of China
| | - Fei Xie
- State Key Laboratory of Oncology in South China, Sun Yat-sen University Cancer Center, Collaborative Innovation Center for Cancer Medicine, Guangzhou 510060, People's Republic of China
| | - Shenghuan Chen
- Department of Radiology, The Sixth Affiliated Hospital of Guangzhou Medical university, Qingyuan People's Hospital, Qingyuan, People's Republic of China
| | - Shanshan Liu
- School of Life & Environmental Science, Guilin University of Electronic Technology, Guilin 541004, People's Republic of China
| | - Haojiang Li
- State Key Laboratory of Oncology in South China, Sun Yat-sen University Cancer Center, Collaborative Innovation Center for Cancer Medicine, Guangzhou 510060, People's Republic of China
| | - Qiong Gong
- School of Life & Environmental Science, Guilin University of Electronic Technology, Guilin 541004, People's Republic of China
| | - Guangying Ruan
- State Key Laboratory of Oncology in South China, Sun Yat-sen University Cancer Center, Collaborative Innovation Center for Cancer Medicine, Guangzhou 510060, People's Republic of China
| | - Lizhi Liu
- State Key Laboratory of Oncology in South China, Sun Yat-sen University Cancer Center, Collaborative Innovation Center for Cancer Medicine, Guangzhou 510060, People's Republic of China
| | - Hongbo Chen
- School of Life & Environmental Science, Guilin University of Electronic Technology, Guilin 541004, People's Republic of China
- Guangxi Human Physiological Information NonInvasive Detection Engineering Technology Research Center, Guilin 541004, People's Republic of China
- Guangxi Colleges and Universities Key Laboratory of Biomedical Sensors and Intelligent Instruments, Guilin 541004, People's Republic of China
- Guangxi Key Laboratory of Metabolic Reprogramming and Intelligent Medical Engineering for Chronic Diseases, Guilin, People's Republic of China
| |
Collapse
|
44
|
Ma F, Li S, Wang S, Guo Y, Wu F, Meng J, Dai C. Deep-learning segmentation method for optical coherence tomography angiography in ophthalmology. JOURNAL OF BIOPHOTONICS 2024; 17:e202300321. [PMID: 37801660 DOI: 10.1002/jbio.202300321] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/10/2023] [Revised: 09/28/2023] [Accepted: 10/04/2023] [Indexed: 10/08/2023]
Abstract
PURPOSE The optic disc and the macular are two major anatomical structures in the human eye. Optic discs are associated with the optic nerve. Macular mainly involves degeneration and impaired function of the macular region. Reliable optic disc and macular segmentation are necessary for the automated screening of retinal diseases. METHODS A swept-source OCTA system was designed to capture OCTA images of human eyes. To address these segmentation tasks, first, we constructed a new Optic Disc and Macula in fundus Image with optical coherence tomography angiography (OCTA) dataset (ODMI). Second, we proposed a Coarse and Fine Attention-Based Network (CFANet). RESULTS The five metrics of our methods on ODMI are 98.91 % , 98.47 % , 89.77 % , 98.49 % , and 89.77 % , respectively. CONCLUSIONS Experimental results show that our CFANet has achieved good performance on segmentation for the optic disc and macula in OCTA.
Collapse
Affiliation(s)
- Fei Ma
- School of Computer Science, Qufu Normal University, Shandong, China
| | - Sien Li
- School of Computer Science, Qufu Normal University, Shandong, China
| | - Shengbo Wang
- School of Computer Science, Qufu Normal University, Shandong, China
| | - Yanfei Guo
- School of Computer Science, Qufu Normal University, Shandong, China
| | - Fei Wu
- School of Automation, Nanjing University of Posts and Telecommunications, Jiangsu, China
| | - Jing Meng
- School of Computer Science, Qufu Normal University, Shandong, China
| | - Cuixia Dai
- College Science, Shanghai Institute of Technology, Shanghai, China
| |
Collapse
|
45
|
Gao W, Fan B, Fang Y, Song N. Lightweight and multi-lesion segmentation model for diabetic retinopathy based on the fusion of mixed attention and ghost feature mapping. Comput Biol Med 2024; 169:107854. [PMID: 38109836 DOI: 10.1016/j.compbiomed.2023.107854] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2023] [Revised: 12/04/2023] [Accepted: 12/11/2023] [Indexed: 12/20/2023]
Abstract
Diabetic retinopathy is the main cause of blindness, and lesion segmentation is an important basic work for the diagnosis of this disease. The main lesions include soft and hard exudates, microaneurysms, and hemorrhages. However, the segmentation of these four types of lesions is difficult because of their variability in size and contrast, and high intertype similarity. Currently, many network models have problems, such as a large number of parameters and complex calculations, and most segmentation models for diabetic retinopathy focus only on one type of lesion. In this study, a lightweight algorithm based on BiSeNet V2 was proposed for the segmentation of multiple lesions in diabetic retinopathy fundus. First, a hybrid attention module was embedded in the semantic branch of BiSeNet V2 for 8- and 16-fold downsampling, which helped reassign deep feature-map weights and enhanced the ability to extract local key features. Second, a ghost feature-mapping unit was used to optimize the traditional convolution layers and further reduce the computational cost. Third, a new loss function based on the dynamic threshold loss function was applied to supervise the training by adjusting the training weights of the high-loss difficult samples, which enhanced the model's attention to small goals. In experiments on the IDRiD dataset, we conducted an ablation study to verify the effectiveness of each component and compared the proposed model, BiSeNet V2-Pro, with several state-of-the-art models. In comparison with the baseline BiSeNet V2, the segmentation performance of BiSeNet V2-Pro improved by 12.17 %, 11.44 %, and 8.49 % in terms of Sensitivity (SEN), Intersection over Union (IoU), and Dice coefficient (DICE), respectively. Specifically, IoU of MA reaches 0.5716. Compared with other methods, the segmentation speed was significantly improved while ensuring segmentation accuracy, and the number of model parameters was lower. These results demonstrate the superiority of BiSeNet V2-Pro in the multi-lesion segmentation of diabetic retinopathy.
Collapse
Affiliation(s)
- Weiwei Gao
- Institute of Mechanical and Automotive Engineering, Shanghai University of Engineering Science, Shanghai 201620, China.
| | - Bo Fan
- Institute of Mechanical and Automotive Engineering, Shanghai University of Engineering Science, Shanghai 201620, China.
| | - Yu Fang
- Institute of Mechanical and Automotive Engineering, Shanghai University of Engineering Science, Shanghai 201620, China.
| | - Nan Song
- Department of Ophthalmology, Eye&Ent Hospital of University, Shanghai 200031, China.
| |
Collapse
|
46
|
Tang S, Song C, Wang D, Gao Y, Liu Y, Lv W. W-Net: A boundary-aware cascade network for robust and accurate optic disc segmentation. iScience 2024; 27:108247. [PMID: 38230262 PMCID: PMC10790032 DOI: 10.1016/j.isci.2023.108247] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/29/2022] [Revised: 03/14/2023] [Accepted: 10/16/2023] [Indexed: 01/18/2024] Open
Abstract
Accurate optic disc (OD) segmentation has a great significance for computer-aided diagnosis of different types of eye diseases. Due to differences in image acquisition equipment and acquisition methods, the resolution, size, contrast, and clarity of images from different datasets show significant differences, resulting in poor generalization performance of deep learning networks. To solve this problem, this study proposes a multi-level segmentation network. The network includes data quality enhancement module (DQEM), coarse segmentation module (CSM), localization module (OLM), and fine segmentation stage module (FSM). In FSM, W-Net is proposed for the first time, and boundary loss is introduced in the loss function, which effectively improves the performance of OD segmentation. We generalized the model in the REFUGE test dataset, GAMMA dataset, Drishti-GS1 dataset, and IDRiD dataset, respectively. The results show that our method has the best OD segmentation performance in different datasets compared with state-of-the-art networks.
Collapse
Affiliation(s)
- Shuo Tang
- School of Instrumentation and Optoelectronic Engineering, Beihang University, Beijing 100191, China
| | - Chongchong Song
- School of Instrumentation and Optoelectronic Engineering, Beihang University, Beijing 100191, China
| | - Defeng Wang
- School of Instrumentation and Optoelectronic Engineering, Beihang University, Beijing 100191, China
| | - Yang Gao
- School of Instrumentation and Optoelectronic Engineering, Beihang University, Beijing 100191, China
| | - Yuchen Liu
- School of Instrumentation and Optoelectronic Engineering, Beihang University, Beijing 100191, China
| | - Wang Lv
- School of Instrumentation and Optoelectronic Engineering, Beihang University, Beijing 100191, China
| |
Collapse
|
47
|
MA S, A J, Perumal T SR. Multi-dimensional dense attention network for pixel-wise segmentation of optic disc in colour fundus images. Technol Health Care 2024; 32:3829-3846. [PMID: 39058458 PMCID: PMC11612978 DOI: 10.3233/thc-230310] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2023] [Accepted: 05/20/2024] [Indexed: 07/28/2024]
Abstract
BACKGROUND Segmentation of retinal fragments like blood vessels, Optic Disc (OD), and Optic Cup (OC) enables the early detection of different retinal pathologies like Diabetic Retinopathy (DR), Glaucoma, etc. OBJECTIVE Accurate segmentation of OD remains challenging due to blurred boundaries, vessel occlusion, and other distractions and limitations. These days, deep learning is rapidly progressing in the segmentation of image pixels, and a number of network models have been proposed for end-to-end image segmentation. However, there are still certain limitations, such as limited ability to represent context, inadequate feature processing, limited receptive field, etc., which lead to the loss of local details and blurred boundaries. METHODS A multi-dimensional dense attention network, or MDDA-Net, is proposed for pixel-wise segmentation of OD in retinal images in order to address the aforementioned issues and produce more thorough and accurate segmentation results. In order to acquire powerful contexts when faced with limited context representation capabilities, a dense attention block is recommended. A triple-attention (TA) block is introduced in order to better extract the relationship between pixels and obtain more comprehensive information, with the goal of addressing the insufficient feature processing. In the meantime, a multi-scale context fusion (MCF) is suggested for acquiring the multi-scale contexts through context improvement. RESULTS Specifically, we provide a thorough assessment of the suggested approach on three difficult datasets. In the MESSIDOR and ORIGA data sets, the suggested MDDA-NET approach obtains accuracy levels of 99.28% and 98.95%, respectively. CONCLUSION The experimental results show that the MDDA-Net can obtain better performance than state-of-the-art deep learning models under the same environmental conditions.
Collapse
Affiliation(s)
- Sreema MA
- Department of Electronics and Communication Engineering, Arunachala College of Engineering for Women, Manavilai, India
| | - Jayachandran A
- Department of Computer Science and Engineering, PSN College of Engineering and Technology, Tirunelveli, India
| | - Sudarson Rama Perumal T
- Department of Computer Science and Engineering, Rohini College of Engineering and Technology, Nagercoil, India
| |
Collapse
|
48
|
Xu Y, Yang W. Editorial: Artificial intelligence applications in chronic ocular diseases. Front Cell Dev Biol 2023; 11:1295850. [PMID: 38143924 PMCID: PMC10740206 DOI: 10.3389/fcell.2023.1295850] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/17/2023] [Accepted: 11/28/2023] [Indexed: 12/26/2023] Open
Affiliation(s)
- Yanwu Xu
- School of Future Technology, South China University of Technology, Guangzhou, Guangdong Province, China
- Pazhou Lab, Guangzhou, Guangdong Province, China
| | - Weihua Yang
- Shenzhen Eye Institute, Shenzhen Eye Hospital, Jinan University, Shenzhen, Guangdong Province, China
| |
Collapse
|
49
|
Wang S, Yu X, Jia W, Chi J, Lv P, Wang J, Wu C. Optic disc detection based on fully convolutional network and weighted matrix recovery model. Med Biol Eng Comput 2023; 61:3319-3333. [PMID: 37668892 DOI: 10.1007/s11517-023-02891-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2023] [Accepted: 07/09/2023] [Indexed: 09/06/2023]
Abstract
Eye diseases often affect human health. Accurate detection of the optic disc contour is one of the important steps in diagnosing and treating eye diseases. However, the structure of fundus images is complex, and the optic disc region is often disturbed by blood vessels. Considering that the optic disc is usually a saliency region in fundus images, we propose a weakly-supervised optic disc detection method based on the fully convolution neural network (FCN) combined with the weighted low-rank matrix recovery model (WLRR). Firstly, we extract the low-level features of the fundus image and cluster the pixels using the Simple Linear Iterative Clustering (SLIC) algorithm to generate the feature matrix. Secondly, the top-down semantic prior information provided by FCN and bottom-up background prior information of the optic disc region are used to jointly construct the prior information weighting matrix, which more accurately guides the decomposition of the feature matrix into a sparse matrix representing the optic disc and a low-rank matrix representing the background. Experimental results on the DRISHTI-GS dataset and IDRiD dataset show that our method can segment the optic disc region accurately, and its performance is better than existing weakly-supervised optic disc segmentation methods. Graphical abstract of optic disc segmentation.
Collapse
Affiliation(s)
- Siqi Wang
- Faculty of Robot Science and Engineering, Northeastern University, 110170, Shen Yang, Liao Ning, China
| | - Xiaosheng Yu
- Faculty of Robot Science and Engineering, Northeastern University, 110170, Shen Yang, Liao Ning, China.
| | - Wenzhuo Jia
- Art School, HE University, 110163, Shen Yang, Liao Ning, China
| | - Jianning Chi
- Faculty of Robot Science and Engineering, Northeastern University, 110170, Shen Yang, Liao Ning, China
- Key Laboratory of Intelligent Computing in Medical Image of Ministry of Education, Northeastern University, 110170, Shen Yang, Liao Ning, China
| | - Pengfei Lv
- Faculty of Robot Science and Engineering, Northeastern University, 110170, Shen Yang, Liao Ning, China
| | - Junxiang Wang
- Faculty of Robot Science and Engineering, Northeastern University, 110170, Shen Yang, Liao Ning, China
| | - Chengdong Wu
- Faculty of Robot Science and Engineering, Northeastern University, 110170, Shen Yang, Liao Ning, China
| |
Collapse
|
50
|
Kim S, Yoon H, Lee J, Yoo S. Facial wrinkle segmentation using weighted deep supervision and semi-automatic labeling. Artif Intell Med 2023; 145:102679. [PMID: 37925209 DOI: 10.1016/j.artmed.2023.102679] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/15/2023] [Revised: 07/28/2023] [Accepted: 10/03/2023] [Indexed: 11/06/2023]
Abstract
Facial wrinkles are important indicators of human aging. Recently, a method using deep learning and a semi-automatic labeling was proposed to segment facial wrinkles, which showed much better performance than conventional image-processing-based methods. However, the difficulty of wrinkle segmentation remains challenging due to the thinness of wrinkles and their small proportion in the entire image. Therefore, performance improvement in wrinkle segmentation is still necessary. To address this issue, we propose a novel loss function that takes into account the thickness of wrinkles based on the semi-automatic labeling approach. First, considering the different spatial dimensions of the decoder in the U-Net architecture, we generated weighted wrinkle maps from ground truth. These weighted wrinkle maps were used to calculate the training losses more accurately than the existing deep supervision approach. This new loss computation approach is defined as weighted deep supervision in our study. The proposed method was evaluated using an image dataset obtained from a professional skin analysis device and labeled using semi-automatic labeling. In our experiment, the proposed weighted deep supervision showed higher Jaccard Similarity Index (JSI) performance for wrinkle segmentation compared to conventional deep supervision and traditional image processing methods. Additionally, we conducted experiments on the labeling using a semi-automatic labeling approach, which had not been explored in previous research, and compared it with human labeling. The semi-automatic labeling technology showed more consistent wrinkle labels than human-made labels. Furthermore, to assess the scalability of the proposed method to other domains, we applied it to retinal vessel segmentation. The results demonstrated superior performance of the proposed method compared to existing retinal vessel segmentation approaches. In conclusion, the proposed method offers high performance and can be easily applied to various biomedical domains and U-Net-based architectures. Therefore, the proposed approach will be beneficial for various biomedical imaging approaches. To facilitate this, we have made the source code of the proposed method publicly available at: https://github.com/resemin/WeightedDeepSupervision.
Collapse
Affiliation(s)
- Semin Kim
- AI R&D Center, Lululab Inc., 318, Dosan-daero, Gangnam-gu, Seoul, Republic of Korea.
| | - Huisu Yoon
- AI R&D Center, Lululab Inc., 318, Dosan-daero, Gangnam-gu, Seoul, Republic of Korea.
| | - Jongha Lee
- AI R&D Center, Lululab Inc., 318, Dosan-daero, Gangnam-gu, Seoul, Republic of Korea.
| | - Sangwook Yoo
- AI R&D Center, Lululab Inc., 318, Dosan-daero, Gangnam-gu, Seoul, Republic of Korea.
| |
Collapse
|