1
|
Medhi JP, S R N, Borah K, Bhattacharjee D, Dandapat S. Sharper insights: Adaptive ellipse-template for robust fovea localization in challenging retinal landscapes. Comput Biol Med 2025; 191:110125. [PMID: 40258324 DOI: 10.1016/j.compbiomed.2025.110125] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/12/2024] [Revised: 03/14/2025] [Accepted: 03/31/2025] [Indexed: 04/23/2025]
Abstract
Automated identification of retinal landmarks, particularly the fovea is crucial for diagnosing diabetic retinopathy and other ocular diseases. But accurate identification is challenging due to varying contrast, color irregularities, anatomical structure and the presence of lesions near the macula in fundus images. Existing methods often struggle to maintain accuracy in these complex conditions, particularly when lesions obscure vital regions. To overcome these limitations, this paper introduces a novel adaptive ellipse-template-based approach for fovea localization, leveraging mathematical modeling of blood vessel (BV) trajectories and optic disc (OD) positioning. Unlike traditional fixed-template model, our method dynamically adjusts the ellipse parameters based on OD diameter, ensuring a generalized and adaptable template. This flexibility enables consistent detection performance, even in challenging images with significant lesion interference. Extensive validation on ten publicly available databases, including MESSIDOR, DRIVE, DIARETDB0, DIARETDB1, HRF, IDRiD, HEIMED, ROC, GEI, and NETRALAYA, demonstrates a superior detection efficiency of 99.5%. Additionally, the method achieves a low mean Euclidean distance of 13.48 pixels with a standard deviation of 15.5 pixels between the actual and detected fovea locations, highlighting its precision and reliability. The proposed approach significantly outperforms conventional template-based and deep learning methods, particularly in lesion-rich and low-contrast conditions. It is computationally efficient, interpretable, and robust, making it a valuable tool for automated retinal image analysis in clinical settings.
Collapse
Affiliation(s)
- Jyoti Prakash Medhi
- Department of Electronics and Communication Engineering, Gauhati University, Guwahati- 781014, Assam, India.
| | - Nirmala S R
- Department of Electronics and Communication Engineering, KLE Technological University, Hubballi, Karnataka, India.
| | - Kuntala Borah
- Department of Electronics and Communication Engineering, Gauhati University, Guwahati- 781014, Assam, India.
| | | | - Samarendra Dandapat
- Department of Electronics and Electrical Engineering, Indian Institute of Technology Guwahati, Assam, India.
| |
Collapse
|
2
|
Silva-Rodríguez J, Chakor H, Kobbi R, Dolz J, Ben Ayed I. A Foundation Language-Image Model of the Retina (FLAIR): encoding expert knowledge in text supervision. Med Image Anal 2025; 99:103357. [PMID: 39418828 DOI: 10.1016/j.media.2024.103357] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2023] [Revised: 05/06/2024] [Accepted: 09/23/2024] [Indexed: 10/19/2024]
Abstract
Foundation vision-language models are currently transforming computer vision, and are on the rise in medical imaging fueled by their very promising generalization capabilities. However, the initial attempts to transfer this new paradigm to medical imaging have shown less impressive performances than those observed in other domains, due to the significant domain shift and the complex, expert domain knowledge inherent to medical-imaging tasks. Motivated by the need for domain-expert foundation models, we present FLAIR, a pre-trained vision-language model for universal retinal fundus image understanding. To this end, we compiled 38 open-access, mostly categorical fundus imaging datasets from various sources, with up to 101 different target conditions and 288,307 images. We integrate the expert's domain knowledge in the form of descriptive textual prompts, during both pre-training and zero-shot inference, enhancing the less-informative categorical supervision of the data. Such a textual expert's knowledge, which we compiled from the relevant clinical literature and community standards, describes the fine-grained features of the pathologies as well as the hierarchies and dependencies between them. We report comprehensive evaluations, which illustrate the benefit of integrating expert knowledge and the strong generalization capabilities of FLAIR under difficult scenarios with domain shifts or unseen categories. When adapted with a lightweight linear probe, FLAIR outperforms fully-trained, dataset-focused models, more so in the few-shot regimes. Interestingly, FLAIR outperforms by a wide margin larger-scale generalist image-language models and retina domain-specific self-supervised networks, which emphasizes the potential of embedding experts' domain knowledge and the limitations of generalist models in medical imaging. The pre-trained model is available at: https://github.com/jusiro/FLAIR.
Collapse
Affiliation(s)
| | | | | | - Jose Dolz
- ÉTS Montréal, Québec, Canada; Centre de Recherche du Centre Hospitalier de l'Université de Montréal (CR-CHUM), Québec, Canada
| | - Ismail Ben Ayed
- ÉTS Montréal, Québec, Canada; Centre de Recherche du Centre Hospitalier de l'Université de Montréal (CR-CHUM), Québec, Canada
| |
Collapse
|
3
|
Pachade S, Porwal P, Kokare M, Deshmukh G, Sahasrabuddhe V, Luo Z, Han F, Sun Z, Qihan L, Kamata SI, Ho E, Wang E, Sivajohan A, Youn S, Lane K, Chun J, Wang X, Gu Y, Lu S, Oh YT, Park H, Lee CY, Yeh H, Cheng KW, Wang H, Ye J, He J, Gu L, Müller D, Soto-Rey I, Kramer F, Arai H, Ochi Y, Okada T, Giancardo L, Quellec G, Mériaudeau F. RFMiD: Retinal Image Analysis for multi-Disease Detection challenge. Med Image Anal 2025; 99:103365. [PMID: 39395210 DOI: 10.1016/j.media.2024.103365] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/10/2022] [Revised: 07/16/2024] [Accepted: 10/02/2024] [Indexed: 10/14/2024]
Abstract
In the last decades, many publicly available large fundus image datasets have been collected for diabetic retinopathy, glaucoma, and age-related macular degeneration, and a few other frequent pathologies. These publicly available datasets were used to develop a computer-aided disease diagnosis system by training deep learning models to detect these frequent pathologies. One challenge limiting the adoption of a such system by the ophthalmologist is, computer-aided disease diagnosis system ignores sight-threatening rare pathologies such as central retinal artery occlusion or anterior ischemic optic neuropathy and others that ophthalmologists currently detect. Aiming to advance the state-of-the-art in automatic ocular disease classification of frequent diseases along with the rare pathologies, a grand challenge on "Retinal Image Analysis for multi-Disease Detection" was organized in conjunction with the IEEE International Symposium on Biomedical Imaging (ISBI - 2021). This paper, reports the challenge organization, dataset, top-performing participants solutions, evaluation measures, and results based on a new "Retinal Fundus Multi-disease Image Dataset" (RFMiD). There were two principal sub-challenges: disease screening (i.e. presence versus absence of pathology - a binary classification problem) and disease/pathology classification (a 28-class multi-label classification problem). It received a positive response from the scientific community with 74 submissions by individuals/teams that effectively entered in this challenge. The top-performing methodologies utilized a blend of data-preprocessing, data augmentation, pre-trained model, and model ensembling. This multi-disease (frequent and rare pathologies) detection will enable the development of generalizable models for screening the retina, unlike the previous efforts that focused on the detection of specific diseases.
Collapse
Affiliation(s)
- Samiksha Pachade
- Center of Excellence in Signal and Image Processing, Shri Guru Gobind Singhji Institute of Engineering and Technology, Nanded 431606, India.
| | - Prasanna Porwal
- Center of Excellence in Signal and Image Processing, Shri Guru Gobind Singhji Institute of Engineering and Technology, Nanded 431606, India
| | - Manesh Kokare
- Center of Excellence in Signal and Image Processing, Shri Guru Gobind Singhji Institute of Engineering and Technology, Nanded 431606, India
| | | | - Vivek Sahasrabuddhe
- Department of Ophthalmology, Shankarrao Chavan Government Medical College, Nanded 431606, India
| | - Zhengbo Luo
- Graduate School of Information Production and Systems, Waseda University, Japan
| | - Feng Han
- University of Shanghai for Science and Technology, Shanghai, China
| | - Zitang Sun
- Graduate School of Information Production and Systems, Waseda University, Japan
| | - Li Qihan
- Graduate School of Information Production and Systems, Waseda University, Japan
| | - Sei-Ichiro Kamata
- Graduate School of Information Production and Systems, Waseda University, Japan
| | - Edward Ho
- Schulich Applied Computing in Medicine, University of Western Ontario, Schulich School of Medicine and Dentistry, Canada
| | - Edward Wang
- Schulich Applied Computing in Medicine, University of Western Ontario, Schulich School of Medicine and Dentistry, Canada
| | - Asaanth Sivajohan
- Schulich Applied Computing in Medicine, University of Western Ontario, Schulich School of Medicine and Dentistry, Canada
| | - Saerom Youn
- Schulich Applied Computing in Medicine, University of Western Ontario, Schulich School of Medicine and Dentistry, Canada
| | - Kevin Lane
- Schulich Applied Computing in Medicine, University of Western Ontario, Schulich School of Medicine and Dentistry, Canada
| | - Jin Chun
- Schulich Applied Computing in Medicine, University of Western Ontario, Schulich School of Medicine and Dentistry, Canada
| | - Xinliang Wang
- Beihang University School of Computer Science, China
| | - Yunchao Gu
- Beihang University School of Computer Science, China
| | - Sixu Lu
- Beijing Normal University School of Artificial Intelligence, China
| | - Young-Tack Oh
- Department of Electrical and Computer Engineering, Sungkyunkwan University, Suwon, Republic of Korea
| | - Hyunjin Park
- Center for Neuroscience Imaging Research, Institute for Basic Science, Suwon, Republic of Korea; School of Electronic and Electrical Engineering, Sungkyunkwan University, Suwon, Republic of Korea
| | - Chia-Yen Lee
- Department of Electrical Engineering, National United University, Miaoli 360001, Taiwan, ROC
| | - Hung Yeh
- Department of Electrical Engineering, National United University, Miaoli 360001, Taiwan, ROC; Institute of Biomedical Engineering, National Yang Ming Chiao Tung University, 1001 Ta-Hsueh Road, Hsinchu, Taiwan, ROC
| | - Kai-Wen Cheng
- Department of Electrical Engineering, National United University, Miaoli 360001, Taiwan, ROC
| | - Haoyu Wang
- School of Biomedical Engineering, the Institute of Medical Robotics, Shanghai Jiao Tong University, Shanghai, China
| | - Jin Ye
- ShenZhen Key Lab of Computer Vision and Pattern Recognition, Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Junjun He
- School of Biomedical Engineering, the Institute of Medical Robotics, Shanghai Jiao Tong University, Shanghai, China; ShenZhen Key Lab of Computer Vision and Pattern Recognition, Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Lixu Gu
- School of Biomedical Engineering, the Institute of Medical Robotics, Shanghai Jiao Tong University, Shanghai, China
| | - Dominik Müller
- IT-Infrastructure for Translational Medical Research, University of Augsburg, Germany; Medical Data Integration Center, University Hospital Augsburg, Germany
| | - Iñaki Soto-Rey
- IT-Infrastructure for Translational Medical Research, University of Augsburg, Germany; Medical Data Integration Center, University Hospital Augsburg, Germany
| | - Frank Kramer
- IT-Infrastructure for Translational Medical Research, University of Augsburg, Germany
| | | | - Yuma Ochi
- National Institute of Technology, Kisarazu College, Japan
| | - Takami Okada
- Institute of Industrial Ecological Sciences, University of Occupational and Environmental Health, Japan
| | - Luca Giancardo
- Center for Precision Health, School of Biomedical Informatics, University of Texas Health Science Center at Houston (UTHealth), Houston, TX 77030, USA
| | | | | |
Collapse
|
4
|
Naz H, Ahuja NJ. A novel contrast enhancement technique for diabetic retinal image pre-processing and classification. Int Ophthalmol 2024; 45:11. [PMID: 39680225 DOI: 10.1007/s10792-024-03377-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/26/2024] [Accepted: 12/03/2024] [Indexed: 12/17/2024]
Abstract
BACKGROUND Diabetic Retinopathy (DR) is a leading cause of blindness among individuals aged 18 to 65 with diabetes, affecting 35-60% of this population, according to the International Diabetes Federation. Early diagnosis is critical for preventing vision loss, yet processing raw fundus images using machine learning faces significant challenges, particularly in accurately identifying microaneurysm lesions, which are crucial for diagnosis. METHODS This study proposes a novel pre-processing technique utilizing the Modified Fuzzy C-means Clustering approach combined with a Support Vector Machine classifier. The method includes converting RGB images to HSI colour space, applying median filtering to reduce noise, enhancing contrast through Intensity Histogram Equalization, and identifying false microaneurysm candidates using connected components. Additionally, morphological operations are performed to remove the optic disc from the enhanced images due to its similarity to microaneurysms. RESULTS The proposed method was evaluated using publicly available datasets, demonstrating superior performance compared to existing state-of-the-art algorithms. The approach achieved an accuracy rate of 99.31%, significantly improving the detection of microaneurysms and reducing false detections. CONCLUSIONS The findings indicate that the proposed pre-processing technique effectively enhances diabetic retinopathy classification by addressing the challenges of false microaneurysm detection. The comparative analysis against state-of-the-art algorithms highlights the effectiveness of the proposed method, particularly in addressing the challenges associated with false microaneurysms.
Collapse
Affiliation(s)
- Huma Naz
- School of Computer Science, UPES, Dehradun, India.
| | | |
Collapse
|
5
|
Qian B, Sheng B, Chen H, Wang X, Li T, Jin Y, Guan Z, Jiang Z, Wu Y, Wang J, Chen T, Guo Z, Chen X, Yang D, Hou J, Feng R, Xiao F, Li Y, El Habib Daho M, Lu L, Ding Y, Liu D, Yang B, Zhu W, Wang Y, Kim H, Nam H, Li H, Wu WC, Wu Q, Dai R, Li H, Ang M, Ting DSW, Cheung CY, Wang X, Cheng CY, Tan GSW, Ohno-Matsui K, Jonas JB, Zheng Y, Tham YC, Wong TY, Wang YX. A Competition for the Diagnosis of Myopic Maculopathy by Artificial Intelligence Algorithms. JAMA Ophthalmol 2024; 142:1006-1015. [PMID: 39325442 PMCID: PMC11428027 DOI: 10.1001/jamaophthalmol.2024.3707] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/25/2024] [Accepted: 07/11/2024] [Indexed: 09/27/2024]
Abstract
Importance Myopic maculopathy (MM) is a major cause of vision impairment globally. Artificial intelligence (AI) and deep learning (DL) algorithms for detecting MM from fundus images could potentially improve diagnosis and assist screening in a variety of health care settings. Objectives To evaluate DL algorithms for MM classification and segmentation and compare their performance with that of ophthalmologists. Design, Setting, and Participants The Myopic Maculopathy Analysis Challenge (MMAC) was an international competition to develop automated solutions for 3 tasks: (1) MM classification, (2) segmentation of MM plus lesions, and (3) spherical equivalent (SE) prediction. Participants were provided 3 subdatasets containing 2306, 294, and 2003 fundus images, respectively, with which to build algorithms. A group of 5 ophthalmologists evaluated the same test sets for tasks 1 and 2 to ascertain performance. Results from model ensembles, which combined outcomes from multiple algorithms submitted by MMAC participants, were compared with each individual submitted algorithm. This study was conducted from March 1, 2023, to March 30, 2024, and data were analyzed from January 15, 2024, to March 30, 2024. Exposure DL algorithms submitted as part of the MMAC competition or ophthalmologist interpretation. Main Outcomes and Measures MM classification was evaluated by quadratic-weighted κ (QWK), F1 score, sensitivity, and specificity. MM plus lesions segmentation was evaluated by dice similarity coefficient (DSC), and SE prediction was evaluated by R2 and mean absolute error (MAE). Results The 3 tasks were completed by 7, 4, and 4 teams, respectively. MM classification algorithms achieved a QWK range of 0.866 to 0.901, an F1 score range of 0.675 to 0.781, a sensitivity range of 0.667 to 0.778, and a specificity range of 0.931 to 0.945. MM plus lesions segmentation algorithms achieved a DSC range of 0.664 to 0.687 for lacquer cracks (LC), 0.579 to 0.673 for choroidal neovascularization, and 0.768 to 0.841 for Fuchs spot (FS). SE prediction algorithms achieved an R2 range of 0.791 to 0.874 and an MAE range of 0.708 to 0.943. Model ensemble results achieved the best performance compared to each submitted algorithms, and the model ensemble outperformed ophthalmologists at MM classification in sensitivity (0.801; 95% CI, 0.764-0.840 vs 0.727; 95% CI, 0.684-0.768; P = .006) and specificity (0.946; 95% CI, 0.939-0.954 vs 0.933; 95% CI, 0.925-0.941; P = .009), LC segmentation (DSC, 0.698; 95% CI, 0.649-0.745 vs DSC, 0.570; 95% CI, 0.515-0.625; P < .001), and FS segmentation (DSC, 0.863; 95% CI, 0.831-0.888 vs DSC, 0.790; 95% CI, 0.742-0.830; P < .001). Conclusions and Relevance In this diagnostic study, 15 AI models for MM classification and segmentation on a public dataset made available for the MMAC competition were validated and evaluated, with some models achieving better diagnostic performance than ophthalmologists.
Collapse
Affiliation(s)
- Bo Qian
- Shanghai Belt and Road International Joint Laboratory of Intelligent Prevention and Treatment for Metabolic Diseases, Department of Computer Science and Engineering, School of Electronic, Information, and Electrical Engineering, Shanghai Jiao Tong University, Department of Endocrinology and Metabolism, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai Diabetes Institute, Shanghai Clinical Center for Diabetes, Shanghai, China
- Ministry of Education Key Laboratory of Artificial Intelligence, School of Electronic, Information, and Electrical Engineering, Shanghai Jiao Tong University, Shanghai, China
| | - Bin Sheng
- Shanghai Belt and Road International Joint Laboratory of Intelligent Prevention and Treatment for Metabolic Diseases, Department of Computer Science and Engineering, School of Electronic, Information, and Electrical Engineering, Shanghai Jiao Tong University, Department of Endocrinology and Metabolism, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai Diabetes Institute, Shanghai Clinical Center for Diabetes, Shanghai, China
- Ministry of Education Key Laboratory of Artificial Intelligence, School of Electronic, Information, and Electrical Engineering, Shanghai Jiao Tong University, Shanghai, China
| | - Hao Chen
- Department of Computer Science and Engineering, The Hong Kong University of Science and Technology, Hong Kong, China
- Department of Chemical and Biological Engineering, The Hong Kong University of Science and Technology, Hong Kong, China
| | - Xiangning Wang
- Department of Ophthalmology, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Tingyao Li
- Shanghai Belt and Road International Joint Laboratory of Intelligent Prevention and Treatment for Metabolic Diseases, Department of Computer Science and Engineering, School of Electronic, Information, and Electrical Engineering, Shanghai Jiao Tong University, Department of Endocrinology and Metabolism, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai Diabetes Institute, Shanghai Clinical Center for Diabetes, Shanghai, China
- Ministry of Education Key Laboratory of Artificial Intelligence, School of Electronic, Information, and Electrical Engineering, Shanghai Jiao Tong University, Shanghai, China
| | - Yixiao Jin
- School of Clinical Medicine, Tsinghua Medicine, Tsinghua University, Beijing, China
- School of Clinical Medicine, Beijing Tsinghua Changgung Hospital, Beijing, China
| | - Zhouyu Guan
- Shanghai Belt and Road International Joint Laboratory of Intelligent Prevention and Treatment for Metabolic Diseases, Department of Computer Science and Engineering, School of Electronic, Information, and Electrical Engineering, Shanghai Jiao Tong University, Department of Endocrinology and Metabolism, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai Diabetes Institute, Shanghai Clinical Center for Diabetes, Shanghai, China
| | - Zehua Jiang
- School of Clinical Medicine, Tsinghua Medicine, Tsinghua University, Beijing, China
- School of Clinical Medicine, Beijing Tsinghua Changgung Hospital, Beijing, China
| | - Yilan Wu
- School of Clinical Medicine, Tsinghua Medicine, Tsinghua University, Beijing, China
| | - Jinyuan Wang
- School of Clinical Medicine, Tsinghua Medicine, Tsinghua University, Beijing, China
- School of Clinical Medicine, Beijing Tsinghua Changgung Hospital, Beijing, China
| | - Tingli Chen
- Department of Ophthalmology, Shanghai Health and Medical Center, Wuxi, China
| | - Zhengrui Guo
- Department of Computer Science and Engineering, The Hong Kong University of Science and Technology, Hong Kong, China
| | - Xiang Chen
- Shanghai Belt and Road International Joint Laboratory of Intelligent Prevention and Treatment for Metabolic Diseases, Department of Computer Science and Engineering, School of Electronic, Information, and Electrical Engineering, Shanghai Jiao Tong University, Department of Endocrinology and Metabolism, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai Diabetes Institute, Shanghai Clinical Center for Diabetes, Shanghai, China
- Ministry of Education Key Laboratory of Artificial Intelligence, School of Electronic, Information, and Electrical Engineering, Shanghai Jiao Tong University, Shanghai, China
| | - Dawei Yang
- Department of Ophthalmology and Visual Sciences, The Chinese University of Hong Kong, Hong Kong, China
| | - Junlin Hou
- Shanghai Key Laboratory of Intelligent Information Processing, School of Computer Science, Fudan University, Shanghai, China
| | - Rui Feng
- Shanghai Key Laboratory of Intelligent Information Processing, School of Computer Science, Fudan University, Shanghai, China
- Academy for Engineering and Technology, Fudan University, Shanghai, China
| | - Fan Xiao
- Academy for Engineering and Technology, Fudan University, Shanghai, China
| | - Yihao Li
- Laboratoire de Traitement de l'Information Médicale UMR 1101, Inserm, Brest, France
- Université de Bretagne Occidentale, Brest, France
| | - Mostafa El Habib Daho
- Laboratoire de Traitement de l'Information Médicale UMR 1101, Inserm, Brest, France
- Université de Bretagne Occidentale, Brest, France
| | - Li Lu
- School of Computer Science and Technology, Dongguan University of Technology, Dongguan, China
| | - Ye Ding
- School of Computer Science and Technology, Dongguan University of Technology, Dongguan, China
| | - Di Liu
- AIFUTURE Laboratory, Beijing, China
- National Digital Health Center of China Top Think Tanks, Beijing Normal University, Beijing, China
- School of Journalism and Communication, Beijing Normal University, Beijing, China
| | - Bo Yang
- AIFUTURE Laboratory, Beijing, China
| | - Wenhui Zhu
- School of Computing and Augmented Intelligence, Arizona State University, Tempe
| | - Yalin Wang
- School of Computing and Augmented Intelligence, Arizona State University, Tempe
| | - Hyeonmin Kim
- Mediwhale, Seoul, South Korea
- Pohang University of Science and Technology, Pohang, South Korea
| | | | - Huayu Li
- Department of Electrical and Computer Engineering, University of Arizona, Tucson
| | - Wei-Chi Wu
- Department of Ophthalmology, Linkou Chang Gung Memorial Hospital, College of Medicine, Chang Gung University, Taoyuan, Taiwan
| | - Qiang Wu
- Department of Ophthalmology, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Rongping Dai
- Department of Ophthalmology, Peking Union Medical College Hospital, Peking Union Medical College, Chinese Academy of Medical Sciences, Beijing, China
| | - Huating Li
- Shanghai Belt and Road International Joint Laboratory of Intelligent Prevention and Treatment for Metabolic Diseases, Department of Computer Science and Engineering, School of Electronic, Information, and Electrical Engineering, Shanghai Jiao Tong University, Department of Endocrinology and Metabolism, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai Diabetes Institute, Shanghai Clinical Center for Diabetes, Shanghai, China
| | - Marcus Ang
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore
| | | | - Carol Y Cheung
- Department of Ophthalmology and Visual Sciences, The Chinese University of Hong Kong, Hong Kong, China
| | - Xiaofei Wang
- Key Laboratory for Biomechanics and Mechanobiology of Ministry of Education, Beijing Advanced Innovation Center for Biomedical Engineering, School of Biological Science and Medical Engineering, Beihang University, Beijing, China
| | - Ching-Yu Cheng
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore
| | - Gavin Siew Wei Tan
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore
| | - Kyoko Ohno-Matsui
- Department of Ophthalmology and Visual Science, Tokyo Medical and Dental University, Tokyo, Japan
| | - Jost B Jonas
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore
- Institut Français de Myopie, Rothschild Foundation Hospital, Paris, France
| | | | - Yih-Chung Tham
- Center for Innovation and Precision Eye Health, Department of Ophthalmology, Yong Loo Lin School of Medicine, National University of Singapore, Singapore
- Ophthalmology and Visual Science Academic Clinical Program, Duke-National University of Singapore Medical School, Singapore
| | - Tien Yin Wong
- School of Clinical Medicine, Tsinghua Medicine, Tsinghua University, Beijing, China
- School of Clinical Medicine, Beijing Tsinghua Changgung Hospital, Beijing, China
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore
- Zhongshan Ophthalmic Center, Guangzhou, China
| | - Ya Xing Wang
- Beijing Institute of Ophthalmology, Beijing Ophthalmology and Visual Sciences Key Laboratory, Beijing Tongren Eye Center, Beijing Tongren Hospital, Capital Medical University, Beijing, China
| |
Collapse
|
6
|
La Franca L, Rutigliani C, Checchin L, Lattanzio R, Bandello F, Cicinelli MV. Rate and Predictors of Misclassification of Active Diabetic Macular Edema as Detected by an Automated Retinal Image Analysis System. Ophthalmol Ther 2024; 13:1553-1567. [PMID: 38587776 PMCID: PMC11109071 DOI: 10.1007/s40123-024-00929-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/27/2023] [Accepted: 03/07/2024] [Indexed: 04/09/2024] Open
Abstract
INTRODUCTION The aim of this work is to estimate the sensitivity, specificity, and misclassification rate of an automated retinal image analysis system (ARIAS) in diagnosing active diabetic macular edema (DME) and to identify factors associated with true and false positives. METHODS We conducted a cross-sectional study of prospectively enrolled patients with diabetes mellitus (DM) referred to a tertiary medical retina center for screening or management of DME. All patients underwent two-field fundus photography (macula- and disc-centered) with a true-color confocal camera; images were processed by EyeArt V.2.1.0 (Woodland Hills, CA, USA). Active DME was defined as the presence of intraretinal or subretinal fluid on spectral-domain optical coherence tomography (SD-OCT). Sensitivity and specificity and their 95% confidence intervals (CIs) were calculated. Variables associated with true (i.e., DME labeled as present by ARIAS + fluid on SD-OCT) and false positives (i.e., DME labeled as present by ARIAS + no fluid on SD-OCT) of active DME were explored. RESULTS A total of 298 eyes were included; 92 eyes (31%) had active DME. ARIAS sensitivity and specificity were 82.61% (95% CI 72.37-89.60) and 84.47% (95% CI 78.34-89.10). The misclassification rate was 16%. Factors associated with true positives included younger age (p = 0.01), shorter DM duration (p = 0.006), presence of hard exudates (p = 0.005), and microaneurysms (p = 0.002). Factors associated with false positives included longer DM duration (p = 0.01), worse diabetic retinopathy severity (p = 0.008), history of inactivated DME (p < 0.001), and presence of hard exudates (p < 0.001), microaneurysms (p < 0.001), or epiretinal membrane (p = 0.06). CONCLUSIONS The sensitivity of ARIAS was diminished in older patients and those without DME-related fundus lesions, while the specificity was reduced in cases with a history of inactivated DME. ARIAS performed well in screening for naïve DME but is not effective in surveillance inactivated DME.
Collapse
Affiliation(s)
- Lamberto La Franca
- Department of Ophthalmology, IRCCS San Raffaele Scientific Institute, IRCCS Ospedale San Raffaele, University Vita-Salute, Via Olgettina 60, 20132, Milan, Italy
| | - Carola Rutigliani
- School of Medicine, Vita-Salute San Raffaele University, Milan, Italy
| | - Lisa Checchin
- Department of Ophthalmology, IRCCS San Raffaele Scientific Institute, IRCCS Ospedale San Raffaele, University Vita-Salute, Via Olgettina 60, 20132, Milan, Italy
| | - Rosangela Lattanzio
- Department of Ophthalmology, IRCCS San Raffaele Scientific Institute, IRCCS Ospedale San Raffaele, University Vita-Salute, Via Olgettina 60, 20132, Milan, Italy
| | - Francesco Bandello
- Department of Ophthalmology, IRCCS San Raffaele Scientific Institute, IRCCS Ospedale San Raffaele, University Vita-Salute, Via Olgettina 60, 20132, Milan, Italy
- School of Medicine, Vita-Salute San Raffaele University, Milan, Italy
| | - Maria Vittoria Cicinelli
- Department of Ophthalmology, IRCCS San Raffaele Scientific Institute, IRCCS Ospedale San Raffaele, University Vita-Salute, Via Olgettina 60, 20132, Milan, Italy.
- School of Medicine, Vita-Salute San Raffaele University, Milan, Italy.
| |
Collapse
|
7
|
Russo C, Bria A, Marrocco C. GravityNet for end-to-end small lesion detection. Artif Intell Med 2024; 150:102842. [PMID: 38553147 DOI: 10.1016/j.artmed.2024.102842] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/27/2023] [Revised: 03/01/2024] [Accepted: 03/11/2024] [Indexed: 04/02/2024]
Abstract
This paper introduces a novel one-stage end-to-end detector specifically designed to detect small lesions in medical images. Precise localization of small lesions presents challenges due to their appearance and the diverse contextual backgrounds in which they are found. To address this, our approach introduces a new type of pixel-based anchor that dynamically moves towards the targeted lesion for detection. We refer to this new architecture as GravityNet, and the novel anchors as gravity points since they appear to be "attracted" by the lesions. We conducted experiments on two well-established medical problems involving small lesions to evaluate the performance of the proposed approach: microcalcifications detection in digital mammograms and microaneurysms detection in digital fundus images. Our method demonstrates promising results in effectively detecting small lesions in these medical imaging tasks.
Collapse
Affiliation(s)
- Ciro Russo
- Department of Electrical and Information Engineering, University of Cassino and L.M., Via G. Di Biasio 43, 03043 Cassino (FR), Italy.
| | - Alessandro Bria
- Department of Electrical and Information Engineering, University of Cassino and L.M., Via G. Di Biasio 43, 03043 Cassino (FR), Italy.
| | - Claudio Marrocco
- Department of Electrical and Information Engineering, University of Cassino and L.M., Via G. Di Biasio 43, 03043 Cassino (FR), Italy.
| |
Collapse
|
8
|
Nakamura H, Fukuda M, Matsuda A, Makino N, Kimura H, Ohtaki Y, Nawa Y, Oyama S, Suzuki Y, Kobayashi T, Ishizawa T, Kakizaki Y, Ueno Y. Differentiating localized autoimmune pancreatitis and pancreatic ductal adenocarcinoma using endoscopic ultrasound images with deep learning. DEN OPEN 2024; 4:e344. [PMID: 38434146 PMCID: PMC10908399 DOI: 10.1002/deo2.344] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/12/2023] [Revised: 01/29/2024] [Accepted: 02/10/2024] [Indexed: 03/05/2024]
Abstract
Objectives Localized autoimmune pancreatitis is difficult to differentiate from pancreatic ductal adenocarcinoma on endoscopic ultrasound images. In recent years, deep learning methods have improved the diagnosis of diseases. Hence, we developed a special cross-validation framework to search for effective methodologies of deep learning in distinguishing autoimmune pancreatitis from pancreatic ductal adenocarcinoma on endoscopic ultrasound images. Methods Data from 24 patients diagnosed with localized autoimmune pancreatitis (8751 images) and 61 patients diagnosed with pancreatic ductal adenocarcinoma (20,584 images) were collected from 2016 to 2022. We applied transfer learning to a convolutional neural network called ResNet152, together with our innovative imaging method contributing to data augmentation and temporal data process. We divided patients into five groups according to different factors for 5-fold cross-validation, where the ordered and balanced datasets were created for the performance evaluations. Results ResNet152 surpassed the endoscopists in all evaluation metrics with almost all datasets. Interestingly, when the dataset is balanced according to the factor of the endoscopists' diagnostic accuracy, the area under the receiver operating characteristic curve and accuracy were highest at 0.85 and 0.80, respectively. Conclusions It is deduced that image features useful for ResNet152 correlate with those used by endoscopists for their diagnoses. This finding may contribute to sample-efficient dataset preparation to train convolutional neural networks for endoscopic ultrasonography-imaging diagnosis.
Collapse
Affiliation(s)
- Hitomi Nakamura
- Department of GastroenterologyFaculty of MedicineYamagata UniversityYamagataJapan
| | - Motohisa Fukuda
- Department of ScienceFaculty of ScienceYamagata UniversityYamagataJapan
| | - Akiko Matsuda
- Department of GastroenterologyFaculty of MedicineYamagata UniversityYamagataJapan
| | - Naohiko Makino
- Department of GastroenterologyFaculty of MedicineYamagata UniversityYamagataJapan
| | | | - Yu Ohtaki
- Department of GastroenterologyFaculty of MedicineYamagata UniversityYamagataJapan
| | - Yoshihito Nawa
- Department of GastroenterologyFaculty of MedicineYamagata UniversityYamagataJapan
| | - Soushi Oyama
- Department of GastroenterologyFaculty of MedicineYamagata UniversityYamagataJapan
| | - Yuya Suzuki
- Department of GastroenterologyFaculty of MedicineYamagata UniversityYamagataJapan
| | - Toshikazu Kobayashi
- Department of GastroenterologyFaculty of MedicineYamagata UniversityYamagataJapan
| | - Tetsuya Ishizawa
- Department of GastroenterologyFaculty of MedicineYamagata UniversityYamagataJapan
| | - Yasuharu Kakizaki
- Department of GastroenterologyFaculty of MedicineYamagata UniversityYamagataJapan
| | - Yoshiyuki Ueno
- Department of GastroenterologyFaculty of MedicineYamagata UniversityYamagataJapan
| |
Collapse
|
9
|
Qian B, Chen H, Wang X, Guan Z, Li T, Jin Y, Wu Y, Wen Y, Che H, Kwon G, Kim J, Choi S, Shin S, Krause F, Unterdechler M, Hou J, Feng R, Li Y, El Habib Daho M, Yang D, Wu Q, Zhang P, Yang X, Cai Y, Tan GSW, Cheung CY, Jia W, Li H, Tham YC, Wong TY, Sheng B. DRAC 2022: A public benchmark for diabetic retinopathy analysis on ultra-wide optical coherence tomography angiography images. PATTERNS (NEW YORK, N.Y.) 2024; 5:100929. [PMID: 38487802 PMCID: PMC10935505 DOI: 10.1016/j.patter.2024.100929] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/25/2023] [Revised: 12/09/2023] [Accepted: 01/15/2024] [Indexed: 03/17/2024]
Abstract
We described a challenge named "DRAC - Diabetic Retinopathy Analysis Challenge" in conjunction with the 25th International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI 2022). Within this challenge, we provided the DRAC datset, an ultra-wide optical coherence tomography angiography (UW-OCTA) dataset (1,103 images), addressing three primary clinical tasks: diabetic retinopathy (DR) lesion segmentation, image quality assessment, and DR grading. The scientific community responded positively to the challenge, with 11, 12, and 13 teams submitting different solutions for these three tasks, respectively. This paper presents a concise summary and analysis of the top-performing solutions and results across all challenge tasks. These solutions could provide practical guidance for developing accurate classification and segmentation models for image quality assessment and DR diagnosis using UW-OCTA images, potentially improving the diagnostic capabilities of healthcare professionals. The dataset has been released to support the development of computer-aided diagnostic systems for DR evaluation.
Collapse
Affiliation(s)
- Bo Qian
- Shanghai Belt and Road International Joint Laboratory for Intelligent Prevention and Treatment of Metabolic Disorders, Department of Computer Science and Engineering, School of Electronic, Information, and Electrical Engineering, Shanghai Jiao Tong University, Department of Endocrinology and Metabolism, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai Diabetes Institute, Shanghai Clinical Center for Diabetes, Shanghai 200240, China
- MOE Key Laboratory of AI, School of Electronic, Information, and Electrical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
| | - Hao Chen
- Department of Computer Science and Engineering, The Hong Kong University of Science and Technology, Hong Kong 999077, China
- Department of Chemical and Biological Engineering, The Hong Kong University of Science and Technology, Hong Kong 999077, China
| | - Xiangning Wang
- Shanghai Belt and Road International Joint Laboratory for Intelligent Prevention and Treatment of Metabolic Disorders, Department of Computer Science and Engineering, School of Electronic, Information, and Electrical Engineering, Shanghai Jiao Tong University, Department of Endocrinology and Metabolism, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai Diabetes Institute, Shanghai Clinical Center for Diabetes, Shanghai 200240, China
- Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai 200233, China
| | - Zhouyu Guan
- Shanghai Belt and Road International Joint Laboratory for Intelligent Prevention and Treatment of Metabolic Disorders, Department of Computer Science and Engineering, School of Electronic, Information, and Electrical Engineering, Shanghai Jiao Tong University, Department of Endocrinology and Metabolism, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai Diabetes Institute, Shanghai Clinical Center for Diabetes, Shanghai 200240, China
| | - Tingyao Li
- Shanghai Belt and Road International Joint Laboratory for Intelligent Prevention and Treatment of Metabolic Disorders, Department of Computer Science and Engineering, School of Electronic, Information, and Electrical Engineering, Shanghai Jiao Tong University, Department of Endocrinology and Metabolism, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai Diabetes Institute, Shanghai Clinical Center for Diabetes, Shanghai 200240, China
- MOE Key Laboratory of AI, School of Electronic, Information, and Electrical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
| | - Yixiao Jin
- Tsinghua Medicine, Tsinghua University, Beijing 100084, China
| | - Yilan Wu
- Tsinghua Medicine, Tsinghua University, Beijing 100084, China
| | - Yang Wen
- School of Electronic and Information Engineering, Shenzhen University, Shenzhen 518060, China
| | - Haoxuan Che
- Department of Computer Science and Engineering, The Hong Kong University of Science and Technology, Hong Kong 999077, China
| | | | | | - Sungjin Choi
- AI/DX Convergence Business Group, KT, Seongnam 13606, Korea
| | - Seoyoung Shin
- AI/DX Convergence Business Group, KT, Seongnam 13606, Korea
| | - Felix Krause
- Johannes Kepler University Linz, Linz 4040, Austria
| | | | - Junlin Hou
- School of Computer Science, Shanghai Key Laboratory of Intelligent Information Processing, Fudan University, Shanghai 200433, China
| | - Rui Feng
- School of Computer Science, Shanghai Key Laboratory of Intelligent Information Processing, Fudan University, Shanghai 200433, China
- Academy for Engineering and Technology, Fudan University, Shanghai 200433, China
| | - Yihao Li
- LaTIM UMR 1101, INSERM, 29609 Brest, France
- University of Western Brittany, 29238 Brest, France
| | - Mostafa El Habib Daho
- LaTIM UMR 1101, INSERM, 29609 Brest, France
- University of Western Brittany, 29238 Brest, France
| | - Dawei Yang
- Department of Ophthalmology and Visual Sciences, The Chinese University of Hong Kong, Hong Kong 999077, China
| | - Qiang Wu
- Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai 200233, China
| | - Ping Zhang
- Department of Computer Science and Engineering, The Ohio State University, Columbus, OH 43210, USA
- Department of Biomedical Informatics, The Ohio State University, Columbus, OH 43210, USA
- Translational Data Analytics Institute, The Ohio State University, Columbus, OH 43210, USA
| | - Xiaokang Yang
- MOE Key Laboratory of AI, School of Electronic, Information, and Electrical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
| | - Yiyu Cai
- School of Mechanical and Aerospace Engineering, Nanyang Technological University, Singapore 639798, Singapore
| | - Gavin Siew Wei Tan
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore 168751, Singapore
| | - Carol Y. Cheung
- Department of Ophthalmology and Visual Sciences, The Chinese University of Hong Kong, Hong Kong 999077, China
| | - Weiping Jia
- Shanghai Belt and Road International Joint Laboratory for Intelligent Prevention and Treatment of Metabolic Disorders, Department of Computer Science and Engineering, School of Electronic, Information, and Electrical Engineering, Shanghai Jiao Tong University, Department of Endocrinology and Metabolism, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai Diabetes Institute, Shanghai Clinical Center for Diabetes, Shanghai 200240, China
| | - Huating Li
- Shanghai Belt and Road International Joint Laboratory for Intelligent Prevention and Treatment of Metabolic Disorders, Department of Computer Science and Engineering, School of Electronic, Information, and Electrical Engineering, Shanghai Jiao Tong University, Department of Endocrinology and Metabolism, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai Diabetes Institute, Shanghai Clinical Center for Diabetes, Shanghai 200240, China
| | - Yih Chung Tham
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore 168751, Singapore
- Centre for Innovation and Precision Eye Health; and Department of Ophthalmology, Yong Loo Lin School of Medicine, National University of Singapore, Singapore 119228, Singapore
- Ophthalmology and Visual Sciences Academic Clinical Program, Duke-NUS Medical School, Singapore 169857, Singapore
| | - Tien Yin Wong
- Tsinghua Medicine, Tsinghua University, Beijing 100084, China
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore 168751, Singapore
- School of Clinical Medicine, Beijing Tsinghua Changgung Hospital, Beijing 102218, China
| | - Bin Sheng
- Shanghai Belt and Road International Joint Laboratory for Intelligent Prevention and Treatment of Metabolic Disorders, Department of Computer Science and Engineering, School of Electronic, Information, and Electrical Engineering, Shanghai Jiao Tong University, Department of Endocrinology and Metabolism, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, Shanghai Diabetes Institute, Shanghai Clinical Center for Diabetes, Shanghai 200240, China
- MOE Key Laboratory of AI, School of Electronic, Information, and Electrical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
| |
Collapse
|
10
|
Steffi S, Sam Emmanuel WR. Resilient back-propagation machine learning-based classification on fundus images for retinal microaneurysm detection. Int Ophthalmol 2024; 44:91. [PMID: 38367192 DOI: 10.1007/s10792-024-02982-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/23/2023] [Accepted: 10/29/2023] [Indexed: 02/19/2024]
Abstract
BACKGROUND The timely diagnosis of medical conditions, particularly diabetic retinopathy, relies on the identification of retinal microaneurysms. However, the commonly used retinography method poses a challenge due to the diminutive dimensions and limited differentiation of microaneurysms in images. PROBLEM STATEMENT Automated identification of microaneurysms becomes crucial, necessitating the use of comprehensive ad-hoc processing techniques. Although fluorescein angiography enhances detectability, its invasiveness limits its suitability for routine preventative screening. OBJECTIVE This study proposes a novel approach for detecting retinal microaneurysms using a fundus scan, leveraging circular reference-based shape features (CR-SF) and radial gradient-based texture features (RG-TF). METHODOLOGY The proposed technique involves extracting CR-SF and RG-TF for each candidate microaneurysm, employing a robust back-propagation machine learning method for training. During testing, extracted features from test images are compared with training features to categorize microaneurysm presence. RESULTS The experimental assessment utilized four datasets (MESSIDOR, Diaretdb1, e-ophtha-MA, and ROC), employing various measures. The proposed approach demonstrated high accuracy (98.01%), sensitivity (98.74%), specificity (97.12%), and area under the curve (91.72%). CONCLUSION The presented approach showcases a successful method for detecting retinal microaneurysms using a fundus scan, providing promising accuracy and sensitivity. This non-invasive technique holds potential for effective screening in diabetic retinopathy and other related medical conditions.
Collapse
Affiliation(s)
- S Steffi
- Department of Computer Science, Nesamony Memorial Christian College Affiliated to Manonmaniam Sundaranar University, Abishekapatti, Tirunelveli, Tamil Nadu, 627012, India.
| | - W R Sam Emmanuel
- Department of PG Computer Science, Nesamony Memorial Christian College Affiliated to Manonmaniam Sundaranar University, Abishekapatti, Tirunelveli, Tamil Nadu, 627012, India
| |
Collapse
|
11
|
Gonçalves MB, Nakayama LF, Ferraz D, Faber H, Korot E, Malerbi FK, Regatieri CV, Maia M, Celi LA, Keane PA, Belfort R. Image quality assessment of retinal fundus photographs for diabetic retinopathy in the machine learning era: a review. Eye (Lond) 2024; 38:426-433. [PMID: 37667028 PMCID: PMC10858054 DOI: 10.1038/s41433-023-02717-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/08/2023] [Revised: 06/26/2023] [Accepted: 08/25/2023] [Indexed: 09/06/2023] Open
Abstract
This study aimed to evaluate the image quality assessment (IQA) and quality criteria employed in publicly available datasets for diabetic retinopathy (DR). A literature search strategy was used to identify relevant datasets, and 20 datasets were included in the analysis. Out of these, 12 datasets mentioned performing IQA, but only eight specified the quality criteria used. The reported quality criteria varied widely across datasets, and accessing the information was often challenging. The findings highlight the importance of IQA for AI model development while emphasizing the need for clear and accessible reporting of IQA information. The study suggests that automated quality assessments can be a valid alternative to manual labeling and emphasizes the importance of establishing quality standards based on population characteristics, clinical use, and research purposes. In conclusion, image quality assessment is important for AI model development; however, strict data quality standards must not limit data sharing. Given the importance of IQA for developing, validating, and implementing deep learning (DL) algorithms, it's recommended that this information be reported in a clear, specific, and accessible way whenever possible. Automated quality assessments are a valid alternative to the traditional manual labeling process, and quality standards should be determined according to population characteristics, clinical use, and research purpose.
Collapse
Affiliation(s)
- Mariana Batista Gonçalves
- Department of Ophthalmology, Sao Paulo Federal University, São Paulo, SP, Brazil
- Instituto Paulista de Estudos e Pesquisas em Oftalmologia, IPEPO, Vision Institute, São Paulo, SP, Brazil
- NIHR Biomedical Research Centre for Ophthalmology, Moorfield Eye Hospital, NHS Foundation Trust, and UCL Institute of Ophthalmology, London, UK
| | - Luis Filipe Nakayama
- Department of Ophthalmology, Sao Paulo Federal University, São Paulo, SP, Brazil.
- Massachusetts Institute of Technology, Laboratory for Computational Physiology, Cambridge, MA, USA.
| | - Daniel Ferraz
- Department of Ophthalmology, Sao Paulo Federal University, São Paulo, SP, Brazil
- Instituto Paulista de Estudos e Pesquisas em Oftalmologia, IPEPO, Vision Institute, São Paulo, SP, Brazil
- NIHR Biomedical Research Centre for Ophthalmology, Moorfield Eye Hospital, NHS Foundation Trust, and UCL Institute of Ophthalmology, London, UK
| | - Hanna Faber
- Department of Ophthalmology, University Medical Center Hamburg-Eppendorf, Hamburg, Germany
- Department of Ophthalmology, University of Tuebingen, Tuebingen, Germany
| | - Edward Korot
- Retina Specialists of Michigan, Grand Rapids, MI, USA
- Stanford University Byers Eye Institute Palo Alto, Palo Alto, CA, USA
| | | | | | - Mauricio Maia
- Department of Ophthalmology, Sao Paulo Federal University, São Paulo, SP, Brazil
| | - Leo Anthony Celi
- Massachusetts Institute of Technology, Laboratory for Computational Physiology, Cambridge, MA, USA
- Harvard TH Chan School of Public Health, Department of Biostatistics, Boston, MA, USA
- Beth Israel Deaconess Medical Center, Department of Medicine, Boston, MA, USA
| | - Pearse A Keane
- NIHR Biomedical Research Centre for Ophthalmology, Moorfield Eye Hospital, NHS Foundation Trust, and UCL Institute of Ophthalmology, London, UK
| | - Rubens Belfort
- Department of Ophthalmology, Sao Paulo Federal University, São Paulo, SP, Brazil
- Instituto Paulista de Estudos e Pesquisas em Oftalmologia, IPEPO, Vision Institute, São Paulo, SP, Brazil
| |
Collapse
|
12
|
Gao M, Hormel TT, Guo Y, Tsuboi K, Flaxel CJ, Huang D, Hwang TS, Jia Y. Perfused and Nonperfused Microaneurysms Identified and Characterized by Structural and Angiographic OCT. Ophthalmol Retina 2024; 8:108-115. [PMID: 37673397 DOI: 10.1016/j.oret.2023.08.019] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/05/2023] [Revised: 08/18/2023] [Accepted: 08/28/2023] [Indexed: 09/08/2023]
Abstract
PURPOSE Microaneurysms (MAs) have distinct, oval-shaped, hyperreflective walls on structural OCT, and inconsistent flow signal in the lumen with OCT angiography (OCTA). Their relationship to regional macular edema in diabetic retinopathy (DR) has not been quantitatively explored. DESIGN Retrospective, cross-sectional study. PARTICIPANTS A total of 99 participants, including 23 with mild, nonproliferative DR (NPDR), 25 with moderate NPDR, 34 with severe NPDR, and 17 with proliferative DR. METHODS We obtained 3 × 3-mm scans with a commercial device (Solix, Visionix/Optovue) in 99 patients with DR. Trained graders manually identified MAs and their location relative to the anatomic layers from cross-sectional OCT. Microaneurysms were first classified as perfused if flow signal was present in the OCTA channel. Then, perfused MAs were further classified into fully and partially perfused MAs based on the flow characteristics in en face OCTA. The presence of retinal fluid based on OCT near MAs was compared between perfused and nonperfused types. We also compared OCT-based MA detection to fundus photography (FP)- and fluorescein angiography (FA)-based detection. MAIN OUTCOME MEASURES OCT-identified MAs can be classified according to colocalized OCTA flow signal into fully perfused, partially perfused, and nonperfused types. Fully perfused MAs may be more likely to be associated with diabetic macular edema (DME) than those without flow. RESULTS We identified 308 MAs (166 fully perfused, 88 partially perfused, 54 nonperfused) in 42 eyes using OCT and OCTA. Nearly half of the MAs identified in this study straddle the inner nuclear layer and outer plexiform layer. Compared with partially perfused and nonperfused MAs, fully perfused MAs were more likely to be associated with local retinal fluid. The associated fluid volumes were larger with fully perfused MAs compared with other types. OCT/OCTA detected all MAs found on FP. Although not all MAs seen with FA were identified with OCT, some MAs seen with OCT were not visible with FA or FP. CONCLUSIONS OCT-identified MAs with colocalized flow on OCTA are more likely to be associated with DME than those without flow. FINANCIAL DISCLOSURE(S) Proprietary or commercial disclosure may be found in the Footnotes and Disclosures at the end of this article.
Collapse
Affiliation(s)
- Min Gao
- Casey Eye Institute, Oregon Health & Science University, Portland, Oregon
| | - Tristan T Hormel
- Casey Eye Institute, Oregon Health & Science University, Portland, Oregon
| | - Yukun Guo
- Casey Eye Institute, Oregon Health & Science University, Portland, Oregon; Department of Biomedical Engineering, Oregon Health & Science University, Portland, Oregon
| | - Kotaro Tsuboi
- Casey Eye Institute, Oregon Health & Science University, Portland, Oregon
| | - Christina J Flaxel
- Casey Eye Institute, Oregon Health & Science University, Portland, Oregon
| | - David Huang
- Casey Eye Institute, Oregon Health & Science University, Portland, Oregon; Department of Biomedical Engineering, Oregon Health & Science University, Portland, Oregon
| | - Thomas S Hwang
- Casey Eye Institute, Oregon Health & Science University, Portland, Oregon
| | - Yali Jia
- Casey Eye Institute, Oregon Health & Science University, Portland, Oregon; Department of Biomedical Engineering, Oregon Health & Science University, Portland, Oregon.
| |
Collapse
|
13
|
Naz H, Nijhawan R, Ahuja NJ, Saba T, Alamri FS, Rehman A. Micro-segmentation of retinal image lesions in diabetic retinopathy using energy-based fuzzy C-Means clustering (EFM-FCM). Microsc Res Tech 2024; 87:78-94. [PMID: 37681440 DOI: 10.1002/jemt.24413] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/04/2023] [Revised: 08/06/2023] [Accepted: 08/24/2023] [Indexed: 09/09/2023]
Abstract
Diabetic retinopathy (DR) is a prevalent cause of global visual impairment, contributing to approximately 4.8% of blindness cases worldwide as reported by the World Health Organization (WHO). The condition is characterized by pathological abnormalities in the retinal layer, including microaneurysms, vitreous hemorrhages, and exudates. Microscopic analysis of retinal images is crucial in diagnosing and treating DR. This article proposes a novel method for early DR screening using segmentation and unsupervised learning techniques. The approach integrates a neural network energy-based model into the Fuzzy C-Means (FCM) algorithm to enhance convergence criteria, aiming to improve the accuracy and efficiency of automated DR screening tools. The evaluation of results includes the primary dataset from the Shiva Netralaya Centre, IDRiD, and DIARETDB1. The performance of the proposed method is compared against FCM, EFCM, FLICM, and M-FLICM techniques, utilizing metrics such as accuracy in noiseless and noisy conditions and average execution time. The results showcase auspicious performance on both primary and secondary datasets, achieving accuracy rates of 99.03% in noiseless conditions and 93.13% in noisy images, with an average execution time of 16.1 s. The proposed method holds significant potential in medical image analysis and could pave the way for future advancements in automated DR diagnosis and management. RESEARCH HIGHLIGHTS: A novel approach is proposed in the article, integrating a neural network energy-based model into the FCM algorithm to enhance the convergence criteria and the accuracy of automated DR screening tools. By leveraging the microscopic characteristics of retinal images, the proposed method significantly improves the accuracy of lesion segmentation, facilitating early detection and monitoring of DR. The evaluation of the method's performance includes primary datasets from reputable sources such as the Shiva Netralaya Centre, IDRiD, and DIARETDB1, demonstrating its effectiveness in comparison to other techniques (FCM, EFCM, FLICM, and M-FLICM) in terms of accuracy in both noiseless and noisy conditions. It achieves impressive accuracy rates of 99.03% in noiseless conditions and 93.13% in noisy images, with an average execution time of 16.1 s.
Collapse
Affiliation(s)
- Huma Naz
- Department of Computer Science, University of Petroleum and Energy Studies, Dehradun, India
| | - Rahul Nijhawan
- Thapar Institute of Engineering and Technology, Patiala, Punjab, India
| | - Neelu Jyothi Ahuja
- Department of Computer Science, University of Petroleum and Energy Studies, Dehradun, India
| | - Tanzila Saba
- Artificial Intelligence and Data Analytics Lab, Prince Sultan University, Riyadh, Saudi Arabia
| | - Faten S Alamri
- Department of Mathematical Sciences, College of Science, Princess Nourah Bint Abdulrahman University, Riyadh, Saudi Arabia
| | - Amjad Rehman
- Artificial Intelligence and Data Analytics Lab, Prince Sultan University, Riyadh, Saudi Arabia
| |
Collapse
|
14
|
Dao QT, Trinh HQ, Nguyen VA. An effective and comprehensible method to detect and evaluate retinal damage due to diabetes complications. PeerJ Comput Sci 2023; 9:e1585. [PMID: 37810367 PMCID: PMC10557496 DOI: 10.7717/peerj-cs.1585] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/08/2023] [Accepted: 08/20/2023] [Indexed: 10/10/2023]
Abstract
The leading cause of vision loss globally is diabetic retinopathy. Researchers are making great efforts to automatically detect and diagnose correctly diabetic retinopathy. Diabetic retinopathy includes five stages: no diabetic retinopathy, mild diabetic retinopathy, moderate diabetic retinopathy, severe diabetic retinopathy and proliferative diabetic retinopathy. Recent studies have offered several multi-tasking deep learning models to detect and assess the level of diabetic retinopathy. However, the explanation for the assessment of disease severity of these models is limited, and only stops at showing lesions through images. These studies have not explained on what basis the appraisal of disease severity is based. In this article, we present a system for assessing and interpreting the five stages of diabetic retinopathy. The proposed system is built from internal models including a deep learning model that detects lesions and an explanatory model that assesses disease stage. The deep learning model that detects lesions uses the Mask R-CNN deep learning network to specify the location and shape of the lesion and classify the lesion types. This model is a combination of two networks: one used to detect hemorrhagic and exudative lesions, and one used to detect vascular lesions like aneurysm and proliferation. The explanatory model appraises disease severity based on the severity of each type of lesion and the association between types. The severity of the disease will be decided by the model based on the number of lesions, the density and the area of the lesions. The experimental results on real-world datasets show that our proposed method achieves high accuracy of assessing five stages of diabetic retinopathy comparable to existing state-of-the-art methods and is capable of explaining the causes of disease severity.
Collapse
Affiliation(s)
- Quang Toan Dao
- Institute of Information Technology, Vietnam Academy of Science and Technology, Hanoi, Vietnam
| | - Hoang Quan Trinh
- Vietnam Space Center, Vietnam Academy of Science and Technology, Hanoi, Vietnam
| | - Viet Anh Nguyen
- Institute of Information Technology, Vietnam Academy of Science and Technology, Hanoi, Vietnam
| |
Collapse
|
15
|
Oganov AC, Seddon I, Jabbehdari S, Uner OE, Fonoudi H, Yazdanpanah G, Outani O, Arevalo JF. Artificial intelligence in retinal image analysis: Development, advances, and challenges. Surv Ophthalmol 2023; 68:905-919. [PMID: 37116544 DOI: 10.1016/j.survophthal.2023.04.001] [Citation(s) in RCA: 11] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2022] [Revised: 04/20/2023] [Accepted: 04/24/2023] [Indexed: 04/30/2023]
Abstract
Modern advances in diagnostic technologies offer the potential for unprecedented insight into ophthalmic conditions relating to the retina. We discuss the current landscape of artificial intelligence in retina with respect to screening, diagnosis, and monitoring of retinal pathologies such as diabetic retinopathy, diabetic macular edema, central serous chorioretinopathy, and age-related macular degeneration. We review the methods used in these models and evaluate their performance in both research and clinical contexts and discuss potential future directions for investigation, use of multiple imaging modalities in artificial intelligence algorithms, and challenges in the application of artificial intelligence in retinal pathologies.
Collapse
Affiliation(s)
- Anthony C Oganov
- Department of Ophthalmology, Renaissance School of Medicine, Stony Brook, NY, USA
| | - Ian Seddon
- College of Osteopathic Medicine, Nova Southeastern University, Fort Lauderdale, FL, USA
| | - Sayena Jabbehdari
- Jones Eye Institute, University of Arkansas for Medical Sciences, Little Rock, AR, USA.
| | - Ogul E Uner
- Casey Eye Institute, Department of Ophthalmology, Oregon Health and Science University, Portland, OR, USA
| | - Hossein Fonoudi
- Eye Research Center, Farabi Eye Hospital, Tehran University of Medical Sciences, Iranshahr University of Medical Sciences, Iranshahr, Sistan and Baluchestan, Iran
| | - Ghasem Yazdanpanah
- Department of Ophthalmology and Visual Sciences, Illinois Eye and Ear Infirmary, University of Illinois at Chicago, Chicago, IL, USA
| | - Oumaima Outani
- Faculty of Medicine and Pharmacy of Rabat, Mohammed 5 University, Rabat, Rabat, Morocco
| | - J Fernando Arevalo
- Wilmer Eye Institute, Johns Hopkins University School of Medicine, Baltimore, MD, USA
| |
Collapse
|
16
|
Mohanty C, Mahapatra S, Acharya B, Kokkoras F, Gerogiannis VC, Karamitsos I, Kanavos A. Using Deep Learning Architectures for Detection and Classification of Diabetic Retinopathy. SENSORS (BASEL, SWITZERLAND) 2023; 23:5726. [PMID: 37420891 PMCID: PMC10301863 DOI: 10.3390/s23125726] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/21/2023] [Revised: 06/07/2023] [Accepted: 06/14/2023] [Indexed: 07/09/2023]
Abstract
Diabetic retinopathy (DR) is a common complication of long-term diabetes, affecting the human eye and potentially leading to permanent blindness. The early detection of DR is crucial for effective treatment, as symptoms often manifest in later stages. The manual grading of retinal images is time-consuming, prone to errors, and lacks patient-friendliness. In this study, we propose two deep learning (DL) architectures, a hybrid network combining VGG16 and XGBoost Classifier, and the DenseNet 121 network, for DR detection and classification. To evaluate the two DL models, we preprocessed a collection of retinal images obtained from the APTOS 2019 Blindness Detection Kaggle Dataset. This dataset exhibits an imbalanced image class distribution, which we addressed through appropriate balancing techniques. The performance of the considered models was assessed in terms of accuracy. The results showed that the hybrid network achieved an accuracy of 79.50%, while the DenseNet 121 model achieved an accuracy of 97.30%. Furthermore, a comparative analysis with existing methods utilizing the same dataset revealed the superior performance of the DenseNet 121 network. The findings of this study demonstrate the potential of DL architectures for the early detection and classification of DR. The superior performance of the DenseNet 121 model highlights its effectiveness in this domain. The implementation of such automated methods can significantly improve the efficiency and accuracy of DR diagnosis, benefiting both healthcare providers and patients.
Collapse
Affiliation(s)
- Cheena Mohanty
- Department of Electronics and Telecommunication, Biju Patnaik University of Technology, Rourkela 769012, Odisha, India;
| | - Sakuntala Mahapatra
- Department of Electronics and Telecommunication Engineering, Trident Academy of Technology, Bhubaneswar 751016, Odisha, India
| | - Biswaranjan Acharya
- Department of Computer Engineering-AI, Marwadi University, Rajkot 360003, Gujarat, India
| | - Fotis Kokkoras
- Department of Digital Systems, University of Thessaly, 41500 Larissa, Greece; (F.K.); (V.C.G.)
| | - Vassilis C. Gerogiannis
- Department of Digital Systems, University of Thessaly, 41500 Larissa, Greece; (F.K.); (V.C.G.)
| | - Ioannis Karamitsos
- Department of Graduate and Research, Rochester Institute of Technology, Dubai 341055, United Arab Emirates;
| | - Andreas Kanavos
- Department of Informatics, Ionian University, 49100 Corfu, Greece;
| |
Collapse
|
17
|
Chłopowiec AR, Karanowski K, Skrzypczak T, Grzesiuk M, Chłopowiec AB, Tabakov M. Counteracting Data Bias and Class Imbalance-Towards a Useful and Reliable Retinal Disease Recognition System. Diagnostics (Basel) 2023; 13:diagnostics13111904. [PMID: 37296756 DOI: 10.3390/diagnostics13111904] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/27/2023] [Revised: 05/22/2023] [Accepted: 05/25/2023] [Indexed: 06/12/2023] Open
Abstract
Multiple studies presented satisfactory performances for the treatment of various ocular diseases. To date, there has been no study that describes a multiclass model, medically accurate, and trained on large diverse dataset. No study has addressed a class imbalance problem in one giant dataset originating from multiple large diverse eye fundus image collections. To ensure a real-life clinical environment and mitigate the problem of biased medical image data, 22 publicly available datasets were merged. To secure medical validity only Diabetic Retinopathy (DR), Age-Related Macular Degeneration (AMD) and Glaucoma (GL) were included. The state-of-the-art models ConvNext, RegNet and ResNet were utilized. In the resulting dataset, there were 86,415 normal, 3787 GL, 632 AMD and 34,379 DR fundus images. ConvNextTiny achieved the best results in terms of recognizing most of the examined eye diseases with the most metrics. The overall accuracy was 80.46 ± 1.48. Specific accuracy values were: 80.01 ± 1.10 for normal eye fundus, 97.20 ± 0.66 for GL, 98.14 ± 0.31 for AMD, 80.66 ± 1.27 for DR. A suitable screening model for the most prevalent retinal diseases in ageing societies was designed. The model was developed on a diverse, combined large dataset which made the obtained results less biased and more generalizable.
Collapse
Affiliation(s)
- Adam R Chłopowiec
- Department of Artificial Intelligence, Wroclaw University of Science and Technology, Wybrzeże Wyspianskiego 27, 50-370 Wroclaw, Poland
| | - Konrad Karanowski
- Department of Artificial Intelligence, Wroclaw University of Science and Technology, Wybrzeże Wyspianskiego 27, 50-370 Wroclaw, Poland
| | - Tomasz Skrzypczak
- Faculty of Medicine, Wroclaw Medical University, Wybrzeże Ludwika Pasteura 1, 50-367 Wroclaw, Poland
| | - Mateusz Grzesiuk
- Department of Artificial Intelligence, Wroclaw University of Science and Technology, Wybrzeże Wyspianskiego 27, 50-370 Wroclaw, Poland
| | - Adrian B Chłopowiec
- Department of Artificial Intelligence, Wroclaw University of Science and Technology, Wybrzeże Wyspianskiego 27, 50-370 Wroclaw, Poland
| | - Martin Tabakov
- Department of Artificial Intelligence, Wroclaw University of Science and Technology, Wybrzeże Wyspianskiego 27, 50-370 Wroclaw, Poland
| |
Collapse
|
18
|
ExpACVO-Hybrid Deep learning: Exponential Anti Corona Virus Optimization enabled Hybrid Deep learning for tongue image segmentation towards diabetes mellitus detection. Biomed Signal Process Control 2023; 83:104635. [PMID: 36741196 PMCID: PMC9886667 DOI: 10.1016/j.bspc.2023.104635] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/16/2022] [Revised: 12/26/2022] [Accepted: 01/25/2023] [Indexed: 02/01/2023]
Abstract
A metabolic disease known as diabetes mellitus (DM) is primarily brought on by an increase in blood sugar levels. On the other hand, DM and the complications it causes, such as diabetic Retinopathy (DR), will quickly emerge as one of the major health challenges of the twenty-first century. This indicates a huge economic burden on health-related authorities and governments. The detection of DM in the earlier stage can lead to early diagnosis and a considerable drop in mortality. Therefore, in order to detect DM at an early stage, an efficient detection system having the ability to detect DM is required. An effective classification method, named Exponential Anti Corona Virus Optimization (ExpACVO) is devised in this research work for Diabetes Mellitus (DM) detection using tongue images. Here, the UNet-Conditional Random Field-Recurrent Neural Network (UNet-CRF-RNN) is used to segment the images, and the proposed ExpACVO algorithm is used to train the UNet-CRF-RNN. Deep Q Network (DQN) classifier is used for DM detection, and the proposed ExpACVO is used for DQN training. The proposed ExpACVO algorithm is a newly created formula that combines Anti Corona Virus Optimization(ACVO) with Exponential Weighted Moving Average (EWMA). With maximum testing accuracy, sensitivity, and specificity values of 0.932, 0.950, and 0.914, respectively, the developed technique thus achieved improved performance.
Collapse
|
19
|
Muchuchuti S, Viriri S. Retinal Disease Detection Using Deep Learning Techniques: A Comprehensive Review. J Imaging 2023; 9:84. [PMID: 37103235 PMCID: PMC10145952 DOI: 10.3390/jimaging9040084] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2023] [Revised: 04/02/2023] [Accepted: 04/07/2023] [Indexed: 04/28/2023] Open
Abstract
Millions of people are affected by retinal abnormalities worldwide. Early detection and treatment of these abnormalities could arrest further progression, saving multitudes from avoidable blindness. Manual disease detection is time-consuming, tedious and lacks repeatability. There have been efforts to automate ocular disease detection, riding on the successes of the application of Deep Convolutional Neural Networks (DCNNs) and vision transformers (ViTs) for Computer-Aided Diagnosis (CAD). These models have performed well, however, there remain challenges owing to the complex nature of retinal lesions. This work reviews the most common retinal pathologies, provides an overview of prevalent imaging modalities and presents a critical evaluation of current deep-learning research for the detection and grading of glaucoma, diabetic retinopathy, Age-Related Macular Degeneration and multiple retinal diseases. The work concluded that CAD, through deep learning, will increasingly be vital as an assistive technology. As future work, there is a need to explore the potential impact of using ensemble CNN architectures in multiclass, multilabel tasks. Efforts should also be expended on the improvement of model explainability to win the trust of clinicians and patients.
Collapse
Affiliation(s)
| | - Serestina Viriri
- School of Mathematics, Statistics and Computer Science, University of KwaZulu-Natal, Durban 4001, South Africa
| |
Collapse
|
20
|
Tavana P, Akraminia M, Koochari A, Bagherifard A. Classification of spinal curvature types using radiography images: deep learning versus classical methods. Artif Intell Rev 2023; 56:1-33. [PMID: 37362895 PMCID: PMC10088798 DOI: 10.1007/s10462-023-10480-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/28/2023]
Abstract
Scoliosis is a spinal abnormality that has two types of curves (C-shaped or S-shaped). The vertebrae of the spine reach an equilibrium at different times, which makes it challenging to detect the type of curves. In addition, it may be challenging to detect curvatures due to observer bias and image quality. This paper aims to evaluate spinal deformity by automatically classifying the type of spine curvature. Automatic spinal curvature classification is performed using SVM and KNN algorithms, and pre-trained Xception and MobileNetV2 networks with SVM as the final activation function to avoid vanishing gradient. Different feature extraction methods should be used to investigate the SVM and KNN machine learning methods in detecting the curvature type. Features are extracted through the representation of radiographic images. These representations are of two groups: (i) Low-level image representation techniques such as texture features and (ii) local patch-based representations such as Bag of Words (BoW). Such features are utilized by various algorithms for classification by SVM and KNN. The feature extraction process is automated in pre-trained deep networks. In this study, 1000 anterior-posterior (AP) radiographic images of the spine were collected as a private dataset from Shafa Hospital, Tehran, Iran. The transfer learning was used due to the relatively small private dataset of anterior-posterior radiology images of the spine. Based on the results of these experiments, pre-trained deep networks were found to be approximately 10% more accurate than classical methods in classifying whether the spinal curvature is C-shaped or S-shaped. As a result of automatic feature extraction, it has been found that the pre-trained Xception and mobilenetV2 networks with SVM as the final activation function for controlling the vanishing gradient perform better than the classical machine learning methods of classification of spinal curvature types.
Collapse
Affiliation(s)
- Parisa Tavana
- Department of Computer Engineering, Science and Research Branch, Islamic Azad University, Tehran, Iran
| | - Mahdi Akraminia
- Mechanical Rotary Equipment Research Department, Niroo Research Institute, Tehran, Iran
| | - Abbas Koochari
- Department of Computer Engineering, Science and Research Branch, Islamic Azad University, Tehran, Iran
| | - Abolfazl Bagherifard
- Bone and Joint Reconstruction Research Center, Shafa Orthopedic Hospital, Iran University of Medical Sciences, Tehran, Iran
| |
Collapse
|
21
|
Mohan NJ, Murugan R, Goel T, Tanveer M, Roy P. An efficient microaneurysms detection approach in retinal fundus images. INT J MACH LEARN CYB 2023. [DOI: 10.1007/s13042-022-01696-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/19/2023]
|
22
|
Computational intelligence in eye disease diagnosis: a comparative study. Med Biol Eng Comput 2023; 61:593-615. [PMID: 36595155 DOI: 10.1007/s11517-022-02737-3] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/21/2021] [Accepted: 12/09/2022] [Indexed: 01/04/2023]
Abstract
In recent years, eye disorders are an important health issue among older people. Generally, individuals with eye diseases are unaware of the gradual growth of symptoms. Therefore, routine eye examinations are required for early diagnosis. Usually, eye disorders are identified by an ophthalmologist via a slit-lamp investigation. Slit-lamp interpretations are inadequate due to the differences in the analytical skills of the ophthalmologist, inconsistency in eye disorder analysis, and record maintenance issues. Therefore, digital images of an eye and computational intelligence (CI)-based approaches are preferred as assistive methods for eye disease diagnosis. A comparative study of CI-based decision support models for eye disorder diagnosis is presented in this paper. The CI-based decision support systems used for eye abnormalities diagnosis were grouped as anterior and retinal eye abnormalities diagnostic systems, and numerous algorithms used for diagnosing the eye abnormalities were also briefed. Various eye imaging modalities, pre-processing methods such as reflection removal, contrast enhancement, region of interest segmentation methods, and public eye image databases used for CI-based eye disease diagnosis system development were also discussed in this paper. In this comparative study, the reliability of various CI-based systems used for anterior eye and retinal disorder diagnosis was compared based on the precision, sensitivity, and specificity in eye disease diagnosis. The outcomes of the comparative analysis indicate that the CI-based anterior and retinal disease diagnosis systems attained significant prediction accuracy. Hence, these CI-based diagnosis systems can be used in clinics to reduce the burden on physicians, minimize fatigue-related misdetection, and take precise clinical decisions.
Collapse
|
23
|
Soares I, Castelo-Branco M, Pinheiro A. Microaneurysms detection in retinal images using a multi-scale approach. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104184] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
24
|
Iqbal S, Khan TM, Naveed K, Naqvi SS, Nawaz SJ. Recent trends and advances in fundus image analysis: A review. Comput Biol Med 2022; 151:106277. [PMID: 36370579 DOI: 10.1016/j.compbiomed.2022.106277] [Citation(s) in RCA: 11] [Impact Index Per Article: 3.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/13/2022] [Revised: 10/19/2022] [Accepted: 10/30/2022] [Indexed: 11/05/2022]
Abstract
Automated retinal image analysis holds prime significance in the accurate diagnosis of various critical eye diseases that include diabetic retinopathy (DR), age-related macular degeneration (AMD), atherosclerosis, and glaucoma. Manual diagnosis of retinal diseases by ophthalmologists takes time, effort, and financial resources, and is prone to error, in comparison to computer-aided diagnosis systems. In this context, robust classification and segmentation of retinal images are primary operations that aid clinicians in the early screening of patients to ensure the prevention and/or treatment of these diseases. This paper conducts an extensive review of the state-of-the-art methods for the detection and segmentation of retinal image features. Existing notable techniques for the detection of retinal features are categorized into essential groups and compared in depth. Additionally, a summary of quantifiable performance measures for various important stages of retinal image analysis, such as image acquisition and preprocessing, is provided. Finally, the widely used in the literature datasets for analyzing retinal images are described and their significance is emphasized.
Collapse
Affiliation(s)
- Shahzaib Iqbal
- Department of Electrical and Computer Engineering, COMSATS University Islamabad (CUI), Islamabad, Pakistan
| | - Tariq M Khan
- School of Computer Science and Engineering, University of New South Wales, Sydney, NSW, Australia.
| | - Khuram Naveed
- Department of Electrical and Computer Engineering, COMSATS University Islamabad (CUI), Islamabad, Pakistan; Department of Electrical and Computer Engineering, Aarhus University, Aarhus, Denmark
| | - Syed S Naqvi
- Department of Electrical and Computer Engineering, COMSATS University Islamabad (CUI), Islamabad, Pakistan
| | - Syed Junaid Nawaz
- Department of Electrical and Computer Engineering, COMSATS University Islamabad (CUI), Islamabad, Pakistan
| |
Collapse
|
25
|
Deep Learning-Based Glaucoma Screening Using Regional RNFL Thickness in Fundus Photography. Diagnostics (Basel) 2022; 12:diagnostics12112894. [PMID: 36428954 PMCID: PMC9689347 DOI: 10.3390/diagnostics12112894] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/20/2022] [Revised: 11/14/2022] [Accepted: 11/14/2022] [Indexed: 11/23/2022] Open
Abstract
Since glaucoma is a progressive and irreversible optic neuropathy, accurate screening and/or early diagnosis is critical in preventing permanent vision loss. Recently, optical coherence tomography (OCT) has become an accurate diagnostic tool to observe and extract the thickness of the retinal nerve fiber layer (RNFL), which closely reflects the nerve damage caused by glaucoma. However, OCT is less accessible than fundus photography due to higher cost and expertise required for operation. Though widely used, fundus photography is effective for early glaucoma detection only when used by experts with extensive training. Here, we introduce a deep learning-based approach to predict the RNFL thickness around optic disc regions in fundus photography for glaucoma screening. The proposed deep learning model is based on a convolutional neural network (CNN) and utilizes images taken with fundus photography and with RNFL thickness measured with OCT for model training and validation. Using a dataset acquired from normal tension glaucoma (NTG) patients, the trained model can estimate RNFL thicknesses in 12 optic disc regions from fundus photos. Using intuitive thickness labels to identify localized damage of the optic nerve head and then estimating regional RNFL thicknesses from fundus images, we determine that screening for glaucoma could achieve 92% sensitivity and 86.9% specificity. Receiver operating characteristic (ROC) analysis results for specificity of 80% demonstrate that use of the localized mean over superior and inferior regions reaches 90.7% sensitivity, whereas 71.2% sensitivity is reached using the global RNFL thicknesses for specificity at 80%. This demonstrates that the new approach of using regional RNFL thicknesses in fundus images holds good promise as a potential screening technique for early stage of glaucoma.
Collapse
|
26
|
An automated unsupervised deep learning–based approach for diabetic retinopathy detection. Med Biol Eng Comput 2022; 60:3635-3654. [DOI: 10.1007/s11517-022-02688-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2022] [Accepted: 10/02/2022] [Indexed: 11/07/2022]
|
27
|
Dubey S, Dixit M. Recent developments on computer aided systems for diagnosis of diabetic retinopathy: a review. MULTIMEDIA TOOLS AND APPLICATIONS 2022; 82:14471-14525. [PMID: 36185322 PMCID: PMC9510498 DOI: 10.1007/s11042-022-13841-9] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 09/14/2021] [Revised: 04/27/2022] [Accepted: 09/06/2022] [Indexed: 06/16/2023]
Abstract
Diabetes is a long-term condition in which the pancreas quits producing insulin or the body's insulin isn't utilised properly. One of the signs of diabetes is Diabetic Retinopathy. Diabetic retinopathy is the most prevalent type of diabetes, if remains unaddressed, diabetic retinopathy can affect all diabetics and become very serious, raising the chances of blindness. It is a chronic systemic condition that affects up to 80% of patients for more than ten years. Many researchers believe that if diabetes individuals are diagnosed early enough, they can be rescued from the condition in 90% of cases. Diabetes damages the capillaries, which are microscopic blood vessels in the retina. On images, blood vessel damage is usually noticeable. Therefore, in this study, several traditional, as well as deep learning-based approaches, are reviewed for the classification and detection of this particular diabetic-based eye disease known as diabetic retinopathy, and also the advantage of one approach over the other is also described. Along with the approaches, the dataset and the evaluation metrics useful for DR detection and classification are also discussed. The main finding of this study is to aware researchers about the different challenges occurs while detecting diabetic retinopathy using computer vision, deep learning techniques. Therefore, a purpose of this review paper is to sum up all the major aspects while detecting DR like lesion identification, classification and segmentation, security attacks on the deep learning models, proper categorization of datasets and evaluation metrics. As deep learning models are quite expensive and more prone to security attacks thus, in future it is advisable to develop a refined, reliable and robust model which overcomes all these aspects which are commonly found while designing deep learning models.
Collapse
Affiliation(s)
- Shradha Dubey
- Madhav Institute of Technology & Science (Department of Computer Science and Engineering), Gwalior, M.P. India
| | - Manish Dixit
- Madhav Institute of Technology & Science (Department of Computer Science and Engineering), Gwalior, M.P. India
| |
Collapse
|
28
|
Karsaz A. A modified convolutional neural network architecture for diabetic retinopathy screening using SVDD. Appl Soft Comput 2022. [DOI: 10.1016/j.asoc.2022.109102] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
29
|
Diabetic retinopathy screening using improved support vector domain description: a clinical study. Soft comput 2022. [DOI: 10.1007/s00500-022-07387-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/16/2022]
|
30
|
Detection of microaneurysms in color fundus images based on local Fourier transform. Biomed Signal Process Control 2022. [DOI: 10.1016/j.bspc.2022.103648] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
|
31
|
Liu R, Wang X, Wu Q, Dai L, Fang X, Yan T, Son J, Tang S, Li J, Gao Z, Galdran A, Poorneshwaran J, Liu H, Wang J, Chen Y, Porwal P, Wei Tan GS, Yang X, Dai C, Song H, Chen M, Li H, Jia W, Shen D, Sheng B, Zhang P. DeepDRiD: Diabetic Retinopathy-Grading and Image Quality Estimation Challenge. PATTERNS (NEW YORK, N.Y.) 2022; 3:100512. [PMID: 35755875 PMCID: PMC9214346 DOI: 10.1016/j.patter.2022.100512] [Citation(s) in RCA: 16] [Impact Index Per Article: 5.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 03/14/2022] [Revised: 03/28/2022] [Accepted: 04/25/2022] [Indexed: 12/19/2022]
Abstract
We described a challenge named "Diabetic Retinopathy (DR)-Grading and Image Quality Estimation Challenge" in conjunction with ISBI 2020 to hold three sub-challenges and develop deep learning models for DR image assessment and grading. The scientific community responded positively to the challenge, with 34 submissions from 574 registrations. In the challenge, we provided the DeepDRiD dataset containing 2,000 regular DR images (500 patients) and 256 ultra-widefield images (128 patients), both having DR quality and grading annotations. We discussed details of the top 3 algorithms in each sub-challenges. The weighted kappa for DR grading ranged from 0.93 to 0.82, and the accuracy for image quality evaluation ranged from 0.70 to 0.65. The results showed that image quality assessment can be used as a further target for exploration. We also have released the DeepDRiD dataset on GitHub to help develop automatic systems and improve human judgment in DR screening and diagnosis.
Collapse
Affiliation(s)
- Ruhan Liu
- Department of Computer Science and Engineering, Shanghai Jiao Tong University, Shanghai, China
- MoE Key Lab of Artificial Intelligence, Artificial Intelligence Institute, Shanghai Jiao Tong University, Shanghai, China
| | - Xiangning Wang
- Department of Ophthalmology, Shanghai Jiao Tong University Affiliated Sixth People’s Hospital, Shanghai, China
| | - Qiang Wu
- Department of Ophthalmology, Shanghai Jiao Tong University Affiliated Sixth People’s Hospital, Shanghai, China
| | - Ling Dai
- Department of Computer Science and Engineering, Shanghai Jiao Tong University, Shanghai, China
- MoE Key Lab of Artificial Intelligence, Artificial Intelligence Institute, Shanghai Jiao Tong University, Shanghai, China
| | - Xi Fang
- Department of Electronic Engineering, Shanghai Jiao Tong University, Shanghai, 200240, China
| | - Tao Yan
- Department of Electromechanical Engineering, University of Macau, Macao, China
| | | | - Shiqi Tang
- Department of Mathematics, City University of Hong Kong, Hong Kong, China
| | - Jiang Li
- Institute of Image Processing and Pattern Recognition, Department of Automation, Shanghai Jiao Tong University, Shanghai, China
| | - Zijian Gao
- School of Electronic Information, Hangzhou Dianzi University, Hangzhou, China
| | | | | | - Hao Liu
- School of Electronic Information, Hangzhou Dianzi University, Hangzhou, China
| | - Jie Wang
- School of Computer Science and Engineering, Beihang University, Beijing, China
| | - Yerui Chen
- Nanjing University of Science and Technology, Nanjing, China
| | - Prasanna Porwal
- Shri Guru Gobind Singhji Institute of Engineering and Technology, Nanded, India
| | - Gavin Siew Wei Tan
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore
| | - Xiaokang Yang
- MoE Key Lab of Artificial Intelligence, Artificial Intelligence Institute, Shanghai Jiao Tong University, Shanghai, China
| | - Chao Dai
- Shanghai Zhi Tang Health Technology Co., LTD., China
| | - Haitao Song
- MoE Key Lab of Artificial Intelligence, Artificial Intelligence Institute, Shanghai Jiao Tong University, Shanghai, China
| | - Mingang Chen
- Shanghai Key Laboratory of Computer Software Testing & Evaluating, Shanghai Development Center of Computer Software Technology, Shanghai, China
| | - Huating Li
- Department of Endocrinology and Metabolism, Shanghai Jiao Tong University Affiliated Sixth People’s Hospital, Shanghai, China
- Shanghai Diabetes Institute, Shanghai Clinical Center for Diabetes, Shanghai, China
| | - Weiping Jia
- Department of Endocrinology and Metabolism, Shanghai Jiao Tong University Affiliated Sixth People’s Hospital, Shanghai, China
- Shanghai Diabetes Institute, Shanghai Clinical Center for Diabetes, Shanghai, China
| | - Dinggang Shen
- School of Biomedical Engineering, ShanghaiTech University, Shanghai, China
- Department of Research and Development, Shanghai United Imaging Intelligence Co., Ltd., Shanghai, China
| | - Bin Sheng
- Department of Computer Science and Engineering, Shanghai Jiao Tong University, Shanghai, China
- MoE Key Lab of Artificial Intelligence, Artificial Intelligence Institute, Shanghai Jiao Tong University, Shanghai, China
| | - Ping Zhang
- Department of Computer Science and Engineering, The Ohio State University, Ohio, USA
- Department of Biomedical Informatics, The Ohio State University, Ohio, USA
- Translational Data Analytics Institute, The Ohio State University, Ohio, USA
| |
Collapse
|
32
|
Automated grading of diabetic retinopathy using CNN with hierarchical clustering of image patches by siamese network. Phys Eng Sci Med 2022; 45:623-635. [PMID: 35587313 DOI: 10.1007/s13246-022-01129-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/20/2021] [Accepted: 04/19/2022] [Indexed: 10/18/2022]
Abstract
Diabetic retinopathy (DR) is a progressive vascular complication that affects people who have diabetes. This retinal abnormality can cause irreversible vision loss or permanent blindness; therefore, it is crucial to undergo frequent eye screening for early recognition and treatment. This paper proposes a feature extraction algorithm using discriminative multi-sized patches, based on deep learning convolutional neural network (CNN) for DR grading. This comprehensive algorithm extracts local and global features for efficient decision-making. Each input image is divided into small-sized patches to extract local-level features and then split into clusters or subsets. Hierarchical clustering by Siamese network with pre-trained CNN is proposed in this paper to select clusters with more discriminative patches. The fine-tuned Xception model of CNN is used to extract the global-level features of larger image patches. Local and global features are combined to improve the overall image-wise classification accuracy. The final support vector machine classifier exhibits 96% of classification accuracy with tenfold cross-validation in classifying DR images.
Collapse
|
33
|
Du J, Zou B, Ouyang P, Zhao R. Retinal microaneurysm detection based on transformation splicing and multi-context ensemble learning. Biomed Signal Process Control 2022. [DOI: 10.1016/j.bspc.2022.103536] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/02/2023]
|
34
|
Kako NA, Abdulazeez AM. Peripapillary Atrophy Segmentation and Classification Methodologies for Glaucoma Image Detection: A Review. Curr Med Imaging 2022; 18:1140-1159. [PMID: 35260060 DOI: 10.2174/1573405618666220308112732] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/15/2021] [Revised: 12/04/2021] [Accepted: 12/22/2021] [Indexed: 11/22/2022]
Abstract
Information-based image processing and computer vision methods are utilized in several healthcare organizations to diagnose diseases. The irregularities in the visual system are identified over fundus images shaped over a fundus camera. Among ophthalmology diseases, glaucoma is measured as the most common case that can lead to neurodegenerative illness. The unsuitable fluid pressure inside the eye within the visual system is described as the major cause of those diseases. Glaucoma has no symptoms in the early stages, and if it is not treated, it may result in total blindness. Diagnosing glaucoma at an early stage may prevent permanent blindness. Manual inspection of the human eye may be a solution, but it depends on the skills of the individuals involved. The auto diagnosis of glaucoma by applying a consolidation of computer vision, artificial intelligence, and image processing can aid in the ban and detection of those diseases. In this review article, we aim to introduce a review of the numerous approaches based on peripapillary atrophy segmentation and classification that can detect these diseases, as well as details about the publicly available image benchmarks, datasets, and measurement of performance. The review article introduces the demonstrated research of numerous available study models that objectively diagnose glaucoma via peripapillary atrophy from the lowest level of feature extraction to the current direction based on deep learning. The advantages and disadvantages of each method are addressed in detail, and tabular descriptions are included to highlight the results of each category. Moreover, the frameworks of each approach and fundus image datasets are provided. The improved reporting of our study would help in providing possible future work directions to diagnose glaucoma in conclusion.
Collapse
Affiliation(s)
- Najdavan A Kako
- Duhok Polytechnic University, Technical Institute of Administration, MIS, Duhok, Iraq
| | | |
Collapse
|
35
|
Deep Red Lesion Classification for Early Screening of Diabetic Retinopathy. MATHEMATICS 2022. [DOI: 10.3390/math10050686] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
Abstract
Diabetic retinopathy (DR) is an asymptotic and vision-threatening complication among working-age adults. To prevent blindness, a deep convolutional neural network (CNN) based diagnosis can help to classify less-discriminative and small-sized red lesions in early screening of DR patients. However, training deep models with minimal data is a challenging task. Fine-tuning through transfer learning is a useful alternative, but performance degradation, overfitting, and domain adaptation issues further demand architectural amendments to effectively train deep models. Various pre-trained CNNs are fine-tuned on an augmented set of image patches. The best-performing ResNet50 model is modified by introducing reinforced skip connections, a global max-pooling layer, and the sum-of-squared-error loss function. The performance of the modified model (DR-ResNet50) on five public datasets is found to be better than state-of-the-art methods in terms of well-known metrics. The highest scores (0.9851, 0.991, 0.991, 0.991, 0.991, 0.9939, 0.0029, 0.9879, and 0.9879) for sensitivity, specificity, AUC, accuracy, precision, F1-score, false-positive rate, Matthews’s correlation coefficient, and kappa coefficient are obtained within a 95% confidence interval for unseen test instances from e-Ophtha_MA. This high sensitivity and low false-positive rate demonstrate the worth of a proposed framework. It is suitable for early screening due to its performance, simplicity, and robustness.
Collapse
|
36
|
MicroNet: microaneurysm detection in retinal fundus images using convolutional neural network. Soft comput 2022. [DOI: 10.1007/s00500-022-06752-2] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/24/2022]
|
37
|
Local Structure Awareness-Based Retinal Microaneurysm Detection with Multi-Feature Combination. Biomedicines 2022; 10:biomedicines10010124. [PMID: 35052803 PMCID: PMC8773350 DOI: 10.3390/biomedicines10010124] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/10/2021] [Revised: 12/31/2021] [Accepted: 01/03/2022] [Indexed: 01/02/2023] Open
Abstract
Retinal microaneurysm (MA) is the initial symptom of diabetic retinopathy (DR). The automatic detection of MA is helpful to assist doctors in diagnosis and treatment. Previous algorithms focused on the features of the target itself; however, the local structural features of the target and background are also worth exploring. To achieve MA detection, an efficient local structure awareness-based retinal MA detection with the multi-feature combination (LSAMFC) is proposed in this paper. We propose a novel local structure feature called a ring gradient descriptor (RGD) to describe the structural differences between an object and its surrounding area. Then, a combination of RGD with the salience and texture features is used by a Gradient Boosting Decision Tree (GBDT) for candidate classification. We evaluate our algorithm on two public datasets, i.e., the e-ophtha MA dataset and retinopathy online challenge (ROC) dataset. The experimental results show that the performance of the trained model significantly improved after combining traditional features with RGD, and the area under the receiver operating characteristic curve (AUC) values in the test results of the datasets e-ophtha MA and ROC increased from 0.9615 to 0.9751 and from 0.9066 to 0.9409, respectively.
Collapse
|
38
|
Xu X, Li J, Guan Y, Zhao L, Zhao Q, Zhang L, Li L. GLA-Net: A global-local attention network for automatic cataract classification. J Biomed Inform 2021; 124:103939. [PMID: 34752858 DOI: 10.1016/j.jbi.2021.103939] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/01/2021] [Revised: 10/02/2021] [Accepted: 10/25/2021] [Indexed: 10/19/2022]
Abstract
Cataracts are the most crucial cause of blindness among all ophthalmic diseases. Convenient and cost-effective early cataract screening is urgently needed to reduce the risks of visual loss. To date, many studies have investigated automatic cataract classification based on fundus images. However, existing methods mainly rely on global image information while ignoring various local and subtle features. Notably, these local features are highly helpful for the identification of cataracts with different severities. To avoid this disadvantage, we introduce a deep learning technique to learn multilevel feature representations of the fundus image simultaneously. Specifically, a global-local attention network (GLA-Net) is proposed to handle the cataract classification task, which consists of two levels of subnets: the global-level attention subnet pays attention to the global structure information of the fundus image, while the local-level attention subnet focuses on the local discriminative features of the specific regions. These two types of subnets extract retinal features at different attention levels, which are then combined for final cataract classification. Our GLA-Net achieves the best performance in all metrics (90.65% detection accuracy, 83.47% grading accuracy, and 81.11% classification accuracy of grades 1 and 2). The experimental results on a real clinical dataset show that the combination of global-level and local-level attention models is effective for cataract screening and provides significant potential for other medical tasks.
Collapse
Affiliation(s)
- Xi Xu
- Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China
| | - Jianqiang Li
- Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China
| | - Yu Guan
- Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China
| | - Linna Zhao
- Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China
| | - Qing Zhao
- Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China.
| | - Li Zhang
- Beijing Tongren Eye Center, Beijing Tongren Hospital, Capital Medical University, Beijing, China
| | - Li Li
- National Center for Children's Health, Beijing Children's Hospital, Capital Medical University, Beijing, China
| |
Collapse
|
39
|
Shekar S, Satpute N, Gupta A. Review on diabetic retinopathy with deep learning methods. JOURNAL OF MEDICAL IMAGING (BELLINGHAM, WASH.) 2021; 8:060901. [PMID: 34859116 DOI: 10.1117/1.jmi.8.6.060901] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Subscribe] [Scholar Register] [Received: 05/13/2021] [Accepted: 10/27/2021] [Indexed: 11/14/2022]
Abstract
Purpose: The purpose of our review paper is to examine many existing works of literature presenting the different methods utilized for diabetic retinopathy (DR) recognition employing deep learning (DL) and machine learning (ML) techniques, and also to address the difficulties faced in various datasets used by DR. Approach: DR is a progressive illness and may become a reason for vision loss. Early identification of DR lesions is, therefore, helpful and prevents damage to the retina. However, it is a complex job in view of the fact that it is symptomless earlier, and also ophthalmologists have been needed in traditional approaches. Recently, automated identification of DR-based studies has been stated based on image processing, ML, and DL. We analyze the recent literature and provide a comparative study that also includes the limitations of the literature and future work directions. Results: A relative analysis among the databases used, performance metrics employed, and ML and DL techniques adopted recently in DR detection based on various DR features is presented. Conclusion: Our review paper discusses the methods employed in DR detection along with the technical and clinical challenges that are encountered, which is missing in existing reviews, as well as future scopes to assist researchers in the field of retinal imaging.
Collapse
Affiliation(s)
- Shreya Shekar
- College of Engineering Pune, Department of Electronics and Telecommunication Engineering, Pune, Maharashtra, India
| | - Nitin Satpute
- Aarhus University, Department of Electrical and Computer Engineering, Aarhus, Denmark
| | - Aditya Gupta
- College of Engineering Pune, Department of Electronics and Telecommunication Engineering, Pune, Maharashtra, India
| |
Collapse
|
40
|
Sun S, Cao Z, Liao D, Lv R. A Magnified Adaptive Feature Pyramid Network for automatic microaneurysms detection. Comput Biol Med 2021; 139:105000. [PMID: 34741905 DOI: 10.1016/j.compbiomed.2021.105000] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/22/2021] [Revised: 10/27/2021] [Accepted: 10/27/2021] [Indexed: 10/19/2022]
Abstract
Diabetic retinopathy (DR), as an important complication of diabetes, is the primary cause of blindness in adults. Automatic DR detection poses a challenge which is crucial for early DR screening. Currently, the vast majority of DR is diagnosed through fundus images, where the microaneurysm (MA) has been widely used as the most distinguishable marker. Research works on automatic DR detection have traditionally utilized manually designed operators, while a few recent researchers have explored deep learning techniques for this topic. But due to issues such as the extremely small size of microaneurysms, low resolution of fundus pictures, and insufficient imaging depth, the DR detection problem is quite challenging and remains unsolved. To address these issues, this research proposes a new deep learning model (Magnified Adaptive Feature Pyramid Network, MAFP-Net) for DR detection, which conducts super-resolution on low quality fundus images and integrates an improved feature pyramid structure while utilizing a standard two-stage detection network as the backbone. Our proposed detection model needs no pre-segmented patches to train the CNN network. When tested on the E-ophtha-MA dataset, the sensitivity value of our method reached as high as 83.5% at false positives per image (FPI) of 8 and the F1 value achieved 0.676, exceeding all those of the state-of-the-art algorithms as well as the human performance of experienced physicians. Similar results were achieved on another public dataset of IDRiD.
Collapse
Affiliation(s)
- Song Sun
- Molecular and Neuroimaging Engineering Research Center of Ministry of Education, School of Life Science and Technology, Xidian University, Xi'an, 710071, China
| | - Zhicheng Cao
- Molecular and Neuroimaging Engineering Research Center of Ministry of Education, School of Life Science and Technology, Xidian University, Xi'an, 710071, China
| | - Dingying Liao
- Health Science Center, Xi'an Jiaotong University, Xi'an, 710061, China
| | - Ruichan Lv
- Molecular and Neuroimaging Engineering Research Center of Ministry of Education, School of Life Science and Technology, Xidian University, Xi'an, 710071, China.
| |
Collapse
|
41
|
Xia H, Lan Y, Song S, Li H. A multi-scale segmentation-to-classification network for tiny microaneurysm detection in fundus images. Knowl Based Syst 2021. [DOI: 10.1016/j.knosys.2021.107140] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
42
|
Wang YL, Yang JY, Yang JY, Zhao XY, Chen YX, Yu WH. Progress of artificial intelligence in diabetic retinopathy screening. Diabetes Metab Res Rev 2021; 37:e3414. [PMID: 33010796 DOI: 10.1002/dmrr.3414] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 02/10/2020] [Revised: 08/22/2020] [Accepted: 08/23/2020] [Indexed: 12/29/2022]
Abstract
Diabetic retinopathy (DR) is one of the leading causes of blindness worldwide, and the limited availability of qualified ophthalmologists restricts its early diagnosis. For the past few years, artificial intelligence technology has developed rapidly and has been applied in DR screening. The upcoming technology provides support on DR screening and improves the identification of DR lesions with a high sensitivity and specificity. This review aims to summarize the progress on automatic detection and classification models for the diagnosis of DR.
Collapse
Affiliation(s)
- Yue-Lin Wang
- Department of Ophthalmology, Peking Union Medical College Hospital & Chinese Academy of Medical Sciences, Beijing, China
- Key Laboratory of Ocular Fundus Diseases, Chinese Academy of Medical Sciences & Peking Union Medical College, Beijing, China
| | - Jing-Yun Yang
- Division of Statistics, School of Economics & Research Center of Financial Information, Shanghai University, Shanghai, China
- Rush Alzheimer's Disease Center & Department of Neurological Sciences, Rush University Medical Center, Chicago, Illinois, USA
| | - Jing-Yuan Yang
- Department of Ophthalmology, Peking Union Medical College Hospital & Chinese Academy of Medical Sciences, Beijing, China
- Key Laboratory of Ocular Fundus Diseases, Chinese Academy of Medical Sciences & Peking Union Medical College, Beijing, China
| | - Xin-Yu Zhao
- Department of Ophthalmology, Peking Union Medical College Hospital & Chinese Academy of Medical Sciences, Beijing, China
- Key Laboratory of Ocular Fundus Diseases, Chinese Academy of Medical Sciences & Peking Union Medical College, Beijing, China
| | - You-Xin Chen
- Department of Ophthalmology, Peking Union Medical College Hospital & Chinese Academy of Medical Sciences, Beijing, China
- Key Laboratory of Ocular Fundus Diseases, Chinese Academy of Medical Sciences & Peking Union Medical College, Beijing, China
| | - Wei-Hong Yu
- Department of Ophthalmology, Peking Union Medical College Hospital & Chinese Academy of Medical Sciences, Beijing, China
- Key Laboratory of Ocular Fundus Diseases, Chinese Academy of Medical Sciences & Peking Union Medical College, Beijing, China
| |
Collapse
|
43
|
Deep learning for diabetic retinopathy detection and classification based on fundus images: A review. Comput Biol Med 2021; 135:104599. [PMID: 34247130 DOI: 10.1016/j.compbiomed.2021.104599] [Citation(s) in RCA: 63] [Impact Index Per Article: 15.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/07/2021] [Revised: 06/12/2021] [Accepted: 06/18/2021] [Indexed: 02/02/2023]
Abstract
Diabetic Retinopathy is a retina disease caused by diabetes mellitus and it is the leading cause of blindness globally. Early detection and treatment are necessary in order to delay or avoid vision deterioration and vision loss. To that end, many artificial-intelligence-powered methods have been proposed by the research community for the detection and classification of diabetic retinopathy on fundus retina images. This review article provides a thorough analysis of the use of deep learning methods at the various steps of the diabetic retinopathy detection pipeline based on fundus images. We discuss several aspects of that pipeline, ranging from the datasets that are widely used by the research community, the preprocessing techniques employed and how these accelerate and improve the models' performance, to the development of such deep learning models for the diagnosis and grading of the disease as well as the localization of the disease's lesions. We also discuss certain models that have been applied in real clinical settings. Finally, we conclude with some important insights and provide future research directions.
Collapse
|
44
|
Ashraf MN, Hussain M, Habib Z. Review of Various Tasks Performed in the Preprocessing Phase of a Diabetic Retinopathy Diagnosis System. Curr Med Imaging 2021; 16:397-426. [PMID: 32410541 DOI: 10.2174/1573405615666190219102427] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/04/2018] [Revised: 12/31/2018] [Accepted: 01/20/2019] [Indexed: 12/15/2022]
Abstract
Diabetic Retinopathy (DR) is a major cause of blindness in diabetic patients. The increasing population of diabetic patients and difficulty to diagnose it at an early stage are limiting the screening capabilities of manual diagnosis by ophthalmologists. Color fundus images are widely used to detect DR lesions due to their comfortable, cost-effective and non-invasive acquisition procedure. Computer Aided Diagnosis (CAD) of DR based on these images can assist ophthalmologists and help in saving many sight years of diabetic patients. In a CAD system, preprocessing is a crucial phase, which significantly affects its performance. Commonly used preprocessing operations are the enhancement of poor contrast, balancing the illumination imbalance due to the spherical shape of a retina, noise reduction, image resizing to support multi-resolution, color normalization, extraction of a field of view (FOV), etc. Also, the presence of blood vessels and optic discs makes the lesion detection more challenging because these two artifacts exhibit specific attributes, which are similar to those of DR lesions. Preprocessing operations can be broadly divided into three categories: 1) fixing the native defects, 2) segmentation of blood vessels, and 3) localization and segmentation of optic discs. This paper presents a review of the state-of-the-art preprocessing techniques related to three categories of operations, highlighting their significant aspects and limitations. The survey is concluded with the most effective preprocessing methods, which have been shown to improve the accuracy and efficiency of the CAD systems.
Collapse
Affiliation(s)
| | - Muhammad Hussain
- Department of Computer Science, College of Computer and Information Sciences, King Saud University, Riyadh, Saudi Arabia
| | - Zulfiqar Habib
- Department of Computer Science, COMSATS University Islamabad, Lahore, Pakistan
| |
Collapse
|
45
|
Liao Y, Xia H, Song S, Li H. Microaneurysm detection in fundus images based on a novel end-to-end convolutional neural network. Biocybern Biomed Eng 2021. [DOI: 10.1016/j.bbe.2021.04.005] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/29/2022]
|
46
|
Bhardwaj C, Jain S, Sood M. Deep Learning-Based Diabetic Retinopathy Severity Grading System Employing Quadrant Ensemble Model. J Digit Imaging 2021; 34:440-457. [PMID: 33686525 DOI: 10.1007/s10278-021-00418-5] [Citation(s) in RCA: 22] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Key Words] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/30/2020] [Revised: 12/23/2020] [Accepted: 01/03/2021] [Indexed: 12/23/2022] Open
Abstract
The diabetic retinopathy accounts in the deterioration of retinal blood vessels leading to a serious compilation affecting the eyes. The automated DR diagnosis frameworks are critically important for the early identification and detection of these eye-related problems, helping the ophthalmic experts in providing the second opinion for effectual treatment. The deep learning techniques have evolved as an improvement over the conventional approaches, which are dependent on the handcrafted feature extraction. To address the issue of proficient DR discrimination, the authors have proposed a quadrant ensemble automated DR grading approach by implementing InceptionResnet-V2 deep neural network framework. The presented model incorporates histogram equalization, optical disc localization, and quadrant cropping along with the data augmentation step for improving the network performance. A superior accuracy performance of 93.33% is observed for the proposed framework, and a significant reduction of 0.325 is noticed in the cross-entropy loss function for MESSIDOR benchmark dataset; however, its validation utilizing the latest IDRiD dataset establishes its generalization ability. The accuracy improvement of 13.58% is observed when the proposed QEIRV-2 model is compared with the classical Inception-V3 CNN model. To justify the viability of the proposed framework, its performance is compared with the existing state-of-the-art approaches and 25.23% of accuracy improvement is observed.
Collapse
Affiliation(s)
- Charu Bhardwaj
- Department of Electronics and Communication Engineering, JUIT Waknaghat, Solan, HP, India.
| | - Shruti Jain
- Department of Electronics and Communication Engineering, JUIT Waknaghat, Solan, HP, India
| | | |
Collapse
|
47
|
Gilbert MJ, Sun JK. Artificial Intelligence in the assessment of diabetic retinopathy from fundus photographs. Semin Ophthalmol 2021; 35:325-332. [PMID: 33539253 DOI: 10.1080/08820538.2020.1855358] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 10/22/2022]
Abstract
Background: Over the next 25 years, the global prevalence of diabetes is expected to grow to affect 700 million individuals. Consequently, an unprecedented number of patients will be at risk for vision loss from diabetic eye disease. This demand will almost certainly exceed the supply of eye care professionals to individually evaluate each patient on an annual basis, signaling the need for 21st century tools to assist our profession in meeting this challenge. Methods: Review of available literature on artificial intelligence (AI) as applied to diabetic retinopathy (DR) detection and predictionResults: The field of AI has seen exponential growth in evaluating fundus photographs for DR. AI systems employ machine learning and artificial neural networks to teach themselves how to grade DR from libraries of tens of thousands of images and may be able to predict future DR progression based on baseline fundus photographs. Conclusions: AI algorithms are highly promising for the purposes of DR detection and will likely be able to reliably predict DR worsening in the future. A deeper understanding of these systems and how they interpret images is critical as they transition from the bench into the clinic.
Collapse
Affiliation(s)
- Michael J Gilbert
- Joslin Diabetes Center, Beetham Eye Institute , Boston, MA, United States
| | - Jennifer K Sun
- Joslin Diabetes Center, Beetham Eye Institute , Boston, MA, United States.,Department of Ophthalmology, Harvard Medical School , Boston, MA, United States
| |
Collapse
|
48
|
Bilal A, Sun G, Mazhar S. Survey on recent developments in automatic detection of diabetic retinopathy. J Fr Ophtalmol 2021; 44:420-440. [PMID: 33526268 DOI: 10.1016/j.jfo.2020.08.009] [Citation(s) in RCA: 14] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/15/2020] [Accepted: 08/24/2020] [Indexed: 12/13/2022]
Abstract
Diabetic retinopathy (DR) is a disease facilitated by the rapid spread of diabetes worldwide. DR can blind diabetic individuals. Early detection of DR is essential to restoring vision and providing timely treatment. DR can be detected manually by an ophthalmologist, examining the retinal and fundus images to analyze the macula, morphological changes in blood vessels, hemorrhage, exudates, and/or microaneurysms. This is a time consuming, costly, and challenging task. An automated system can easily perform this function by using artificial intelligence, especially in screening for early DR. Recently, much state-of-the-art research relevant to the identification of DR has been reported. This article describes the current methods of detecting non-proliferative diabetic retinopathy, exudates, hemorrhage, and microaneurysms. In addition, the authors point out future directions in overcoming current challenges in the field of DR research.
Collapse
Affiliation(s)
- A Bilal
- Faculty of Information Technology, Beijing University of Technology, Chaoyang District, Beijing 100124, China.
| | - G Sun
- Faculty of Information Technology, Beijing University of Technology, Chaoyang District, Beijing 100124, China
| | - S Mazhar
- Faculty of Information Technology, Beijing University of Technology, Chaoyang District, Beijing 100124, China
| |
Collapse
|
49
|
Li T, Bo W, Hu C, Kang H, Liu H, Wang K, Fu H. Applications of deep learning in fundus images: A review. Med Image Anal 2021; 69:101971. [PMID: 33524824 DOI: 10.1016/j.media.2021.101971] [Citation(s) in RCA: 99] [Impact Index Per Article: 24.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/11/2020] [Accepted: 01/12/2021] [Indexed: 02/06/2023]
Abstract
The use of fundus images for the early screening of eye diseases is of great clinical importance. Due to its powerful performance, deep learning is becoming more and more popular in related applications, such as lesion segmentation, biomarkers segmentation, disease diagnosis and image synthesis. Therefore, it is very necessary to summarize the recent developments in deep learning for fundus images with a review paper. In this review, we introduce 143 application papers with a carefully designed hierarchy. Moreover, 33 publicly available datasets are presented. Summaries and analyses are provided for each task. Finally, limitations common to all tasks are revealed and possible solutions are given. We will also release and regularly update the state-of-the-art results and newly-released datasets at https://github.com/nkicsl/Fundus_Review to adapt to the rapid development of this field.
Collapse
Affiliation(s)
- Tao Li
- College of Computer Science, Nankai University, Tianjin 300350, China
| | - Wang Bo
- College of Computer Science, Nankai University, Tianjin 300350, China
| | - Chunyu Hu
- College of Computer Science, Nankai University, Tianjin 300350, China
| | - Hong Kang
- College of Computer Science, Nankai University, Tianjin 300350, China
| | - Hanruo Liu
- Beijing Tongren Hospital, Capital Medical University, Address, Beijing 100730 China
| | - Kai Wang
- College of Computer Science, Nankai University, Tianjin 300350, China.
| | - Huazhu Fu
- Inception Institute of Artificial Intelligence (IIAI), Abu Dhabi, UAE
| |
Collapse
|
50
|
Xie Y, Zhang J, Lu H, Shen C, Xia Y. SESV: Accurate Medical Image Segmentation by Predicting and Correcting Errors. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:286-296. [PMID: 32956049 DOI: 10.1109/tmi.2020.3025308] [Citation(s) in RCA: 34] [Impact Index Per Article: 8.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
Medical image segmentation is an essential task in computer-aided diagnosis. Despite their prevalence and success, deep convolutional neural networks (DCNNs) still need to be improved to produce accurate and robust enough segmentation results for clinical use. In this paper, we propose a novel and generic framework called Segmentation-Emendation-reSegmentation-Verification (SESV) to improve the accuracy of existing DCNNs in medical image segmentation, instead of designing a more accurate segmentation model. Our idea is to predict the segmentation errors produced by an existing model and then correct them. Since predicting segmentation errors is challenging, we design two ways to tolerate the mistakes in the error prediction. First, rather than using a predicted segmentation error map to correct the segmentation mask directly, we only treat the error map as the prior that indicates the locations where segmentation errors are prone to occur, and then concatenate the error map with the image and segmentation mask as the input of a re-segmentation network. Second, we introduce a verification network to determine whether to accept or reject the refined mask produced by the re-segmentation network on a region-by-region basis. The experimental results on the CRAG, ISIC, and IDRiD datasets suggest that using our SESV framework can improve the accuracy of DeepLabv3+ substantially and achieve advanced performance in the segmentation of gland cells, skin lesions, and retinal microaneurysms. Consistent conclusions can also be drawn when using PSPNet, U-Net, and FPN as the segmentation network, respectively. Therefore, our SESV framework is capable of improving the accuracy of different DCNNs on different medical image segmentation tasks.
Collapse
|