1
|
Mikhail D, Milad D, Antaki F, Hammamji K, Qian CX, Rezende FA, Duval R. The Role of Artificial Intelligence in Epiretinal Membrane Care: A Scoping Review. OPHTHALMOLOGY SCIENCE 2025; 5:100689. [PMID: 40182981 PMCID: PMC11964620 DOI: 10.1016/j.xops.2024.100689] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/21/2024] [Revised: 12/02/2024] [Accepted: 12/16/2024] [Indexed: 04/05/2025]
Abstract
Topic In ophthalmology, artificial intelligence (AI) demonstrates potential in using ophthalmic imaging across diverse diseases, often matching ophthalmologists' performance. However, the range of machine learning models for epiretinal membrane (ERM) management, which differ in methodology, application, and performance, remains largely unsynthesized. Clinical Relevance Epiretinal membrane management relies on clinical evaluation and imaging, with surgical intervention considered in cases of significant impairment. AI analysis of ophthalmic images and clinical features could enhance ERM detection, characterization, and prognostication, potentially improving clinical decision-making. This scoping review aims to evaluate the methodologies, applications, and reported performance of AI models in ERM diagnosis, characterization, and prognostication. Methods A comprehensive literature search was conducted across 5 electronic databases including Ovid MEDLINE, EMBASE, Cochrane Central Register of Controlled Trials, Cochrane Database of Systematic Reviews, and Web of Science Core Collection from inception to November 14, 2024. Studies pertaining to AI algorithms in the context of ERM were included. The primary outcomes measured will be the reported design, application in ERM management, and performance of each AI model. Results Three hundred ninety articles were retrieved, with 33 studies meeting inclusion criteria. There were 30 studies (91%) reporting their training and validation methods. Altogether, 61 distinct AI models were included. OCT scans and fundus photographs were used in 26 (79%) and 7 (21%) papers, respectively. Supervised learning and both supervised and unsupervised learning were used in 32 (97%) and 1 (3%) studies, respectively. Twenty-seven studies (82%) developed or adapted AI models using images, whereas 5 (15%) had models using both images and clinical features, and 1 (3%) used preoperative and postoperative clinical features without ophthalmic images. Study objectives were categorized into 3 stages of ERM care. Twenty-three studies (70%) implemented AI for diagnosis (stage 1), 1 (3%) identified ERM characteristics (stage 2), and 6 (18%) predicted vision impairment after diagnosis or postoperative vision outcomes (stage 3). No articles studied treatment planning. Three studies (9%) used AI in stages 1 and 2. Of the 16 studies comparing AI performance to human graders (i.e., retinal specialists, general ophthalmologists, and trainees), 10 (63%) reported equivalent or higher performance. Conclusion Artificial intelligence-driven assessments of ophthalmic images and clinical features demonstrated high performance in detecting ERM, identifying its morphological properties, and predicting visual outcomes following ERM surgery. Future research might consider the validation of algorithms for clinical applications in personal treatment plan development, ideally to identify patients who might benefit most from surgery. Financial Disclosures The author(s) have no proprietary or commercial interest in any materials discussed in this article.
Collapse
Affiliation(s)
- David Mikhail
- Temerty Faculty of Medicine, University of Toronto, Toronto, Canada
- Department of Ophthalmology, University of Montreal, Montreal, Canada
| | - Daniel Milad
- Department of Ophthalmology, University of Montreal, Montreal, Canada
- Department of Ophthalmology, Hôpital Maisonneuve-Rosemont, Montreal, Canada
- Department of Ophthalmology, Centre Hospitalier de l'Université de Montréal (CHUM), Montreal, Canada
| | - Fares Antaki
- Department of Ophthalmology, University of Montreal, Montreal, Canada
- Department of Ophthalmology, Centre Hospitalier de l'Université de Montréal (CHUM), Montreal, Canada
| | - Karim Hammamji
- Department of Ophthalmology, University of Montreal, Montreal, Canada
- Department of Ophthalmology, Centre Hospitalier de l'Université de Montréal (CHUM), Montreal, Canada
| | - Cynthia X. Qian
- Department of Ophthalmology, University of Montreal, Montreal, Canada
- Department of Ophthalmology, Hôpital Maisonneuve-Rosemont, Montreal, Canada
| | - Flavio A. Rezende
- Department of Ophthalmology, University of Montreal, Montreal, Canada
- Department of Ophthalmology, Hôpital Maisonneuve-Rosemont, Montreal, Canada
| | - Renaud Duval
- Department of Ophthalmology, University of Montreal, Montreal, Canada
- Department of Ophthalmology, Hôpital Maisonneuve-Rosemont, Montreal, Canada
| |
Collapse
|
2
|
Gou R, Ma X, Su N, Yuan S, Chen Q. Bilateral deformable attention transformer for screening of high myopia using optical coherence tomography. Comput Biol Med 2025; 191:110236. [PMID: 40253920 DOI: 10.1016/j.compbiomed.2025.110236] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/18/2024] [Revised: 04/16/2025] [Accepted: 04/17/2025] [Indexed: 04/22/2025]
Abstract
Myopia is a visual impairment caused by excessive refractive power of the cornea or lens or elongation of the eyeball. Due to the various classification criteria associated with high myopia, such as spherical equivalent (SE) and axial length (AL), existing methods primarily rely on individual classification criteria for model design. In this paper, to comprehensively utilize multiple indicators, we design a multi-label classification model for high myopia. Moreover, image data play a pivotal role in studying high myopia and pathological myopia. Notable features of high myopia, including increased retinal curvature, choroidal thinning, and scleral shadowing, are observable in Optical Coherence Tomography (OCT) images of the retina. We propose a model named Bilateral Deformable Attention Transformer (BDA-Tran) for multi-label screening of high myopia in OCT data. Based on the vision transformer, we introduce a bilateral deformable attention mechanism (BDA) where the queries in self-attention are composed of both the global queries and the data-dependent queries from the left and right sides. This flexible approach allows attention to focus on relevant regions and capture more myopia-related information features, thereby concentrating attention primarily on regions related to the choroid and sclera, among other areas associated with high myopia. BDA-Tran is trained and tested on OCT images of 243 patients, achieving the accuracies of 83.1 % and 87.7 % for SE and AL, respectively. Furthermore, we visualize attention maps to provide transparent and interpretable judgments. Experimental results demonstrate that BDA-Tran outperforms existing methods in terms of effectiveness and reliability under the same experimental conditions.
Collapse
Affiliation(s)
- Ruoxuan Gou
- School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing, China
| | - Xiao Ma
- School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing, China
| | - Na Su
- Department of Ophthalmology, The First Affiliated Hospital of Nanjing Medical University, Nanjing, China
| | - Songtao Yuan
- Department of Ophthalmology, The First Affiliated Hospital of Nanjing Medical University, Nanjing, China
| | - Qiang Chen
- School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing, China.
| |
Collapse
|
3
|
Lin YT, Xiong X, Zheng YP, Zhou Q. Transfer Learning and Multi-Feature Fusion-Based Deep Learning Model for Idiopathic Macular Hole Diagnosis and Grading from Optical Coherence Tomography Images. Clin Ophthalmol 2025; 19:1593-1607. [PMID: 40396157 PMCID: PMC12091069 DOI: 10.2147/opth.s521558] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2025] [Accepted: 05/05/2025] [Indexed: 05/22/2025] Open
Abstract
Background Idiopathic macular hole is an ophthalmic disease that seriously affects vision, and its early diagnosis and treatment have important clinical significance to reduce the occurrence of blindness. At present, OCT is the gold standard for diagnosing this disease, but its application is limited due to the need for professional ophthalmologist to diagnose it. The introduction of artificial intelligence will break this situation and make its diagnosis efficient, and how to build an effective predictive model is the key to the problem, and more clinical trials are still needed to verify it. Objective This study aims to evaluate the role of deep learning systems in Idiopathic Macular Hole diagnosis, grading, and prediction. Methods A single-center, retrospective study used binocular OCT images from IMH patients at the First Affiliated Hospital of Nanchang University (November 2019 - January 2023). A deep learning algorithm, including traditional omics, Resnet101, and fusion models incorporating multi-feature fusion and transfer learning, was developed. Model performance was evaluated using accuracy and AUC. Logistic regression analyzed clinical factors, and a nomogram predicted surgical risk. Analysis was conducted with SPSS 22.0 and R 3.6.3. P < 0.05 was statistically significant. Results Among 229 OCT images, the traditional omics, Resnet101, and fusion models achieved accuracies of 93%, 94%, and 95%, respectively, in the training set. In the test set, the fusion model and Resnet101 correctly identified 39 images, while the traditional omics model identified 35. The nomogram had a C-index of 0.996, with macular hole diameter most strongly associated with surgical risk. Conclusion The deep learning system with transfer learning and multi-feature fusion effectively diagnoses and grades IMH from OCT images.
Collapse
Affiliation(s)
- Ye-Ting Lin
- Department of Ophthalmology, The First Affiliated Hospital, Jiangxi Medical College, Nanchang University, Nanchang, Jiangxi, People’s Republic of China
| | - Xu Xiong
- Department of Ophthalmology, The First Affiliated Hospital, Jiangxi Medical College, Nanchang University, Nanchang, Jiangxi, People’s Republic of China
| | - Ying-Ping Zheng
- Department of Product Design, Jiangxi Normal University, Nanchang, Jiangxi, People’s Republic of China
| | - Qiong Zhou
- Department of Ophthalmology, The First Affiliated Hospital, Jiangxi Medical College, Nanchang University, Nanchang, Jiangxi, People’s Republic of China
| |
Collapse
|
4
|
Fan Y, Jiang Y, Mu Z, Xu Y, Xie P, Liu Q, Pu L, Hu Z. Optical Coherence Tomography Characteristics Between Idiopathic Epiretinal Membranes and Secondary Epiretinal Membranes due to Peripheral Retinal Hole. J Ophthalmol 2025; 2025:9299651. [PMID: 40371012 PMCID: PMC12077978 DOI: 10.1155/joph/9299651] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2024] [Revised: 03/18/2025] [Accepted: 04/17/2025] [Indexed: 05/16/2025] Open
Abstract
Purpose: In clinical practice, some eyes preoperatively diagnosed with "idiopathic epiretinal membranes (iERM)" will be amended to "secondary epiretinal membranes (sERM)" once peripheral retinal hole is detected. This study utilized optical coherence tomography (OCT) images to compare the characteristics between the iERM and sERM due to peripheral retinal hole (PRH). Methods: In this retrospective, cross-sectional study, 635 eyes that had undergone pars plana vitrectomy with membrane peeling were enrolled. A total of 115 eyes (18.1%) detected with peripheral retinal holes were allocated to the sERM-PRH group while the other 520 eyes were to the iERM group. The demographic data and OCT characteristics were compared between the two groups. Besides, all the eyes were evaluated by a double-grading scheme: severity grading of ERM progression into four stages plus anatomical classification into three kinds of part-thickness macular holes associated with ERMs. Results: No significant difference was found in age, gender, symptom duration, axial length, or best-corrected visual acuity between the two groups. There was also no difference concerning the features based on OCT, ranging from central macular thickness, the ratios of the photoreceptor inner/outer segment junction line defect, intraretinal fluid, cotton ball sign, to epiretinal proliferation. However, the native difference in parafoveal thickness between the temporal and nasal quadrants was observed in the iERM group, yet disappeared in the sERM-PRH group. Moreover, eyes between the two groups were distributionally similar in both grading scales. Conclusion: Our results demonstrated that even OCT images could hardly provide effective clues for early differentiating sERM from iERM, which highlighted the necessity of a thorough pre- and intro-operative fundus examination of the peripheral retina for clinicians.
Collapse
Affiliation(s)
- Yuanyuan Fan
- Department of Ophthalmology, The First Affiliated Hospital of Nanjing Medical University, Nanjing, Jiangsu 210029, China
| | - Yingying Jiang
- Department of Ophthalmology, The First Affiliated Hospital of Nanjing Medical University, Nanjing, Jiangsu 210029, China
- Department of Ophthalmology, Zhangjiagang Hospital Affiliated to Soochow University, Suzhou, Jiangsu 215600, China
| | - Zhaoxia Mu
- Department of Ophthalmology, The First Affiliated Hospital of Nanjing Medical University, Nanjing, Jiangsu 210029, China
| | - Yulian Xu
- Department of Ophthalmology, The First Affiliated Hospital of Nanjing Medical University, Nanjing, Jiangsu 210029, China
| | - Ping Xie
- Department of Ophthalmology, The First Affiliated Hospital of Nanjing Medical University, Nanjing, Jiangsu 210029, China
| | - Qinghuai Liu
- Department of Ophthalmology, The First Affiliated Hospital of Nanjing Medical University, Nanjing, Jiangsu 210029, China
| | - Lijun Pu
- Department of Ophthalmology, Zhangjiagang Hospital Affiliated to Soochow University, Suzhou, Jiangsu 215600, China
| | - Zizhong Hu
- Department of Ophthalmology, The First Affiliated Hospital of Nanjing Medical University, Nanjing, Jiangsu 210029, China
| |
Collapse
|
5
|
Zhang L, Li X, Chen W, Gu Y, Wu H, Lu Z, Dong B. A joint learning approach for automated diagnosis of keratinocyte carcinoma using optical attenuation coefficients. NPJ Digit Med 2025; 8:232. [PMID: 40307427 PMCID: PMC12043986 DOI: 10.1038/s41746-025-01634-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/22/2025] [Accepted: 04/10/2025] [Indexed: 05/02/2025] Open
Abstract
Keratinocyte carcinoma, such as Actinic Keratosis (AK) and Basal Cell Carcinoma (BCC), share similar clinical presentations but differ significantly in prognosis and treatment, highlighting the importance of effective screening. Optical coherence tomography (OCT) shows promise for diagnosing AK and BCC using signal intensity and skin layer thickness, but variability due to skin characteristics and system settings underscores the need for a standardized diagnostic method. Here, we propose an automated diagnostic method using the optical attenuation coefficient (OAC) and a joint learning strategy to classify AK, BCC, and normal skin. OAC images extracted from OCT data revealed notable disparities between normal and cancerous tissues. By incorporating probability distribution function (PDF) information alongside OAC images, the model achieved an accuracy of over 80% and approaching 100% by utilizing 3D OAC data to enhance robustness. This approach highlights the potential of OAC-based analysis for automated, intelligent diagnosis of early-stage non-melanoma skin cancers.
Collapse
Grants
- 2022YFF0708700 National Key Research and Development Program of China
- 2022YFF0708700 National Key Research and Development Program of China
- 2022YFF0708700 National Key Research and Development Program of China
- 2022YFF0708700 National Key Research and Development Program of China
- 2022YFF0708700 National Key Research and Development Program of China
- 22TQ020 Shanghai Basic Research Special Zone Program
- 22TQ020 Shanghai Basic Research Special Zone Program
- 22TQ020 Shanghai Basic Research Special Zone Program
- 22TQ020 Shanghai Basic Research Special Zone Program
- 22TQ020 Shanghai Basic Research Special Zone Program
- 22ZR1404300, 22ZR1409500 Natural Science Foundation of Shanghai Municipality
- 22ZR1404300, 22ZR1409500 Natural Science Foundation of Shanghai Municipality
- 22ZR1404300, 22ZR1409500 Natural Science Foundation of Shanghai Municipality
- 22ZR1404300, 22ZR1409500 Natural Science Foundation of Shanghai Municipality
- 22ZR1404300, 22ZR1409500 Natural Science Foundation of Shanghai Municipality
- 22ZR1404300, 22ZR1409500 Natural Science Foundation of Shanghai Municipality
- 22S31905500 Shanghai Science and Technology Innovation Action Plan
- 22S31905500 Shanghai Science and Technology Innovation Action Plan
- 22S31905500 Shanghai Science and Technology Innovation Action Plan
- 22S31905500 Shanghai Science and Technology Innovation Action Plan
- 22S31905500 Shanghai Science and Technology Innovation Action Plan
- yg2021-032, yg2022-2 Medical Engineering Fund of Fudan University
- yg2021-032, yg2022-2 Medical Engineering Fund of Fudan University
- yg2021-032, yg2022-2 Medical Engineering Fund of Fudan University
- yg2021-032, yg2022-2 Medical Engineering Fund of Fudan University
- yg2021-032, yg2022-2 Medical Engineering Fund of Fudan University
- yg2021-032, yg2022-2 Medical Engineering Fund of Fudan University
- yg2021-032, yg2022-2 Medical Engineering Fund of Fudan University
- 2022YQ043 Young Talents of Shanghai Health Commission
- 2024CX06 Huashan Hospital Innovation Fund
- 2024CX06 Huashan Hospital Innovation Fund
- 2024CX06 Huashan Hospital Innovation Fund
Collapse
Affiliation(s)
- Lei Zhang
- The Academy for Engineering and Technology, Yiwu Research Institute, Fudan University, Shanghai, 200433, China
| | - Xiaoran Li
- The Academy for Engineering and Technology, Yiwu Research Institute, Fudan University, Shanghai, 200433, China
| | - Wen Chen
- The Academy for Engineering and Technology, Yiwu Research Institute, Fudan University, Shanghai, 200433, China
| | - Yuanjie Gu
- The Academy for Engineering and Technology, Yiwu Research Institute, Fudan University, Shanghai, 200433, China
| | - Hao Wu
- The Department of Dermatology, Huashan Hospital, Fudan University, Shanghai, 200040, China.
| | - Zhong Lu
- The Department of Dermatology, Huashan Hospital, Fudan University, Shanghai, 200040, China.
| | - Biqin Dong
- The Academy for Engineering and Technology, Yiwu Research Institute, Fudan University, Shanghai, 200433, China.
| |
Collapse
|
6
|
Liu S, Shang B, Yan J, Zhu Z, Ding Y, Zhou Q, Wei C, Shen Y, Zhu X. A method for quantifying and automatic grading of musculoskeletal ultrasound superb microvascular imaging based on dynamic analysis of optical flow model. Sci Rep 2025; 15:13369. [PMID: 40247053 PMCID: PMC12006311 DOI: 10.1038/s41598-025-97924-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/26/2024] [Accepted: 04/08/2025] [Indexed: 04/19/2025] Open
Abstract
Superb microvascular flow signals in joints are important indicators for evaluating inflammation in arthritis diagnosis. Super Microvascular Imaging (SMI), a musculoskeletal ultrasound technique, captures microvascular signals with enhanced resolution, enabling improved quantitative analysis of joint superb microvascular flow. However, existing musculoskeletal ultrasound imaging predominantly relies on static observations for analyzing these signals, which are heavily influenced by subjective factors, thereby limiting diagnostic accuracy for arthritis. This study introduces a novel quantitative and automated grading method utilizing dynamic analysis through an optical flow model. Real-time dynamic quantification of superb microvascular flow signals is achieved via motion estimation and skeleton extraction based on the optical flow model. The Kappa consistency test evaluates the agreement between the automated grading system and physician assessments, with differences between the two methods analyzed. A total of 47 patient samples were included, comprising 20 males and 27 females (p = 0.307 > 0.05, χ2=1.042). The agreement between the automated grading system and physician assessments reached 70.2%, with a Kappa value of 0.627 (p < 0.001), indicating good consistency. Nonetheless, the system displayed a tendency to high-grade cases of moderate inflammation. The proposed quantitative and automated grading method for superb microvascular flow, based on dynamic analysis through an optical flow model, improves the objectivity and consistency of superb microvascular flow grading and demonstrates significant clinical potential. The method shows strong anti-interference performance in noisy signal environments, representing a promising advancement for non-invasive arthritis diagnosis.
Collapse
Affiliation(s)
- Shanna Liu
- Department of Information Technology, the Fourth Affiliated Hospital of School of Medicine, and International School of Medicine, International Institutes of Medicine, Zhejiang University, Yiwu, 322000, China
| | - Bo Shang
- College of Medical Engineering and Technology, Xinjiang Medical University, Urumqi, 830017, Xinjiang, China
| | - Junliang Yan
- Department of Ultrasound in Medicine, the Fourth Affiliated Hospital of School of Medicine, and International School of Medicine, International Institutes of Medicine, Zhejiang University, Yiwu, 322000, China
| | - Zenghua Zhu
- School of Information Engineering, Hangzhou Medical College, Hangzhou, Zhejiang, 311399, China
| | - Yuanhao Ding
- Department of Information Technology, the Fourth Affiliated Hospital of School of Medicine, and International School of Medicine, International Institutes of Medicine, Zhejiang University, Yiwu, 322000, China
| | - Qingli Zhou
- Department of Information Technology, the Fourth Affiliated Hospital of School of Medicine, and International School of Medicine, International Institutes of Medicine, Zhejiang University, Yiwu, 322000, China
| | - Chengjing Wei
- College of Public Health, Xinjiang Medical University, Urumqi, 830017, Xinjiang, China
| | - Yuqiang Shen
- Department of Information Technology, the Fourth Affiliated Hospital of School of Medicine, and International School of Medicine, International Institutes of Medicine, Zhejiang University, Yiwu, 322000, China.
| | - Xinjian Zhu
- Department of Information Technology, the Fourth Affiliated Hospital of School of Medicine, and International School of Medicine, International Institutes of Medicine, Zhejiang University, Yiwu, 322000, China.
| |
Collapse
|
7
|
Beuse A, Wenzel DA, Spitzer MS, Bartz-Schmidt KU, Schultheiss M, Poli S, Grohmann C. Automated Detection of Central Retinal Artery Occlusion Using OCT Imaging via Explainable Deep Learning. OPHTHALMOLOGY SCIENCE 2025; 5:100630. [PMID: 39669299 PMCID: PMC11634984 DOI: 10.1016/j.xops.2024.100630] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/01/2024] [Revised: 09/23/2024] [Accepted: 10/04/2024] [Indexed: 12/14/2024]
Abstract
Objective To demonstrate the capability of a deep learning model to detect central retinal artery occlusion (CRAO), a retinal pathology with significant clinical urgency, using OCT data. Design Retrospective, external validation study analyzing OCT and clinical baseline data of 2 institutions via deep learning classification analysis. Subjects Patients presenting to the University Medical Center Tübingen and the University Medical Center Hamburg-Eppendorf in Germany. Methods OCT data of patients suffering from CRAO, differential diagnosis with (sub) acute visual loss (central retinal vein occlusion, diabetic macular edema, nonarteritic ischemic optic neuropathy), and from controls were expertly graded and distinguished into 3 groups. Our methodological approach involved a nested multiclass five fold cross-validation classification scheme. Main Outcome Measures Area under the curve (AUC). Results The optimal performance of our algorithm was observed using 30 epochs, complemented by an early stopping mechanism to prevent overfitting. Our model followed a multiclass approach, distinguishing among the 3 different classes: control, CRAO, and differential diagnoses. The evaluation was conducted by the "one vs. all" area under the receiver operating characteristics curve (AUC) method. The results demonstrated AUC of 0.96 (95% confidence interval [CI], ± 0.01); 0.99 (95% CI, ± 0.00); and 0.90 (95% CI, ± 0.03) for each class, respectively. Conclusions Our machine learning algorithm (MLA) exhibited a high AUC, as well as sensitivity and specificity in detecting CRAO and the differential classes, respectively. These findings underscore the potential for deploying MLAs in the identification of less common etiologies within an acute emergency clinical setting. Financial Disclosures Proprietary or commercial disclosure may be found in the Footnotes and Disclosures at the end of this article.
Collapse
Affiliation(s)
- Ansgar Beuse
- Department of Ophthalmology, University Medical Center Hamburg-Eppendorf, Hamburg, Germany
| | - Daniel Alexander Wenzel
- University Eye Hospital, Centre for Ophthalmology, University Hospital Tübingen, Tübingen, Germany
| | - Martin Stephan Spitzer
- Department of Ophthalmology, University Medical Center Hamburg-Eppendorf, Hamburg, Germany
| | | | - Maximilian Schultheiss
- Department of Ophthalmology, University Medical Center Hamburg-Eppendorf, Hamburg, Germany
| | - Sven Poli
- Department of Neurology and Stroke, University Hospital Tübingen, Tübingen, Germany
- Hertie Institute for Clinical Brain Research, University Hospital Tübingen, Tübingen, Germany
| | - Carsten Grohmann
- Department of Ophthalmology, University Medical Center Hamburg-Eppendorf, Hamburg, Germany
| |
Collapse
|
8
|
Abbas Y, Hadi HJ, Aziz K, Ahmed N, Akhtar MU, Alshara MA, Chakrabarti P. Reinforcement-based leveraging transfer learning for multiclass optical coherence tomography images classification. Sci Rep 2025; 15:6193. [PMID: 39979354 PMCID: PMC11842753 DOI: 10.1038/s41598-025-89831-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/18/2024] [Accepted: 02/07/2025] [Indexed: 02/22/2025] Open
Abstract
The accurate diagnosis of retinal diseases, such as Diabetic Macular Edema (DME) and Age-related Macular Degeneration (AMD), is essential for preventing vision loss. Optical Coherence Tomography (OCT) imaging plays a crucial role in identifying these conditions, especially given the increasing prevalence of AMD. This study introduces a novel Reinforcement-Based Leveraging Transfer Learning (RBLTL) framework, which integrates reinforcement Q-learning with transfer learning using pre-trained models, including InceptionV3, DenseNet201, and InceptionResNetV2. The RBLTL framework dynamically optimizes hyperparameters, improving classification accuracy and generalization while mitigating overfitting. Experimental evaluations demonstrate remarkable performance, achieving testing accuracies of 98.75%, 98.90%, and 99.20% across three scenarios for multiclass OCT image classification. These results highlight the effectiveness of the RBLTL framework in categorizing OCT images for conditions like DME and AMD, establishing it as a reliable and versatile approach for automated medical image classification with significant implications for clinical diagnostics.
Collapse
Affiliation(s)
- Yawar Abbas
- Key Laboratory of Aerospace Information Security and Trusted Computing, Ministry of Education, School of Cyber Science and Engineering, Wuhan University, Wuhan, China.
| | - Hassan Jalil Hadi
- College of Computer and Information Sciences, Prince Sultan University, Riyadh, Saudi Arabia
| | - Kamran Aziz
- Key Laboratory of Aerospace Information Security and Trusted Computing, Ministry of Education, School of Cyber Science and Engineering, Wuhan University, Wuhan, China.
| | - Naveed Ahmed
- College of Computer and Information Sciences, Prince Sultan University, Riyadh, Saudi Arabia
| | | | - Mohammed Ali Alshara
- College of Computer and Information Sciences, Prince Sultan University, Riyadh, Saudi Arabia
| | - Prasun Chakrabarti
- Department of Computer Science and Engineering, Sir Padampat Singhania University, Udaipur, Rajasthan, 313601, India
| |
Collapse
|
9
|
Mikhail D, Milad D, Antaki F, Hammamji K, Qian CX, Rezende FA, Duval R. The role of artificial intelligence in macular hole management: A scoping review. Surv Ophthalmol 2025; 70:12-27. [PMID: 39357748 DOI: 10.1016/j.survophthal.2024.09.003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/31/2024] [Revised: 09/16/2024] [Accepted: 09/23/2024] [Indexed: 10/04/2024]
Abstract
We focus on the utility of artificial intelligence (AI) in the management of macular hole (MH). We synthesize 25 studies, comprehensively reporting on each AI model's development strategy, validation, tasks, performance, strengths, and limitations. All models analyzed ophthalmic images, and 5 (20 %) also analyzed clinical features. Study objectives were categorized based on 3 stages of MH care: diagnosis, identification of MH characteristics, and postoperative predictions of hole closure and vision recovery. Twenty-two (88 %) AI models underwent supervised learning, and the models were most often deployed to determine a MH diagnosis. None of the articles applied AI to guiding treatment plans. AI model performance was compared to other algorithms and to human graders. Of the 10 studies comparing AI to human graders (i.e., retinal specialists, general ophthalmologists, and ophthalmology trainees), 5 (50 %) reported equivalent or higher performance. Overall, AI analysis of images and clinical characteristics in MH demonstrated high diagnostic and predictive accuracy. Convolutional neural networks comprised the majority of included AI models, including those which were high performing. Future research may consider validating algorithms to propose personalized treatment plans and explore clinical use of the aforementioned algorithms.
Collapse
Affiliation(s)
- David Mikhail
- Temerty Faculty of Medicine, University of Toronto, Toronto, Canada; Department of Ophthalmology, University of Montreal, Montreal, Canada
| | - Daniel Milad
- Department of Ophthalmology, University of Montreal, Montreal, Canada; Department of Ophthalmology, Hôpital Maisonneuve-Rosemont, Montreal, Canada; Department of Ophthalmology, Centre Hospitalier de l'Université de Montréal (CHUM), Montreal, Canada
| | - Fares Antaki
- Department of Ophthalmology, University of Montreal, Montreal, Canada; Department of Ophthalmology, Centre Hospitalier de l'Université de Montréal (CHUM), Montreal, Canada
| | - Karim Hammamji
- Department of Ophthalmology, University of Montreal, Montreal, Canada; Department of Ophthalmology, Centre Hospitalier de l'Université de Montréal (CHUM), Montreal, Canada
| | - Cynthia X Qian
- Department of Ophthalmology, University of Montreal, Montreal, Canada; Department of Ophthalmology, Hôpital Maisonneuve-Rosemont, Montreal, Canada
| | - Flavio A Rezende
- Department of Ophthalmology, University of Montreal, Montreal, Canada; Department of Ophthalmology, Hôpital Maisonneuve-Rosemont, Montreal, Canada
| | - Renaud Duval
- Department of Ophthalmology, University of Montreal, Montreal, Canada; Department of Ophthalmology, Hôpital Maisonneuve-Rosemont, Montreal, Canada.
| |
Collapse
|
10
|
Du K, Shah S, Bollepalli SC, Ibrahim MN, Gadari A, Sutharahan S, Sahel JA, Chhablani J, Vupparaboina KK. Inter-rater reliability in labeling quality and pathological features of retinal OCT scans: A customized annotation software approach. PLoS One 2024; 19:e0314707. [PMID: 39693322 DOI: 10.1371/journal.pone.0314707] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/10/2024] [Accepted: 11/14/2024] [Indexed: 12/20/2024] Open
Abstract
OBJECTIVES Various imaging features on optical coherence tomography (OCT) are crucial for identifying and defining disease progression. Establishing a consensus on these imaging features is essential, particularly for training deep learning models for disease classification. This study aims to analyze the inter-rater reliability in labeling the quality and common imaging signatures of retinal OCT scans. METHODS 500 OCT scans obtained from CIRRUS HD-OCT 5000 devices were displayed at 512x1024x128 resolution on a customizable, in-house annotation software. Each patient's eye was represented by 16 random scans. Two masked reviewers independently labeled the quality and specific pathological features of each scan. Evaluated features included overall image quality, presence of fovea, and disease signatures including subretinal fluid (SRF), intraretinal fluid (IRF), drusen, pigment epithelial detachment (PED), and hyperreflective material. The raw percentage agreement and Cohen's kappa (κ) coefficient were used to evaluate concurrence between the two sets of labels. RESULTS Our analysis revealed κ = 0.60 for the inter-rater reliability of overall scan quality, indicating substantial agreement. In contrast, there was slight agreement in determining the cause of poor image quality (κ = 0.18). The binary determination of presence and absence of retinal disease signatures showed almost complete agreement between reviewers (κ = 0.85). Specific retinal pathologies, such as the foveal location of the scan (0.78), IRF (0.63), drusen (0.73), and PED (0.87), exhibited substantial concordance. However, less agreement was found in identifying SRF (0.52), hyperreflective dots (0.41), and hyperreflective foci (0.33). CONCLUSIONS Our study demonstrates significant inter-rater reliability in labeling the quality and retinal pathologies on OCT scans. While some features show stronger agreement than others, these standardized labels can be utilized to create automated machine learning tools for diagnosing retinal diseases and capturing valuable pathological features in each scan. This standardization will aid in the consistency of medical diagnoses and enhance the accessibility of OCT diagnostic tools.
Collapse
Affiliation(s)
- Katherine Du
- Department of Ophthalmology, University of Pittsburgh Medical Center, Pittsburgh, PA, United States of America
| | - Stavan Shah
- Department of Ophthalmology, University of Pittsburgh Medical Center, Pittsburgh, PA, United States of America
| | - Sandeep Chandra Bollepalli
- Department of Ophthalmology, University of Pittsburgh Medical Center, Pittsburgh, PA, United States of America
| | - Mohammed Nasar Ibrahim
- Department of Ophthalmology, University of Pittsburgh Medical Center, Pittsburgh, PA, United States of America
| | - Adarsh Gadari
- Department of Computer Science, University of North Carolina at Greensboro, Greensboro, NC, United States of America
| | - Shan Sutharahan
- Department of Computer Science, University of North Carolina at Greensboro, Greensboro, NC, United States of America
| | - José-Alain Sahel
- Department of Ophthalmology, University of Pittsburgh Medical Center, Pittsburgh, PA, United States of America
| | - Jay Chhablani
- Department of Ophthalmology, University of Pittsburgh Medical Center, Pittsburgh, PA, United States of America
| | - Kiran Kumar Vupparaboina
- Department of Ophthalmology, University of Pittsburgh Medical Center, Pittsburgh, PA, United States of America
| |
Collapse
|
11
|
Alenezi AM, Aloqalaa DA, Singh SK, Alrabiah R, Habib S, Islam M, Daradkeh YI. Multiscale attention-over-attention network for retinal disease recognition in OCT radiology images. Front Med (Lausanne) 2024; 11:1499393. [PMID: 39582968 PMCID: PMC11583944 DOI: 10.3389/fmed.2024.1499393] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/20/2024] [Accepted: 10/14/2024] [Indexed: 11/26/2024] Open
Abstract
Retinal disease recognition using Optical Coherence Tomography (OCT) images plays a pivotal role in the early diagnosis and treatment of conditions. However, the previous attempts relied on extracting single-scale features often refined by stacked layered attentions. This paper presents a novel deep learning-based Multiscale Feature Enhancement via a Dual Attention Network specifically designed for retinal disease recognition in OCT images. Our approach leverages the EfficientNetB7 backbone to extract multiscale features from OCT images, ensuring a comprehensive representation of global and local retinal structures. To further refine feature extraction, we propose a Pyramidal Attention mechanism that integrates Multi-Head Self-Attention (MHSA) with Dense Atrous Spatial Pyramid Pooling (DASPP), effectively capturing long-range dependencies and contextual information at multiple scales. Additionally, Efficient Channel Attention (ECA) and Spatial Refinement modules are introduced to enhance channel-wise and spatial feature representations, enabling precise localization of retinal abnormalities. A comprehensive ablation study confirms the progressive impact of integrated blocks and attention mechanisms that enhance overall performance. Our findings underscore the potential of advanced attention mechanisms and multiscale processing, highlighting the effectiveness of the network. Extensive experiments on two benchmark datasets demonstrate the superiority of the proposed network over existing state-of-the-art methods.
Collapse
Affiliation(s)
- Abdulmajeed M. Alenezi
- Department of Electrical Engineering, Faculty of Engineering, Islamic University of Madinah, Madinah, Saudi Arabia
| | - Daniyah A. Aloqalaa
- Department of Information Technology, College of Computer, Qassim University, Buraydah, Saudi Arabia
| | - Sushil Kumar Singh
- Department of Computer Engineering, Marwadi University, Rajkot, Gujarat, India
| | - Raqinah Alrabiah
- Department of Information Technology, College of Computer, Qassim University, Buraydah, Saudi Arabia
| | - Shabana Habib
- Department of Information Technology, College of Computer, Qassim University, Buraydah, Saudi Arabia
| | - Muhammad Islam
- Department of Electrical Engineering, College of Engineering, Qassim University, Buraydah, Saudi Arabia
| | - Yousef Ibrahim Daradkeh
- Department of Computer Engineering and Information, College of Engineering in Wadi Alddawasir, Prince Sattam Bin Abdulaziz University, Al-Kharj, Saudi Arabia
| |
Collapse
|
12
|
Hosseini F, Asadi F, Rabiei R, Kiani F, Harari RE. Applications of artificial intelligence in diagnosis of uncommon cystoid macular edema using optical coherence tomography imaging: A systematic review. Surv Ophthalmol 2024; 69:937-944. [PMID: 38942125 DOI: 10.1016/j.survophthal.2024.06.005] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/07/2024] [Revised: 06/20/2024] [Accepted: 06/24/2024] [Indexed: 06/30/2024]
Abstract
Cystoid macular edema (CME) is a sight-threatening condition often associated with inflammatory and diabetic diseases. Early detection is crucial to prevent irreversible vision loss. Artificial intelligence (AI) has shown promise in automating CME diagnosis through optical coherence tomography (OCT) imaging, but its utility needs critical evaluation. This systematic review assesses the application of AI to diagnosis CME, specifically focusing on disorders like postoperative CME (Irvine Gass syndrome) and retinitis pigmentosa without obvious vasculopathy, using OCT imaging. A comprehensive search was conducted across 6 databases (PubMed, Scopus, Web of Science, Wiley, ScienceDirect, and IEEE) from 2018 to November, 2023. Twenty-three articles met the inclusion criteria and were selected for in-depth analysis. We evaluate AI's role in CME diagnosis and its performance in "detection", "classification", and "segmentation" of OCT retinal images. We found that convolutional neural network (CNN)-based methods consistently outperformed other machine learning techniques, achieving an average accuracy of over 96 % in detecting and identifying CME from OCT images. Despite certain limitations such as dataset size and ethical concerns, the synergy between AI and OCT, particularly through CNNs, holds promise for significantly advancing CME diagnostics.
Collapse
Affiliation(s)
- Farhang Hosseini
- Department of Health Information Technology and Management, School of Allied Medical Sciences, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - Farkhondeh Asadi
- Department of Health Information Technology and Management, School of Allied Medical Sciences, Shahid Beheshti University of Medical Sciences, Tehran, Iran.
| | - Reza Rabiei
- Department of Health Information Technology and Management, School of Allied Medical Sciences, Shahid Beheshti University of Medical Sciences, Tehran, Iran.
| | - Fatemeh Kiani
- Department of Health Information Technology and Management, School of Allied Medical Sciences, Shahid Beheshti University of Medical Sciences, Tehran, Iran
| | - Rayan Ebnali Harari
- STRATUS Center for Medical Simulation, Harvard Medical School, Boston, MA, USA.
| |
Collapse
|
13
|
Bell J, Whitney J, Cetin H, Le T, Cardwell N, Srivasatava SK, Ehlers JP. Validation of Inter-Reader Agreement/Consistency for Quantification of Ellipsoid Zone Integrity and Sub-RPE Compartmental Features Across Retinal Diseases. Diagnostics (Basel) 2024; 14:2395. [PMID: 39518363 PMCID: PMC11545794 DOI: 10.3390/diagnostics14212395] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/21/2024] [Revised: 09/24/2024] [Accepted: 10/07/2024] [Indexed: 11/16/2024] Open
Abstract
BACKGROUND An unmet need exists when clinically assessing retinal and layer-based features of retinal diseases. Therefore, quantification of retinal-layer-thicknesses/fluid volumes using deep-learning-augmented platforms to reproduce human-obtained clinical measurements is needed. METHODS In this analysis, 210 spectral-domain optical coherence tomography (SD-OCT) scans (30 without pathology, 60 dry age-related macular degeneration [AMD], 60 wet AMD, and 60 diabetic macular edema [total 23,625 B-scans]) were included. A fully automated segmentation platform segmented four retinal layers for compartmental assessment (internal limiting membrane, ellipsoid zone [EZ], retinal pigment epithelium [RPE], and Bruch's membrane). Two certified OCT readers independently completed manual segmentation and B-scan level validation of automated segmentation, with segmentation correction when needed (semi-automated). Certified reader metrics were compared to gold standard metrics using intraclass correlation coefficients (ICCs) to assess overall agreement. Across different diseases, several metrics generated from automated segmentations approached or matched human readers performance. RESULTS Absolute ICCs for retinal mean thickness measurements showed excellent agreement (range 0.980-0.999) across four cohorts. EZ-RPE thickness values and sub-RPE compartment ICCs demonstrated excellent agreement (ranges of 0.953-0.987 and 0.944-0.997, respectively) for full dataset, dry-AMD, and wet-AMD cohorts. CONCLUSIONS Analyses demonstrated high reliability and consistency of segmentation of outer retinal compartmental features using a completely human/manual approach or a semi-automated approach to segmentation. These results support the critical role that measuring features, such as photoreceptor preservation through EZ integrity, in future clinical trials may optimize clinical care.
Collapse
Affiliation(s)
- Jordan Bell
- Cleveland Clinic Lerner College of Medicine Program, Case Western Reserve University, Cleveland, OH 44106, USA
- The Tony and Leona Campane Center for Excellence in Image-Guided Surgery and Advanced Imaging Research, Cleveland Clinic, Cleveland, OH 44195, USA
| | - Jon Whitney
- The Tony and Leona Campane Center for Excellence in Image-Guided Surgery and Advanced Imaging Research, Cleveland Clinic, Cleveland, OH 44195, USA
| | - Hasan Cetin
- The Tony and Leona Campane Center for Excellence in Image-Guided Surgery and Advanced Imaging Research, Cleveland Clinic, Cleveland, OH 44195, USA
| | - Thuy Le
- The Tony and Leona Campane Center for Excellence in Image-Guided Surgery and Advanced Imaging Research, Cleveland Clinic, Cleveland, OH 44195, USA
| | - Nicole Cardwell
- The Tony and Leona Campane Center for Excellence in Image-Guided Surgery and Advanced Imaging Research, Cleveland Clinic, Cleveland, OH 44195, USA
| | - Sunil K. Srivasatava
- The Tony and Leona Campane Center for Excellence in Image-Guided Surgery and Advanced Imaging Research, Cleveland Clinic, Cleveland, OH 44195, USA
- Vitreoretinal Service, Cole Eye Institute, Cleveland, OH 44195, USA
| | - Justis P. Ehlers
- The Tony and Leona Campane Center for Excellence in Image-Guided Surgery and Advanced Imaging Research, Cleveland Clinic, Cleveland, OH 44195, USA
- Vitreoretinal Service, Cole Eye Institute, Cleveland, OH 44195, USA
| |
Collapse
|
14
|
Akpinar MH, Sengur A, Faust O, Tong L, Molinari F, Acharya UR. Artificial intelligence in retinal screening using OCT images: A review of the last decade (2013-2023). COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 254:108253. [PMID: 38861878 DOI: 10.1016/j.cmpb.2024.108253] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/22/2023] [Revised: 04/22/2024] [Accepted: 05/25/2024] [Indexed: 06/13/2024]
Abstract
BACKGROUND AND OBJECTIVES Optical coherence tomography (OCT) has ushered in a transformative era in the domain of ophthalmology, offering non-invasive imaging with high resolution for ocular disease detection. OCT, which is frequently used in diagnosing fundamental ocular pathologies, such as glaucoma and age-related macular degeneration (AMD), plays an important role in the widespread adoption of this technology. Apart from glaucoma and AMD, we will also investigate pertinent pathologies, such as epiretinal membrane (ERM), macular hole (MH), macular dystrophy (MD), vitreomacular traction (VMT), diabetic maculopathy (DMP), cystoid macular edema (CME), central serous chorioretinopathy (CSC), diabetic macular edema (DME), diabetic retinopathy (DR), drusen, glaucomatous optic neuropathy (GON), neovascular AMD (nAMD), myopia macular degeneration (MMD) and choroidal neovascularization (CNV) diseases. This comprehensive review examines the role that OCT-derived images play in detecting, characterizing, and monitoring eye diseases. METHOD The 2020 PRISMA guideline was used to structure a systematic review of research on various eye conditions using machine learning (ML) or deep learning (DL) techniques. A thorough search across IEEE, PubMed, Web of Science, and Scopus databases yielded 1787 publications, of which 1136 remained after removing duplicates. Subsequent exclusion of conference papers, review papers, and non-open-access articles reduced the selection to 511 articles. Further scrutiny led to the exclusion of 435 more articles due to lower-quality indexing or irrelevance, resulting in 76 journal articles for the review. RESULTS During our investigation, we found that a major challenge for ML-based decision support is the abundance of features and the determination of their significance. In contrast, DL-based decision support is characterized by a plug-and-play nature rather than relying on a trial-and-error approach. Furthermore, we observed that pre-trained networks are practical and especially useful when working on complex images such as OCT. Consequently, pre-trained deep networks were frequently utilized for classification tasks. Currently, medical decision support aims to reduce the workload of ophthalmologists and retina specialists during routine tasks. In the future, it might be possible to create continuous learning systems that can predict ocular pathologies by identifying subtle changes in OCT images.
Collapse
Affiliation(s)
- Muhammed Halil Akpinar
- Department of Electronics and Automation, Vocational School of Technical Sciences, Istanbul University-Cerrahpasa, Istanbul, Turkey
| | - Abdulkadir Sengur
- Electrical-Electronics Engineering Department, Technology Faculty, Firat University, Elazig, Turkey.
| | - Oliver Faust
- School of Computing and Information Science, Anglia Ruskin University Cambridge Campus, United Kingdom
| | - Louis Tong
- Singapore Eye Research Institute, Singapore, Singapore
| | - Filippo Molinari
- Biolab, PolitoBIOMedLab, Department of Electronics and Telecommunications, Politecnico di Torino, Turin, Italy
| | - U Rajendra Acharya
- School of Mathematics, Physics and Computing, University of Southern Queensland, Springfield, Australia
| |
Collapse
|
15
|
Wheeler TW, Hunter K, Garcia PA, Li H, Thomson AC, Hunter A, Mehanian C. Self-supervised contrastive learning improves machine learning discrimination of full thickness macular holes from epiretinal membranes in retinal OCT scans. PLOS DIGITAL HEALTH 2024; 3:e0000411. [PMID: 39186771 PMCID: PMC11346922 DOI: 10.1371/journal.pdig.0000411] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/13/2023] [Accepted: 07/08/2024] [Indexed: 08/28/2024]
Abstract
There is a growing interest in using computer-assisted models for the detection of macular conditions using optical coherence tomography (OCT) data. As the quantity of clinical scan data of specific conditions is limited, these models are typically developed by fine-tuning a generalized network to classify specific macular conditions of interest. Full thickness macular holes (FTMH) present a condition requiring urgent surgical repair to prevent vision loss. Other works on automated FTMH classification have tended to use supervised ImageNet pre-trained networks with good results but leave room for improvement. In this paper, we develop a model for FTMH classification using OCT B-scans around the central foveal region to pre-train a naïve network using contrastive self-supervised learning. We found that self-supervised pre-trained networks outperform ImageNet pre-trained networks despite a small training set size (284 eyes total, 51 FTMH+ eyes, 3 B-scans from each eye). On three replicate data splits, 3D spatial contrast pre-training yields a model with an average F1-score of 1.0 on holdout data (50 eyes total, 10 FTMH+), compared to an average F1-score of 0.831 for FTMH detection by ImageNet pre-trained models. These results demonstrate that even limited data may be applied toward self-supervised pre-training to substantially improve performance for FTMH classification, indicating applicability toward other OCT-based problems.
Collapse
Affiliation(s)
- Timothy William Wheeler
- Department of Bioengineering, University of Oregon, Eugene, Oregon, United States of America
| | - Kaitlyn Hunter
- Oregon Eye Consultants, Eugene, Oregon, United States of America
| | | | - Henry Li
- Oregon Eye Consultants, Eugene, Oregon, United States of America
| | | | - Allan Hunter
- Oregon Eye Consultants, Eugene, Oregon, United States of America
| | - Courosh Mehanian
- Department of Bioengineering, University of Oregon, Eugene, Oregon, United States of America
- Global Health Labs, Bellevue, Washington, United States of America
| |
Collapse
|
16
|
Poh SSJ, Sia JT, Yip MYT, Tsai ASH, Lee SY, Tan GSW, Weng CY, Kadonosono K, Kim M, Yonekawa Y, Ho AC, Toth CA, Ting DSW. Artificial Intelligence, Digital Imaging, and Robotics Technologies for Surgical Vitreoretinal Diseases. Ophthalmol Retina 2024; 8:633-645. [PMID: 38280425 DOI: 10.1016/j.oret.2024.01.018] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/17/2023] [Revised: 01/14/2024] [Accepted: 01/19/2024] [Indexed: 01/29/2024]
Abstract
OBJECTIVE To review recent technological advancement in imaging, surgical visualization, robotics technology, and the use of artificial intelligence in surgical vitreoretinal (VR) diseases. BACKGROUND Technological advancements in imaging enhance both preoperative and intraoperative management of surgical VR diseases. Widefield imaging in fundal photography and OCT can improve assessment of peripheral retinal disorders such as retinal detachments, degeneration, and tumors. OCT angiography provides a rapid and noninvasive imaging of the retinal and choroidal vasculature. Surgical visualization has also improved with intraoperative OCT providing a detailed real-time assessment of retinal layers to guide surgical decisions. Heads-up display and head-mounted display utilize 3-dimensional technology to provide surgeons with enhanced visual guidance and improved ergonomics during surgery. Intraocular robotics technology allows for greater surgical precision and is shown to be useful in retinal vein cannulation and subretinal drug delivery. In addition, deep learning techniques leverage on diverse data including widefield retinal photography and OCT for better predictive accuracy in classification, segmentation, and prognostication of many surgical VR diseases. CONCLUSION This review article summarized the latest updates in these areas and highlights the importance of continuous innovation and improvement in technology within the field. These advancements have the potential to reshape management of surgical VR diseases in the very near future and to ultimately improve patient care. FINANCIAL DISCLOSURE(S) Proprietary or commercial disclosure may be found in the Footnotes and Disclosures at the end of this article.
Collapse
Affiliation(s)
- Stanley S J Poh
- Singapore National Eye Centre, Singapore Eye Research Institute, Singapore; Ophthalmology and Visual Sciences Academic Clinical Program, Duke-NUS Medical School, Singapore
| | - Josh T Sia
- Singapore National Eye Centre, Singapore Eye Research Institute, Singapore
| | - Michelle Y T Yip
- Singapore National Eye Centre, Singapore Eye Research Institute, Singapore
| | - Andrew S H Tsai
- Singapore National Eye Centre, Singapore Eye Research Institute, Singapore; Ophthalmology and Visual Sciences Academic Clinical Program, Duke-NUS Medical School, Singapore
| | - Shu Yen Lee
- Singapore National Eye Centre, Singapore Eye Research Institute, Singapore; Ophthalmology and Visual Sciences Academic Clinical Program, Duke-NUS Medical School, Singapore
| | - Gavin S W Tan
- Singapore National Eye Centre, Singapore Eye Research Institute, Singapore; Ophthalmology and Visual Sciences Academic Clinical Program, Duke-NUS Medical School, Singapore
| | - Christina Y Weng
- Department of Ophthalmology, Baylor College of Medicine, Houston, Texas
| | | | - Min Kim
- Department of Ophthalmology, Gangnam Severance Hospital, Yonsei University College of Medicine, Seoul, South Korea
| | - Yoshihiro Yonekawa
- Wills Eye Hospital, Mid Atlantic Retina, Thomas Jefferson University, Philadelphia, Pennsylvania
| | - Allen C Ho
- Wills Eye Hospital, Mid Atlantic Retina, Thomas Jefferson University, Philadelphia, Pennsylvania
| | - Cynthia A Toth
- Departments of Ophthalmology and Biomedical Engineering, Duke University, Durham, North Carolina
| | - Daniel S W Ting
- Singapore National Eye Centre, Singapore Eye Research Institute, Singapore; Ophthalmology and Visual Sciences Academic Clinical Program, Duke-NUS Medical School, Singapore; Byers Eye Institute, Stanford University, Palo Alto, California.
| |
Collapse
|
17
|
Hill C, Malone J, Liu K, Ng SPY, MacAulay C, Poh C, Lane P. Three-Dimension Epithelial Segmentation in Optical Coherence Tomography of the Oral Cavity Using Deep Learning. Cancers (Basel) 2024; 16:2144. [PMID: 38893263 PMCID: PMC11172075 DOI: 10.3390/cancers16112144] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/03/2024] [Revised: 06/01/2024] [Accepted: 06/02/2024] [Indexed: 06/21/2024] Open
Abstract
This paper aims to simplify the application of optical coherence tomography (OCT) for the examination of subsurface morphology in the oral cavity and reduce barriers towards the adoption of OCT as a biopsy guidance device. The aim of this work was to develop automated software tools for the simplified analysis of the large volume of data collected during OCT. Imaging and corresponding histopathology were acquired in-clinic using a wide-field endoscopic OCT system. An annotated dataset (n = 294 images) from 60 patients (34 male and 26 female) was assembled to train four unique neural networks. A deep learning pipeline was built using convolutional and modified u-net models to detect the imaging field of view (network 1), detect artifacts (network 2), identify the tissue surface (network 3), and identify the presence and location of the epithelial-stromal boundary (network 4). The area under the curve of the image and artifact detection networks was 1.00 and 0.94, respectively. The Dice similarity score for the surface and epithelial-stromal boundary segmentation networks was 0.98 and 0.83, respectively. Deep learning (DL) techniques can identify the location and variations in the epithelial surface and epithelial-stromal boundary in OCT images of the oral mucosa. Segmentation results can be synthesized into accessible en face maps to allow easier visualization of changes.
Collapse
Affiliation(s)
- Chloe Hill
- Department of Integrative Oncology, British Columbia Cancer Research Institute, 675 W 10th Ave., Vancouver, BC V5Z 1L3, Canada; (C.H.); (J.M.); (K.L.); (C.M.); (C.P.)
- School of Engineering Science, Simon Fraser University, 8888 University Drive, Burnaby, BC V5A 1S6, Canada
| | - Jeanie Malone
- Department of Integrative Oncology, British Columbia Cancer Research Institute, 675 W 10th Ave., Vancouver, BC V5Z 1L3, Canada; (C.H.); (J.M.); (K.L.); (C.M.); (C.P.)
- School of Biomedical Engineering, University of British Columbia, 2222 Health Sciences Mall, Vancouver, BC V6T 1Z3, Canada
| | - Kelly Liu
- Department of Integrative Oncology, British Columbia Cancer Research Institute, 675 W 10th Ave., Vancouver, BC V5Z 1L3, Canada; (C.H.); (J.M.); (K.L.); (C.M.); (C.P.)
- School of Biomedical Engineering, University of British Columbia, 2222 Health Sciences Mall, Vancouver, BC V6T 1Z3, Canada
- Faculty of Dentistry, University of British Columbia, 2199 Wesbrook Mall, Vancouver, BC V6T 1Z3, Canada;
| | - Samson Pak-Yan Ng
- Faculty of Dentistry, University of British Columbia, 2199 Wesbrook Mall, Vancouver, BC V6T 1Z3, Canada;
| | - Calum MacAulay
- Department of Integrative Oncology, British Columbia Cancer Research Institute, 675 W 10th Ave., Vancouver, BC V5Z 1L3, Canada; (C.H.); (J.M.); (K.L.); (C.M.); (C.P.)
- Department of Pathology and Laboratory Medicine, University of British Columbia, 2211 Wesbrook Mall, Vancouver, BC V6T 1Z7, Canada
| | - Catherine Poh
- Department of Integrative Oncology, British Columbia Cancer Research Institute, 675 W 10th Ave., Vancouver, BC V5Z 1L3, Canada; (C.H.); (J.M.); (K.L.); (C.M.); (C.P.)
- Faculty of Dentistry, University of British Columbia, 2199 Wesbrook Mall, Vancouver, BC V6T 1Z3, Canada;
| | - Pierre Lane
- Department of Integrative Oncology, British Columbia Cancer Research Institute, 675 W 10th Ave., Vancouver, BC V5Z 1L3, Canada; (C.H.); (J.M.); (K.L.); (C.M.); (C.P.)
- School of Engineering Science, Simon Fraser University, 8888 University Drive, Burnaby, BC V5A 1S6, Canada
- School of Biomedical Engineering, University of British Columbia, 2222 Health Sciences Mall, Vancouver, BC V6T 1Z3, Canada
| |
Collapse
|
18
|
Ayhan MS, Neubauer J, Uzel MM, Gelisken F, Berens P. Interpretable detection of epiretinal membrane from optical coherence tomography with deep neural networks. Sci Rep 2024; 14:8484. [PMID: 38605115 PMCID: PMC11009346 DOI: 10.1038/s41598-024-57798-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2022] [Accepted: 03/21/2024] [Indexed: 04/13/2024] Open
Abstract
This study aimed to automatically detect epiretinal membranes (ERM) in various OCT-scans of the central and paracentral macula region and classify them by size using deep-neural-networks (DNNs). To this end, 11,061 OCT-images were included and graded according to the presence of an ERM and its size (small 100-1000 µm, large > 1000 µm). The data set was divided into training, validation and test sets (75%, 10%, 15% of the data, respectively). An ensemble of DNNs was trained and saliency maps were generated using Guided-Backprob. OCT-scans were also transformed into a one-dimensional-value using t-SNE analysis. The DNNs' receiver-operating-characteristics on the test set showed a high performance for no-ERM, small-ERM and large-ERM cases (AUC: 0.99, 0.92, 0.99, respectively; 3-way accuracy: 89%), with small-ERMs being the most difficult ones to detect. t-SNE analysis sorted cases by size and, in particular, revealed increased classification uncertainty at the transitions between groups. Saliency maps reliably highlighted ERM, regardless of the presence of other OCT features (i.e. retinal-thickening, intraretinal pseudo-cysts, epiretinal-proliferation) and entities such as ERM-retinoschisis, macular-pseudohole and lamellar-macular-hole. This study showed therefore that DNNs can reliably detect and grade ERMs according to their size not only in the fovea but also in the paracentral region. This is also achieved in cases of hard-to-detect, small-ERMs. In addition, the generated saliency maps can be used to highlight small-ERMs that might otherwise be missed. The proposed model could be used for screening-programs or decision-support-systems in the future.
Collapse
Affiliation(s)
- Murat Seçkin Ayhan
- Institute for Ophthalmic Research, University of Tübingen, Elfriede Aulhorn Str. 7, 72076, Tübingen, Germany
| | - Jonas Neubauer
- University Eye Clinic, University of Tübingen, Tübingen, Germany
| | - Mehmet Murat Uzel
- University Eye Clinic, University of Tübingen, Tübingen, Germany
- Department of Ophthalmology, Balıkesir University School of Medicine, Balıkesir, Turkey
| | - Faik Gelisken
- University Eye Clinic, University of Tübingen, Tübingen, Germany.
| | - Philipp Berens
- Institute for Ophthalmic Research, University of Tübingen, Elfriede Aulhorn Str. 7, 72076, Tübingen, Germany.
- Tübingen AI Center, Tübingen, Germany.
| |
Collapse
|
19
|
Yan Y, Huang X, Jiang X, Gao Z, Liu X, Jin K, Ye J. Clinical evaluation of deep learning systems for assisting in the diagnosis of the epiretinal membrane grade in general ophthalmologists. Eye (Lond) 2024; 38:730-736. [PMID: 37848677 PMCID: PMC10920879 DOI: 10.1038/s41433-023-02765-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/09/2023] [Revised: 08/28/2023] [Accepted: 09/18/2023] [Indexed: 10/19/2023] Open
Abstract
BACKGROUND Epiretinal membrane (ERM) is a common age-related retinal disease detected by optical coherence tomography (OCT), with a prevalence of 34.1% among people over 60 years old. This study aims to develop artificial intelligence (AI) systems to assist in the diagnosis of ERM grade using OCT images and to clinically evaluate the potential benefits and risks of our AI systems with a comparative experiment. METHODS A segmentation deep learning (DL) model that segments retinal features associated with ERM severity and a classification DL model that grades the severity of ERM were developed based on an OCT dataset obtained from three hospitals. A comparative experiment was conducted to compare the performance of four general ophthalmologists with and without assistance from the AI in diagnosing ERM severity. RESULTS The segmentation network had a pixel accuracy (PA) of 0.980 and a mean intersection over union (MIoU) of 0.873, while the six-classification network had a total accuracy of 81.3%. The diagnostic accuracy scores of the four ophthalmologists increased with AI assistance from 81.7%, 80.7%, 78.0%, and 80.7% to 87.7%, 86.7%, 89.0%, and 91.3%, respectively, while the corresponding time expenditures were reduced. The specific results of the study as well as the misinterpretations of the AI systems were analysed. CONCLUSION Through our comparative experiment, the AI systems proved to be valuable references for medical diagnosis and demonstrated the potential to accelerate clinical workflows. Systematic efforts are needed to ensure the safe and rapid integration of AI systems into ophthalmic practice.
Collapse
Affiliation(s)
- Yan Yan
- Eye Center, The Second Affiliated Hospital, School of Medicine, Zhejiang University, Hangzhou, 310009, China
| | - Xiaoling Huang
- Eye Center, The Second Affiliated Hospital, School of Medicine, Zhejiang University, Hangzhou, 310009, China
| | - Xiaoyu Jiang
- College of Control Science and Engineering, Zhejiang University, Hangzhou, 310027, China
| | - Zhiyuan Gao
- Eye Center, The Second Affiliated Hospital, School of Medicine, Zhejiang University, Hangzhou, 310009, China
| | - Xindi Liu
- Eye Center, The Second Affiliated Hospital, School of Medicine, Zhejiang University, Hangzhou, 310009, China
| | - Kai Jin
- Eye Center, The Second Affiliated Hospital, School of Medicine, Zhejiang University, Hangzhou, 310009, China.
| | - Juan Ye
- Eye Center, The Second Affiliated Hospital, School of Medicine, Zhejiang University, Hangzhou, 310009, China.
| |
Collapse
|
20
|
Cai L, Wen C, Jiang J, Liang C, Zheng H, Su Y, Chen C. Classification of diabetic maculopathy based on optical coherence tomography images using a Vision Transformer model. BMJ Open Ophthalmol 2023; 8:e001423. [PMID: 38135350 DOI: 10.1136/bmjophth-2023-001423] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2023] [Accepted: 11/22/2023] [Indexed: 12/24/2023] Open
Abstract
PURPOSE To develop a Vision Transformer model to detect different stages of diabetic maculopathy (DM) based on optical coherence tomography (OCT) images. METHODS After removing images with poor quality, a total of 3319 OCT images were extracted from the Eye Center of the Renmin Hospital of Wuhan University and randomly split the images into training and validation sets in a 7:3 ratio. All macular cross-sectional scan OCT images were collected retrospectively from the eyes of DM patients from 2016 to 2022. One of the OCT stages of DM, including early diabetic macular oedema (DME), advanced DME, severe DME and atrophic maculopathy, was labelled on the collected images, respectively. A deep learning (DL) model based on Vision Transformer was trained to detect four OCT grading of DM. RESULTS The model proposed in our paper can provide an impressive detection performance. We achieved an accuracy of 82.00%, an F1 score of 83.11%, an area under the receiver operating characteristic curve (AUC) of 0.96. The AUC for the detection of four OCT grading (ie, early DME, advanced DME, severe DME and atrophic maculopathy) was 0.96, 0.95, 0.87 and 0.98, respectively, with an accuracy of 90.87%, 89.96%, 94.42% and 95.13%, respectively, a precision of 88.46%, 80.31%, 89.42% and 87.74%, respectively, a sensitivity of 87.03%, 88.18%, 63.39% and 89.42%, respectively, a specificity of 93.02%, 90.72%, 98.40% and 96.66%, respectively and an F1 score of 87.74%, 84.06%, 88.18% and 88.57%, respectively. CONCLUSION Our DL model based on Vision Transformer demonstrated a relatively high accuracy in the detection of OCT grading of DM, which can help with patients in a preliminary screening to identify groups with serious conditions. These patients need a further test for an accurate diagnosis, and a timely treatment to obtain a good visual prognosis. These results emphasised the potential of artificial intelligence in assisting clinicians in developing therapeutic strategies with DM in the future.
Collapse
Affiliation(s)
- Liwei Cai
- Department of Ophthalmology, Renmin Hospital of Wuhan University, Wuhan, Hubei, China
| | - Chi Wen
- Wuhan University School of Computer Science, Wuhan, Hubei, China
| | - Jingwen Jiang
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangdong Provincial Key Laboratory of Ophthalmology and Visual Science, Guangdong Provincial Clinical Research Center for Ocular Diseases, Guangzhou, Guangdong, China
| | - Congbi Liang
- Department of Ophthalmology, Renmin Hospital of Wuhan University, Wuhan, Hubei, China
| | - Hongmei Zheng
- Department of Ophthalmology, Renmin Hospital of Wuhan University, Wuhan, Hubei, China
| | - Yu Su
- Department of Ophthalmology, Renmin Hospital of Wuhan University, Wuhan, Hubei, China
| | - Changzheng Chen
- Department of Ophthalmology, Renmin Hospital of Wuhan University, Wuhan, Hubei, China
| |
Collapse
|
21
|
Liu HC, Lin MH, Chang WC, Zeng RC, Wang YM, Sun CW. Rapid On-Site AI-Assisted Grading for Lung Surgery Based on Optical Coherence Tomography. Cancers (Basel) 2023; 15:5388. [PMID: 38001648 PMCID: PMC10670228 DOI: 10.3390/cancers15225388] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/03/2023] [Revised: 11/02/2023] [Accepted: 11/08/2023] [Indexed: 11/26/2023] Open
Abstract
The determination of resection extent traditionally relies on the microscopic invasiveness of frozen sections (FSs) and is crucial for surgery of early lung cancer with preoperatively unknown histology. While previous research has shown the value of optical coherence tomography (OCT) for instant lung cancer diagnosis, tumor grading through OCT remains challenging. Therefore, this study proposes an interactive human-machine interface (HMI) that integrates a mobile OCT system, deep learning algorithms, and attention mechanisms. The system is designed to mark the lesion's location on the image smartly and perform tumor grading in real time, potentially facilitating clinical decision making. Twelve patients with a preoperatively unknown tumor but a final diagnosis of adenocarcinoma underwent thoracoscopic resection, and the artificial intelligence (AI)-designed system mentioned above was used to measure fresh specimens. Results were compared to FSs benchmarked on permanent pathologic reports. Current results show better differentiating power among minimally invasive adenocarcinoma (MIA), invasive adenocarcinoma (IA), and normal tissue, with an overall accuracy of 84.9%, compared to 20% for FSs. Additionally, the sensitivity and specificity, the sensitivity and specificity were 89% and 82.7% for MIA and 94% and 80.6% for IA, respectively. The results suggest that this AI system can potentially produce rapid and efficient diagnoses and ultimately improve patient outcomes.
Collapse
Affiliation(s)
- Hung-Chang Liu
- Section of Thoracic Surgery, Mackay Memorial Hospital, Taipei City 10449, Taiwan;
- Intensive Care Unit, Mackay Memorial Hospital, Taipei City 10449, Taiwan
- Department of Medicine, Mackay Medical College, New Taipei City 25245, Taiwan
- Department of Optometry, Mackay Junior College of Medicine, Nursing, and Management, Taipei City 11260, Taiwan
| | - Miao-Hui Lin
- Biomedical Optical Imaging Lab, Department of Photonics, College of Electrical and Computer Engineering, National Yang Ming Chiao Tung University, Hsinchu City 30010, Taiwan; (M.-H.L.); (R.-C.Z.); (Y.-M.W.)
| | - Wei-Chin Chang
- Department of Pathology, Mackay Memorial Hospital, New Taipei City 25160, Taiwan;
- Department of Pathology, Taipei Medical University Hospital, Taipei City 11030, Taiwan
- Department of Pathology, School of Medicine, College of Medicine, Taipei Medical University, Taipei City 11030, Taiwan
| | - Rui-Cheng Zeng
- Biomedical Optical Imaging Lab, Department of Photonics, College of Electrical and Computer Engineering, National Yang Ming Chiao Tung University, Hsinchu City 30010, Taiwan; (M.-H.L.); (R.-C.Z.); (Y.-M.W.)
| | - Yi-Min Wang
- Biomedical Optical Imaging Lab, Department of Photonics, College of Electrical and Computer Engineering, National Yang Ming Chiao Tung University, Hsinchu City 30010, Taiwan; (M.-H.L.); (R.-C.Z.); (Y.-M.W.)
| | - Chia-Wei Sun
- Biomedical Optical Imaging Lab, Department of Photonics, College of Electrical and Computer Engineering, National Yang Ming Chiao Tung University, Hsinchu City 30010, Taiwan; (M.-H.L.); (R.-C.Z.); (Y.-M.W.)
- Institute of Biomedical Engineering, College of Electrical and Computer Engineering, National Yang Ming Chiao Tung University, Hsinchu City 30010, Taiwan
- Medical Device Innovation and Translation Center, National Yang Ming Chiao Tung University, Taipei City 11259, Taiwan
| |
Collapse
|
22
|
Wang T, Li H, Pu T, Yang L. Microsurgery Robots: Applications, Design, and Development. SENSORS (BASEL, SWITZERLAND) 2023; 23:8503. [PMID: 37896597 PMCID: PMC10611418 DOI: 10.3390/s23208503] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/24/2023] [Revised: 10/07/2023] [Accepted: 10/09/2023] [Indexed: 10/29/2023]
Abstract
Microsurgical techniques have been widely utilized in various surgical specialties, such as ophthalmology, neurosurgery, and otolaryngology, which require intricate and precise surgical tool manipulation on a small scale. In microsurgery, operations on delicate vessels or tissues require high standards in surgeons' skills. This exceptionally high requirement in skills leads to a steep learning curve and lengthy training before the surgeons can perform microsurgical procedures with quality outcomes. The microsurgery robot (MSR), which can improve surgeons' operation skills through various functions, has received extensive research attention in the past three decades. There have been many review papers summarizing the research on MSR for specific surgical specialties. However, an in-depth review of the relevant technologies used in MSR systems is limited in the literature. This review details the technical challenges in microsurgery, and systematically summarizes the key technologies in MSR with a developmental perspective from the basic structural mechanism design, to the perception and human-machine interaction methods, and further to the ability in achieving a certain level of autonomy. By presenting and comparing the methods and technologies in this cutting-edge research, this paper aims to provide readers with a comprehensive understanding of the current state of MSR research and identify potential directions for future development in MSR.
Collapse
Affiliation(s)
- Tiexin Wang
- ZJU-UIUC Institute, International Campus, Zhejiang University, Haining 314400, China; (T.W.); (H.L.); (T.P.)
- School of Mechanical Engineering, Zhejiang University, Hangzhou 310058, China
| | - Haoyu Li
- ZJU-UIUC Institute, International Campus, Zhejiang University, Haining 314400, China; (T.W.); (H.L.); (T.P.)
| | - Tanhong Pu
- ZJU-UIUC Institute, International Campus, Zhejiang University, Haining 314400, China; (T.W.); (H.L.); (T.P.)
| | - Liangjing Yang
- ZJU-UIUC Institute, International Campus, Zhejiang University, Haining 314400, China; (T.W.); (H.L.); (T.P.)
- School of Mechanical Engineering, Zhejiang University, Hangzhou 310058, China
- Department of Mechanical Engineering, University of Illinois Urbana-Champaign, Urbana, IL 61801, USA
| |
Collapse
|
23
|
Leandro I, Lorenzo B, Aleksandar M, Dario M, Rosa G, Agostino A, Daniele T. OCT-based deep-learning models for the identification of retinal key signs. Sci Rep 2023; 13:14628. [PMID: 37670066 PMCID: PMC10480174 DOI: 10.1038/s41598-023-41362-4] [Citation(s) in RCA: 11] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/15/2023] [Accepted: 08/25/2023] [Indexed: 09/07/2023] Open
Abstract
A new system based on binary Deep Learning (DL) convolutional neural networks has been developed to recognize specific retinal abnormality signs on Optical Coherence Tomography (OCT) images useful for clinical practice. Images from the local hospital database were retrospectively selected from 2017 to 2022. Images were labeled by two retinal specialists and included central fovea cross-section OCTs. Nine models were developed using the Visual Geometry Group 16 architecture to distinguish healthy versus abnormal retinas and to identify eight different retinal abnormality signs. A total of 21,500 OCT images were screened, and 10,770 central fovea cross-section OCTs were included in the study. The system achieved high accuracy in identifying healthy retinas and specific pathological signs, ranging from 93 to 99%. Accurately detecting abnormal retinal signs from OCT images is crucial for patient care. This study aimed to identify specific signs related to retinal pathologies, aiding ophthalmologists in diagnosis. The high-accuracy system identified healthy retinas and pathological signs, making it a useful diagnostic aid. Labelled OCT images remain a challenge, but our approach reduces dataset creation time and shows DL models' potential to improve ocular pathology diagnosis and clinical decision-making.
Collapse
Affiliation(s)
- Inferrera Leandro
- Department of Medicine, Surgery and Health Sciences, Eye Clinic, Ophthalmology Clinic, University of Trieste, Piazza Dell'Ospitale 1, 34125, Trieste, Italy.
| | - Borsatti Lorenzo
- Department of Medicine, Surgery and Health Sciences, Eye Clinic, Ophthalmology Clinic, University of Trieste, Piazza Dell'Ospitale 1, 34125, Trieste, Italy
| | | | - Marangoni Dario
- Department of Medicine, Surgery and Health Sciences, Eye Clinic, Ophthalmology Clinic, University of Trieste, Piazza Dell'Ospitale 1, 34125, Trieste, Italy
| | - Giglio Rosa
- Department of Medicine, Surgery and Health Sciences, Eye Clinic, Ophthalmology Clinic, University of Trieste, Piazza Dell'Ospitale 1, 34125, Trieste, Italy
| | - Accardo Agostino
- Department of Engineering and Architecture, University of Trieste, Trieste, Italy
| | - Tognetto Daniele
- Department of Medicine, Surgery and Health Sciences, Eye Clinic, Ophthalmology Clinic, University of Trieste, Piazza Dell'Ospitale 1, 34125, Trieste, Italy
| |
Collapse
|
24
|
Chen S, Wu Z, Li M, Zhu Y, Xie H, Yang P, Zhao C, Zhang Y, Zhang S, Zhao X, Lu L, Zhang G, Lei B. FIT-Net: Feature Interaction Transformer Network for Pathologic Myopia Diagnosis. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:2524-2538. [PMID: 37030824 DOI: 10.1109/tmi.2023.3260990] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Automatic and accurate classification of retinal optical coherence tomography (OCT) images is essential to assist physicians in diagnosing and grading pathological changes in pathologic myopia (PM). Clinically, due to the obvious differences in the position, shape, and size of the lesion structure in different scanning directions, ophthalmologists usually need to combine the lesion structure in the OCT images in the horizontal and vertical scanning directions to diagnose the type of pathological changes in PM. To address these challenges, we propose a novel feature interaction Transformer network (FIT-Net) to diagnose PM using OCT images, which consists of two dual-scale Transformer (DST) blocks and an interactive attention (IA) unit. Specifically, FIT-Net divides image features of different scales into a series of feature block sequences. In order to enrich the feature representation, we propose an IA unit to realize the interactive learning of class token in feature sequences of different scales. The interaction between feature sequences of different scales can effectively integrate different scale image features, and hence FIT-Net can focus on meaningful lesion regions to improve the PM classification performance. Finally, by fusing the dual-view image features in the horizontal and vertical scanning directions, we propose six dual-view feature fusion methods for PM diagnosis. The extensive experimental results based on the clinically obtained datasets and three publicly available datasets demonstrate the effectiveness and superiority of the proposed method. Our code is avaiable at: https://github.com/chenshaobin/FITNet.
Collapse
|
25
|
Wang J, Zong Y, He Y, Shi G, Jiang C. Domain Adaptation-Based Automated Detection of Retinal Diseases from Optical Coherence Tomography Images. Curr Eye Res 2023; 48:836-842. [PMID: 37203787 DOI: 10.1080/02713683.2023.2212878] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/09/2022] [Revised: 05/03/2023] [Accepted: 05/05/2023] [Indexed: 05/20/2023]
Abstract
PURPOSE To verify the effectiveness of domain adaptation in generalizing a deep learning-based anomaly detection model to unseen optical coherence tomography (OCT) images. METHODS Two datasets (source and target, where labelled training data was only available for the source) captured by two different OCT facilities were collected to train the model. We defined the model containing a feature extractor and a classifier as Model One and trained it with only labeled source data. The proposed domain adaptation model was defined as Model Two, which has the same feature extractor and classifier as Model One but has an additional domain critic in the training phase. We trained the Model Two with both the source and target datasets; the feature extractor was trained to extract domain-invariant features while the domain critic learned to capture the domain discrepancy. Finally, a well-trained feature extractor was used to extract domain-invariant features and a classifier was used to detect images with retinal pathologies in the two domains. RESULTS The target data consisted of 3,058 OCT B-scans captured from 163 participants. Model One achieved an area under the curve (AUC) of 0.912 [95% confidence interval (CI), 0.895-0.962], while Model Two achieved an overall AUC of 0.989 [95% CI, 0.982-0.993] for detecting pathological retinas from healthy samples. Moreover, Model Two achieved an average retinopathies detection accuracy of 94.52%. Heat maps showed that the algorithm focused on the area with pathological changes during processing, similar to manual grading in daily clinical work. CONCLUSIONS The proposed domain adaptation model showed a strong ability in reducing the domain distance between different OCT datasets.
Collapse
Affiliation(s)
- Jing Wang
- Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Science, Suzhou, People's Republic of China
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Suzhou, People's Republic of China
| | - Yuan Zong
- Department of Ophthalmology and Vision Science, Eye and ENT Hospital, Fudan University, Shanghai, People's Republic of China
- Key Laboratory of Myopia of State Health Ministry and Key Laboratory of Visual Impairment and Restoration, Shanghai, People's Republic of China
| | - Yi He
- Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Science, Suzhou, People's Republic of China
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Suzhou, People's Republic of China
| | - Guohua Shi
- Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Science, Suzhou, People's Republic of China
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Suzhou, People's Republic of China
| | - Chunhui Jiang
- Department of Ophthalmology and Vision Science, Eye and ENT Hospital, Fudan University, Shanghai, People's Republic of China
- Key Laboratory of Myopia of State Health Ministry and Key Laboratory of Visual Impairment and Restoration, Shanghai, People's Republic of China
| |
Collapse
|
26
|
Zhao PY, Bommakanti N, Yu G, Aaberg MT, Patel TP, Paulus YM. Deep learning for automated detection of neovascular leakage on ultra-widefield fluorescein angiography in diabetic retinopathy. Sci Rep 2023; 13:9165. [PMID: 37280345 DOI: 10.1038/s41598-023-36327-6] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/21/2022] [Accepted: 06/01/2023] [Indexed: 06/08/2023] Open
Abstract
Diabetic retinopathy is a leading cause of blindness in working-age adults worldwide. Neovascular leakage on fluorescein angiography indicates progression to the proliferative stage of diabetic retinopathy, which is an important distinction that requires timely ophthalmic intervention with laser or intravitreal injection treatment to reduce the risk of severe, permanent vision loss. In this study, we developed a deep learning algorithm to detect neovascular leakage on ultra-widefield fluorescein angiography images obtained from patients with diabetic retinopathy. The algorithm, an ensemble of three convolutional neural networks, was able to accurately classify neovascular leakage and distinguish this disease marker from other angiographic disease features. With additional real-world validation and testing, our algorithm could facilitate identification of neovascular leakage in the clinical setting, allowing timely intervention to reduce the burden of blinding diabetic eye disease.
Collapse
Affiliation(s)
- Peter Y Zhao
- Department of Ophthalmology and Visual Sciences, W.K. Kellogg Eye Center, University of Michigan, 1000 Wall Street, Ann Arbor, MI, 48105, USA
| | - Nikhil Bommakanti
- Department of Ophthalmology and Visual Sciences, W.K. Kellogg Eye Center, University of Michigan, 1000 Wall Street, Ann Arbor, MI, 48105, USA
| | - Gina Yu
- Department of Ophthalmology and Visual Sciences, W.K. Kellogg Eye Center, University of Michigan, 1000 Wall Street, Ann Arbor, MI, 48105, USA
| | - Michael T Aaberg
- Department of Ophthalmology and Visual Sciences, W.K. Kellogg Eye Center, University of Michigan, 1000 Wall Street, Ann Arbor, MI, 48105, USA
| | - Tapan P Patel
- Department of Ophthalmology and Visual Sciences, W.K. Kellogg Eye Center, University of Michigan, 1000 Wall Street, Ann Arbor, MI, 48105, USA
| | - Yannis M Paulus
- Department of Ophthalmology and Visual Sciences, W.K. Kellogg Eye Center, University of Michigan, 1000 Wall Street, Ann Arbor, MI, 48105, USA.
| |
Collapse
|
27
|
Kayadibi İ, Güraksın GE. An Explainable Fully Dense Fusion Neural Network with Deep Support Vector Machine for Retinal Disease Determination. INT J COMPUT INT SYS 2023. [DOI: 10.1007/s44196-023-00210-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/11/2023] Open
Abstract
AbstractRetinal issues are crucial because they result in visual loss. Early diagnosis can aid physicians in initiating treatment and preventing visual loss. Optical coherence tomography (OCT), which portrays retinal morphology cross-sectionally and noninvasively, is used to identify retinal abnormalities. The process of analyzing OCT images, on the other hand, takes time. This study has proposed a hybrid approach based on a fully dense fusion neural network (FD-CNN) and dual preprocessing to identify retinal diseases, such as choroidal neovascularization, diabetic macular edema, drusen from OCT images. A dual preprocessing methodology, in other words, a hybrid speckle reduction filter was initially used to diminish speckle noise present in OCT images. Secondly, the FD-CNN architecture was trained, and the features obtained from this architecture were extracted. Then Deep Support Vector Machine (D-SVM) and Deep K-Nearest Neighbor (D-KNN) classifiers were proposed to reclassify those features and tested on University of California San Diego (UCSD) and Duke OCT datasets. D-SVM demonstrated the best performance in both datasets. D-SVM achieved 99.60% accuracy, 99.60% sensitivity, 99.87% specificity, 99.60% precision and 99.60% F1 score in the UCSD dataset. It achieved 97.50% accuracy, 97.64% sensitivity, 98.91% specificity, 96.61% precision, and 97.03% F1 score in Duke dataset. Additionally, the results were compared to state-of-the-art works on the both datasets. The D-SVM was demonstrated to be an efficient and productive strategy for improving the robustness of automatic retinal disease classification. Also, in this study, it is shown that the unboxing of how AI systems' black-box choices is made by generating heat maps using the local interpretable model-agnostic explanation method, which is an explainable artificial intelligence (XAI) technique. Heat maps, in particular, may contribute to the development of more stable deep learning-based systems, as well as enhancing the confidence in the diagnosis of retinal disease in the analysis of OCT image for ophthalmologists.
Collapse
|
28
|
He J, Wang J, Han Z, Ma J, Wang C, Qi M. An interpretable transformer network for the retinal disease classification using optical coherence tomography. Sci Rep 2023; 13:3637. [PMID: 36869160 PMCID: PMC9984386 DOI: 10.1038/s41598-023-30853-z] [Citation(s) in RCA: 12] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/22/2022] [Accepted: 03/02/2023] [Indexed: 03/05/2023] Open
Abstract
Retinal illnesses such as age-related macular degeneration and diabetic macular edema will lead to irreversible blindness. With optical coherence tomography (OCT), doctors are able to see cross-sections of the retinal layers and provide patients with a diagnosis. Manual reading of OCT images is time-consuming, labor-intensive and even error-prone. Computer-aided diagnosis algorithms improve efficiency by automatically analyzing and diagnosing retinal OCT images. However, the accuracy and interpretability of these algorithms can be further improved through effective feature extraction, loss optimization and visualization analysis. In this paper, we propose an interpretable Swin-Poly Transformer network for performing automatically retinal OCT image classification. By shifting the window partition, the Swin-Poly Transformer constructs connections between neighboring non-overlapping windows in the previous layer and thus has the flexibility to model multi-scale features. Besides, the Swin-Poly Transformer modifies the importance of polynomial bases to refine cross entropy for better retinal OCT image classification. In addition, the proposed method also provides confidence score maps, assisting medical practitioners to understand the models' decision-making process. Experiments in OCT2017 and OCT-C8 reveal that the proposed method outperforms both the convolutional neural network approach and ViT, with an accuracy of 99.80% and an AUC of 99.99%.
Collapse
Affiliation(s)
- Jingzhen He
- Department of Radiology, Qilu Hospital of Shandong University, Jinan, 250012, China.
| | - Junxia Wang
- School of Information Science and Engineering, Shandong Normal University, Jinan, 250358, China
| | - Zeyu Han
- School of Mathematics and Statistics, Shandong University, Weihai, 264209, China
| | - Jun Ma
- School of Cyber Science and Engineering, Southeast University, Nanjing, 211189, China
| | - Chongjing Wang
- China Academy of Information and Communications Technology, Beijing, 100191, China
| | - Meng Qi
- School of Information Science and Engineering, Shandong Normal University, Jinan, 250358, China.
| |
Collapse
|
29
|
A review of deep learning-based multiple-lesion recognition from medical images: classification, detection and segmentation. Comput Biol Med 2023; 157:106726. [PMID: 36924732 DOI: 10.1016/j.compbiomed.2023.106726] [Citation(s) in RCA: 40] [Impact Index Per Article: 20.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/15/2022] [Revised: 02/07/2023] [Accepted: 02/27/2023] [Indexed: 03/05/2023]
Abstract
Deep learning-based methods have become the dominant methodology in medical image processing with the advancement of deep learning in natural image classification, detection, and segmentation. Deep learning-based approaches have proven to be quite effective in single lesion recognition and segmentation. Multiple-lesion recognition is more difficult than single-lesion recognition due to the little variation between lesions or the too wide range of lesions involved. Several studies have recently explored deep learning-based algorithms to solve the multiple-lesion recognition challenge. This paper includes an in-depth overview and analysis of deep learning-based methods for multiple-lesion recognition developed in recent years, including multiple-lesion recognition in diverse body areas and recognition of whole-body multiple diseases. We discuss the challenges that still persist in the multiple-lesion recognition tasks by critically assessing these efforts. Finally, we outline existing problems and potential future research areas, with the hope that this review will help researchers in developing future approaches that will drive additional advances.
Collapse
|
30
|
Mohamad Almustafa K, Kumar Sharma A, Bhardwaj S. STARC: Deep learning Algorithms’ modelling for STructured analysis of retina classification. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104357] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/12/2022]
|
31
|
Chen X, Xue Y, Wu X, Zhong Y, Rao H, Luo H, Weng Z. Deep Learning-Based System for Disease Screening and Pathologic Region Detection From Optical Coherence Tomography Images. Transl Vis Sci Technol 2023; 12:29. [PMID: 36716039 PMCID: PMC9896901 DOI: 10.1167/tvst.12.1.29] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/31/2023] Open
Abstract
Purpose This study was designed to apply deep learning models in retinal disease screening and lesion detection based on optical coherence tomography (OCT) images. Methods We collected 37,138 OCT images from 775 patients and labelled by ophthalmologists. Multiple deep learning models including ResNet50 and YOLOv3 were developed to identify the types and locations of diseases or lesions based on the images. Results The model were evaluated using patient-based independent holdout set. For binary classification of OCT images with or without lesions, the performance accuracy was 98.5%, sensitivity was 98.7%, specificity was 98.4%, and the F1 score was 97.7%. For multiclass multilabel disease classification, the models was able to detect vitreomacular traction syndrome and age-related macular degeneration both with an accuracy of more than 99%, sensitivity of more than 98%, specificity of more than 98%, and an F1 score of more than 97%. For lesion location detection, the recalls for different lesion types ranged from 87.0% (epiretinal membrane) to 98.2% (macular pucker). Conclusions Deep learning-based models have potentials to aid retinal disease screening, classification and diagnosis with excellent performance, which may serve as useful references for ophthalmologists. Translational Relevance The deep learning-based models are capable of identifying and predicting different eye diseases and lesions from OCT images and may have potential clinical application to assist the ophthalmologists for fast and accuracy retinal disease screening.
Collapse
Affiliation(s)
- Xiaoming Chen
- College of Mathematics and Computer Science, Fuzhou University, Fujian province, China,The Centre for Big Data Research in Burns and Trauma, College of Mathematics and Computer Science, Fuzhou University, Fujian province, China
| | - Ying Xue
- Department of Ophthalmology, Fujian Provincial Hospital, Fuzhou, China
| | - Xiaoyan Wu
- Department of Ophthalmology, Fujian Provincial Hospital, Fuzhou, China
| | - Yi Zhong
- The Centre for Big Data Research in Burns and Trauma, College of Mathematics and Computer Science, Fuzhou University, Fujian province, China,College of Biological Science and Engineering, Fuzhou University, Fujian province, China
| | - Huiying Rao
- Department of Ophthalmology, Fujian Provincial Hospital, Fuzhou, China
| | - Heng Luo
- The Centre for Big Data Research in Burns and Trauma, College of Mathematics and Computer Science, Fuzhou University, Fujian province, China,College of Biological Science and Engineering, Fuzhou University, Fujian province, China,MetaNovas Biotech Inc., Foster City, CA, USA
| | - Zuquan Weng
- The Centre for Big Data Research in Burns and Trauma, College of Mathematics and Computer Science, Fuzhou University, Fujian province, China,College of Biological Science and Engineering, Fuzhou University, Fujian province, China
| |
Collapse
|
32
|
Gan F, Wu FP, Zhong YL. Artificial intelligence method based on multi-feature fusion for automatic macular edema (ME) classification on spectral-domain optical coherence tomography (SD-OCT) images. Front Neurosci 2023; 17:1097291. [PMID: 36793539 PMCID: PMC9922866 DOI: 10.3389/fnins.2023.1097291] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/13/2022] [Accepted: 01/09/2023] [Indexed: 01/31/2023] Open
Abstract
Purpose A common ocular manifestation, macular edema (ME) is the primary cause of visual deterioration. In this study, an artificial intelligence method based on multi-feature fusion was introduced to enable automatic ME classification on spectral-domain optical coherence tomography (SD-OCT) images, to provide a convenient method of clinical diagnosis. Methods First, 1,213 two-dimensional (2D) cross-sectional OCT images of ME were collected from the Jiangxi Provincial People's Hospital between 2016 and 2021. According to OCT reports of senior ophthalmologists, there were 300 images with diabetic (DME), 303 images with age-related macular degeneration (AMD), 304 images with retinal-vein occlusion (RVO), and 306 images with central serous chorioretinopathy (CSC). Then, traditional omics features of the images were extracted based on the first-order statistics, shape, size, and texture. After extraction by the alexnet, inception_v3, resnet34, and vgg13 models and selected by dimensionality reduction using principal components analysis (PCA), the deep-learning features were fused. Next, the gradient-weighted class-activation map (Grad-CAM) was used to visualize the-deep-learning process. Finally, the fusion features set, which was fused from the traditional omics features and the deep-fusion features, was used to establish the final classification models. The performance of the final models was evaluated by accuracy, confusion matrix, and the receiver operating characteristic (ROC) curve. Results Compared with other classification models, the performance of the support vector machine (SVM) model was best, with an accuracy of 93.8%. The area under curves AUC of micro- and macro-averages were 99%, and the AUC of the AMD, DME, RVO, and CSC groups were 100, 99, 98, and 100%, respectively. Conclusion The artificial intelligence model in this study could be used to classify DME, AME, RVO, and CSC accurately from SD-OCT images.
Collapse
Affiliation(s)
- Fan Gan
- Medical College of Nanchang University, Nanchang, China,Department of Ophthalmology, Jiangxi Provincial People’s Hospital, The First Affiliated Hospital of Nanchang Medical College, Nanchang, China
| | - Fei-Peng Wu
- Department of Ophthalmology, Jiangxi Provincial People’s Hospital, The First Affiliated Hospital of Nanchang Medical College, Nanchang, China
| | - Yu-Lin Zhong
- Department of Ophthalmology, Jiangxi Provincial People’s Hospital, The First Affiliated Hospital of Nanchang Medical College, Nanchang, China,*Correspondence: Yu-Lin Zhong,
| |
Collapse
|
33
|
Choudhary A, Ahlawat S, Urooj S, Pathak N, Lay-Ekuakille A, Sharma N. A Deep Learning-Based Framework for Retinal Disease Classification. Healthcare (Basel) 2023; 11:healthcare11020212. [PMID: 36673578 PMCID: PMC9859538 DOI: 10.3390/healthcare11020212] [Citation(s) in RCA: 12] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/21/2022] [Revised: 12/23/2022] [Accepted: 12/29/2022] [Indexed: 01/12/2023] Open
Abstract
This study addresses the problem of the automatic detection of disease states of the retina. In order to solve the abovementioned problem, this study develops an artificially intelligent model. The model is based on a customized 19-layer deep convolutional neural network called VGG-19 architecture. The model (VGG-19 architecture) is empowered by transfer learning. The model is designed so that it can learn from a large set of images taken with optical coherence tomography (OCT) and classify them into four conditions of the retina: (1) choroidal neovascularization, (2) drusen, (3) diabetic macular edema, and (4) normal form. The training datasets (taken from publicly available sources) consist of 84,568 instances of OCT retinal images. The datasets exhibit all four classes of retinal disease mentioned above. The proposed model achieved a 99.17% classification accuracy with 0.995 specificities and 0.99 sensitivity, making it better than the existing models. In addition, the proper statistical evaluation is done on the predictions using such performance measures as (1) area under the receiver operating characteristic curve, (2) Cohen's kappa parameter, and (3) confusion matrix. Experimental results show that the proposed VGG-19 architecture coupled with transfer learning is an effective technique for automatically detecting the disease state of a retina.
Collapse
Affiliation(s)
- Amit Choudhary
- University School of Automation and Robotics, G.G.S. Indraprastha University, New Delhi 110092, India
| | - Savita Ahlawat
- Maharaja Surajmal Institute of Technology, G.G.S. Indraprastha University, New Delhi 110058, India
- Correspondence: (S.A.); (S.U.)
| | - Shabana Urooj
- Department of Electrical Engineering, College of Engineering, Princess Nourah Bint Abdulrahman University, P.O. Box 84428, Riyadh 11671, Saudi Arabia
- Correspondence: (S.A.); (S.U.)
| | - Nitish Pathak
- Department of Information Technology, Bhagwan Parshuram Institute of Technology (BPIT), G.G.S. Indraprastha University, New Delhi 110078, India
| | - Aimé Lay-Ekuakille
- Department of Innovation Engineering, University of Salento, 73100 Lecce, Italy
| | - Neelam Sharma
- Department of Artificial Intelligence and Machine Learning, Maharaja Agrasen Institute of Technology (MAIT), G.G.S. Indraprastha University, New Delhi 110086, India
| |
Collapse
|
34
|
Jin K, Yan Y, Wang S, Yang C, Chen M, Liu X, Terasaki H, Yeo TH, Singh NG, Wang Y, Ye J. iERM: An Interpretable Deep Learning System to Classify Epiretinal Membrane for Different Optical Coherence Tomography Devices: A Multi-Center Analysis. J Clin Med 2023; 12:400. [PMID: 36675327 PMCID: PMC9862104 DOI: 10.3390/jcm12020400] [Citation(s) in RCA: 22] [Impact Index Per Article: 11.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/14/2022] [Revised: 12/29/2022] [Accepted: 01/03/2023] [Indexed: 01/06/2023] Open
Abstract
Background: Epiretinal membranes (ERM) have been found to be common among individuals >50 years old. However, the severity grading assessment for ERM based on optical coherence tomography (OCT) images has remained a challenge due to lacking reliable and interpretable analysis methods. Thus, this study aimed to develop a two-stage deep learning (DL) system named iERM to provide accurate automatic grading of ERM for clinical practice. Methods: The iERM was trained based on human segmentation of key features to improve classification performance and simultaneously provide interpretability to the classification results. We developed and tested iERM using a total of 4547 OCT B-Scans of four different commercial OCT devices that were collected from nine international medical centers. Results: As per the results, the integrated network effectively improved the grading performance by 1−5.9% compared with the traditional classification DL model and achieved high accuracy scores of 82.9%, 87.0%, and 79.4% in the internal test dataset and two external test datasets, respectively. This is comparable to retinal specialists whose average accuracy scores are 87.8% and 79.4% in two external test datasets. Conclusion: This study proved to be a benchmark method to improve the performance and enhance the interpretability of the traditional DL model with the implementation of segmentation based on prior human knowledge. It may have the potential to provide precise guidance for ERM diagnosis and treatment.
Collapse
Affiliation(s)
- Kai Jin
- Department of Ophthalmology, The Second Affiliated Hospital of Zhejiang University, College of Medicine, Hangzhou 310009, China
| | - Yan Yan
- Department of Ophthalmology, The Second Affiliated Hospital of Zhejiang University, College of Medicine, Hangzhou 310009, China
| | - Shuai Wang
- School of Mechanical, Electrical and Information Engineering, Shandong University, Weihai 264209, China
| | - Ce Yang
- School of Mechanical, Electrical and Information Engineering, Shandong University, Weihai 264209, China
| | - Menglu Chen
- Department of Ophthalmology, The Second Affiliated Hospital of Zhejiang University, College of Medicine, Hangzhou 310009, China
| | - Xindi Liu
- Department of Ophthalmology, The Second Affiliated Hospital of Zhejiang University, College of Medicine, Hangzhou 310009, China
| | - Hiroto Terasaki
- Department of Ophthalmology, Kagoshima University Graduate School of Medical and Dental Sciences, Kagoshima 890-8520, Japan
| | - Tun-Hang Yeo
- Ophthalmology and Visual Sciences, Khoo Teck Puat Hospital, National Healthcare Group, Singapore 768828, Singapore
| | - Neha Gulab Singh
- Ophthalmology and Visual Sciences, Khoo Teck Puat Hospital, National Healthcare Group, Singapore 768828, Singapore
| | - Yao Wang
- Department of Ophthalmology, The Second Affiliated Hospital of Zhejiang University, College of Medicine, Hangzhou 310009, China
| | - Juan Ye
- Department of Ophthalmology, The Second Affiliated Hospital of Zhejiang University, College of Medicine, Hangzhou 310009, China
| |
Collapse
|
35
|
Hsia Y, Lin YY, Wang BS, Su CY, Lai YH, Hsieh YT. Prediction of Visual Impairment in Epiretinal Membrane and Feature Analysis: A Deep Learning Approach Using Optical Coherence Tomography. Asia Pac J Ophthalmol (Phila) 2023; 12:21-28. [PMID: 36706331 DOI: 10.1097/apo.0000000000000576] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/17/2022] [Accepted: 09/14/2022] [Indexed: 01/28/2023] Open
Abstract
PURPOSE The aim was to develop a deep learning model for predicting the extent of visual impairment in epiretinal membrane (ERM) using optical coherence tomography (OCT) images, and to analyze the associated features. METHODS Six hundred macular OCT images from eyes with ERM and no visually significant media opacity or other retinal diseases were obtained. Those with best-corrected visual acuity ≤20/50 were classified as "profound visual impairment," while those with best-corrected visual acuity >20/50 were classified as "less visual impairment." Ninety percent of images were used as the training data set and 10% were used for testing. Two convolutional neural network models (ResNet-50 and ResNet-18) were adopted for training. The t-distributed stochastic neighbor-embedding approach was used to compare their performances. The Grad-CAM technique was used in the heat map generative phase for feature analysis. RESULTS During the model development, the training accuracy was 100% in both convolutional neural network models, while the testing accuracy was 70% and 80% for ResNet-18 and ResNet-50, respectively. The t-distributed stochastic neighbor-embedding approach found that the deeper structure (ResNet-50) had better discrimination on OCT characteristics for visual impairment than the shallower structure (ResNet-18). The heat maps indicated that the key features for visual impairment were located mostly in the inner retinal layers of the fovea and parafoveal regions. CONCLUSIONS Deep learning algorithms could assess the extent of visual impairment from OCT images in patients with ERM. Changes in inner retinal layers were found to have a greater impact on visual acuity than the outer retinal changes.
Collapse
Affiliation(s)
- Yun Hsia
- National Taiwan University Biomedical Park Hospital, Hsin-Chu
- Department of Ophthalmology, National Taiwan University Hospital, Taipei, Taiwan
| | - Yu-Yi Lin
- Department of Biomedical Engineering, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Bo-Sin Wang
- Department of Biomedical Engineering, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Chung-Yen Su
- Department of Electrical Engineering, National Taiwan Normal University, Taipei, Taiwan
| | - Ying-Hui Lai
- Department of Biomedical Engineering, National Yang Ming Chiao Tung University, Taipei, Taiwan
- Medical Device Innovation & Translation Center, National Yang Ming Chiao Tung University, Taipei, Taiwan
| | - Yi-Ting Hsieh
- Department of Ophthalmology, National Taiwan University Hospital, Taipei, Taiwan
| |
Collapse
|
36
|
Tvenning A, Hanssen SR, Austeng D, Morken TS. Deep learning identify retinal nerve fibre and choroid layers as markers of age-related macular degeneration in the classification of macular spectral-domain optical coherence tomography volumes. Acta Ophthalmol 2022; 100:937-945. [PMID: 35233918 PMCID: PMC9790497 DOI: 10.1111/aos.15126] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/11/2021] [Revised: 02/08/2022] [Accepted: 02/18/2022] [Indexed: 12/30/2022]
Affiliation(s)
- Arnt‐Ole Tvenning
- Department of Ophthalmology, St. Olav HospitalTrondheim University HospitalTrondheimNorway
| | - Stian Rikstad Hanssen
- Department of Computer ScienceNorwegian University of Science and Technology (NTNU)TrondheimNorway
| | - Dordi Austeng
- Department of Ophthalmology, St. Olav HospitalTrondheim University HospitalTrondheimNorway,Department of Neuromedicine and Movement ScienceNTNUTrondheimNorway
| | - Tora Sund Morken
- Department of Ophthalmology, St. Olav HospitalTrondheim University HospitalTrondheimNorway,Department of Neuromedicine and Movement ScienceNTNUTrondheimNorway
| |
Collapse
|
37
|
Puneet, Kumar R, Gupta M. Optical coherence tomography image based eye disease detection using deep convolutional neural network. Health Inf Sci Syst 2022; 10:13. [PMID: 35756852 PMCID: PMC9213631 DOI: 10.1007/s13755-022-00182-y] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2022] [Accepted: 06/08/2022] [Indexed: 12/23/2022] Open
Abstract
Over the past few decades, health care industries and medical practitioners faced a lot of obstacles to diagnosing medical-related problems due to inadequate technology and availability of equipment. In the present era, computer science technologies such as IoT, Cloud Computing, Artificial Intelligence and its allied techniques, etc. play a crucial role in the identification of medical diseases, especially in the domain of Ophthalmology. Despite this, ophthalmologists have to perform the various disease diagnosis task manually which is time-consuming and the chances of error are also very high because some of the abnormalities of eye diseases possess the same symptoms. Furthermore, multiple autonomous systems also exist to categorize the diseases but their prediction rate does not accomplish state-of-art accuracy. In the proposed approach by implementing the concept of Attention, Transfer Learning with the Deep Convolution Neural Network, the model accomplished an accuracy of 97.79% and 95.6% on the training and testing data respectively. This autonomous model efficiently classifies the various oscular disorders namely Choroidal Neovascularization, Diabetic Macular Edema, Drusen from the Optical Coherence Tomography images. It may provide a realistic solution to the healthcare sector to bring down the ophthalmologist burden in the screening of Diabetic Retinopathy.
Collapse
Affiliation(s)
- Puneet
- Department of Computer Science and Engineering, Chandigarh University, Mohali, Punjab India
| | - Rakesh Kumar
- Department of Computer Science and Engineering, Chandigarh University, Mohali, Punjab India
| | - Meenu Gupta
- Department of Computer Science and Engineering, Chandigarh University, Mohali, Punjab India
| |
Collapse
|
38
|
Sun LC, Pao SI, Huang KH, Wei CY, Lin KF, Chen PN. Generative adversarial network-based deep learning approach in classification of retinal conditions with optical coherence tomography images. Graefes Arch Clin Exp Ophthalmol 2022; 261:1399-1412. [PMID: 36441228 DOI: 10.1007/s00417-022-05919-9] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/19/2022] [Revised: 09/21/2022] [Accepted: 11/22/2022] [Indexed: 11/29/2022] Open
Abstract
PURPOSE To determine whether a deep learning approach using generative adversarial networks (GANs) is beneficial for the classification of retinal conditions with Optical coherence tomography (OCT) images. METHODS Our study utilized 84,452 retinal OCT images obtained from a publicly available dataset (Kermany Dataset). Employing GAN, synthetic OCT images are produced to balance classes of retinal disorders. A deep learning classification model is constructed using pretrained deep neural networks (DNNs), and outcomes are evaluated using 2082 images collected from patients who visited the Department of Ophthalmology and the Department of Endocrinology and Metabolism at the Tri-service General Hospital in Taipei from January 2017 to December 2021. RESULTS The highest classification accuracies accomplished by deep learning machines trained on the unbalanced dataset for its training set, validation set, fivefold cross validation (CV), Kermany test set, and TSGH test set were 97.73%, 96.51%, 97.14%, 99.59%, and 81.03%, respectively. The highest classification accuracies accomplished by deep learning machines trained on the synthesis-balanced dataset for its training set, validation set, fivefold CV, Kermany test set, and TSGH test set were 98.60%, 98.41%, 98.52%, 99.38%, and 84.92%, respectively. In comparing the highest accuracies, deep learning machines trained on the synthesis-balanced dataset outperformed deep learning machines trained on the unbalanced dataset for the training set, validation set, fivefold CV, and TSGH test set. CONCLUSIONS Overall, deep learning machines on a synthesis-balanced dataset demonstrated to be advantageous over deep learning machines trained on an unbalanced dataset for the classification of retinal conditions.
Collapse
Affiliation(s)
- Ling-Chun Sun
- School of Medicine, National Defense Medical Center, Taipei, Taiwan
| | - Shu-I Pao
- Department of Ophthalmology, Tri-Service General Hospital, National Defense Medical Center, Taipei, Taiwan
| | - Ke-Hao Huang
- Department of Ophthalmology, Song-Shan Branch of Tri-Service General Hospital, National Defense Medical Center, Taipei, Taiwan
| | - Chih-Yuan Wei
- Graduate Institute of Life Sciences, National Defense Medical Center, Taipei, Taiwan
| | - Ke-Feng Lin
- Medical Informatics Office, Tri-Service General Hospital, National Defense Medical Center, Taipei, Taiwan
- School of Public Health, National Defense Medical Center, Taipei, Taiwan
| | - Ping-Nan Chen
- Department of Biomedical Engineering, National Defense Medical Center, No.161, Sec.6, Minchiuan E. Rd., Neihu Dist, Taipei, 11490, Taiwan.
| |
Collapse
|
39
|
Kugelman J, Alonso-Caneiro D, Read SA, Collins MJ. A review of generative adversarial network applications in optical coherence tomography image analysis. JOURNAL OF OPTOMETRY 2022; 15 Suppl 1:S1-S11. [PMID: 36241526 PMCID: PMC9732473 DOI: 10.1016/j.optom.2022.09.004] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/31/2022] [Revised: 08/19/2022] [Accepted: 09/20/2022] [Indexed: 06/16/2023]
Abstract
Optical coherence tomography (OCT) has revolutionized ophthalmic clinical practice and research, as a result of the high-resolution images that the method is able to capture in a fast, non-invasive manner. Although clinicians can interpret OCT images qualitatively, the ability to quantitatively and automatically analyse these images represents a key goal for eye care by providing clinicians with immediate and relevant metrics to inform best clinical practice. The range of applications and methods to analyse OCT images is rich and rapidly expanding. With the advent of deep learning methods, the field has experienced significant progress with state-of-the-art-performance for several OCT image analysis tasks. Generative adversarial networks (GANs) represent a subfield of deep learning that allows for a range of novel applications not possible in most other deep learning methods, with the potential to provide more accurate and robust analyses. In this review, the progress in this field and clinical impact are reviewed and the potential future development of applications of GANs to OCT image processing are discussed.
Collapse
Affiliation(s)
- Jason Kugelman
- Queensland University of Technology (QUT), Contact Lens and Visual Optics Laboratory, Centre for Vision and Eye Research, School of Optometry and Vision Science, Kelvin Grove, QLD 4059, Australia.
| | - David Alonso-Caneiro
- Queensland University of Technology (QUT), Contact Lens and Visual Optics Laboratory, Centre for Vision and Eye Research, School of Optometry and Vision Science, Kelvin Grove, QLD 4059, Australia
| | - Scott A Read
- Queensland University of Technology (QUT), Contact Lens and Visual Optics Laboratory, Centre for Vision and Eye Research, School of Optometry and Vision Science, Kelvin Grove, QLD 4059, Australia
| | - Michael J Collins
- Queensland University of Technology (QUT), Contact Lens and Visual Optics Laboratory, Centre for Vision and Eye Research, School of Optometry and Vision Science, Kelvin Grove, QLD 4059, Australia
| |
Collapse
|
40
|
Wongchaisuwat P, Thamphithak R, Jitpukdee P, Wongchaisuwat N. Application of Deep Learning for Automated Detection of Polypoidal Choroidal Vasculopathy in Spectral Domain Optical Coherence Tomography. Transl Vis Sci Technol 2022; 11:16. [PMID: 36219163 PMCID: PMC9580222 DOI: 10.1167/tvst.11.10.16] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/03/2022] [Accepted: 08/29/2022] [Indexed: 11/25/2022] Open
Abstract
Objective To develop an automated polypoidal choroidal vasculopathy (PCV) screening model to distinguish PCV from wet age-related macular degeneration (wet AMD). Methods A retrospective review of spectral domain optical coherence tomography (SD-OCT) images was undertaken. The included SD-OCT images were classified into two distinct categories (PCV or wet AMD) prior to the development of the PCV screening model. The automated detection of PCV using the developed model was compared with the results of gold-standard fundus fluorescein angiography and indocyanine green (FFA + ICG) angiography. A framework of SHapley Additive exPlanations was used to interpret the results from the model. Results A total of 2334 SD-OCT images were enrolled for training purposes, and an additional 1171 SD-OCT images were used for external validation. The ResNet attention model yielded superior performance with average area under the curve values of 0.8 and 0.81 for the training and external validation data sets, respectively. The sensitivity/specificity calculated at a patient level was 100%/60% and 85%/71% for the training and external validation data sets, respectively. Conclusions A conventional FFA + ICG investigation to differentiate PCV from wet AMD requires intense health care resources and adversely affects patients. A deep learning algorithm is proposed to automatically distinguish PCV from wet AMD. The developed algorithm exhibited promising performance for further development into an alternative PCV screening tool. Enhancement of the model's performance with additional data is needed prior to implementation of this diagnostic tool in real-world clinical practice. The invisibility of disease signs within SD-OCT images is the main limitation of the proposed model. Translational Relevance Basic research of deep learning algorithms was applied to differentiate PCV from wet AMD based on OCT images, benefiting a diagnosis process and minimizing a risk of ICG angiogram.
Collapse
Affiliation(s)
- Papis Wongchaisuwat
- Department of Industrial Engineering, Faculty of Engineering, Kasetsart University, Bangkok, Thailand
| | - Ranida Thamphithak
- Department of Ophthalmology, Faculty of Medicine Siriraj Hospital, Mahidol University, Bangkok, Thailand
| | - Peerakarn Jitpukdee
- Department of Industrial Engineering, Faculty of Engineering, Kasetsart University, Bangkok, Thailand
| | - Nida Wongchaisuwat
- Department of Ophthalmology, Faculty of Medicine Siriraj Hospital, Mahidol University, Bangkok, Thailand
| |
Collapse
|
41
|
End-to-End Multi-Task Learning Approaches for the Joint Epiretinal Membrane Segmentation and Screening in OCT Images. Comput Med Imaging Graph 2022; 98:102068. [DOI: 10.1016/j.compmedimag.2022.102068] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/19/2022] [Revised: 03/28/2022] [Accepted: 04/18/2022] [Indexed: 02/07/2023]
|
42
|
Lachance A, Godbout M, Antaki F, Hébert M, Bourgault S, Caissie M, Tourville É, Durand A, Dirani A. Predicting Visual Improvement After Macular Hole Surgery: A Combined Model Using Deep Learning and Clinical Features. Transl Vis Sci Technol 2022; 11:6. [PMID: 35385045 PMCID: PMC8994199 DOI: 10.1167/tvst.11.4.6] [Citation(s) in RCA: 18] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/02/2022] Open
Abstract
Purpose The purpose of this study was to assess the feasibility of deep learning (DL) methods to enhance the prediction of visual acuity (VA) improvement after macular hole (MH) surgery from a combined model using DL on high-definition optical coherence tomography (HD-OCT) B-scans and clinical features. Methods We trained a DL convolutional neural network (CNN) using pre-operative HD-OCT B-scans of the macula and combined with a logistic regression model of pre-operative clinical features to predict VA increase ≥15 Early Treatment Diabetic Retinopathy Study (ETDRS) letters at 6 months post-vitrectomy in closed MHs. A total of 121 MHs with 242 HD-OCT B-scans and 484 clinical data points were used to train, validate, and test the model. Prediction of VA increase was evaluated using the area under the receiver operating characteristic curve (AUROC) and F1 scores. We also extracted the weight of each input feature in the hybrid model. Results All performances are reported on the held-out test set, matching results obtained with cross-validation. Using a regression on clinical features, the AUROC was 80.6, with an F1 score of 79.7. For the CNN, relying solely on the HD-OCT B-scans, the AUROC was 72.8 ± 14.6, with an F1 score of 61.5 ± 23.7. For our hybrid regression model using clinical features and CNN prediction, the AUROC was 81.9 ± 5.2, with an F1 score of 80.4 ± 7.7. In the hybrid model, the baseline VA was the most important feature (weight = 59.1 ± 6.9%), while the weight of HD-OCT prediction was 9.6 ± 4.2%. Conclusions Both the clinical data and HD-OCT models can predict postoperative VA improvement in patients undergoing vitrectomy for a MH with good discriminative performances. Combining them into a hybrid model did not significantly improve performance. Translational Relevance OCT-based DL models can predict postoperative VA improvement following vitrectomy for MH but fusing those models with clinical data might not provide improved predictive performance.
Collapse
Affiliation(s)
- Alexandre Lachance
- Faculté de Médecine, Université Laval, Québec, QC, Canada.,Département d'Ophtalmologie et d'oto-Rhino-Laryngologie - Chirurgie Cervico-Faciale, Centre Universitaire d'Ophtalmologie, Hôpital du Saint-Sacrement, CHU de Québec - Université Laval, Québec, QC, Canada
| | - Mathieu Godbout
- Département d'informatique et de Génie Logiciel, Université Laval, Québec, QC, Canada
| | - Fares Antaki
- Département d'ophtalmologie, Centre Hospitalier de l'Université de Montréal (CHUM), Montréal, Québec, QC, Canada
| | - Mélanie Hébert
- Faculté de Médecine, Université Laval, Québec, QC, Canada.,Département d'Ophtalmologie et d'oto-Rhino-Laryngologie - Chirurgie Cervico-Faciale, Centre Universitaire d'Ophtalmologie, Hôpital du Saint-Sacrement, CHU de Québec - Université Laval, Québec, QC, Canada
| | - Serge Bourgault
- Faculté de Médecine, Université Laval, Québec, QC, Canada.,Département d'Ophtalmologie et d'oto-Rhino-Laryngologie - Chirurgie Cervico-Faciale, Centre Universitaire d'Ophtalmologie, Hôpital du Saint-Sacrement, CHU de Québec - Université Laval, Québec, QC, Canada
| | - Mathieu Caissie
- Faculté de Médecine, Université Laval, Québec, QC, Canada.,Département d'Ophtalmologie et d'oto-Rhino-Laryngologie - Chirurgie Cervico-Faciale, Centre Universitaire d'Ophtalmologie, Hôpital du Saint-Sacrement, CHU de Québec - Université Laval, Québec, QC, Canada
| | - Éric Tourville
- Faculté de Médecine, Université Laval, Québec, QC, Canada.,Département d'Ophtalmologie et d'oto-Rhino-Laryngologie - Chirurgie Cervico-Faciale, Centre Universitaire d'Ophtalmologie, Hôpital du Saint-Sacrement, CHU de Québec - Université Laval, Québec, QC, Canada
| | - Audrey Durand
- Département d'informatique et de Génie Logiciel, Université Laval, Québec, QC, Canada.,Département de Génie Électrique et de Génie Informatique, Université Laval, Québec, QC, Canada
| | - Ali Dirani
- Faculté de Médecine, Université Laval, Québec, QC, Canada.,Département d'Ophtalmologie et d'oto-Rhino-Laryngologie - Chirurgie Cervico-Faciale, Centre Universitaire d'Ophtalmologie, Hôpital du Saint-Sacrement, CHU de Québec - Université Laval, Québec, QC, Canada
| |
Collapse
|
43
|
Chueh KM, Hsieh YT, Chen HH, Ma IH, Huang SL. Identification of Sex and Age from Macular Optical Coherence Tomography and Feature Analysis Using Deep Learning. Am J Ophthalmol 2022; 235:221-228. [PMID: 34582766 DOI: 10.1016/j.ajo.2021.09.015] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/27/2021] [Revised: 09/09/2021] [Accepted: 09/15/2021] [Indexed: 11/01/2022]
Abstract
PURPOSE To develop deep learning models for identification of sex and age from macular optical coherence tomography (OCT) and to analyze the features for differentiation of sex and age. DESIGN Algorithm development using database of macular OCT. METHODS We reviewed 6147 sets of macular OCT images from the healthy eyes of 3134 individuals from a single eye center in Taiwan. Deep learning-based algorithms were used to develop models for the identification of sex and age, and 10-fold cross-validation was applied. Gradient-weighted class activation mapping was used for feature analysis. RESULTS The accuracy for sex prediction using deep learning from macular OCT was 85.6% ± 2.1% compared with accuracy of 61.9% using macular thickness and 61.4% ± 4.0% using deep learning from infrared fundus photography (P < .001 for both). The mean absolute error for age prediction using deep learning from macular OCT was 5.78 ± 0.29 years. A thorough analysis of the prediction accuracy and the gradient-weighted class activation mapping showed that the cross-sectional foveal contour lead to a better sex distinction than macular thickness or fundus photography, and the age-related characteristics of macula were on the whole layers of retina rather than the choroid. CONCLUSIONS Sex and age could be identified from macular OCT using deep learning with good accuracy. The main sexual difference of macula lies in the foveal contour, and the whole layers of retina differ with aging. These novel findings provide useful information for further investigation in the pathogenesis of sex- and age-related macular structural diseases.
Collapse
|
44
|
Zgolli H, El Zarrug HHK, Meddeb M, Mabrouk S, Khlifa N. Anatomical prognosis after idiopathic macular hole surgery: machine learning based-predection. Libyan J Med 2022; 17:2034334. [PMID: 35180831 PMCID: PMC8865103 DOI: 10.1080/19932820.2022.2034334] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/04/2022] Open
Abstract
To develop a machine learning (ML) model for the prediction of the idiopathic macular hole (MH) status at 9 months after vitrectomy and inverted flap internal limiting membrane (ILM) peeling surgery. This single center was conducted at Department A, Institute Hedi Raies of Ophthalmology, Tunis, Tunisia. The study included 114 patients. In total, 120 eyes underwent optical coherence tomography (OCT) and inverted flap ILM peeling for surgery. Then 510 B scan of macular OCT was acquired 9 months after surgery. MH diameter, basal MH diameter (b), nasal and temporal arm lengths and macular hole angle were measured. Indices including hole form factor, MH index, diameter hole index (DHI) and tractional hole, MH area index and MH volume index were calculated. Receiver operating characteristic (ROC) curves and cut‑off values were derived for each indices predicting closure or not of the MH. The area under the receiver operating characteristic curve (AUC) and kappa value were calculated to evaluate performance of the medical decision support system (MDSS) in predicting the MH closure. From the ROC curve analysis, it was derived that MH indices like MH diameter, diameter hole index (DHI), MH index, and hole formation factor were capable of successfully predicting MH closure while basal diameter, DHI and MH area index predicted none closure MH. The MDSS achieved an AUC of 0.984 with a kappa value of 0.934. Based on the preoperative OCT parameters, our ML model achieved remarkable accuracy in predicting MH outcomes after pars plana vitrectomy and inverted flap ILM peeling. Therefore, MDSS may help optimize surgical planning for full thickness macular hole patients in the future.
Collapse
Affiliation(s)
- Hsouna Zgolli
- Department A, Institute Hedi Raies of Ophthalmology, Tunis, Tunisia
| | - Hamad H K El Zarrug
- Department of Ophthalmology University of Benghazi, Faculty of Medicine Lybia Benghazi
| | - Moufid Meddeb
- Laboratory of Biophysics and Medical Technologies, Higher Institute of Medical Technologies of Tunis, University of Tuins El Manar, Tunis, Tunisia
| | - Sonya Mabrouk
- Department A, Institute Hedi Raies of Ophthalmology, Tunis, Tunisia
| | - Nawres Khlifa
- Laboratory of Biophysics and Medical Technologies, Higher Institute of Medical Technologies of Tunis, University of Tuins El Manar, Tunis, Tunisia
| |
Collapse
|
45
|
Al Turk L, Georgieva D, Alsawadi H, Wang S, Krause P, Alsawadi H, Alshamrani AZ, Saleh GM, Tang HL. Learning to Discover Explainable Clinical Features With Minimum Supervision. Transl Vis Sci Technol 2022; 11:11. [PMID: 35015061 PMCID: PMC8762682 DOI: 10.1167/tvst.11.1.11] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022] Open
Abstract
Purpose To compare supervised transfer learning to semisupervised learning for their ability to learn in-depth knowledge with limited data in the optical coherence tomography (OCT) domain. Methods Transfer learning with EfficientNet-B4 and semisupervised learning with SimCLR are used in this work. The largest public OCT dataset, consisting of 108,312 images and four categories (choroidal neovascularization, diabetic macular edema, drusen, and normal) is used. In addition, two smaller datasets are constructed, containing 31,200 images for the limited version and 4000 for the mini version of the dataset. To illustrate the effectiveness of the developed models, local interpretable model-agnostic explanations and class activation maps are used as explainability techniques. Results The proposed transfer learning approach using the EfficientNet-B4 model trained on the limited dataset achieves an accuracy of 0.976 (95% confidence interval [CI], 0.963, 0.983), sensitivity of 0.973 and specificity of 0.991. The semisupervised based solution with SimCLR using 10% labeled data and the limited dataset performs with an accuracy of 0.946 (95% CI, 0.932, 0.960), sensitivity of 0.941, and specificity of 0.983. Conclusions Semisupervised learning has a huge potential for datasets that contain both labeled and unlabeled inputs, generally, with a significantly smaller number of labeled samples. The semisupervised based solution provided with merely 10% labeled data achieves very similar performance to the supervised transfer learning that uses 100% labeled samples. Translational Relevance Semisupervised learning enables building performant models while requiring less expertise effort and time by using to good advantage the abundant amount of available unlabeled data along with the labeled samples.
Collapse
Affiliation(s)
- Lutfiah Al Turk
- Department of Statistics, King Abdulaziz University, Jeddah, Kingdom of Saudi Arabia
| | - Darina Georgieva
- Department of Computer Science, University of Surrey, Guildford, Surrey, UK
| | - Hassan Alsawadi
- Department of Electrical and Computer Engineering, King Abdulaziz, University, Jeddah, Kingdom of Saudi Arabia
| | - Su Wang
- Department of Computer Science, University of Surrey, Guildford, Surrey, UK
| | - Paul Krause
- Department of Computer Science, University of Surrey, Guildford, Surrey, UK
| | - Hend Alsawadi
- Faculty of Medicine, King Abdulaziz University, Jeddah, Kingdom of Saudi Arabia
| | | | - George M Saleh
- NIHR Biomedical Research Centre at Moorfields Eye Hospital and the UCL Institute of Ophthalmology, UK
| | | |
Collapse
|
46
|
Development and Validation of an Explainable Artificial Intelligence Framework for Macular Disease Diagnosis Based on OCT Images. Retina 2021; 42:456-464. [PMID: 34723902 DOI: 10.1097/iae.0000000000003325] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
Abstract
PURPOSE To develop and validate an artificial intelligence framework for identifying multiple retinal lesions at image-level and performing an explainable macular disease diagnosis at eye-level in optical coherence tomography (OCT) images. METHODS 26,815 OCT images were collected from 865 eyes, and 9 retinal lesions and 3 macular diseases were labelled by ophthalmologists, including diabetic macular edema (DME) and dry/wet age-related macular degeneration (dry/wet AMD). We applied deep learning to classify retinal lesion at image-level and random forests to achieve an explainable disease diagnosis at eye-level. The performance of the integrated two-stage framework was evaluated and compared with human experts. RESULTS On testing dataset of 2,480 OCT images from 80 eyes, deep learning model achieved average Area Under Curve (AUC) of 0.978 (95% CI, 0.971-0.983) for lesion classification. And random forests performed accurate disease diagnosis with 0% error rate, which achieved the same accuracy as one of human experts and was better than the other 3 experts. It also revealed that the detection of specific lesions in the center of macular region had more contribution to macular disease diagnosis. CONCLUSIONS The integrated method achieved high accuracy and interpretability in retinal lesion classification and macular disease diagnosis in OCT images, and could have the potential to facilitate the clinical diagnosis.
Collapse
|
47
|
Zhang Y, Li M, Ji Z, Fan W, Yuan S, Liu Q, Chen Q. Twin self-supervision based semi-supervised learning (TS-SSL): Retinal anomaly classification in SD-OCT images. Neurocomputing 2021. [DOI: 10.1016/j.neucom.2021.08.051] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/29/2023]
|
48
|
Shao E, Liu C, Wang L, Song D, Guo L, Yao X, Xiong J, Wang B, Hu Y. Artificial intelligence-based detection of epimacular membrane from color fundus photographs. Sci Rep 2021; 11:19291. [PMID: 34588493 PMCID: PMC8481557 DOI: 10.1038/s41598-021-98510-x] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/30/2020] [Accepted: 09/01/2021] [Indexed: 12/25/2022] Open
Abstract
Epiretinal membrane (ERM) is a common ophthalmological disorder of high prevalence. Its symptoms include metamorphopsia, blurred vision, and decreased visual acuity. Early diagnosis and timely treatment of ERM is crucial to preventing vision loss. Although optical coherence tomography (OCT) is regarded as a de facto standard for ERM diagnosis due to its intuitiveness and high sensitivity, ophthalmoscopic examination or fundus photographs still have the advantages of price and accessibility. Artificial intelligence (AI) has been widely applied in the health care industry for its robust and significant performance in detecting various diseases. In this study, we validated the use of a previously trained deep neural network based-AI model in ERM detection based on color fundus photographs. An independent test set of fundus photographs was labeled by a group of ophthalmologists according to their corresponding OCT images as the gold standard. Then the test set was interpreted by other ophthalmologists and AI model without knowing their OCT results. Compared with manual diagnosis based on fundus photographs alone, the AI model had comparable accuracy (AI model 77.08% vs. integrated manual diagnosis 75.69%, χ2 = 0.038, P = 0.845, McNemar’s test), higher sensitivity (75.90% vs. 63.86%, χ2 = 4.500, P = 0.034, McNemar’s test), under the cost of lower but reasonable specificity (78.69% vs. 91.80%, χ2 = 6.125, P = 0.013, McNemar’s test). Thus our AI model can serve as a possible alternative for manual diagnosis in ERM screening.
Collapse
Affiliation(s)
- Enhua Shao
- Department of Ophthalmology, Beijing Tisnghua Changgung Hospital, School of Clinical Medicine, Tsinghua University, Beijing, China
| | - Congxin Liu
- Beijing Eaglevision Technology Co., Ltd, Beijing, China
| | - Lei Wang
- Department of Ophthalmology, Beijing Tisnghua Changgung Hospital, School of Clinical Medicine, Tsinghua University, Beijing, China
| | - Dan Song
- Department of Ophthalmology, Beijing Tisnghua Changgung Hospital, School of Clinical Medicine, Tsinghua University, Beijing, China
| | - Libin Guo
- Department of Ophthalmology, Beijing Tisnghua Changgung Hospital, School of Clinical Medicine, Tsinghua University, Beijing, China
| | - Xuan Yao
- Beijing Eaglevision Technology Co., Ltd, Beijing, China
| | - Jianhao Xiong
- Beijing Eaglevision Technology Co., Ltd, Beijing, China
| | - Bin Wang
- Beijing Eaglevision Technology Co., Ltd, Beijing, China
| | - Yuntao Hu
- Department of Ophthalmology, Beijing Tisnghua Changgung Hospital, School of Clinical Medicine, Tsinghua University, Beijing, China.
| |
Collapse
|
49
|
Owen JP, Blazes M, Manivannan N, Lee GC, Yu S, Durbin MK, Nair A, Singh RP, Talcott KE, Melo AG, Greenlee T, Chen ER, Conti TF, Lee CS, Lee AY. Student becomes teacher: training faster deep learning lightweight networks for automated identification of optical coherence tomography B-scans of interest using a student-teacher framework. BIOMEDICAL OPTICS EXPRESS 2021; 12:5387-5399. [PMID: 34692189 PMCID: PMC8515993 DOI: 10.1364/boe.433432] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/09/2021] [Revised: 07/10/2021] [Accepted: 07/19/2021] [Indexed: 06/13/2023]
Abstract
This work explores a student-teacher framework that leverages unlabeled images to train lightweight deep learning models with fewer parameters to perform fast automated detection of optical coherence tomography B-scans of interest. Twenty-seven lightweight models (LWMs) from four families of models were trained on expert-labeled B-scans (∼70 K) as either "abnormal" or "normal", which established a baseline performance for the models. Then the LWMs were trained from random initialization using a student-teacher framework to incorporate a large number of unlabeled B-scans (∼500 K). A pre-trained ResNet50 model served as the teacher network. The ResNet50 teacher model achieved 96.0% validation accuracy and the validation accuracy achieved by the LWMs ranged from 89.6% to 95.1%. The best performing LWMs were 2.53 to 4.13 times faster than ResNet50 (0.109s to 0.178s vs. 0.452s). All LWMs benefitted from increasing the training set by including unlabeled B-scans in the student-teacher framework, with several models achieving validation accuracy of 96.0% or higher. The three best-performing models achieved comparable sensitivity and specificity in two hold-out test sets to the teacher network. We demonstrated the effectiveness of a student-teacher framework for training fast LWMs for automated B-scan of interest detection leveraging unlabeled, routinely-available data.
Collapse
Affiliation(s)
- Julia P. Owen
- Department of Ophthalmology, University of Washington, Seattle, WA 98195, USA
| | - Marian Blazes
- Department of Ophthalmology, University of Washington, Seattle, WA 98195, USA
| | | | - Gary C. Lee
- Carl Zeiss Meditec, Inc., Dublin, CA 94568, USA
| | - Sophia Yu
- Carl Zeiss Meditec, Inc., Dublin, CA 94568, USA
| | | | - Aditya Nair
- Carl Zeiss Meditec, Inc., Dublin, CA 94568, USA
| | - Rishi P. Singh
- Center for Ophthalmic Bioinformatics, Cole Eye Institute, Cleveland Clinic, Cleveland, OH 44195, USA
| | - Katherine E. Talcott
- Center for Ophthalmic Bioinformatics, Cole Eye Institute, Cleveland Clinic, Cleveland, OH 44195, USA
| | - Alline G. Melo
- Center for Ophthalmic Bioinformatics, Cole Eye Institute, Cleveland Clinic, Cleveland, OH 44195, USA
| | - Tyler Greenlee
- Center for Ophthalmic Bioinformatics, Cole Eye Institute, Cleveland Clinic, Cleveland, OH 44195, USA
| | - Eric R. Chen
- Center for Ophthalmic Bioinformatics, Cole Eye Institute, Cleveland Clinic, Cleveland, OH 44195, USA
| | - Thais F. Conti
- Center for Ophthalmic Bioinformatics, Cole Eye Institute, Cleveland Clinic, Cleveland, OH 44195, USA
| | - Cecilia S. Lee
- Department of Ophthalmology, University of Washington, Seattle, WA 98195, USA
| | - Aaron Y. Lee
- Department of Ophthalmology, University of Washington, Seattle, WA 98195, USA
| |
Collapse
|
50
|
Contactless optical coherence tomography of the eyes of freestanding individuals with a robotic scanner. Nat Biomed Eng 2021; 5:726-736. [PMID: 34253888 PMCID: PMC9272353 DOI: 10.1038/s41551-021-00753-6] [Citation(s) in RCA: 23] [Impact Index Per Article: 5.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/06/2020] [Accepted: 05/21/2021] [Indexed: 02/02/2023]
Abstract
Clinical systems for optical coherence tomography (OCT) are used routinely to diagnose and monitor patients with a range of ocular diseases. They are large tabletop instruments operated by trained staff, and require mechanical stabilization of the head of the patient for positioning and motion reduction. Here we report the development and performance of a robot-mounted OCT scanner for the autonomous contactless imaging, at safe distances, of the eyes of freestanding individuals without the need for operator intervention or head stabilization. The scanner uses robotic positioning to align itself with the eye to be imaged, as well as optical active scanning to locate the pupil and to attenuate physiological eye motion. We show that the scanner enables the acquisition of OCT volumetric datasets, comparable in quality to those of clinical tabletop systems, that resolve key anatomic structures relevant for the management of common eye conditions. Robotic OCT scanners may enable the diagnosis and monitoring of patients with eye conditions in non-specialist clinics.
Collapse
|