1
|
Greenfield JA, Scherer R, Alba D, De Arrigunaga S, Alvarez O, Palioura S, Nanji A, Bayyat GA, da Costa DR, Herskowitz W, Antonietti M, Jammal A, Al-Khersan H, Wu W, Shousha MA, O'Brien R, Galor A, Medeiros FA, Karp CL. Detection of Ocular Surface Squamous Neoplasia Using Artificial Intelligence With Anterior Segment Optical Coherence Tomography. Am J Ophthalmol 2025; 273:182-191. [PMID: 39983942 PMCID: PMC11985264 DOI: 10.1016/j.ajo.2025.02.019] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/19/2024] [Revised: 02/12/2025] [Accepted: 02/13/2025] [Indexed: 02/23/2025]
Abstract
PURPOSE To develop and validate a deep learning (DL) model to differentiate ocular surface squamous neoplasia (OSSN) from pterygium and pinguecula using high-resolution anterior segment optical coherence tomography (AS-OCT). DESIGN Retrospective Diagnostic Accuracy Study. METHODS Setting: Single-center. STUDY POPULATION All eyes with a clinical or biopsy-proven diagnosis of OSSN, pterygium, or pinguecula that received AS-OCT imaging. PROCEDURES Imaging data was extracted from Optovue AS-OCT (Fremont, CA) and patients' clinical or biopsy-proven diagnoses were collected from electronic medical records. A DL classification model was developed using two methodologies: (1) a masked autoencoder was trained with unlabeled data from 105,859 AS-OCT images of 5746 eyes and (2) a Vision Transformer supervised model coupled to the autoencoder used labeled data for fine-tuning a binary classifier (OSSN vs non-OSSN lesions). A sample of 2022 AS-OCT images from 523 eyes (427 patients) were classified by expert graders into "OSSN or suspicious for OSSN" and "pterygium or pinguecula." The algorithm's diagnostic performance was evaluated in a separate test sample using 566 scans (62 eyes, 48 patients) with biopsy-proven OSSN and compared with expert clinicians who were masked to the diagnosis. Analysis was conducted at the scan-level for both the DL model and expert clinicians, who were not provided with clinical images or supporting clinical data. MAIN OUTCOME Diagnostic performance of expert clinicians and the DL model in identifying OSSN on AS-OCT scans. RESULTS The DL model had an accuracy of 90.3% (95% confidence intervals [CI]: 87.5%-92.6%), with sensitivity of 86.4% (95% CI: 81.4%-90.4%) and specificity of 93.2% (95% CI: 89.9%-95.7%) compared to the biopsy-proven diagnosis. Expert graders had a lower sensitivity 69.8% (95% CI: 63.6%-75.5%) and slightly higher specificity 98.5% (95% CI: 96.4%-99.5%) than the DL model. The area under the receiver operating characteristic curve for the DL model was 0.945 (95% CI: 0.918-0.972) and significantly greater than expert graders (area under the receiver operating characteristic curve = 0.688, P < .001). CONCLUSIONS A DL model applied to AS-OCT scans demonstrated high accuracy, sensitivity, and specificity in differentiating OSSN from pterygium and pinguecula. Interestingly, the model had comparable diagnostic performance to expert clinicians in this study and shows promise for enhancing clinical decision-making. Further research is warranted to explore the integration of this artificial intelligence-driven approach in routine screening and diagnostic protocols for OSSN.
Collapse
Affiliation(s)
- Jason A Greenfield
- From the Bascom Palmer Eye Institute (J.A.G., R.S., D.A., S.D.A., O.A., S.P., D.R.C., W.H., M.A., A.J., H.A.K., W.W., M.A.S., R.O., A.G., F.A.M., C.L.K.), University of Miami Miller School of Medicine, Miami, Florida, USA
| | - Rafael Scherer
- From the Bascom Palmer Eye Institute (J.A.G., R.S., D.A., S.D.A., O.A., S.P., D.R.C., W.H., M.A., A.J., H.A.K., W.W., M.A.S., R.O., A.G., F.A.M., C.L.K.), University of Miami Miller School of Medicine, Miami, Florida, USA
| | - Diego Alba
- From the Bascom Palmer Eye Institute (J.A.G., R.S., D.A., S.D.A., O.A., S.P., D.R.C., W.H., M.A., A.J., H.A.K., W.W., M.A.S., R.O., A.G., F.A.M., C.L.K.), University of Miami Miller School of Medicine, Miami, Florida, USA
| | - Sofia De Arrigunaga
- From the Bascom Palmer Eye Institute (J.A.G., R.S., D.A., S.D.A., O.A., S.P., D.R.C., W.H., M.A., A.J., H.A.K., W.W., M.A.S., R.O., A.G., F.A.M., C.L.K.), University of Miami Miller School of Medicine, Miami, Florida, USA
| | - Osmel Alvarez
- From the Bascom Palmer Eye Institute (J.A.G., R.S., D.A., S.D.A., O.A., S.P., D.R.C., W.H., M.A., A.J., H.A.K., W.W., M.A.S., R.O., A.G., F.A.M., C.L.K.), University of Miami Miller School of Medicine, Miami, Florida, USA
| | - Sotiria Palioura
- From the Bascom Palmer Eye Institute (J.A.G., R.S., D.A., S.D.A., O.A., S.P., D.R.C., W.H., M.A., A.J., H.A.K., W.W., M.A.S., R.O., A.G., F.A.M., C.L.K.), University of Miami Miller School of Medicine, Miami, Florida, USA
| | - Afshan Nanji
- Oregon Health & Science University (A.N.), Portland, Oregon, USA
| | | | - Douglas Rodrigues da Costa
- From the Bascom Palmer Eye Institute (J.A.G., R.S., D.A., S.D.A., O.A., S.P., D.R.C., W.H., M.A., A.J., H.A.K., W.W., M.A.S., R.O., A.G., F.A.M., C.L.K.), University of Miami Miller School of Medicine, Miami, Florida, USA
| | - William Herskowitz
- From the Bascom Palmer Eye Institute (J.A.G., R.S., D.A., S.D.A., O.A., S.P., D.R.C., W.H., M.A., A.J., H.A.K., W.W., M.A.S., R.O., A.G., F.A.M., C.L.K.), University of Miami Miller School of Medicine, Miami, Florida, USA
| | - Michael Antonietti
- From the Bascom Palmer Eye Institute (J.A.G., R.S., D.A., S.D.A., O.A., S.P., D.R.C., W.H., M.A., A.J., H.A.K., W.W., M.A.S., R.O., A.G., F.A.M., C.L.K.), University of Miami Miller School of Medicine, Miami, Florida, USA
| | - Alessandro Jammal
- From the Bascom Palmer Eye Institute (J.A.G., R.S., D.A., S.D.A., O.A., S.P., D.R.C., W.H., M.A., A.J., H.A.K., W.W., M.A.S., R.O., A.G., F.A.M., C.L.K.), University of Miami Miller School of Medicine, Miami, Florida, USA
| | - Hasenin Al-Khersan
- From the Bascom Palmer Eye Institute (J.A.G., R.S., D.A., S.D.A., O.A., S.P., D.R.C., W.H., M.A., A.J., H.A.K., W.W., M.A.S., R.O., A.G., F.A.M., C.L.K.), University of Miami Miller School of Medicine, Miami, Florida, USA
| | - Winfred Wu
- From the Bascom Palmer Eye Institute (J.A.G., R.S., D.A., S.D.A., O.A., S.P., D.R.C., W.H., M.A., A.J., H.A.K., W.W., M.A.S., R.O., A.G., F.A.M., C.L.K.), University of Miami Miller School of Medicine, Miami, Florida, USA
| | - Mohamed Abou Shousha
- From the Bascom Palmer Eye Institute (J.A.G., R.S., D.A., S.D.A., O.A., S.P., D.R.C., W.H., M.A., A.J., H.A.K., W.W., M.A.S., R.O., A.G., F.A.M., C.L.K.), University of Miami Miller School of Medicine, Miami, Florida, USA
| | - Robert O'Brien
- From the Bascom Palmer Eye Institute (J.A.G., R.S., D.A., S.D.A., O.A., S.P., D.R.C., W.H., M.A., A.J., H.A.K., W.W., M.A.S., R.O., A.G., F.A.M., C.L.K.), University of Miami Miller School of Medicine, Miami, Florida, USA
| | - Anat Galor
- From the Bascom Palmer Eye Institute (J.A.G., R.S., D.A., S.D.A., O.A., S.P., D.R.C., W.H., M.A., A.J., H.A.K., W.W., M.A.S., R.O., A.G., F.A.M., C.L.K.), University of Miami Miller School of Medicine, Miami, Florida, USA; Department of Ophthalmology (A.G.), Miami Veterans Administration Medical Center, Miami, Florida, USA
| | - Felipe A Medeiros
- From the Bascom Palmer Eye Institute (J.A.G., R.S., D.A., S.D.A., O.A., S.P., D.R.C., W.H., M.A., A.J., H.A.K., W.W., M.A.S., R.O., A.G., F.A.M., C.L.K.), University of Miami Miller School of Medicine, Miami, Florida, USA
| | - Carol L Karp
- From the Bascom Palmer Eye Institute (J.A.G., R.S., D.A., S.D.A., O.A., S.P., D.R.C., W.H., M.A., A.J., H.A.K., W.W., M.A.S., R.O., A.G., F.A.M., C.L.K.), University of Miami Miller School of Medicine, Miami, Florida, USA.
| |
Collapse
|
2
|
Phipps B, Hadoux X, Sheng B, Campbell JP, Liu TYA, Keane PA, Cheung CY, Chung TY, Wong TY, van Wijngaarden P. AI image generation technology in ophthalmology: Use, misuse and future applications. Prog Retin Eye Res 2025; 106:101353. [PMID: 40107410 DOI: 10.1016/j.preteyeres.2025.101353] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2024] [Revised: 03/12/2025] [Accepted: 03/13/2025] [Indexed: 03/22/2025]
Abstract
BACKGROUND AI-powered image generation technology holds the potential to reshape medical practice, yet it remains an unfamiliar technology for both medical researchers and clinicians alike. Given the adoption of this technology relies on clinician understanding and acceptance, we sought to demystify its use in ophthalmology. To this end, we present a literature review on image generation technology in ophthalmology, examining both its theoretical applications and future role in clinical practice. METHODS First, we consider the key model designs used for image synthesis, including generative adversarial networks, autoencoders, and diffusion models. We then perform a survey of the literature for image generation technology in ophthalmology prior to September 2024, presenting both the type of model used and its clinical application. Finally, we discuss the limitations of this technology, the risks of its misuse and the future directions of research in this field. RESULTS Applications of this technology include improving AI diagnostic models, inter-modality image transformation, more accurate treatment and disease prognostication, image denoising, and individualised education. Key barriers to its adoption include bias in generative models, risks to patient data security, computational and logistical barriers to development, challenges with model explainability, inconsistent use of validation metrics between studies and misuse of synthetic images. Looking forward, researchers are placing a further emphasis on clinically grounded metrics, the development of image generation foundation models and the implementation of methods to ensure data provenance. CONCLUSION Compared to other medical applications of AI, image generation is still in its infancy. Yet, it holds the potential to revolutionise ophthalmology across research, education and clinical practice. This review aims to guide ophthalmic researchers wanting to leverage this technology, while also providing an insight for clinicians on how it may change ophthalmic practice in the future.
Collapse
Affiliation(s)
- Benjamin Phipps
- Centre for Eye Research Australia, Royal Victorian Eye and Ear Hospital, East Melbourne, 3002, VIC, Australia; Ophthalmology, Department of Surgery, University of Melbourne, Parkville, 3010, VIC, Australia.
| | - Xavier Hadoux
- Centre for Eye Research Australia, Royal Victorian Eye and Ear Hospital, East Melbourne, 3002, VIC, Australia; Ophthalmology, Department of Surgery, University of Melbourne, Parkville, 3010, VIC, Australia
| | - Bin Sheng
- Department of Computer Science and Engineering, Shanghai Jiao Tong University, Shanghai, China
| | - J Peter Campbell
- Department of Ophthalmology, Casey Eye Institute, Oregon Health and Science University, Portland, USA
| | - T Y Alvin Liu
- Retina Division, Wilmer Eye Institute, Johns Hopkins University, Baltimore, MD, 21287, USA
| | - Pearse A Keane
- NIHR Biomedical Research Centre at Moorfields Eye Hospital NHS Foundation Trust, London, UK; Institute of Ophthalmology, University College London, London, UK
| | - Carol Y Cheung
- Department of Ophthalmology and Visual Sciences, The Chinese University of Hong Kong, 999077, China
| | - Tham Yih Chung
- Yong Loo Lin School of Medicine, National University of Singapore, Singapore; Centre for Innovation and Precision Eye Health, Department of Ophthalmology, Yong Loo Lin School of Medicine, National University of Singapore and National University Health System, Singapore; Singapore Eye Research Institute, Singapore National Eye Centre, Singapore; Eye Academic Clinical Program (Eye ACP), Duke NUS Medical School, Singapore
| | - Tien Y Wong
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore; Tsinghua Medicine, Tsinghua University, Beijing, China; Beijing Visual Science and Translational Eye Research Institute, Beijing Tsinghua Changgung Hospital, Beijing, China
| | - Peter van Wijngaarden
- Centre for Eye Research Australia, Royal Victorian Eye and Ear Hospital, East Melbourne, 3002, VIC, Australia; Ophthalmology, Department of Surgery, University of Melbourne, Parkville, 3010, VIC, Australia; Florey Institute of Neuroscience & Mental Health, Parkville, VIC, Australia
| |
Collapse
|
3
|
D N S, Pai RM, Bhat SN, Pai M M M. Assessment of perceived realism in AI-generated synthetic spine fracture CT images. Technol Health Care 2025; 33:931-944. [PMID: 40105176 DOI: 10.1177/09287329241291368] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/20/2025]
Abstract
BackgroundDeep learning-based decision support systems require synthetic images generated by adversarial networks, which require clinical evaluation to ensure their quality.ObjectiveThe study evaluates perceived realism of high-dimension synthetic spine fracture CT images generated Progressive Growing Generative Adversarial Networks (PGGANs).Method: The study used 2820 spine fracture CT images from 456 patients to train an PGGAN model. The model synthesized images up to 512 × 512 pixels, and the realism of the generated images was assessed using Visual Turing Tests and Fracture Identification Test. Three spine surgeons evaluated the images, and clinical evaluation results were statistically analysed.Result: Spine surgeons have an average prediction accuracy of nearly 50% during clinical evaluations, indicating difficulty in distinguishing between real and generated images. The accuracy varies for different dimensions, with synthetic images being more realistic, especially in 512 × 512-dimension images. During FIT, among 16 generated images of each fracture type, 13-15 images were correctly identified, indicating images are more realistic and clearly depict fracture lines in 512 × 512 dimensions.ConclusionThe study reveals that AI-based PGGAN can generate realistic synthetic spine fracture CT images up to 512 × 512 pixels, making them difficult to distinguish from real images, and improving the automatic spine fracture type detection system.
Collapse
Affiliation(s)
- Sindhura D N
- Department of Data Science and Computer Applications, Manipal Institute of Technology, Manipal, Manipal Academy of Higher Education, Manipal, India
| | - Radhika M Pai
- Department of Data Science and Computer Applications, Manipal Institute of Technology, Manipal, Manipal Academy of Higher Education, Manipal, India
| | - Shyamasunder N Bhat
- Department of Orthopaedics, Kasturba Medical College, Manipal, Manipal Academy of Higher Education, Manipal, India
| | - Manohara Pai M M
- Department of Information and Communication Technology, Manipal Institute of Technology, Manipal, Manipal Academy of Higher Education, Manipal, India
| |
Collapse
|
4
|
Behara K, Bhero E, Agee JT. AI in dermatology: a comprehensive review into skin cancer detection. PeerJ Comput Sci 2024; 10:e2530. [PMID: 39896358 PMCID: PMC11784784 DOI: 10.7717/peerj-cs.2530] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/04/2024] [Accepted: 10/28/2024] [Indexed: 02/04/2025]
Abstract
Background Artificial Intelligence (AI) is significantly transforming dermatology, particularly in early skin cancer detection and diagnosis. This technological advancement addresses a crucial public health issue by enhancing diagnostic accuracy, efficiency, and accessibility. AI integration in medical imaging and diagnostic procedures offers promising solutions to the limitations of traditional methods, which often rely on subjective clinical evaluations and histopathological analyses. This study systematically reviews current AI applications in skin cancer classification, providing a comprehensive overview of their advantages, challenges, methodologies, and functionalities. Methodology In this study, we conducted a comprehensive analysis of artificial intelligence (AI) applications in the classification of skin cancer. We evaluated publications from three prominent journal databases: Scopus, IEEE, and MDPI. We conducted a thorough selection process using the PRISMA guidelines, collecting 1,156 scientific articles. Our methodology included evaluating the titles and abstracts and thoroughly examining the full text to determine their relevance and quality. Consequently, we included a total of 95 publications in the final study. We analyzed and categorized the articles based on four key dimensions: advantages, difficulties, methodologies, and functionalities. Results AI-based models exhibit remarkable performance in skin cancer detection by leveraging advanced deep learning algorithms, image processing techniques, and feature extraction methods. The advantages of AI integration include significantly improved diagnostic accuracy, faster turnaround times, and increased accessibility to dermatological expertise, particularly benefiting underserved areas. However, several challenges remain, such as concerns over data privacy, complexities in integrating AI systems into existing workflows, and the need for large, high-quality datasets. AI-based methods for skin cancer detection, including CNNs, SVMs, and ensemble learning techniques, aim to improve lesion classification accuracy and increase early detection. AI systems enhance healthcare by enabling remote consultations, continuous patient monitoring, and supporting clinical decision-making, leading to more efficient care and better patient outcomes. Conclusions This comprehensive review highlights the transformative potential of AI in dermatology, particularly in skin cancer detection and diagnosis. While AI technologies have significantly improved diagnostic accuracy, efficiency, and accessibility, several challenges remain. Future research should focus on ensuring data privacy, developing robust AI systems that can generalize across diverse populations, and creating large, high-quality datasets. Integrating AI tools into clinical workflows is critical to maximizing their utility and effectiveness. Continuous innovation and interdisciplinary collaboration will be essential for fully realizing the benefits of AI in skin cancer detection and diagnosis.
Collapse
Affiliation(s)
- Kavita Behara
- Department of Electrical Engineering, Mangosuthu University of Technology, Durban, Kwazulu- Natal, South Africa
| | - Ernest Bhero
- Discipline of Computer Engineering, University of KwaZulu Natal, Durban, KwaZulu-Natal, South Africa
| | - John Terhile Agee
- Discipline of Computer Engineering, University of KwaZulu Natal, Durban, KwaZulu-Natal, South Africa
| |
Collapse
|
5
|
Sonmez SC, Sevgi M, Antaki F, Huemer J, Keane PA. Generative artificial intelligence in ophthalmology: current innovations, future applications and challenges. Br J Ophthalmol 2024; 108:1335-1340. [PMID: 38925907 PMCID: PMC11503064 DOI: 10.1136/bjo-2024-325458] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/02/2024] [Accepted: 06/03/2024] [Indexed: 06/28/2024]
Abstract
The rapid advancements in generative artificial intelligence are set to significantly influence the medical sector, particularly ophthalmology. Generative adversarial networks and diffusion models enable the creation of synthetic images, aiding the development of deep learning models tailored for specific imaging tasks. Additionally, the advent of multimodal foundational models, capable of generating images, text and videos, presents a broad spectrum of applications within ophthalmology. These range from enhancing diagnostic accuracy to improving patient education and training healthcare professionals. Despite the promising potential, this area of technology is still in its infancy, and there are several challenges to be addressed, including data bias, safety concerns and the practical implementation of these technologies in clinical settings.
Collapse
Affiliation(s)
| | - Mertcan Sevgi
- Institute of Ophthalmology, University College London, London, UK
- Moorfields Eye Hospital, NIHR Moorfields Biomedical Research Centre, London, UK
| | - Fares Antaki
- Institute of Ophthalmology, University College London, London, UK
- Moorfields Eye Hospital, NIHR Moorfields Biomedical Research Centre, London, UK
- The CHUM School of Artificial Intelligence in Healthcare, Montreal, Quebec, Canada
| | - Josef Huemer
- Moorfields Eye Hospital, NIHR Moorfields Biomedical Research Centre, London, UK
- Department of Ophthalmology and Optometry, Kepler University Hospital, Linz, Austria
| | - Pearse A Keane
- Institute of Ophthalmology, University College London, London, UK
- Moorfields Eye Hospital, NIHR Moorfields Biomedical Research Centre, London, UK
| |
Collapse
|
6
|
Li Z, Wang L, Qiang W, Chen K, Wang Z, Zhang Y, Xie H, Wu S, Jiang J, Chen W. DeepMonitoring: a deep learning-based monitoring system for assessing the quality of cornea images captured by smartphones. Front Cell Dev Biol 2024; 12:1447067. [PMID: 39258227 PMCID: PMC11385315 DOI: 10.3389/fcell.2024.1447067] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/11/2024] [Accepted: 08/19/2024] [Indexed: 09/12/2024] Open
Abstract
Smartphone-based artificial intelligence (AI) diagnostic systems could assist high-risk patients to self-screen for corneal diseases (e.g., keratitis) instead of detecting them in traditional face-to-face medical practices, enabling the patients to proactively identify their own corneal diseases at an early stage. However, AI diagnostic systems have significantly diminished performance in low-quality images which are unavoidable in real-world environments (especially common in patient-recorded images) due to various factors, hindering the implementation of these systems in clinical practice. Here, we construct a deep learning-based image quality monitoring system (DeepMonitoring) not only to discern low-quality cornea images created by smartphones but also to identify the underlying factors contributing to the generation of such low-quality images, which can guide operators to acquire high-quality images in a timely manner. This system performs well across validation, internal, and external testing sets, with AUCs ranging from 0.984 to 0.999. DeepMonitoring holds the potential to filter out low-quality cornea images produced by smartphones, facilitating the application of smartphone-based AI diagnostic systems in real-world clinical settings, especially in the context of self-screening for corneal diseases.
Collapse
Affiliation(s)
- Zhongwen Li
- Ningbo Key Laboratory of Medical Research on Blinding Eye Diseases, Ningbo Eye Institute, Ningbo Eye Hospital, Wenzhou Medical University, Ningbo, China
- National Clinical Research Center for Ocular Diseases, Eye Hospital, Wenzhou Medical University, Wenzhou, China
| | - Lei Wang
- National Clinical Research Center for Ocular Diseases, Eye Hospital, Wenzhou Medical University, Wenzhou, China
| | - Wei Qiang
- Ningbo Key Laboratory of Medical Research on Blinding Eye Diseases, Ningbo Eye Institute, Ningbo Eye Hospital, Wenzhou Medical University, Ningbo, China
| | - Kuan Chen
- Cangnan Hospital, Wenzhou Medical University, Wenzhou, China
| | - Zhouqian Wang
- National Clinical Research Center for Ocular Diseases, Eye Hospital, Wenzhou Medical University, Wenzhou, China
| | - Yi Zhang
- School of Electronic Engineering, Xi'an University of Posts and Telecommunications, Xi'an, China
| | - He Xie
- National Clinical Research Center for Ocular Diseases, Eye Hospital, Wenzhou Medical University, Wenzhou, China
| | - Shanjun Wu
- Ningbo Key Laboratory of Medical Research on Blinding Eye Diseases, Ningbo Eye Institute, Ningbo Eye Hospital, Wenzhou Medical University, Ningbo, China
| | - Jiewei Jiang
- School of Electronic Engineering, Xi'an University of Posts and Telecommunications, Xi'an, China
| | - Wei Chen
- Ningbo Key Laboratory of Medical Research on Blinding Eye Diseases, Ningbo Eye Institute, Ningbo Eye Hospital, Wenzhou Medical University, Ningbo, China
- National Clinical Research Center for Ocular Diseases, Eye Hospital, Wenzhou Medical University, Wenzhou, China
| |
Collapse
|
7
|
Qurban Q, Cassidy L. Artificial intelligence and machine learning a new frontier in the diagnosis of ocular adnexal tumors: A review. SAGE Open Med 2024; 12:20503121241274197. [PMID: 39206232 PMCID: PMC11350536 DOI: 10.1177/20503121241274197] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/04/2023] [Accepted: 07/11/2024] [Indexed: 09/04/2024] Open
Abstract
In our article, we explore the transformative potential of Artificial Intelligence and Machine Learning in oculo-oncology, focusing on the diagnosis and management of ocular adnexal tumors. Delving into the intricacies of adnexal conditions such as conjunctival melanoma and squamous conjunctival carcinoma, the study emphasizes recent breakthroughs, such as Artificial Intelligence-driven early detection methods. While acknowledging challenges like the scarcity of specialized datasets and issues in standardizing image capture, the research underscores encouraging patient acceptance, as demonstrated in melanoma diagnosis studies. The abstract calls for overcoming obstacles, conducting clinical trials, establishing global regulatory norms and fostering collaboration between ophthalmologists and Artificial Intelligence experts. Overall, the article envisions Artificial Intelligence's imminent transformative impact on ocular and periocular cancer diagnosis.
Collapse
Affiliation(s)
- Qirat Qurban
- Department of Ophthalmology and Oculoplastic, Royal Victoria Eye and Ear Hospital, Dublin, Ireland
- Trinity College Dublin, Dublin, Ireland
| | - Lorraine Cassidy
- Department of Ophthalmology and Oculoplastic, Royal Victoria Eye and Ear Hospital, Dublin, Ireland
| |
Collapse
|
8
|
Feng X, Xu K, Luo MJ, Chen H, Yang Y, He Q, Song C, Li R, Wu Y, Wang H, Tham YC, Ting DSW, Lin H, Wong TY, Lam DSC. Latest developments of generative artificial intelligence and applications in ophthalmology. Asia Pac J Ophthalmol (Phila) 2024; 13:100090. [PMID: 39128549 DOI: 10.1016/j.apjo.2024.100090] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/08/2024] [Revised: 07/30/2024] [Accepted: 08/07/2024] [Indexed: 08/13/2024] Open
Abstract
The emergence of generative artificial intelligence (AI) has revolutionized various fields. In ophthalmology, generative AI has the potential to enhance efficiency, accuracy, personalization and innovation in clinical practice and medical research, through processing data, streamlining medical documentation, facilitating patient-doctor communication, aiding in clinical decision-making, and simulating clinical trials. This review focuses on the development and integration of generative AI models into clinical workflows and scientific research of ophthalmology. It outlines the need for development of a standard framework for comprehensive assessments, robust evidence, and exploration of the potential of multimodal capabilities and intelligent agents. Additionally, the review addresses the risks in AI model development and application in clinical service and research of ophthalmology, including data privacy, data bias, adaptation friction, over interdependence, and job replacement, based on which we summarized a risk management framework to mitigate these concerns. This review highlights the transformative potential of generative AI in enhancing patient care, improving operational efficiency in the clinical service and research in ophthalmology. It also advocates for a balanced approach to its adoption.
Collapse
Affiliation(s)
- Xiaoru Feng
- School of Biomedical Engineering, Tsinghua Medicine, Tsinghua University, Beijing, China; Institute for Hospital Management, Tsinghua Medicine, Tsinghua University, Beijing, China
| | - Kezheng Xu
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangdong Provincial Key Laboratory of Ophthalmology and Visual Science, Guangdong Provincial Clinical Research Center for Ocular Diseases, Guangzhou, China
| | - Ming-Jie Luo
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangdong Provincial Key Laboratory of Ophthalmology and Visual Science, Guangdong Provincial Clinical Research Center for Ocular Diseases, Guangzhou, China
| | - Haichao Chen
- School of Clinical Medicine, Beijing Tsinghua Changgung Hospital, Tsinghua Medicine, Tsinghua University, Beijing, China
| | - Yangfan Yang
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangdong Provincial Key Laboratory of Ophthalmology and Visual Science, Guangdong Provincial Clinical Research Center for Ocular Diseases, Guangzhou, China
| | - Qi He
- Research Centre of Big Data and Artificial Research for Medicine, The First Affiliated Hospital, Sun Yat-Sen University, Guangzhou, China
| | - Chenxin Song
- Research Centre of Big Data and Artificial Research for Medicine, The First Affiliated Hospital, Sun Yat-Sen University, Guangzhou, China
| | - Ruiyao Li
- Research Centre of Big Data and Artificial Research for Medicine, The First Affiliated Hospital, Sun Yat-Sen University, Guangzhou, China
| | - You Wu
- Institute for Hospital Management, Tsinghua Medicine, Tsinghua University, Beijing, China; School of Basic Medical Sciences, Tsinghua Medicine, Tsinghua University, Beijing, China; Department of Health Policy and Management, Bloomberg School of Public Health, Johns Hopkins University, Baltimore, MD, USA.
| | - Haibo Wang
- Research Centre of Big Data and Artificial Research for Medicine, The First Affiliated Hospital, Sun Yat-Sen University, Guangzhou, China.
| | - Yih Chung Tham
- Department of Ophthalmology, Yong Loo Lin School of Medicine, National University of Singapore, Singapore; Centre for Innovation and Precision Eye Health, Yong Loo Lin School of Medicine, National University of Singapore, Singapore; Singapore Eye Research Institute, Singapore National Eye Centre, Singapore; Ophthalmology and Visual Science Academic Clinical Program, Duke-NUS Medical School, Singapore
| | - Daniel Shu Wei Ting
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore; Ophthalmology and Visual Science Academic Clinical Program, Duke-NUS Medical School, Singapore; Byers Eye Institute, Stanford University, Palo Alto, CA, USA
| | - Haotian Lin
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangdong Provincial Key Laboratory of Ophthalmology and Visual Science, Guangdong Provincial Clinical Research Center for Ocular Diseases, Guangzhou, China; Center for Precision Medicine and Department of Genetics and Biomedical Informatics, Zhongshan School of Medicine, Sun Yat-sen University, Guangzhou, China; Hainan Eye Hospital and Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Haikou, China
| | - Tien Yin Wong
- School of Clinical Medicine, Beijing Tsinghua Changgung Hospital, Tsinghua Medicine, Tsinghua University, Beijing, China; Singapore Eye Research Institute, Singapore National Eye Centre, Singapore; Tsinghua Medicine, Tsinghua University, Beijing, China
| | - Dennis Shun-Chiu Lam
- The International Eye Research Institute, The Chinese University of Hong Kong (Shenzhen), Shenzhen, China; The C-MER International Eye Care Group, Hong Kong, Hong Kong, China
| |
Collapse
|
9
|
Chen R, Zhang W, Song F, Yu H, Cao D, Zheng Y, He M, Shi D. Translating color fundus photography to indocyanine green angiography using deep-learning for age-related macular degeneration screening. NPJ Digit Med 2024; 7:34. [PMID: 38347098 PMCID: PMC10861476 DOI: 10.1038/s41746-024-01018-7] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/22/2023] [Accepted: 01/18/2024] [Indexed: 02/15/2024] Open
Abstract
Age-related macular degeneration (AMD) is the leading cause of central vision impairment among the elderly. Effective and accurate AMD screening tools are urgently needed. Indocyanine green angiography (ICGA) is a well-established technique for detecting chorioretinal diseases, but its invasive nature and potential risks impede its routine clinical application. Here, we innovatively developed a deep-learning model capable of generating realistic ICGA images from color fundus photography (CF) using generative adversarial networks (GANs) and evaluated its performance in AMD classification. The model was developed with 99,002 CF-ICGA pairs from a tertiary center. The quality of the generated ICGA images underwent objective evaluation using mean absolute error (MAE), peak signal-to-noise ratio (PSNR), structural similarity measures (SSIM), etc., and subjective evaluation by two experienced ophthalmologists. The model generated realistic early, mid and late-phase ICGA images, with SSIM spanned from 0.57 to 0.65. The subjective quality scores ranged from 1.46 to 2.74 on the five-point scale (1 refers to the real ICGA image quality, Kappa 0.79-0.84). Moreover, we assessed the application of translated ICGA images in AMD screening on an external dataset (n = 13887) by calculating area under the ROC curve (AUC) in classifying AMD. Combining generated ICGA with real CF images improved the accuracy of AMD classification with AUC increased from 0.93 to 0.97 (P < 0.001). These results suggested that CF-to-ICGA translation can serve as a cross-modal data augmentation method to address the data hunger often encountered in deep-learning research, and as a promising add-on for population-based AMD screening. Real-world validation is warranted before clinical usage.
Collapse
Affiliation(s)
- Ruoyu Chen
- Experimental Ophthalmology, School of Optometry, The Hong Kong Polytechnic University, Kowloon, Hong Kong SAR, China
- Research Centre for SHARP Vision, The Hong Kong Polytechnic University, Kowloon, Hong Kong SAR, China
| | - Weiyi Zhang
- Experimental Ophthalmology, School of Optometry, The Hong Kong Polytechnic University, Kowloon, Hong Kong SAR, China
- Research Centre for SHARP Vision, The Hong Kong Polytechnic University, Kowloon, Hong Kong SAR, China
| | - Fan Song
- Experimental Ophthalmology, School of Optometry, The Hong Kong Polytechnic University, Kowloon, Hong Kong SAR, China
- Research Centre for SHARP Vision, The Hong Kong Polytechnic University, Kowloon, Hong Kong SAR, China
| | - Honghua Yu
- Department of Ophthalmology, Guangdong Academy of Medical Sciences, Guangdong Provincial People's Hospital, Southern Medical University, Guangzhou, China
| | - Dan Cao
- Department of Ophthalmology, Guangdong Academy of Medical Sciences, Guangdong Provincial People's Hospital, Southern Medical University, Guangzhou, China
| | - Yingfeng Zheng
- State Key Laboratory of Ophthalmology, Zhongshan Ophthalmic Center, Sun Yat-sen University, Guangdong Provincial Key Laboratory of Ophthalmology and Visual Science, Guangdong Provincial Clinical Research Center for Ocular Diseases, Guangzhou, China.
| | - Mingguang He
- Experimental Ophthalmology, School of Optometry, The Hong Kong Polytechnic University, Kowloon, Hong Kong SAR, China.
- Research Centre for SHARP Vision, The Hong Kong Polytechnic University, Kowloon, Hong Kong SAR, China.
- Centre for Eye and Vision Research (CEVR), 17W Hong Kong Science Park, Hong Kong SAR, China.
| | - Danli Shi
- Experimental Ophthalmology, School of Optometry, The Hong Kong Polytechnic University, Kowloon, Hong Kong SAR, China.
- Research Centre for SHARP Vision, The Hong Kong Polytechnic University, Kowloon, Hong Kong SAR, China.
| |
Collapse
|
10
|
Choi JY, Kim H, Kim JK, Lee IS, Ryu IH, Kim JS, Yoo TK. Deep learning prediction of steep and flat corneal curvature using fundus photography in post-COVID telemedicine era. Med Biol Eng Comput 2024; 62:449-463. [PMID: 37889431 DOI: 10.1007/s11517-023-02952-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/21/2023] [Accepted: 10/14/2023] [Indexed: 10/28/2023]
Abstract
Recently, fundus photography (FP) is being increasingly used. Corneal curvature is an essential factor in refractive errors and is associated with several pathological corneal conditions. As FP-based examination systems have already been widely distributed, it would be helpful for telemedicine to extract information such as corneal curvature using FP. This study aims to develop a deep learning model based on FP for corneal curvature prediction by categorizing corneas into steep, regular, and flat groups. The EfficientNetB0 architecture with transfer learning was used to learn FP patterns to predict flat, regular, and steep corneas. In validation, the model achieved a multiclass accuracy of 0.727, a Matthews correlation coefficient of 0.519, and an unweighted Cohen's κ of 0.590. The areas under the receiver operating characteristic curves for binary prediction of flat and steep corneas were 0.863 and 0.848, respectively. The optic nerve and its peripheral areas were the main focus of the model. The developed algorithm shows that FP can potentially be used as an imaging modality to estimate corneal curvature in the post-COVID-19 era, whereby patients may benefit from the detection of abnormal corneal curvatures using FP in the telemedicine setting.
Collapse
Affiliation(s)
- Joon Yul Choi
- Department of Biomedical Engineering, Yonsei University, Wonju, South Korea
| | | | - Jin Kuk Kim
- Department of Refractive Surgery, B&VIIT Eye Center, B2 GT Tower, 1317-23 Seocho-Dong, Seocho-Gu, Seoul, South Korea
| | - In Sik Lee
- Department of Refractive Surgery, B&VIIT Eye Center, B2 GT Tower, 1317-23 Seocho-Dong, Seocho-Gu, Seoul, South Korea
| | - Ik Hee Ryu
- Department of Refractive Surgery, B&VIIT Eye Center, B2 GT Tower, 1317-23 Seocho-Dong, Seocho-Gu, Seoul, South Korea
- Research and Development Department, VISUWORKS, Seoul, South Korea
| | - Jung Soo Kim
- Research and Development Department, VISUWORKS, Seoul, South Korea
| | - Tae Keun Yoo
- Department of Refractive Surgery, B&VIIT Eye Center, B2 GT Tower, 1317-23 Seocho-Dong, Seocho-Gu, Seoul, South Korea.
- Research and Development Department, VISUWORKS, Seoul, South Korea.
| |
Collapse
|
11
|
Ryyppö R, Häyrynen S, Joutsijoki H, Juhola M, Seppänen MRJ. Comparison of machine learning methods in the early identification of vasculitides, myositides and glomerulonephritides. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 243:107917. [PMID: 37948909 DOI: 10.1016/j.cmpb.2023.107917] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/15/2023] [Revised: 09/29/2023] [Accepted: 11/05/2023] [Indexed: 11/12/2023]
Abstract
BACKGROUND Rare disease diagnoses are often delayed by years, including multiple doctor visits, and potential imprecise or incorrect diagnoses before receiving the correct one. Machine learning could solve this problem by flagging potential patients that doctors should examine more closely. METHODS Making the prediction situation as close as possible to real situation, we tested different masking sizes. In the masking phase, data was removed, and it was applied to all data points following the first rare disease diagnosis, including the day when the diagnosis was received, and in addition applied to selected number of days before initial diagnosis. Performance of machine learning models were compared with positive predictive value (PPV), negative predictive value (NPV), prevalence PPV (pPPV), prevalence NPV (pNPV), accuracy (ACC) and area under the receiver operation characteristics curve (AUC). RESULTS XGBoost had PPVs over 90 % in all masking settings, and InceptionVasGloMyotides had most of the PPVs over 90 %, but not as consistently. When the prevalence of the diseases was considered XGBoost achieved highest value of 8.8 % in binary classification with 30 days masking and InceptionVasGloMyotides achieved the best value of 6 % in the binary classification as well, but with 2160 days and 4320 days masking. ACC were varying between 89 % and 98 % with XGBoost and InceptionVasGloMyotides having variation between 79 % and 94 %. AUC on the other hand varied between 72.6 % and 94.5 % with InceptionVasGloMyotides and for XGBoost it varied between 69.9 % and 96.4 %. CONCLUSIONS XGBoost and InceptionVasGloMyotides could successfully predict rare diseases for patients at least 30 days prior to initial rare disease diagnose. In addition, we managed to build performative custom deep learning model.
Collapse
Affiliation(s)
- Rasmus Ryyppö
- Faculty of Information Technology and Communication Sciences, Tampere University, Kanslerinrinne 1, Tampere 33014, Finland; Tietoevry Ltd, Espoo, Finland.
| | | | - Henry Joutsijoki
- Faculty of Information Technology and Communication Sciences, Tampere University, Kanslerinrinne 1, Tampere 33014, Finland
| | - Martti Juhola
- Faculty of Information Technology and Communication Sciences, Tampere University, Kanslerinrinne 1, Tampere 33014, Finland
| | - Mikko R J Seppänen
- Rare Disease Center and Pediatric Research Center, New Children's Hospital, University of Helsinki and HUS Helsinki University Hospital, Helsinki, Finland
| |
Collapse
|
12
|
Jain R, Yoo TK, Ryu IH, Song J, Kolte N, Nariani A. Deep Transfer Learning for Ethnically Distinct Populations: Prediction of Refractive Error Using Optical Coherence Tomography. Ophthalmol Ther 2024; 13:305-319. [PMID: 37955835 PMCID: PMC10776546 DOI: 10.1007/s40123-023-00842-6] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2023] [Accepted: 10/20/2023] [Indexed: 11/14/2023] Open
Abstract
INTRODUCTION The mismatch between training and testing data distribution causes significant degradation in the deep learning model performance in multi-ethnic scenarios. To reduce the performance differences between ethnic groups and image domains, we built a deep transfer learning model with adaptation training to predict uncorrected refractive errors using posterior segment optical coherence tomography (OCT) images of the macula and optic nerve. METHODS Observational, cross-sectional, multicenter study design. We pre-trained a deep learning model on OCT images from the B&VIIT Eye Center (Seoul, South Korea) (N = 2602 eyes of 1301 patients). OCT images from Poona Eye Care (Pune, India) were chronologically sorted into adaptation training data (N = 60 eyes of 30 patients) for transfer learning and test data (N = 142 eyes of 71 patients) for validation. Deep learning models were trained to predict spherical equivalent (SE) and mean keratometry (K) values via transfer learning for domain adaptation. RESULTS Both adaptation models for SE and K were significantly better than those without adaptation (P < 0.001). In myopia/hyperopia classification, the model trained on circular optic disc OCT images yielded the best performance (accuracy = 74.7%). It also performed best in estimating SE with the lowest mean absolute error (MAE) of 1.58 D. For classifying the degree of corneal curvature, the optic nerve vertical algorithm performed best (accuracy = 65.7%). The optic nerve horizontal model achieved the lowest MAE (1.85 D) when predicting the K value. Saliency maps frequently highlighted the retinal nerve fiber layers. CONCLUSIONS Adaptation training via transfer learning is an effective technique for estimating refractive errors and K values using macular and optic nerve OCT images from ethnically heterogeneous populations. Further studies with larger sample sizes and various data sources are needed to confirm the feasibility of the proposed algorithm.
Collapse
Affiliation(s)
- Rishabh Jain
- Department of Biomedical Engineering, Duke University, Durham, NC, USA
| | - Tae Keun Yoo
- Department of Refractive Surgery, B&VIIT Eye Center, B2 GT Tower, 1317-23 Seocho-Dong, Seocho-Gu, Seoul, Republic of Korea.
- Research and Development Department, VISUWORKS, Seoul, South Korea.
| | - Ik Hee Ryu
- Department of Refractive Surgery, B&VIIT Eye Center, B2 GT Tower, 1317-23 Seocho-Dong, Seocho-Gu, Seoul, Republic of Korea
- Research and Development Department, VISUWORKS, Seoul, South Korea
| | - Joanna Song
- Research and Development Department, VISUWORKS, Seoul, South Korea
| | | | - Ashiyana Nariani
- Department of Ophthalmology, King Edward Memorial Hospital and Seth Gordhandas Sunderdas Medical College, Mumbai, Maharashtra, India
| |
Collapse
|
13
|
Martin R, Segars P, Samei E, Miró J, Duong L. Unsupervised synthesis of realistic coronary artery X-ray angiogram. Int J Comput Assist Radiol Surg 2023; 18:2329-2338. [PMID: 37336801 PMCID: PMC10786317 DOI: 10.1007/s11548-023-02982-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2022] [Accepted: 06/01/2023] [Indexed: 06/21/2023]
Abstract
PURPOSE Medical image analysis suffers from a sparsity of annotated data necessary in learning-based models. Cardiorespiratory simulators have been developed to counter the lack of data. However, the resulting data often lack realism. Hence, the proposed method aims to synthesize realistic and fully customizable angiograms of coronary arteries for the training of learning-based biomedical tasks, for cardiologists performing interventions, and for cardiologist trainees. METHODS 3D models of coronary arteries are generated with a fully customizable realistic cardiorespiratory simulator. The transfer of X-ray angiography style to simulator-generated images is performed using a new vessel-specific adaptation of the CycleGAN model. The CycleGAN model is paired with a vesselness-based loss function that is designed as a vessel-specific structural integrity constraint. RESULTS Validation is performed both on the style and on the preservation of the shape of the arteries of the images. The results show a PSNR of 14.125, an SSIM of 0.898, and an overlapping of 89.5% using the Dice coefficient. CONCLUSION We proposed a novel fluoroscopy-based style transfer method for the enhancement of the realism of simulated coronary artery angiograms. The results show that the proposed model is capable of accurately transferring the style of X-ray angiograms to the simulations while keeping the integrity of the structures of interest (i.e., the topology of the coronary arteries).
Collapse
Affiliation(s)
- Rémi Martin
- Department of Software and Information Technology Engineering, École de Technologie Supérieure, 1100 Notre-Dame, Montréal, QC, H3C 1K3, Canada.
| | - Paul Segars
- Carl E. Ravin Advanced Imaging Laboratories, Department of Radiology, Duke University Medical Center, 2424 Erwin Road, Durham, NC, 27705, USA
| | - Ehsan Samei
- Carl E. Ravin Advanced Imaging Laboratories, Department of Radiology, Duke University Medical Center, 2424 Erwin Road, Durham, NC, 27705, USA
| | - Joaquim Miró
- Department of Pediatrics, CHU Sainte-Justine, 3175 Chem. de la Côte-Sainte-Catherine, Montréal, QC, H3T 1C5, Canada
| | - Luc Duong
- Department of Software and Information Technology Engineering, École de Technologie Supérieure, 1100 Notre-Dame, Montréal, QC, H3C 1K3, Canada
| |
Collapse
|
14
|
Agharezaei Z, Firouzi R, Hassanzadeh S, Zarei-Ghanavati S, Bahaadinbeigy K, Golabpour A, Akbarzadeh R, Agharezaei L, Bakhshali MA, Sedaghat MR, Eslami S. Computer-aided diagnosis of keratoconus through VAE-augmented images using deep learning. Sci Rep 2023; 13:20586. [PMID: 37996439 PMCID: PMC10667539 DOI: 10.1038/s41598-023-46903-5] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/17/2023] [Accepted: 11/07/2023] [Indexed: 11/25/2023] Open
Abstract
Detecting clinical keratoconus (KCN) poses a challenging and time-consuming task. During the diagnostic process, ophthalmologists are required to review demographic and clinical ophthalmic examinations in order to make an accurate diagnosis. This study aims to develop and evaluate the accuracy of deep convolutional neural network (CNN) models for the detection of keratoconus (KCN) using corneal topographic maps. We retrospectively collected 1758 corneal images (978 normal and 780 keratoconus) from 1010 subjects of the KCN group with clinically evident keratoconus and the normal group with regular astigmatism. To expand the dataset, we developed a model using Variational Auto Encoder (VAE) to generate and augment images, resulting in a dataset of 4000 samples. Four deep learning models were used to extract and identify deep corneal features of original and synthesized images. We demonstrated that the utilization of synthesized images during training process increased classification performance. The overall average accuracy of the deep learning models ranged from 99% for VGG16 to 95% for EfficientNet-B0. All CNN models exhibited sensitivity and specificity above 0.94, with the VGG16 model achieving an AUC of 0.99. The customized CNN model achieved satisfactory results with an accuracy and AUC of 0.97 at a much faster processing speed compared to other models. In conclusion, the DL models showed high accuracy in screening for keratoconus based on corneal topography images. This is a development toward the potential clinical implementation of a more enhanced computer-aided diagnosis (CAD) system for KCN detection, which would aid ophthalmologists in validating the clinical decision and carrying out prompt and precise KCN treatment.
Collapse
Affiliation(s)
- Zhila Agharezaei
- Pharmaceutical Research Center, Pharmaceutical Technology Institute, Mashhad University of Medical Sciences, Mashhad, Iran
- Department of Medical Informatics, Faculty of Medicine, Mashhad University of Medical Sciences, Mashhad, Iran
- Medical Informatics Research Center, Institute for Future Studies in Health, Kerman University of Medical Sciences, Kerman, Iran
| | - Reza Firouzi
- Department of Computer Engineering, Ferdowsi University of Mashhad, Mashhad, Iran
| | - Samira Hassanzadeh
- School of Paramedical Sciences and Rehabilitation, Mashhad University of Medical Sciences, Mashhad, Iran
| | | | - Kambiz Bahaadinbeigy
- Medical Informatics Research Center, Institute for Future Studies in Health, Kerman University of Medical Sciences, Kerman, Iran
| | - Amin Golabpour
- School of Medicine, Shahroud University of Medical Sciences, Shahroud, Iran
| | - Reyhaneh Akbarzadeh
- Department of Optometry, School of Paramedical Sciences, Mashhad University of Medical Sciences, Mashhad, Iran
| | - Laleh Agharezaei
- Modeling in Health Research Center, Institute for Future Studies in Health, Kerman University of Medical Sciences, Kerman, Iran
| | - Mohamad Amin Bakhshali
- Department of Medical Informatics, Faculty of Medicine, Mashhad University of Medical Sciences, Mashhad, Iran
| | | | - Saeid Eslami
- Pharmaceutical Research Center, Pharmaceutical Technology Institute, Mashhad University of Medical Sciences, Mashhad, Iran.
- Department of Medical Informatics, Faculty of Medicine, Mashhad University of Medical Sciences, Mashhad, Iran.
| |
Collapse
|
15
|
Cabrera-Aguas M, Watson SL. Updates in Diagnostic Imaging for Infectious Keratitis: A Review. Diagnostics (Basel) 2023; 13:3358. [PMID: 37958254 PMCID: PMC10647798 DOI: 10.3390/diagnostics13213358] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/16/2023] [Revised: 10/15/2023] [Accepted: 10/16/2023] [Indexed: 11/15/2023] Open
Abstract
Infectious keratitis (IK) is among the top five leading causes of blindness globally. Early diagnosis is needed to guide appropriate therapy to avoid complications such as vision impairment and blindness. Slit lamp microscopy and culture of corneal scrapes are key to diagnosing IK. Slit lamp photography was transformed when digital cameras and smartphones were invented. The digital camera or smartphone camera sensor's resolution, the resolution of the slit lamp and the focal length of the smartphone camera system are key to a high-quality slit lamp image. Alternative diagnostic tools include imaging, such as optical coherence tomography (OCT) and in vivo confocal microscopy (IVCM). OCT's advantage is its ability to accurately determine the depth and extent of the corneal ulceration, infiltrates and haze, therefore characterizing the severity and progression of the infection. However, OCT is not a preferred choice in the diagnostic tool package for infectious keratitis. Rather, IVCM is a great aid in the diagnosis of fungal and Acanthamoeba keratitis with overall sensitivities of 66-74% and 80-100% and specificity of 78-100% and 84-100%, respectively. Recently, deep learning (DL) models have been shown to be promising aids for the diagnosis of IK via image recognition. Most of the studies that have developed DL models to diagnose the different types of IK have utilised slit lamp photographs. Some studies have used extremely efficient single convolutional neural network algorithms to train their models, and others used ensemble approaches with variable results. Limitations of DL models include the need for large image datasets to train the models, the difficulty in finding special features of the different types of IK, the imbalance of training models, the lack of image protocols and misclassification bias, which need to be overcome to apply these models into real-world settings. Newer artificial intelligence technology that generates synthetic data, such as generative adversarial networks, may assist in overcoming some of these limitations of CNN models.
Collapse
Affiliation(s)
- Maria Cabrera-Aguas
- Save Sight Institute, Discipline of Ophthalmology, Faculty of Medicine and Health, The University of Sydney, Sydney, NSW 2000, Australia;
- Sydney Eye Hospital, Sydney, NSW 2000, Australia
| | - Stephanie L Watson
- Save Sight Institute, Discipline of Ophthalmology, Faculty of Medicine and Health, The University of Sydney, Sydney, NSW 2000, Australia;
- Sydney Eye Hospital, Sydney, NSW 2000, Australia
| |
Collapse
|
16
|
She Z, Marzullo A, Destito M, Spadea MF, Leone R, Anzalone N, Steffanoni S, Erbella F, Ferreri AJM, Ferrigno G, Calimeri T, De Momi E. Deep learning-based overall survival prediction model in patients with rare cancer: a case study for primary central nervous system lymphoma. Int J Comput Assist Radiol Surg 2023; 18:1849-1856. [PMID: 37083973 PMCID: PMC10497660 DOI: 10.1007/s11548-023-02886-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/09/2022] [Accepted: 03/27/2023] [Indexed: 04/22/2023]
Abstract
PURPOSE Primary central nervous system lymphoma (PCNSL) is a rare, aggressive form of extranodal non-Hodgkin lymphoma. To predict the overall survival (OS) in advance is of utmost importance as it has the potential to aid clinical decision-making. Though radiomics-based machine learning (ML) has demonstrated the promising performance in PCNSL, it demands large amounts of manual feature extraction efforts from magnetic resonance images beforehand. deep learning (DL) overcomes this limitation. METHODS In this paper, we tailored the 3D ResNet to predict the OS of patients with PCNSL. To overcome the limitation of data sparsity, we introduced data augmentation and transfer learning, and we evaluated the results using r stratified k-fold cross-validation. To explain the results of our model, gradient-weighted class activation mapping was applied. RESULTS We obtained the best performance (the standard error) on post-contrast T1-weighted (T1Gd)-area under curve [Formula: see text], accuracy [Formula: see text], precision [Formula: see text], recall [Formula: see text] and F1-score [Formula: see text], while compared with ML-based models on clinical data and radiomics data, respectively, further confirming the stability of our model. Also, we observed that PCNSL is a whole-brain disease and in the cases where the OS is less than 1 year, it is more difficult to distinguish the tumor boundary from the normal part of the brain, which is consistent with the clinical outcome. CONCLUSIONS All these findings indicate that T1Gd can improve prognosis predictions of patients with PCNSL. To the best of our knowledge, this is the first time to use DL to explain model patterns in OS classification of patients with PCNSL. Future work would involve collecting more data of patients with PCNSL, or additional retrospective studies on different patient populations with rare diseases, to further promote the clinical role of our model.
Collapse
Affiliation(s)
- Ziyu She
- Department of Electronics, Information and Bioengineering, Politecnico di Milano, Milan, Italy.
| | - Aldo Marzullo
- Department of Mathematics and Computer Science, University of Calabria, Rende, Italy
| | - Michela Destito
- Department of Experimental and Clinical Medicine, University of Catanzaro, Catanzaro, Italy
| | - Maria Francesca Spadea
- Department of Experimental and Clinical Medicine, University of Catanzaro, Catanzaro, Italy
| | - Riccardo Leone
- Neuroradiology Unit, IRCCS San Raffaele Scientific Institute, Milan, Italy
| | - Nicoletta Anzalone
- Neuroradiology Unit, IRCCS San Raffaele Scientific Institute, Milan, Italy
| | - Sara Steffanoni
- Lymphoma Unit, IRCCS San Raffaele Scientific Institute, Milan, Italy
| | - Federico Erbella
- Lymphoma Unit, IRCCS San Raffaele Scientific Institute, Milan, Italy
| | | | - Giancarlo Ferrigno
- Department of Electronics, Information and Bioengineering, Politecnico di Milano, Milan, Italy
| | - Teresa Calimeri
- Lymphoma Unit, IRCCS San Raffaele Scientific Institute, Milan, Italy
| | - Elena De Momi
- Department of Electronics, Information and Bioengineering, Politecnico di Milano, Milan, Italy
| |
Collapse
|
17
|
Paladugu PS, Ong J, Nelson N, Kamran SA, Waisberg E, Zaman N, Kumar R, Dias RD, Lee AG, Tavakkoli A. Generative Adversarial Networks in Medicine: Important Considerations for this Emerging Innovation in Artificial Intelligence. Ann Biomed Eng 2023; 51:2130-2142. [PMID: 37488468 DOI: 10.1007/s10439-023-03304-z] [Citation(s) in RCA: 24] [Impact Index Per Article: 12.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/01/2023] [Accepted: 07/03/2023] [Indexed: 07/26/2023]
Abstract
The advent of artificial intelligence (AI) and machine learning (ML) has revolutionized the field of medicine. Although highly effective, the rapid expansion of this technology has created some anticipated and unanticipated bioethical considerations. With these powerful applications, there is a necessity for framework regulations to ensure equitable and safe deployment of technology. Generative Adversarial Networks (GANs) are emerging ML techniques that have immense applications in medical imaging due to their ability to produce synthetic medical images and aid in medical AI training. Producing accurate synthetic images with GANs can address current limitations in AI development for medical imaging and overcome current dataset type and size constraints. Offsetting these constraints can dramatically improve the development and implementation of AI medical imaging and restructure the practice of medicine. As observed with its other AI predecessors, considerations must be taken into place to help regulate its development for clinical use. In this paper, we discuss the legal, ethical, and technical challenges for future safe integration of this technology in the healthcare sector.
Collapse
Affiliation(s)
- Phani Srivatsav Paladugu
- Brigham and Women's Hospital, Harvard Medical School, Boston, MA, USA
- Sidney Kimmel Medical College, Thomas Jefferson University, Philadelphia, PA, USA
| | - Joshua Ong
- Michigan Medicine, University of Michigan, Ann Arbor, MI, USA
| | - Nicolas Nelson
- Sidney Kimmel Medical College, Thomas Jefferson University, Philadelphia, PA, USA
| | - Sharif Amit Kamran
- Human-Machine Perception Laboratory, Department of Computer Science and Engineering, University of Nevada, Reno, Reno, NV, USA
| | - Ethan Waisberg
- University College Dublin School of Medicine, Belfield, Dublin, Ireland
| | - Nasif Zaman
- Human-Machine Perception Laboratory, Department of Computer Science and Engineering, University of Nevada, Reno, Reno, NV, USA
| | | | - Roger Daglius Dias
- Department of Emergency Medicine, Harvard Medical School, Boston, MA, USA
- STRATUS Center for Medical Simulation, Brigham and Women's Hospital, Boston, MA, USA
| | - Andrew Go Lee
- Center for Space Medicine, Baylor College of Medicine, Houston, TX, USA
- Department of Ophthalmology, Blanton Eye Institute, Houston Methodist Hospital, Houston, TX, USA
- The Houston Methodist Research Institute, Houston Methodist Hospital, Houston, TX, USA
- Departments of Ophthalmology, Neurology, and Neurosurgery, Weill Cornell Medicine, New York, NY, USA
- Department of Ophthalmology, University of Texas Medical Branch, Galveston, TX, USA
- University of Texas MD Anderson Cancer Center, Houston, TX, USA
- Texas A&M College of Medicine, Bryan, TX, USA
- Department of Ophthalmology, The University of Iowa Hospitals and Clinics, Iowa City, IA, USA
| | - Alireza Tavakkoli
- Human-Machine Perception Laboratory, Department of Computer Science and Engineering, University of Nevada, Reno, Reno, NV, USA.
| |
Collapse
|
18
|
Koseoglu ND, Corrêa ZM, Liu TA. Artificial intelligence for ocular oncology. Curr Opin Ophthalmol 2023; 34:437-440. [PMID: 37326226 PMCID: PMC10399931 DOI: 10.1097/icu.0000000000000982] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/17/2023]
Abstract
PURPOSE OF REVIEW The aim of this article is to provide an update on the latest applications of deep learning (DL) and classical machine learning (ML) techniques to the detection and prognostication of intraocular and ocular surface malignancies. RECENT FINDINGS Most recent studies focused on using DL and classical ML techniques for prognostication purposes in patients with uveal melanoma (UM). SUMMARY DL has emerged as the leading ML technique for prognostication in ocular oncological conditions, particularly in UM. However, the application of DL may be limited by the relatively rarity of these conditions.
Collapse
Affiliation(s)
| | - Zélia Maria Corrêa
- Ocular Oncology, Bascom Palmer Eye Institute
- Sylvester Comprehensive Cancer Center, University of Miami Miller School of Medicine, Miami, Florida, USA
| | - T.Y. Alvin Liu
- Wilmer Eye Institute, Johns Hopkins University, Baltimore, Maryland
| |
Collapse
|
19
|
Nouri H, Nasri R, Abtahi SH. Addressing inter-device variations in optical coherence tomography angiography: will image-to-image translation systems help? Int J Retina Vitreous 2023; 9:51. [PMID: 37644613 PMCID: PMC10466880 DOI: 10.1186/s40942-023-00491-8] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/08/2023] [Accepted: 08/17/2023] [Indexed: 08/31/2023] Open
Abstract
BACKGROUND Optical coherence tomography angiography (OCTA) is an innovative technology providing visual and quantitative data on retinal microvasculature in a non-invasive manner. MAIN BODY Due to variations in the technical specifications of different OCTA devices, there are significant inter-device differences in OCTA data, which can limit their comparability and generalizability. These variations can also result in a domain shift problem that may interfere with applicability of machine learning models on data obtained from different OCTA machines. One possible approach to address this issue may be unsupervised deep image-to-image translation leveraging systems such as Cycle-Consistent Generative Adversarial Networks (Cycle-GANs) and Denoising Diffusion Probabilistic Models (DDPMs). Through training on unpaired images from different device domains, Cycle-GANs and DDPMs may enable cross-domain translation of images. They have been successfully applied in various medical imaging tasks, including segmentation, denoising, and cross-modality image-to-image translation. In this commentary, we briefly describe how Cycle-GANs and DDPMs operate, and review the recent experiments with these models on medical and ocular imaging data. We then discuss the benefits of applying such techniques for inter-device translation of OCTA data and the potential challenges ahead. CONCLUSION Retinal imaging technologies and deep learning-based domain adaptation techniques are rapidly evolving. We suggest exploring the potential of image-to-image translation methods in improving the comparability of OCTA data from different centers or devices. This may facilitate more efficient analysis of heterogeneous data and broader applicability of machine learning models trained on limited datasets in this field.
Collapse
Affiliation(s)
- Hosein Nouri
- Ophthalmic Research Center, Research Institute for Ophthalmology and Vision Science, Shahid Beheshti University of Medical Sciences, Tehran, Iran.
- School of Medicine, Isfahan University of Medical Sciences, Isfahan, Iran.
| | - Reza Nasri
- School of Engineering, University of Isfahan, Isfahan, Iran
| | - Seyed-Hossein Abtahi
- Ophthalmic Research Center, Research Institute for Ophthalmology and Vision Science, Shahid Beheshti University of Medical Sciences, Tehran, Iran.
- Department of Ophthalmology, Torfe Medical Center, Shahid Beheshti University of Medical Sciences, Tehran, Iran.
| |
Collapse
|
20
|
Zhang S, He X, Xia X, Xiao P, Wu Q, Zheng F, Lu Q. Machine-Learning-Enabled Framework in Engineering Plastics Discovery: A Case Study of Designing Polyimides with Desired Glass-Transition Temperature. ACS APPLIED MATERIALS & INTERFACES 2023; 15:37893-37902. [PMID: 37490394 DOI: 10.1021/acsami.3c05376] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/27/2023]
Abstract
Great and continuous efforts have been made to discover high-performance engineering plastics with specific properties to replace traditional engineering materials in many fields. The utilization of machine learning (ML) has brought more opportunities for the discovery of high-performing engineering plastics. However, hindered by either the relatively small database or a lack of accurate structure descriptors with clear physical and chemical meanings relating to polymer properties, the current ML studies show some flaws in the accuracy and efficiency in polymer development. Herein, we collected a dataset of 878 polyimides (PI), one of the best engineering plastics, with experimentally measured glass-transition temperature (Tg) values, and developed a rapid and accurate ML approach to design PI candidates with the desired Tg value. After the conversion from PI structures into "mechanically identifiable" SMILES (Simplified molecular input line entry system) language, the eight most critical descriptors were ultimately obtained by multiple analysis methods. The physiochemical meaning of the key descriptors was further analyzed carefully to translate the implicit "machine language" to chemical knowledge. The artificial neural network (ANN)-based model gave the most accurate results with a root-mean-square error of ∼11 K among the studied ML methods. More importantly, three potential PI candidates with desired Tg (DPIs) were designed according to the chemical insight of the key descriptors, which were then verified by experiments. The experimental and predicted Tg values of DPIs have an acceptable average deviation of ca. 3.66%. This accuracy has reached the level of the traditional molecular simulation, but the time consumption and hold-up computing resource are tremendously reduced. Furthermore, the current ML approach could offer a scalable and adaptable framework in future engineer plastics innovation.
Collapse
Affiliation(s)
- Songyang Zhang
- School of Chemical Science and Engineering, Tongji University, Shanghai 200092, China
| | - Xiaojie He
- School of Chemical Science and Engineering, Tongji University, Shanghai 200092, China
| | - Xuejian Xia
- School of Chemical Science and Engineering, Tongji University, Shanghai 200092, China
| | - Peng Xiao
- School of Chemical Science and Engineering, Tongji University, Shanghai 200092, China
| | - Qi Wu
- Shanghai Key Lab of Electrical & Thermal Aging, School of Chemistry and Chemical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
| | - Feng Zheng
- School of Chemical Science and Engineering, Tongji University, Shanghai 200092, China
| | - Qinghua Lu
- Shanghai Key Lab of Electrical & Thermal Aging, School of Chemistry and Chemical Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
| |
Collapse
|
21
|
Zhang Z, Wang Y, Zhang H, Samusak A, Rao H, Xiao C, Abula M, Cao Q, Dai Q. Artificial intelligence-assisted diagnosis of ocular surface diseases. Front Cell Dev Biol 2023; 11:1133680. [PMID: 36875760 PMCID: PMC9981656 DOI: 10.3389/fcell.2023.1133680] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/29/2022] [Accepted: 02/08/2023] [Indexed: 02/19/2023] Open
Abstract
With the rapid development of computer technology, the application of artificial intelligence (AI) in ophthalmology research has gained prominence in modern medicine. Artificial intelligence-related research in ophthalmology previously focused on the screening and diagnosis of fundus diseases, particularly diabetic retinopathy, age-related macular degeneration, and glaucoma. Since fundus images are relatively fixed, their standards are easy to unify. Artificial intelligence research related to ocular surface diseases has also increased. The main issue with research on ocular surface diseases is that the images involved are complex, with many modalities. Therefore, this review aims to summarize current artificial intelligence research and technologies used to diagnose ocular surface diseases such as pterygium, keratoconus, infectious keratitis, and dry eye to identify mature artificial intelligence models that are suitable for research of ocular surface diseases and potential algorithms that may be used in the future.
Collapse
Affiliation(s)
- Zuhui Zhang
- The First People's Hospital of Aksu District in Xinjiang, Aksu City, China.,National Clinical Research Center for Ocular Diseases, Eye Hospital, Wenzhou Medical University, Wenzhou, China
| | - Ying Wang
- The First People's Hospital of Aksu District in Xinjiang, Aksu City, China
| | - Hongzhen Zhang
- The First People's Hospital of Aksu District in Xinjiang, Aksu City, China
| | - Arzigul Samusak
- The First People's Hospital of Aksu District in Xinjiang, Aksu City, China
| | - Huimin Rao
- The First People's Hospital of Aksu District in Xinjiang, Aksu City, China
| | - Chun Xiao
- The First People's Hospital of Aksu District in Xinjiang, Aksu City, China
| | - Muhetaer Abula
- The First People's Hospital of Aksu District in Xinjiang, Aksu City, China
| | - Qixin Cao
- Huzhou Traditional Chinese Medicine Hospital Affiliated to Zhejiang University of Traditional Chinese Medicine, Huzhou, China
| | - Qi Dai
- The First People's Hospital of Aksu District in Xinjiang, Aksu City, China.,National Clinical Research Center for Ocular Diseases, Eye Hospital, Wenzhou Medical University, Wenzhou, China
| |
Collapse
|
22
|
Zhang Z, Cheng N, Liu Y, Song J, Liu X, Zhang S, Zhang G. Prediction of corneal astigmatism based on corneal tomography after femtosecond laser arcuate keratotomy using a pix2pix conditional generative adversarial network. Front Public Health 2022; 10:1012929. [PMID: 36187623 PMCID: PMC9523441 DOI: 10.3389/fpubh.2022.1012929] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/06/2022] [Accepted: 08/29/2022] [Indexed: 01/27/2023] Open
Abstract
Purpose This study aimed to develop a deep learning model to generate a postoperative corneal axial curvature map of femtosecond laser arcuate keratotomy (FLAK) based on corneal tomography using a pix2pix conditional generative adversarial network (pix2pix cGAN) for surgical planning. Methods A total of 451 eyes of 318 nonconsecutive patients were subjected to FLAK for corneal astigmatism correction during cataract surgery. Paired or single anterior penetrating FLAKs were performed at an 8.0-mm optical zone with a depth of 90% using a femtosecond laser (LenSx laser, Alcon Laboratories, Inc.). Corneal tomography images were acquired from Oculus Pentacam HR (Optikgeräte GmbH, Wetzlar, Germany) before and 3 months after the surgery. The raw data required for analysis consisted of the anterior corneal curvature for a range of ± 3.5 mm around the corneal apex in 0.1-mm steps, which the pseudo-color corneal curvature map synthesized was based on. The deep learning model used was a pix2pix conditional generative adversarial network. The prediction accuracy of synthetic postoperative corneal astigmatism in zones of different diameters centered on the corneal apex was assessed using vector analysis. The synthetic postoperative corneal axial curvature maps were compared with the real postoperative corneal axial curvature maps using the structural similarity index measure (SSIM) and peak signal-to-noise ratio (PSNR). Results A total of 386 pairs of preoperative and postoperative corneal tomography data were included in the training set, whereas 65 preoperative data were retrospectively included in the test set. The correlation coefficient between synthetic and real postoperative astigmatism (difference vector) in the 3-mm zone was 0.89, and that between surgically induced astigmatism (SIA) was 0.93. The mean absolute errors of SIA for real and synthetic postoperative corneal axial curvature maps in the 1-, 3-, and 5-mm zone were 0.20 ± 0.25, 0.12 ± 0.17, and 0.09 ± 0.13 diopters, respectively. The average SSIM and PSNR of the 3-mm zone were 0.86 ± 0.04 and 18.24 ± 5.78, respectively. Conclusion Our results showed that the application of pix2pix cGAN can synthesize plausible postoperative corneal tomography for FLAK, showing the possibility of using GAN to predict corneal tomography, with the potential of applying artificial intelligence to construct surgical planning models.
Collapse
Affiliation(s)
- Zhe Zhang
- Shenzhen Eye Hospital, Jinan University, Shenzhen Eye Institute, Shenzhen, China,Department of Cataract, Shanxi Eye Hospital, Taiyuan, China,First Hospital of Shanxi Medical University, Taiyuan, China
| | - Nan Cheng
- College of Biomedical Engineering, Taiyuan University of Technology, Taiyuan, China
| | - Yunfang Liu
- Department of Ophthalmology, First Affiliated Hospital of Huzhou University, Huzhou, China
| | - Junyang Song
- Department of Ophthalmology, First Affiliated Hospital of Huzhou University, Huzhou, China,*Correspondence: Junyang Song
| | - Xinhua Liu
- Shenzhen Eye Hospital, Jinan University, Shenzhen Eye Institute, Shenzhen, China
| | - Suhua Zhang
- Department of Cataract, Shanxi Eye Hospital, Taiyuan, China,Taiyuan Central Hospital of Shanxi Medical University, Taiyuan, China,Suhua Zhang
| | - Guanghua Zhang
- Department of Intelligence and Automation, Taiyuan University, Taiyuan, China,Graphics and Imaging Laboratory, University of Girona, Girona, Spain,Guanghua Zhang
| |
Collapse
|
23
|
Yoo TK, Ryu IH, Kim JK, Lee IS, Kim HK. A deep learning approach for detection of shallow anterior chamber depth based on the hidden features of fundus photographs. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 219:106735. [PMID: 35305492 DOI: 10.1016/j.cmpb.2022.106735] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/14/2021] [Revised: 02/15/2022] [Accepted: 03/04/2022] [Indexed: 06/14/2023]
Abstract
BACKGROUND AND OBJECTIVES Patients with angle-closure glaucoma (ACG) are asymptomatic until they experience a painful attack. Shallow anterior chamber depth (ACD) is considered a significant risk factor for ACG. We propose a deep learning approach to detect shallow ACD using fundus photographs and to identify the hidden features of shallow ACD. METHODS This retrospective study assigned healthy subjects to the training (n = 1188 eyes) and test (n = 594) datasets (prospective validation design). We used a deep learning approach to estimate ACD and build a classification model to identify eyes with a shallow ACD. The proposed method, including subtraction of the input and output images of CycleGAN and a thresholding algorithm, was adopted to visualize the characteristic features of fundus photographs with a shallow ACD. RESULTS The deep learning model integrating fundus photographs and clinical variables achieved areas under the receiver operating characteristic curve of 0.978 (95% confidence interval [CI], 0.963-0.988) for an ACD ≤ 2.60 mm and 0.895 (95% CI, 0.868-0.919) for an ACD ≤ 2.80 mm, and outperformed the regression model using only clinical variables. However, the difference between shallow and deep ACD classes on fundus photographs was difficult to be detected with the naked eye. We were unable to identify the features of shallow ACD using the Grad-CAM. The CycleGAN-based feature images showed that area around the macula and optic disk significantly contributed to the classification of fundus photographs with a shallow ACD. CONCLUSIONS We demonstrated the feasibility of a novel deep learning model to detect a shallow ACD as a screening tool for ACG using fundus photographs. The CycleGAN-based feature map showed the hidden characteristic features of shallow ACD that were previously undetectable by conventional techniques and ophthalmologists. This framework will facilitate the early detection of shallow ACD to prevent overlooking the risks associated with ACG.
Collapse
Affiliation(s)
- Tae Keun Yoo
- B&VIIT Eye Center, Seoul, South Korea; Department of Ophthalmology, Aerospace Medical Center, Republic of Korea Air Force, Cheongju, South Korea.
| | - Ik Hee Ryu
- B&VIIT Eye Center, Seoul, South Korea; VISUWORKS, Seoul, South Korea
| | - Jin Kuk Kim
- B&VIIT Eye Center, Seoul, South Korea; VISUWORKS, Seoul, South Korea
| | | | - Hong Kyu Kim
- Department of Ophthalmology, Dankook University Hospital, Dankook University College of Medicine, Cheonan, South Korea
| |
Collapse
|
24
|
Han Z, Huang H, Fan Q, Li Y, Li Y, Chen X. SMD-YOLO: An efficient and lightweight detection method for mask wearing status during the COVID-19 pandemic. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 221:106888. [PMID: 35598435 PMCID: PMC9098810 DOI: 10.1016/j.cmpb.2022.106888] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 02/13/2022] [Revised: 04/30/2022] [Accepted: 05/11/2022] [Indexed: 06/15/2023]
Abstract
BACKGROUND AND OBJECTIVE At present, the COVID-19 epidemic is still spreading worldwide and wearing a mask in public areas is an effective way to prevent the spread of the respiratory virus. Although there are many deep learning methods used for detecting the face masks, there are few lightweight detectors having a good effect on small or medium-size face masks detection in the complicated environments. METHODS In this work we propose an efficient and lightweight detection method based on YOLOv4-tiny, and a face mask detection and monitoring system for mask wearing status. Two feasible improvement strategies, network structure optimization and K-means++ clustering algorithm, are utilized for improving the detection accuracy on the premise of ensuring the real-time face masks recognition. Particularly, the improved residual module and cross fusion module are designed to aim at extracting the features of small or medium-size targets effectively. Moreover, the enhanced dual attention mechanism and the improved spatial pyramid pooling module are employed for merging sufficiently the deep and shallow semantic information and expanding the receptive field. Afterwards, the detection accuracy is compensated through the combination of activation functions. Finally, the depthwise separable convolution module is used to reduce the quantity of parameters and improve the detection efficiency. Our proposed detector is evaluated on a public face mask dataset, and an ablation experiment is also provided to verify the effectiveness of our proposed model, which is compared with the state-of-the-art (SOTA) models as well. RESULTS Our proposed detector increases the AP (average precision) values in each category of the public face mask dataset compared with the original YOLOv4-tiny. The mAP (mean average precision) is improved by 4.56% and the speed reaches 92.81 FPS. Meanwhile, the quantity of parameters and the FLOPs (floating-point operations) are reduced by 1/3, 16.48%, respectively. CONCLUSIONS The proposed detector achieves better overall detection performance compared with other SOTA detectors for real-time mask detection, demonstrated the superiority with both theoretical value and practical significance. The developed system also brings greater flexibility to the application of face mask detection in hospitals, campuses, communities, etc.
Collapse
Affiliation(s)
- Zhenggong Han
- Key Laboratory of Advanced Manufacturing Technology, Ministry of Education, Guizhou University, Guiyang 550025, Guizhou, China
| | - Haisong Huang
- Key Laboratory of Advanced Manufacturing Technology, Ministry of Education, Guizhou University, Guiyang 550025, Guizhou, China; Chongqing Vocational and Technical University of Mechatronics, Chongqing 400036, China
| | - Qingsong Fan
- Key Laboratory of Advanced Manufacturing Technology, Ministry of Education, Guizhou University, Guiyang 550025, Guizhou, China
| | - Yiting Li
- College of Big Data Statistics, GuiZhou University of Finance and Economics, Guiyang 550025, Guizhou, China
| | - Yuqin Li
- Stomotological Hospital of Guizhou Medical University, Guiyang 550004, Guizhou, China
| | - Xingran Chen
- Key Laboratory of Advanced Manufacturing Technology, Ministry of Education, Guizhou University, Guiyang 550025, Guizhou, China
| |
Collapse
|
25
|
You A, Kim JK, Ryu IH, Yoo TK. Application of generative adversarial networks (GAN) for ophthalmology image domains: a survey. EYE AND VISION (LONDON, ENGLAND) 2022; 9:6. [PMID: 35109930 PMCID: PMC8808986 DOI: 10.1186/s40662-022-00277-3] [Citation(s) in RCA: 57] [Impact Index Per Article: 19.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/15/2021] [Accepted: 01/11/2022] [Indexed: 12/12/2022]
Abstract
BACKGROUND Recent advances in deep learning techniques have led to improved diagnostic abilities in ophthalmology. A generative adversarial network (GAN), which consists of two competing types of deep neural networks, including a generator and a discriminator, has demonstrated remarkable performance in image synthesis and image-to-image translation. The adoption of GAN for medical imaging is increasing for image generation and translation, but it is not familiar to researchers in the field of ophthalmology. In this work, we present a literature review on the application of GAN in ophthalmology image domains to discuss important contributions and to identify potential future research directions. METHODS We performed a survey on studies using GAN published before June 2021 only, and we introduced various applications of GAN in ophthalmology image domains. The search identified 48 peer-reviewed papers in the final review. The type of GAN used in the analysis, task, imaging domain, and the outcome were collected to verify the usefulness of the GAN. RESULTS In ophthalmology image domains, GAN can perform segmentation, data augmentation, denoising, domain transfer, super-resolution, post-intervention prediction, and feature extraction. GAN techniques have established an extension of datasets and modalities in ophthalmology. GAN has several limitations, such as mode collapse, spatial deformities, unintended changes, and the generation of high-frequency noises and artifacts of checkerboard patterns. CONCLUSIONS The use of GAN has benefited the various tasks in ophthalmology image domains. Based on our observations, the adoption of GAN in ophthalmology is still in a very early stage of clinical validation compared with deep learning classification techniques because several problems need to be overcome for practical use. However, the proper selection of the GAN technique and statistical modeling of ocular imaging will greatly improve the performance of each image analysis. Finally, this survey would enable researchers to access the appropriate GAN technique to maximize the potential of ophthalmology datasets for deep learning research.
Collapse
Affiliation(s)
- Aram You
- School of Architecture, Kumoh National Institute of Technology, Gumi, Gyeongbuk, South Korea
| | - Jin Kuk Kim
- B&VIIT Eye Center, Seoul, South Korea
- VISUWORKS, Seoul, South Korea
| | - Ik Hee Ryu
- B&VIIT Eye Center, Seoul, South Korea
- VISUWORKS, Seoul, South Korea
| | - Tae Keun Yoo
- B&VIIT Eye Center, Seoul, South Korea.
- Department of Ophthalmology, Aerospace Medical Center, Republic of Korea Air Force, 635 Danjae-ro, Namil-myeon, Cheongwon-gun, Cheongju, Chungcheongbuk-do, 363-849, South Korea.
| |
Collapse
|