1
|
Hassan T, Raja H, Belwafi K, Akcay S, Jleli M, Samet B, Werghi N, Yousaf J, Ghazal M. A Vision Language Correlation Framework for Screening Disabled Retina. IEEE J Biomed Health Inform 2025; 29:1283-1296. [PMID: 39298306 DOI: 10.1109/jbhi.2024.3462653] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/21/2024]
Abstract
Retinopathy is a group of retinal disabilities that causes severe visual impairments or complete blindness. Due to the capability of optical coherence tomography to reveal early retinal abnormalities, many researchers have utilized it to develop autonomous retinal screening systems. However, to the best of our knowledge, most of these systems rely only on mathematical features, which might not be helpful to clinicians since they do not encompass the clinical manifestations of screening the underlying diseases. Such clinical manifestations are critically important to be considered within the autonomous screening systems to match the grading of ophthalmologists within the clinical settings. To overcome these limitations, we present a novel framework that exploits the fusion of vision language correlation between the retinal imagery and the set of clinical prompts to recognize the different types of retinal disabilities. The proposed framework is rigorously tested on six public datasets, where, across each dataset, the proposed framework outperformed state-of-the-art methods in various metrics. Moreover, the clinical significance of the proposed framework is also tested under strict blind testing experiments, where the proposed system achieved a statistically significant correlation coefficient of 0.9185 and 0.9529 with the two expert clinicians. These blind test experiments highlight the potential of the proposed framework to be deployed in the real world for accurate screening of retinal diseases.
Collapse
|
2
|
Kang C, Lo JE, Zhang H, Ng SM, Lin JC, Scott IU, Kalpathy-Cramer J, Liu SHA, Greenberg PB. Artificial intelligence for diagnosing exudative age-related macular degeneration. Cochrane Database Syst Rev 2024; 10:CD015522. [PMID: 39417312 PMCID: PMC11483348 DOI: 10.1002/14651858.cd015522.pub2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Indexed: 10/19/2024]
Abstract
BACKGROUND Age-related macular degeneration (AMD) is a retinal disorder characterized by central retinal (macular) damage. Approximately 10% to 20% of non-exudative AMD cases progress to the exudative form, which may result in rapid deterioration of central vision. Individuals with exudative AMD (eAMD) need prompt consultation with retinal specialists to minimize the risk and extent of vision loss. Traditional methods of diagnosing ophthalmic disease rely on clinical evaluation and multiple imaging techniques, which can be resource-consuming. Tests leveraging artificial intelligence (AI) hold the promise of automatically identifying and categorizing pathological features, enabling the timely diagnosis and treatment of eAMD. OBJECTIVES To determine the diagnostic accuracy of artificial intelligence (AI) as a triaging tool for exudative age-related macular degeneration (eAMD). SEARCH METHODS We searched CENTRAL, MEDLINE, Embase, three clinical trials registries, and Data Archiving and Networked Services (DANS) for gray literature. We did not restrict searches by language or publication date. The date of the last search was April 2024. SELECTION CRITERIA Included studies compared the test performance of algorithms with that of human readers to detect eAMD on retinal images collected from people with AMD who were evaluated at eye clinics in community or academic medical centers, and who were not receiving treatment for eAMD when the images were taken. We included algorithms that were either internally or externally validated or both. DATA COLLECTION AND ANALYSIS Pairs of review authors independently extracted data and assessed study quality using the Quality Assessment of Diagnostic Accuracy Studies-2 (QUADAS-2) tool with revised signaling questions. For studies that reported more than one set of performance results, we extracted only one set of diagnostic accuracy data per study based on the last development stage or the optimal algorithm as indicated by the study authors. For two-class algorithms, we collected data from the 2x2 table whenever feasible. For multi-class algorithms, we first consolidated data from all classes other than eAMD before constructing the corresponding 2x2 tables. Assuming a common positivity threshold applied by the included studies, we chose random-effects, bivariate logistic models to estimate summary sensitivity and specificity as the primary performance metrics. MAIN RESULTS We identified 36 eligible studies that reported 40 sets of algorithm performance data, encompassing over 16,000 participants and 62,000 images. We included 28 studies (78%) that reported 31 algorithms with performance data in the meta-analysis. The remaining nine studies (25%) reported eight algorithms that lacked usable performance data; we reported them in the qualitative synthesis. Study characteristics and risk of bias Most studies were conducted in Asia, followed by Europe, the USA, and collaborative efforts spanning multiple countries. Most studies identified study participants from the hospital setting, while others used retinal images from public repositories; a few studies did not specify image sources. Based on four of the 36 studies reporting demographic information, the age of the study participants ranged from 62 to 82 years. The included algorithms used various retinal image types as model input, such as optical coherence tomography (OCT) images (N = 15), fundus images (N = 6), and multi-modal imaging (N = 7). The predominant core method used was deep neural networks. All studies that reported externally validated algorithms were at high risk of bias mainly due to potential selection bias from either a two-gate design or the inappropriate exclusion of potentially eligible retinal images (or participants). Findings Only three of the 40 included algorithms were externally validated (7.5%, 3/40). The summary sensitivity and specificity were 0.94 (95% confidence interval (CI) 0.90 to 0.97) and 0.99 (95% CI 0.76 to 1.00), respectively, when compared to human graders (3 studies; 27,872 images; low-certainty evidence). The prevalence of images with eAMD ranged from 0.3% to 49%. Twenty-eight algorithms were reportedly either internally validated (20%, 8/40) or tested on a development set (50%, 20/40); the pooled sensitivity and specificity were 0.93 (95% CI 0.89 to 0.96) and 0.96 (95% CI 0.94 to 0.98), respectively, when compared to human graders (28 studies; 33,409 images; low-certainty evidence). We did not identify significant sources of heterogeneity among these 28 algorithms. Although algorithms using OCT images appeared more homogeneous and had the highest summary specificity (0.97, 95% CI 0.93 to 0.98), they were not superior to algorithms using fundus images alone (0.94, 95% CI 0.89 to 0.97) or multimodal imaging (0.96, 95% CI 0.88 to 0.99; P for meta-regression = 0.239). The median prevalence of images with eAMD was 30% (interquartile range [IQR] 22% to 39%). We did not include eight studies that described nine algorithms (one study reported two sets of algorithm results) to distinguish eAMD from normal images, images of other AMD, or other non-AMD retinal lesions in the meta-analysis. Five of these algorithms were generally based on smaller datasets (range 21 to 218 participants per study) yet with a higher prevalence of eAMD images (range 33% to 66%). Relative to human graders, the reported sensitivity in these studies ranged from 0.95 and 0.97, while the specificity ranged from 0.94 to 0.99. Similarly, using small datasets (range 46 to 106), an additional four algorithms for detecting eAMD from other retinal lesions showed high sensitivity (range 0.96 to 1.00) and specificity (range 0.77 to 1.00). AUTHORS' CONCLUSIONS Low- to very low-certainty evidence suggests that an algorithm-based test may correctly identify most individuals with eAMD without increasing unnecessary referrals (false positives) in either the primary or the specialty care settings. There were significant concerns for applying the review findings due to variations in the eAMD prevalence in the included studies. In addition, among the included algorithm-based tests, diagnostic accuracy estimates were at risk of bias due to study participants not reflecting real-world characteristics, inadequate model validation, and the likelihood of selective results reporting. Limited quality and quantity of externally validated algorithms highlighted the need for high-certainty evidence. This evidence will require a standardized definition for eAMD on different imaging modalities and external validation of the algorithm to assess generalizability.
Collapse
Affiliation(s)
- Chaerim Kang
- Division of Ophthalmology, Brown University, Providence, RI, USA
| | - Jui-En Lo
- Department of Internal Medicine, MetroHealth Medical Center/Case Western Reserve University, Cleveland, USA
| | - Helen Zhang
- Program in Liberal Medical Education, Brown University, Providence, RI, USA
| | - Sueko M Ng
- Department of Ophthalmology, University of Colorado Anschutz Medical Campus, Aurora, Colorado, USA
| | - John C Lin
- Department of Medicine, University of Pennsylvania Perelman School of Medicine, Philadelphia, PA, USA
| | - Ingrid U Scott
- Department of Ophthalmology and Public Health Sciences, Penn State College of Medicine, Hershey, PA, USA
| | | | - Su-Hsun Alison Liu
- Department of Ophthalmology, University of Colorado Anschutz Medical Campus, Aurora, Colorado, USA
- Department of Epidemiology, University of Colorado Anschutz Medical Campus, Aurora, Colorado, USA
| | - Paul B Greenberg
- Division of Ophthalmology, Brown University, Providence, RI, USA
- Section of Ophthalmology, VA Providence Healthcare System, Providence, RI, USA
| |
Collapse
|
3
|
Kalupahana D, Kahatapitiya NS, Silva BN, Kim J, Jeon M, Wijenayake U, Wijesinghe RE. Dense Convolutional Neural Network-Based Deep Learning Pipeline for Pre-Identification of Circular Leaf Spot Disease of Diospyros kaki Leaves Using Optical Coherence Tomography. SENSORS (BASEL, SWITZERLAND) 2024; 24:5398. [PMID: 39205092 PMCID: PMC11359294 DOI: 10.3390/s24165398] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 06/21/2024] [Revised: 07/30/2024] [Accepted: 08/05/2024] [Indexed: 09/04/2024]
Abstract
Circular leaf spot (CLS) disease poses a significant threat to persimmon cultivation, leading to substantial harvest reductions. Existing visual and destructive inspection methods suffer from subjectivity, limited accuracy, and considerable time consumption. This study presents an automated pre-identification method of the disease through a deep learning (DL) based pipeline integrated with optical coherence tomography (OCT), thereby addressing the highlighted issues with the existing methods. The investigation yielded promising outcomes by employing transfer learning with pre-trained DL models, specifically DenseNet-121 and VGG-16. The DenseNet-121 model excels in differentiating among three stages of CLS disease (healthy (H), apparently healthy (or healthy-infected (HI)), and infected (I)). The model achieved precision values of 0.7823 for class-H, 0.9005 for class-HI, and 0.7027 for class-I, supported by recall values of 0.8953 for class-HI and 0.8387 for class-I. Moreover, the performance of CLS detection was enhanced by a supplemental quality inspection model utilizing VGG-16, which attained an accuracy of 98.99% in discriminating between low-detail and high-detail images. Moreover, this study employed a combination of LAMP and A-scan for the dataset labeling process, significantly enhancing the accuracy of the models. Overall, this study underscores the potential of DL techniques integrated with OCT to enhance disease identification processes in agricultural settings, particularly in persimmon cultivation, by offering efficient and objective pre-identification of CLS and enabling early intervention and management strategies.
Collapse
Affiliation(s)
- Deshan Kalupahana
- Department of Computer Engineering, Faculty of Engineering, University of Sri Jayewardenepura, Nugegoda 10250, Sri Lanka; (D.K.); (N.S.K.)
| | - Nipun Shantha Kahatapitiya
- Department of Computer Engineering, Faculty of Engineering, University of Sri Jayewardenepura, Nugegoda 10250, Sri Lanka; (D.K.); (N.S.K.)
| | - Bhagya Nathali Silva
- Department of Information Technology, Faculty of Computing, Sri Lanka Institute of Information Technology, Malabe 10115, Sri Lanka;
- Center for Excellence in Informatics, Electronics & Transmission (CIET), Sri Lanka Institute of Information Technology, Malabe 10115, Sri Lanka
| | - Jeehyun Kim
- School of Electronic and Electrical Engineering, College of IT Engineering, Kyungpook National University, 80 Daehak-ro, Buk-gu, Daegu 41566, Republic of Korea; (J.K.); (M.J.)
| | - Mansik Jeon
- School of Electronic and Electrical Engineering, College of IT Engineering, Kyungpook National University, 80 Daehak-ro, Buk-gu, Daegu 41566, Republic of Korea; (J.K.); (M.J.)
| | - Udaya Wijenayake
- Department of Computer Engineering, Faculty of Engineering, University of Sri Jayewardenepura, Nugegoda 10250, Sri Lanka; (D.K.); (N.S.K.)
| | - Ruchire Eranga Wijesinghe
- Center for Excellence in Informatics, Electronics & Transmission (CIET), Sri Lanka Institute of Information Technology, Malabe 10115, Sri Lanka
- Department of Electrical and Electronic Engineering, Faculty of Engineering, Sri Lanka Institute of Information Technology, Malabe 10115, Sri Lanka
| |
Collapse
|
4
|
Salimi M, Roshanfar M, Tabatabaei N, Mosadegh B. Machine Learning-Assisted Short-Wave InfraRed (SWIR) Techniques for Biomedical Applications: Towards Personalized Medicine. J Pers Med 2023; 14:33. [PMID: 38248734 PMCID: PMC10817559 DOI: 10.3390/jpm14010033] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/24/2023] [Revised: 12/08/2023] [Accepted: 12/20/2023] [Indexed: 01/23/2024] Open
Abstract
Personalized medicine transforms healthcare by adapting interventions to individuals' unique genetic, molecular, and clinical profiles. To maximize diagnostic and/or therapeutic efficacy, personalized medicine requires advanced imaging devices and sensors for accurate assessment and monitoring of individual patient conditions or responses to therapeutics. In the field of biomedical optics, short-wave infrared (SWIR) techniques offer an array of capabilities that hold promise to significantly enhance diagnostics, imaging, and therapeutic interventions. SWIR techniques provide in vivo information, which was previously inaccessible, by making use of its capacity to penetrate biological tissues with reduced attenuation and enable researchers and clinicians to delve deeper into anatomical structures, physiological processes, and molecular interactions. Combining SWIR techniques with machine learning (ML), which is a powerful tool for analyzing information, holds the potential to provide unprecedented accuracy for disease detection, precision in treatment guidance, and correlations of complex biological features, opening the way for the data-driven personalized medicine field. Despite numerous biomedical demonstrations that utilize cutting-edge SWIR techniques, the clinical potential of this approach has remained significantly underexplored. This paper demonstrates how the synergy between SWIR imaging and ML is reshaping biomedical research and clinical applications. As the paper showcases the growing significance of SWIR imaging techniques that are empowered by ML, it calls for continued collaboration between researchers, engineers, and clinicians to boost the translation of this technology into clinics, ultimately bridging the gap between cutting-edge technology and its potential for personalized medicine.
Collapse
Affiliation(s)
| | - Majid Roshanfar
- Department of Mechanical Engineering, Concordia University, Montreal, QC H3G 1M8, Canada;
| | - Nima Tabatabaei
- Department of Mechanical Engineering, York University, Toronto, ON M3J 1P3, Canada;
| | - Bobak Mosadegh
- Dalio Institute of Cardiovascular Imaging, Department of Radiology, Weill Cornell Medicine, New York, NY 10021, USA
| |
Collapse
|
5
|
Tripathi A, Kumar P, Mayya V, Tulsani A. Generating OCT B-Scan DME images using optimized Generative Adversarial Networks (GANs). Heliyon 2023; 9:e18773. [PMID: 37609420 PMCID: PMC10440457 DOI: 10.1016/j.heliyon.2023.e18773] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/03/2023] [Revised: 07/26/2023] [Accepted: 07/27/2023] [Indexed: 08/24/2023] Open
Abstract
Diabetic Macular Edema (DME) represents a significant visual impairment among individuals with diabetes, leading to a dramatic reduction in visual acuity and potentially resulting in irreversible vision loss. Optical Coherence Tomography (OCT), a technique that produces high-resolution retinal images, plays a vital role in the clinical assessment of this condition. Physicians typically rely on OCT B-Scan images to evaluate DME severity. However, manual interpretation of these images is susceptible to errors, which can lead to detrimental consequences, such as misdiagnosis and improper treatment strategies. Hence, there is a critical need for more reliable diagnostic methods. This study aims to address this gap by proposing an automated model based on Generative Adversarial Networks (GANs) to generate OCT B-Scan images of DME. The model synthesizes images from patients' baseline OCT B-Scan images, which could potentially enhance the robustness of DME detection systems. We employ five distinct GANs in this study: Deep Convolutional GAN, Conditional GAN, CycleGAN, StyleGAN2, and StyleGAN3, drawing comparisons across their performance. Subsequently, the hyperparameters of the best-performing GAN are fine-tuned using Particle Swarm Optimization (PSO) to produce more realistic OCT images. This comparative analysis not only serves to improve the detection of DME severity using OCT images but also provides insights into the appropriate choice of GANs for the effective generation of realistic OCT images from the baseline OCT datasets.
Collapse
Affiliation(s)
- Aditya Tripathi
- Department of Information & Communication Technology, Manipal Institute of Technology, Manipal, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India
| | - Preetham Kumar
- Department of Information & Communication Technology, Manipal Institute of Technology, Manipal, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India
| | - Veena Mayya
- Department of Information & Communication Technology, Manipal Institute of Technology, Manipal, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India
| | - Akshat Tulsani
- Department of Information & Communication Technology, Manipal Institute of Technology, Manipal, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India
| |
Collapse
|
6
|
Zhang L, Tang L, Xia M, Cao G. The application of artificial intelligence in glaucoma diagnosis and prediction. Front Cell Dev Biol 2023; 11:1173094. [PMID: 37215077 PMCID: PMC10192631 DOI: 10.3389/fcell.2023.1173094] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/24/2023] [Accepted: 04/24/2023] [Indexed: 05/24/2023] Open
Abstract
Artificial intelligence is a multidisciplinary and collaborative science, the ability of deep learning for image feature extraction and processing gives it a unique advantage in dealing with problems in ophthalmology. The deep learning system can assist ophthalmologists in diagnosing characteristic fundus lesions in glaucoma, such as retinal nerve fiber layer defects, optic nerve head damage, optic disc hemorrhage, etc. Early detection of these lesions can help delay structural damage, protect visual function, and reduce visual field damage. The development of deep learning led to the emergence of deep convolutional neural networks, which are pushing the integration of artificial intelligence with testing devices such as visual field meters, fundus imaging and optical coherence tomography to drive more rapid advances in clinical glaucoma diagnosis and prediction techniques. This article details advances in artificial intelligence combined with visual field, fundus photography, and optical coherence tomography in the field of glaucoma diagnosis and prediction, some of which are familiar and some not widely known. Then it further explores the challenges at this stage and the prospects for future clinical applications. In the future, the deep cooperation between artificial intelligence and medical technology will make the datasets and clinical application rules more standardized, and glaucoma diagnosis and prediction tools will be simplified in a single direction, which will benefit multiple ethnic groups.
Collapse
Affiliation(s)
- Linyu Zhang
- The Affiliated Eye Hospital of Nanjing Medical University, Nanjing, China
- The Fourth School of Clinical Medicine, Nanjing Medical University, Nanjing, China
| | - Li Tang
- The Affiliated Eye Hospital of Nanjing Medical University, Nanjing, China
| | - Min Xia
- The Affiliated Eye Hospital of Nanjing Medical University, Nanjing, China
- The Fourth School of Clinical Medicine, Nanjing Medical University, Nanjing, China
| | - Guofan Cao
- The Affiliated Eye Hospital of Nanjing Medical University, Nanjing, China
- The Fourth School of Clinical Medicine, Nanjing Medical University, Nanjing, China
| |
Collapse
|
7
|
Thapa P, Singh V, Gupta K, Shrivastava A, Kumar V, Kataria K, Mishra PR, Mehta DS. Point-of-care devices based on fluorescence imaging and spectroscopy for tumor margin detection during breast cancer surgery: Towards breast conservation treatment. Lasers Surg Med 2023; 55:423-436. [PMID: 36884000 DOI: 10.1002/lsm.23651] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/06/2022] [Revised: 02/22/2023] [Accepted: 02/23/2023] [Indexed: 03/09/2023]
Abstract
OBJECTIVE Fluorescence-based methods are highly specific and sensitive and have potential in breast cancer detection. Simultaneous fluorescence imaging and spectroscopy during intraoperative procedures of breast cancer have great advantages in detection of tumor margin as well as in classification of tumor to healthy tissues. Intra-operative real-time confirmation of breast cancer tumor margin is the aim of surgeons, and therefore, there is an urgent need for such techniques and devices which fulfill the surgeon's priorities. METHODS In this article, we propose the development of fluorescence-based smartphone imaging and spectroscopic point-of-care multi-modal devices for detection of invasive ductal carcinoma in tumor margin during removal of tumor. These multimodal devices are portable, cost-effective, noninvasive, and user-friendly. Molecular level sensitivity of fluorescence process shows different behavior in normal, cancerous and marginal tissues. We observed significant spectral changes, such as, red-shift, full-width half maximum (FWHM), and increased intensity as we go towards tumor center from normal tissue. High contrast in fluorescence images and spectra are also recorded for cancer tissues compared to healthy tissues. Preliminary results for the initial trial of the devices are reported in this article. RESULTS A total 44 spectra from 11 patients of invasive ductal carcinoma (11 spectra for invasive ductal carcinoma and rest are normal and negative margins) are used. Principle component analysis is used for the classification of invasive ductal carcinoma with an accuracy of 93%, specificity of 75% and sensitivity of 92.8%. We obtained an average 6.17 ± 1.66 nm red shift for IDC with respect to normal tissue. The red shift and maximum fluorescence intensity indicates p < 0.01. These results described here are supported by histopathological examination of the same sample. CONCLUSION In the present manuscript, simultaneous fluorescence-based imaging and spectroscopy is accomplished for the classification of IDC tissues and breast cancer margin detection.
Collapse
Affiliation(s)
- Pramila Thapa
- Department of Physics, Bio-photonics and Green-photonics Laboratory, Indian Institute of Technology Delhi, New Delhi, India
| | - Veena Singh
- Department of Physics, Bio-photonics and Green-photonics Laboratory, Indian Institute of Technology Delhi, New Delhi, India
| | - Komal Gupta
- Department of Surgical Disciplines, All India Institute of Medical Sciences (AIIMS), New Delhi, India
| | - Anurag Shrivastava
- Department of Surgical Disciplines, All India Institute of Medical Sciences (AIIMS), New Delhi, India
| | - Virendra Kumar
- Department of Physics, Bio-photonics and Green-photonics Laboratory, Indian Institute of Technology Delhi, New Delhi, India
| | - Kamal Kataria
- Department of Surgical Disciplines, All India Institute of Medical Sciences (AIIMS), New Delhi, India
| | - Piyush R Mishra
- Department of Surgical Disciplines, All India Institute of Medical Sciences (AIIMS), New Delhi, India
| | - Dalip S Mehta
- Department of Physics, Bio-photonics and Green-photonics Laboratory, Indian Institute of Technology Delhi, New Delhi, India
| |
Collapse
|
8
|
Tampu IE, Eklund A, Johansson K, Gimm O, Haj-Hosseini N. Diseased thyroid tissue classification in OCT images using deep learning: Towards surgical decision support. JOURNAL OF BIOPHOTONICS 2023; 16:e202200227. [PMID: 36203247 DOI: 10.1002/jbio.202200227] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/14/2022] [Revised: 09/15/2022] [Accepted: 09/16/2022] [Indexed: 06/16/2023]
Abstract
Intraoperative guidance tools for thyroid surgery based on optical coherence tomography (OCT) could aid distinguish between normal and diseased tissue. However, OCT images are difficult to interpret, thus, real-time automatic analysis could support the clinical decision-making. In this study, several deep learning models were investigated for thyroid disease classification on 2D and 3D OCT data obtained from ex vivo specimens of 22 patients undergoing surgery and diagnosed with several thyroid pathologies. Additionally, two open-access datasets were used to evaluate the custom models. On the thyroid dataset, the best performance was achieved by the 3D vision transformer model with a Matthew's correlation coefficient (MCC) of 0.79 (accuracy = 0.90) for the normal-versus-abnormal classification. On the open-access datasets, the custom models achieved the best performance (MCC > 0.88, accuracy > 0.96). Results obtained for the normal-versus-abnormal classification suggest OCT, complemented with deep learning-based analysis, as a tool for real-time automatic diseased tissue identification in thyroid surgery.
Collapse
Affiliation(s)
- Iulian Emil Tampu
- Department of Biomedical Engineering, Linköping University, Linköping, Sweden
- Center for Medical Image Science and Visualization, Linköping University, Linköping, Sweden
| | - Anders Eklund
- Department of Biomedical Engineering, Linköping University, Linköping, Sweden
- Center for Medical Image Science and Visualization, Linköping University, Linköping, Sweden
- Division of Statistics & Machine Learning, Department of Computer and Information Science, Linköping University, Linköping, Sweden
| | - Kenth Johansson
- Department of Surgery, Västervik Hospital, Västervik, Sweden
- Department of Surgery, Örebro University Hospital, Örebro, Sweden
| | - Oliver Gimm
- Department of Surgery, Linköping University Hospital, Linköping, Sweden
- Department of Biomedical and Clinical Sciences, Linköping University, Linköping, Sweden
| | - Neda Haj-Hosseini
- Department of Biomedical Engineering, Linköping University, Linköping, Sweden
- Center for Medical Image Science and Visualization, Linköping University, Linköping, Sweden
| |
Collapse
|
9
|
Tampu IE, Eklund A, Haj-Hosseini N. Inflation of test accuracy due to data leakage in deep learning-based classification of OCT images. Sci Data 2022; 9:580. [PMID: 36138025 PMCID: PMC9500039 DOI: 10.1038/s41597-022-01618-6] [Citation(s) in RCA: 14] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2022] [Accepted: 08/09/2022] [Indexed: 11/10/2022] Open
Abstract
In the application of deep learning on optical coherence tomography (OCT) data, it is common to train classification networks using 2D images originating from volumetric data. Given the micrometer resolution of OCT systems, consecutive images are often very similar in both visible structures and noise. Thus, an inappropriate data split can result in overlap between the training and testing sets, with a large portion of the literature overlooking this aspect. In this study, the effect of improper dataset splitting on model evaluation is demonstrated for three classification tasks using three OCT open-access datasets extensively used, Kermany's and Srinivasan's ophthalmology datasets, and AIIMS breast tissue dataset. Results show that the classification performance is inflated by 0.07 up to 0.43 in terms of Matthews Correlation Coefficient (accuracy: 5% to 30%) for models tested on datasets with improper splitting, highlighting the considerable effect of dataset handling on model evaluation. This study intends to raise awareness on the importance of dataset splitting given the increased research interest in implementing deep learning on OCT data.
Collapse
Affiliation(s)
- Iulian Emil Tampu
- Department of Biomedical Engineering, Linköping University, 581 85, Linköping, Sweden. .,Center for Medical Image Science and Visualization, Linköping University, 581 85, Linköping, Sweden.
| | - Anders Eklund
- Department of Biomedical Engineering, Linköping University, 581 85, Linköping, Sweden.,Center for Medical Image Science and Visualization, Linköping University, 581 85, Linköping, Sweden.,Division of Statistics & Machine Learning, Department of Computer and Information Science, Linköping University, 581 83, Linköping, Sweden
| | - Neda Haj-Hosseini
- Department of Biomedical Engineering, Linköping University, 581 85, Linköping, Sweden.,Center for Medical Image Science and Visualization, Linköping University, 581 85, Linköping, Sweden
| |
Collapse
|
10
|
Foo KY, Newman K, Fang Q, Gong P, Ismail HM, Lakhiani DD, Zilkens R, Dessauvagie BF, Latham B, Saunders CM, Chin L, Kennedy BF. Multi-class classification of breast tissue using optical coherence tomography and attenuation imaging combined via deep learning. BIOMEDICAL OPTICS EXPRESS 2022; 13:3380-3400. [PMID: 35781967 PMCID: PMC9208580 DOI: 10.1364/boe.455110] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/04/2022] [Revised: 04/23/2022] [Accepted: 04/25/2022] [Indexed: 05/27/2023]
Abstract
We demonstrate a convolutional neural network (CNN) for multi-class breast tissue classification as adipose tissue, benign dense tissue, or malignant tissue, using multi-channel optical coherence tomography (OCT) and attenuation images, and a novel Matthews correlation coefficient (MCC)-based loss function that correlates more strongly with performance metrics than the commonly used cross-entropy loss. We hypothesized that using multi-channel images would increase tumor detection performance compared to using OCT alone. 5,804 images from 29 patients were used to fine-tune a pre-trained ResNet-18 network. Adding attenuation images to OCT images yields statistically significant improvements in several performance metrics, including benign dense tissue sensitivity (68.0% versus 59.6%), malignant tissue positive predictive value (PPV) (79.4% versus 75.5%), and total accuracy (85.4% versus 83.3%), indicating that the additional contrast from attenuation imaging is most beneficial for distinguishing between benign dense tissue and malignant tissue.
Collapse
Affiliation(s)
- Ken Y. Foo
- BRITElab, Harry Perkins Institute of Medical Research, QEII Medical Centre, Nedlands, and Centre for Medical Research, The University of Western Australia, Perth, WA 6009, Australia
- Department of Electrical, Electronic & Computer Engineering, School of Engineering, The University of Western Australia, Perth, WA 6009, Australia
| | - Kyle Newman
- BRITElab, Harry Perkins Institute of Medical Research, QEII Medical Centre, Nedlands, and Centre for Medical Research, The University of Western Australia, Perth, WA 6009, Australia
- Department of Electrical, Electronic & Computer Engineering, School of Engineering, The University of Western Australia, Perth, WA 6009, Australia
| | - Qi Fang
- BRITElab, Harry Perkins Institute of Medical Research, QEII Medical Centre, Nedlands, and Centre for Medical Research, The University of Western Australia, Perth, WA 6009, Australia
- Department of Electrical, Electronic & Computer Engineering, School of Engineering, The University of Western Australia, Perth, WA 6009, Australia
| | - Peijun Gong
- BRITElab, Harry Perkins Institute of Medical Research, QEII Medical Centre, Nedlands, and Centre for Medical Research, The University of Western Australia, Perth, WA 6009, Australia
- Department of Electrical, Electronic & Computer Engineering, School of Engineering, The University of Western Australia, Perth, WA 6009, Australia
| | - Hina M. Ismail
- BRITElab, Harry Perkins Institute of Medical Research, QEII Medical Centre, Nedlands, and Centre for Medical Research, The University of Western Australia, Perth, WA 6009, Australia
- Department of Electrical, Electronic & Computer Engineering, School of Engineering, The University of Western Australia, Perth, WA 6009, Australia
| | - Devina D. Lakhiani
- BRITElab, Harry Perkins Institute of Medical Research, QEII Medical Centre, Nedlands, and Centre for Medical Research, The University of Western Australia, Perth, WA 6009, Australia
- Department of Electrical, Electronic & Computer Engineering, School of Engineering, The University of Western Australia, Perth, WA 6009, Australia
| | - Renate Zilkens
- BRITElab, Harry Perkins Institute of Medical Research, QEII Medical Centre, Nedlands, and Centre for Medical Research, The University of Western Australia, Perth, WA 6009, Australia
- Division of Surgery, Medical School, The University of Western Australia, Perth, WA 6009, Australia
| | - Benjamin F. Dessauvagie
- Division of Pathology and Laboratory Medicine, Medical School, The University of Western Australia, Perth, WA 6009, Australia
- PathWest, Fiona Stanley Hospital, Murdoch, WA 6150, Australia
| | - Bruce Latham
- PathWest, Fiona Stanley Hospital, Murdoch, WA 6150, Australia
- School of Medicine, The University of Notre Dame, Fremantle, WA 6160, Australia
| | - Christobel M. Saunders
- Division of Surgery, Medical School, The University of Western Australia, Perth, WA 6009, Australia
- Breast Centre, Fiona Stanley Hospital, Murdoch, WA 6150, Australia
- Breast Clinic, Royal Perth Hospital, Perth, WA 6000, Australia
- Department of Surgery, Melbourne Medical School, The University of Melbourne, Parkville, VIC 3010, Australia
| | - Lixin Chin
- BRITElab, Harry Perkins Institute of Medical Research, QEII Medical Centre, Nedlands, and Centre for Medical Research, The University of Western Australia, Perth, WA 6009, Australia
- Department of Electrical, Electronic & Computer Engineering, School of Engineering, The University of Western Australia, Perth, WA 6009, Australia
| | - Brendan F. Kennedy
- BRITElab, Harry Perkins Institute of Medical Research, QEII Medical Centre, Nedlands, and Centre for Medical Research, The University of Western Australia, Perth, WA 6009, Australia
- Department of Electrical, Electronic & Computer Engineering, School of Engineering, The University of Western Australia, Perth, WA 6009, Australia
- Australian Research Council Centre for Personalised Therapeutics Technologies, Perth, WA 6000, Australia
| |
Collapse
|
11
|
CDC-Net: Cascaded decoupled convolutional network for lesion-assisted detection and grading of retinopathy using optical coherence tomography (OCT) scans. Biomed Signal Process Control 2021. [DOI: 10.1016/j.bspc.2021.103030] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/12/2023]
|
12
|
Arefin R, Samad MD, Akyelken FA, Davanian A. Non-transfer Deep Learning of Optical Coherence Tomography for Post-hoc Explanation of Macular Disease Classification. IEEE INTERNATIONAL CONFERENCE ON HEALTHCARE INFORMATICS. IEEE INTERNATIONAL CONFERENCE ON HEALTHCARE INFORMATICS 2021; 2021:48-52. [PMID: 36168324 PMCID: PMC9511893 DOI: 10.1109/ichi52183.2021.00020] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/14/2023]
Abstract
Deep transfer learning is a popular choice for classifying monochromatic medical images using models that are pretrained by natural images with color channels. This choice may introduce unnecessarily redundant model complexity that can limit explanations of such model behavior and outcomes in the context of medical imaging. To investigate this hypothesis, we develop a configurable deep convolutional neural network (CNN) to classify four macular disease conditions using retinal optical coherence tomography (OCT) images. Our proposed non-transfer deep CNN model (acc: 97.9%) outperforms existing transfer learning models such as ResNet-50 (acc: 89.0%), ResNet-101 (acc: 96.7%), VGG-19 (acc: 93.3%), Inception-V3 (acc: 95.8%) in the same retinal OCT image classification task. We perform post-hoc analysis of the trained model and model extracted image features, which reveals that only eight out of 256 filter kernels are active at our final convolutional layer. The convolutional responses of these selective eight filters yield image features that efficiently separate four macular disease classes even when projected onto two-dimensional principal component space. Our findings suggest that many deep learning parameters and their computations are redundant and expensive for retinal OCT image classification, which are expected to be more intense when using transfer learning. Additionally, we provide clinical interpretations of our misclassified test images identifying manifest artifacts, shadowing of useful texture, false texture representing fluids, and other confounding factors. These clinical explanations along with model optimization via kernel selection can improve the classification accuracy, computational costs, and explainability of model outcomes.
Collapse
Affiliation(s)
- Raisul Arefin
- Dept. of Computer Science Auburn University Auburn, AB USA
| | - Manar D Samad
- Dept. of Computer Science Tennessee State University Nashville, TN USA
| | - Furkan A Akyelken
- Dept. of Computer Science Tennessee State University Nashville, TN, USA
| | - Arash Davanian
- Vanderbilt Eye Institute Vanderbilt University Medical Center Nashville, TN, USA
| |
Collapse
|
13
|
Zhu D, Wang J, Marjanovic M, Chaney EJ, Cradock KA, Higham AM, Liu ZG, Gao Z, Boppart SA. Differentiation of breast tissue types for surgical margin assessment using machine learning and polarization-sensitive optical coherence tomography. BIOMEDICAL OPTICS EXPRESS 2021; 12:3021-3036. [PMID: 34168912 PMCID: PMC8194620 DOI: 10.1364/boe.423026] [Citation(s) in RCA: 19] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/22/2021] [Revised: 04/10/2021] [Accepted: 04/12/2021] [Indexed: 05/04/2023]
Abstract
We report an automated differentiation model for classifying malignant tumor, fibro-adipose, and stroma in human breast tissues based on polarization-sensitive optical coherence tomography (PS-OCT). A total of 720 PS-OCT images from 72 sites of 41 patients with H&E histology-confirmed diagnoses as the gold standard were employed in this study. The differentiation model is trained by the features extracted from both one standard OCT-based metric (i.e., intensity) and four PS-OCT-based metrics (i.e., phase difference between two channels (PD), phase retardation (PR), local phase retardation (LPR), and degree of polarization uniformity (DOPU)). Further optimized by forward searching and validated by leave-one-site-out-cross-validation (LOSOCV) method, the best feature subset was acquired with the highest overall accuracy of 93.5% for the model. Furthermore, to show the superiority of our differentiation model based on PS-OCT images over standard OCT images, the best model trained by intensity-only features (usually obtained by standard OCT systems) was also obtained with an overall accuracy of 82.9%, demonstrating the significance of the polarization information in breast tissue differentiation. The high performance of our differentiation model suggests the potential of using PS-OCT for intraoperative human breast tissue differentiation during the surgical resection of breast cancer.
Collapse
Affiliation(s)
- Dan Zhu
- Beckman Institute for Advanced Science and Technology, University of Illinois at Urbana-Champaign, Urbana, Illinois 61801, USA
- School of Electronic and Optical Engineering, Nanjing University of Science and Technology, Nanjing 210094, China
- These authors contributed equally to this work
| | - Jianfeng Wang
- Beckman Institute for Advanced Science and Technology, University of Illinois at Urbana-Champaign, Urbana, Illinois 61801, USA
- These authors contributed equally to this work
| | - Marina Marjanovic
- Beckman Institute for Advanced Science and Technology, University of Illinois at Urbana-Champaign, Urbana, Illinois 61801, USA
- Department of Bioengineering, University of Illinois at Urbana-Champaign, Urbana, Illinois 61801, USA
| | - Eric J Chaney
- Beckman Institute for Advanced Science and Technology, University of Illinois at Urbana-Champaign, Urbana, Illinois 61801, USA
| | - Kimberly A Cradock
- Department of Surgery, Carle Foundation Hospital, Urbana, Illinois 61801, USA
- Carle Illinois College of Medicine, University of Illinois at Urbana-Champaign, Champaign, Illinois 61801, USA
| | - Anna M Higham
- Department of Surgery, Carle Foundation Hospital, Urbana, Illinois 61801, USA
- Carle Illinois College of Medicine, University of Illinois at Urbana-Champaign, Champaign, Illinois 61801, USA
| | - Zheng G Liu
- Carle Illinois College of Medicine, University of Illinois at Urbana-Champaign, Champaign, Illinois 61801, USA
- Department of Pathology, Carle Foundation Hospital, Urbana, Illinois 61801, USA
| | - Zhishan Gao
- School of Electronic and Optical Engineering, Nanjing University of Science and Technology, Nanjing 210094, China
| | - Stephen A Boppart
- Beckman Institute for Advanced Science and Technology, University of Illinois at Urbana-Champaign, Urbana, Illinois 61801, USA
- Department of Bioengineering, University of Illinois at Urbana-Champaign, Urbana, Illinois 61801, USA
- Carle Illinois College of Medicine, University of Illinois at Urbana-Champaign, Champaign, Illinois 61801, USA
- Department of Electrical and Computer Engineering, University of Illinois at Urbana-Champaign, Urbana, Illinois 61801, USA
- Cancer Center at Illinois, University of Illinois at Urbana-Champaign, Urbana, Illinois 61801, USA
| |
Collapse
|