1
|
Liu X, Zhang J, Zhang Y, Chen L, Luo L, Tang J. Weakly supervised segmentation of retinal layers on OCT images with AMD using uncertainty prototype and boundary regression. Med Image Anal 2025; 102:103572. [PMID: 40179629 DOI: 10.1016/j.media.2025.103572] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/02/2024] [Revised: 02/16/2025] [Accepted: 03/27/2025] [Indexed: 04/05/2025]
Abstract
Retinal layer segmentation for optical coherence tomography (OCT) images of eyes is a critical step in the diagnosis and treatment of age-related macular degeneration (AMD) eye disease. In recent years, dense annotation supervised OCT layer segmentation methods have made significant progress. However, obtaining pixel-by-pixel labeled masks from OCT retinal images is time-consuming and labor-intensive. To reduce dependence on dense annotations, this paper proposes a novel weakly supervised layer segmentation method with Uncertainty Prototype module and Boundary Regression loss (W-UPBR), which only requires scribble annotations. Specifically, we first propose a feature enhancement U-Net (FEU-Net) to alleviate the severe layer distortion problem in OCT images with AMD. And this model serves as the backbone of a dual-branch network framework to enhance features. Within FEU-Net, in addition to the basic U-Net, two modules have been proposed: the global-local context-aware (GLCA) module, which captures both global and local contextual information, and the multi-scale fusion (MSF) module, designed for fusing multi-scale features. Secondly, we propose an uncertainty prototype module that combines the uncertainty-guided prototype and distance optimization loss. This module aims to exploit the similarities and dissimilarities between OCT images, thereby reducing mis-segmentation in layers caused by interference factors. Furthermore, a mixed pseudo-label strategy is incorporated to mix different predictions to alleviate the limitations posed by insufficient supervision and further promote network training. Finally, we design a boundary regression loss that constrains the boundaries in both 1D and 2D dimensions to enhance boundary under the supervision of generated mixed pseudo-labels, thereby reducing topological errors. The proposed method was evaluated on three datasets, and the results show that the proposed method outperformed other state-of-the-art weakly supervised methods and could achieve comparable performance to fully supervised methods.
Collapse
Affiliation(s)
- Xiaoming Liu
- School of Computer Science and Technology, Wuhan University of Science and Technology, Wuhan 430065, China; Hubei Province Key Laboratory of Intelligent Information Processing and Real-time Industrial System, Wuhan 430065, China.
| | - Jia Zhang
- School of Computer Science and Technology, Wuhan University of Science and Technology, Wuhan 430065, China; Hubei Province Key Laboratory of Intelligent Information Processing and Real-time Industrial System, Wuhan 430065, China
| | - Ying Zhang
- Aier Eye Hospital of Wuhan University, Wuhan 430014, China
| | - Li Chen
- School of Computer Science and Technology, Wuhan University of Science and Technology, Wuhan 430065, China; Hubei Province Key Laboratory of Intelligent Information Processing and Real-time Industrial System, Wuhan 430065, China
| | - Liangfu Luo
- School of Computer Science Institute, Wuhan Qingchuan University, Wuhan 430205, China
| | - Jinshan Tang
- Department of Health Administration and Policy, George Mason University, Fairfax, VA 22030, USA
| |
Collapse
|
2
|
Xu X, Wang H, Lu Y, Zhang H, Tan T, Xu F, Lei J. Joint segmentation of retinal layers and fluid lesions in optical coherence tomography with cross-dataset learning. Artif Intell Med 2025; 162:103096. [PMID: 39999658 DOI: 10.1016/j.artmed.2025.103096] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2024] [Revised: 12/24/2024] [Accepted: 02/19/2025] [Indexed: 02/27/2025]
Abstract
BACKGROUND AND OBJECTIVES Age-related macular degeneration (AMD) is the leading cause of irreversible vision loss among people over 50 years old, which manifests in the retina through various changes of retinal layers and pathological lesions. The accurate segmentation of optical coherence tomography (OCT) image features is crucial for the identification and tracking of AMD. Although the recent developments in deep neural network have brought profound progress in this area, accurately segmenting retinal layers and pathological lesions remains a challenging task because of the interaction between these two tasks. METHODS In this study, we propose a three-branch, hierarchical multi-task framework that enables joint segmentation of seven retinal layers and three types of pathological lesions. A regression guidance module is introduced to provide explicit shape guidance between sub-tasks. We also propose a cross-dataset learning strategy to leverage public datasets with partial labels. The proposed framework was evaluated on a clinical dataset consisting of 140 OCT B-scans with pixel-level annotations of seven retinal layers and three types of lesions. Additionally, we compared its performance with the state-of-the-art methods on two public datasets. RESULTS Comprehensive ablation showed that the proposed hierarchical architecture significantly improved performance for most retinal layers and pathological lesions, achieving the highest mean DSC of 76.88 %. The IRF also achieved the best performance with a DSC of 68.15 %. Comparative studies demonstrated that the hierarchical multi-task architecture could significantly enhance segmentation accuracy and outperform state-of-the-art methods. CONCLUSION The proposed framework could also be generalized to other medical image segmentation tasks with interdependent relationships.
Collapse
Affiliation(s)
- Xiayu Xu
- The Key Laboratory of Biomedical Information Engineering of Ministry of Education, School of Life Science and Technology, Xi'an Jiaotong University, Xi'an 710049, PR China; Bioinspired Engineering and Biomechanics Center (BEBC), Xi'an Jiaotong University, Xi'an 710049, PR China.
| | - Hualin Wang
- The Key Laboratory of Biomedical Information Engineering of Ministry of Education, School of Life Science and Technology, Xi'an Jiaotong University, Xi'an 710049, PR China; Bioinspired Engineering and Biomechanics Center (BEBC), Xi'an Jiaotong University, Xi'an 710049, PR China
| | - Yulei Lu
- The Key Laboratory of Biomedical Information Engineering of Ministry of Education, School of Life Science and Technology, Xi'an Jiaotong University, Xi'an 710049, PR China; Bioinspired Engineering and Biomechanics Center (BEBC), Xi'an Jiaotong University, Xi'an 710049, PR China
| | - Hanze Zhang
- Department of Ophthalmology, First Affiliated Hospital of Xi'an Jiaotong University, Xi'an 710049, PR China
| | - Tao Tan
- Faculty of Applied Sciences, Macao Polytechnic University, 999078, Macao
| | - Feng Xu
- The Key Laboratory of Biomedical Information Engineering of Ministry of Education, School of Life Science and Technology, Xi'an Jiaotong University, Xi'an 710049, PR China; Bioinspired Engineering and Biomechanics Center (BEBC), Xi'an Jiaotong University, Xi'an 710049, PR China
| | - Jianqin Lei
- Department of Ophthalmology, First Affiliated Hospital of Xi'an Jiaotong University, Xi'an 710049, PR China.
| |
Collapse
|
3
|
Wang S, Liu L, Wang J, Peng X, Liu B. MSR-UNet: enhancing multi-scale and long-range dependencies in medical image segmentation. PeerJ Comput Sci 2024; 10:e2563. [PMID: 39650414 PMCID: PMC11623095 DOI: 10.7717/peerj-cs.2563] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/04/2024] [Accepted: 11/08/2024] [Indexed: 12/11/2024]
Abstract
Transformer-based technology has attracted widespread attention in medical image segmentation. Due to the diversity of organs, effective modelling of multi-scale information and establishing long-range dependencies between pixels are crucial for successful medical image segmentation. However, most studies rely on a fixed single-scale window for modeling, which ignores the potential impact of window size on performance. This limitation can hinder window-based models' ability to fully explore multi-scale and long-range relationships within medical images. To address this issue, we propose a multi-scale reconfiguration self-attention (MSR-SA) module that accurately models multi-scale information and long-range dependencies in medical images. The MSR-SA module first divides the attention heads into multiple groups, each assigned an ascending dilation rate. These groups are then uniformly split into several non-overlapping local windows. Using dilated sampling, we gather the same number of keys to obtain both long-range and multi-scale information. Finally, dynamic information fusion is achieved by integrating features from the sampling points at corresponding positions across different windows. Based on the MSR-SA module, we propose a multi-scale reconfiguration U-Net (MSR-UNet) framework for medical image segmentation. Experiments on the Synapse and automated cardiac diagnosis challenge (ACDC) datasets show that MSR-UNet can achieve satisfactory segmentation results. The code is available at https://github.com/davidsmithwj/MSR-UNet (DOI: 10.5281/zenodo.13969855).
Collapse
Affiliation(s)
- Shuai Wang
- School of Computer Science and Technology, Huaibei Normal University, Huaibei, China
| | - Lei Liu
- School of Computer Science and Technology, Huaibei Normal University, Huaibei, China
- Huaibei Key Laboratory of Digital Multimedia Intelligent Information Processing, Huaibei, China
| | - Jun Wang
- College of Electronic and Information Engineering, Hebei University, Baoding, China
| | - Xinyue Peng
- School of Computer Science and Technology, Huaibei Normal University, Huaibei, China
| | - Baosen Liu
- Huaibei People’s Hospital, Huaibei, China
| |
Collapse
|
4
|
Sendecki A, Ledwoń D, Nycz J, Wąsowska A, Boguszewska-Chachulska A, Mitas AW, Wylęgała E, Teper S. A deep learning approach to explore the association of age-related macular degeneration polygenic risk score with retinal optical coherence tomography: A preliminary study. Acta Ophthalmol 2024; 102:e1029-e1039. [PMID: 38761033 DOI: 10.1111/aos.16710] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/11/2024] [Accepted: 05/06/2024] [Indexed: 05/20/2024]
Abstract
PURPOSE Age-related macular degeneration (AMD) is a complex eye disorder affecting millions worldwide. This article uses deep learning techniques to investigate the relationship between AMD, genetics and optical coherence tomography (OCT) scans. METHODS The cohort consisted of 332 patients, of which 235 were diagnosed with AMD and 97 were controls with no signs of AMD. The genome-wide association studies summary statistics utilized to establish the polygenic risk score (PRS) in relation to AMD were derived from the GERA European study. A PRS estimation based on OCT volumes for both eyes was performed using a proprietary convolutional neural network (CNN) model supported by machine learning models. The method's performance was assessed using numerical evaluation metrics, and the Grad-CAM technique was used to evaluate the results by visualizing the features learned by the model. RESULTS The best results were obtained with the CNN and the Extra Tree regressor (MAE = 0.55, MSE = 0.49, RMSE = 0.70, R2 = 0.34). Extending the feature vector with additional information on AMD diagnosis, age and smoking history improved the results slightly, with mainly AMD diagnosis used by the model (MAE = 0.54, MSE = 0.44, RMSE = 0.66, R2 = 0.42). Grad-CAM heatmap evaluation showed that the model decisions rely on retinal morphology factors relevant to AMD diagnosis. CONCLUSION The developed method allows an efficient PRS estimation from OCT images. A new technique for analysing the association of OCT images with PRS of AMD, using a deep learning approach, may provide an opportunity to discover new associations between genotype-based AMD risk and retinal morphology.
Collapse
Affiliation(s)
- Adam Sendecki
- Chair and Clinical Department of Ophthalmology, Faculty of Medical Sciences in Zabrze, Medical University of Silesia, Katowice, Poland
| | - Daniel Ledwoń
- Faculty of Biomedical Engineering, Silesian University of Technology, Zabrze, Poland
| | - Julia Nycz
- Institute of Biomedical Engineering and Informatics, Technische Universität Ilmenau, Ilmenau, Germany
| | - Anna Wąsowska
- Chair and Clinical Department of Ophthalmology, Faculty of Medical Sciences in Zabrze, Medical University of Silesia, Katowice, Poland
- Genomed S.A., Warszawa, Poland
| | | | - Andrzej W Mitas
- Faculty of Biomedical Engineering, Silesian University of Technology, Zabrze, Poland
| | - Edward Wylęgała
- Chair and Clinical Department of Ophthalmology, Faculty of Medical Sciences in Zabrze, Medical University of Silesia, Katowice, Poland
| | - Sławomir Teper
- Chair and Clinical Department of Ophthalmology, Faculty of Medical Sciences in Zabrze, Medical University of Silesia, Katowice, Poland
| |
Collapse
|
5
|
Sekimitsu S, Shweikh Y, Shareef S, Zhao Y, Elze T, Segrè A, Wiggs J, Zebardast N. Association of retinal optical coherence tomography metrics and polygenic risk scores with cognitive function and future cognitive decline. Br J Ophthalmol 2024; 108:599-606. [PMID: 36990674 DOI: 10.1136/bjo-2022-322762] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/19/2022] [Accepted: 03/15/2023] [Indexed: 03/31/2023]
Abstract
PURPOSE To evaluate the potential of retinal optical coherence tomography (OCT) measurements and polygenic risk scores (PRS) to identify people at risk of cognitive impairment. METHODS Using OCT images from 50 342 UK Biobank participants, we examined associations between retinal layer thickness and genetic risk for neurodegenerative disease and combined these metrics with PRS to predict baseline cognitive function and future cognitive deterioration. Multivariate Cox proportional hazard models were used to predict cognitive performance. P values for retinal thickness analyses are false-discovery-rate-adjusted. RESULTS Higher Alzheimer's disease PRS was associated with a thicker inner nuclear layer (INL), chorio-scleral interface (CSI) and inner plexiform layer (IPL) (all p<0.05). Higher Parkinson's disease PRS was associated with thinner outer plexiform layer (p<0.001). Worse baseline cognitive performance was associated with thinner retinal nerve fibre layer (RNFL) (aOR=1.038, 95% CI (1.029 to 1.047), p<0.001) and photoreceptor (PR) segment (aOR=1.035, 95% CI (1.019 to 1.051), p<0.001), ganglion cell complex (aOR=1.007, 95% CI (1.002 to 1.013), p=0.004) and thicker ganglion cell layer (aOR=0.981, 95% CI (0.967 to 0.995), p=0.009), IPL (aOR=0.976, 95% CI (0.961 to 0.992), p=0.003), INL (aOR=0.923, 95% CI (0.905 to 0.941), p<0.001) and CSI (aOR=0.998, 95% CI (0.997 to 0.999), p<0.001). Worse future cognitive performance was associated with thicker IPL (aOR=0.945, 95% CI (0.915 to 0.999), p=0.045) and CSI (aOR=0.996, 95% CI (0.993 to 0.999) 95% CI, p=0.014). Prediction of cognitive decline was significantly improved with the addition of PRS and retinal measurements. CONCLUSIONS AND RELEVANCE Retinal OCT measurements are significantly associated with genetic risk of neurodegenerative disease and may serve as biomarkers predictive of future cognitive impairment.
Collapse
Affiliation(s)
| | - Yusrah Shweikh
- Department of Ophthalmology, Massachusetts Eye and Ear Infirmary, Boston, Massachusetts, USA
- Sussex Eye Hospital, University Hospitals Sussex NHS Foundation Trust, Sussex, UK
| | - Sarah Shareef
- Department of Ophthalmology, Massachusetts Eye and Ear Infirmary, Boston, Massachusetts, USA
| | - Yan Zhao
- Department of Ophthalmology, Massachusetts Eye and Ear Infirmary, Boston, Massachusetts, USA
- Broad Institute of Harvard and MIT, Cambridge, Massachusetts, USA
| | - Tobias Elze
- Schepens Eye Research Institute, Harvard Medical School, Boston, Massachusetts, USA
| | - Ayellet Segrè
- Department of Ophthalmology, Massachusetts Eye and Ear Infirmary, Boston, Massachusetts, USA
- Broad Institute of Harvard and MIT, Cambridge, Massachusetts, USA
| | - Janey Wiggs
- Department of Ophthalmology, Massachusetts Eye and Ear Infirmary, Boston, Massachusetts, USA
- Broad Institute of Harvard and MIT, Cambridge, Massachusetts, USA
| | - Nazlee Zebardast
- Department of Ophthalmology, Massachusetts Eye and Ear Infirmary, Boston, Massachusetts, USA
| |
Collapse
|
6
|
Tan Y, Shen WD, Wu MY, Liu GN, Zhao SX, Chen Y, Yang KF, Li YJ. Retinal Layer Segmentation in OCT Images With Boundary Regression and Feature Polarization. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:686-700. [PMID: 37725718 DOI: 10.1109/tmi.2023.3317072] [Citation(s) in RCA: 3] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/21/2023]
Abstract
The geometry of retinal layers is an important imaging feature for the diagnosis of some ophthalmic diseases. In recent years, retinal layer segmentation methods for optical coherence tomography (OCT) images have emerged one after another, and huge progress has been achieved. However, challenges due to interference factors such as noise, blurring, fundus effusion, and tissue artifacts remain in existing methods, primarily manifesting as intra-layer false positives and inter-layer boundary deviation. To solve these problems, we propose a method called Tightly combined Cross-Convolution and Transformer with Boundary regression and feature Polarization (TCCT-BP). This method uses a hybrid architecture of CNN and lightweight Transformer to improve the perception of retinal layers. In addition, a feature grouping and sampling method and the corresponding polarization loss function are designed to maximize the differentiation of the feature vectors of different retinal layers, and a boundary regression loss function is devised to constrain the retinal boundary distribution for a better fit to the ground truth. Extensive experiments on four benchmark datasets demonstrate that the proposed method achieves state-of-the-art performance in dealing with problems of false positives and boundary distortion. The proposed method ranked first in the OCT Layer Segmentation task of GOALS challenge held by MICCAI 2022. The source code is available at https://www.github.com/tyb311/TCCT.
Collapse
|
7
|
Zekavat SM, Jorshery SD, Rauscher FG, Horn K, Sekimitsu S, Koyama S, Nguyen TT, Costanzo MC, Jang D, Burtt NP, Kühnapfel A, Shweikh Y, Ye Y, Raghu V, Zhao H, Ghassemi M, Elze T, Segrè AV, Wiggs JL, Del Priore L, Scholz M, Wang JC, Natarajan P, Zebardast N. Phenome- and genome-wide analyses of retinal optical coherence tomography images identify links between ocular and systemic health. Sci Transl Med 2024; 16:eadg4517. [PMID: 38266105 DOI: 10.1126/scitranslmed.adg4517] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2022] [Accepted: 01/03/2024] [Indexed: 01/26/2024]
Abstract
The human retina is a multilayered tissue that offers a unique window into systemic health. Optical coherence tomography (OCT) is widely used in eye care and allows the noninvasive, rapid capture of retinal anatomy in exquisite detail. We conducted genotypic and phenotypic analyses of retinal layer thicknesses using macular OCT images from 44,823 UK Biobank participants. We performed OCT layer cross-phenotype association analyses (OCT-XWAS), associating retinal thicknesses with 1866 incident conditions (median 10-year follow-up) and 88 quantitative traits and blood biomarkers. We performed genome-wide association studies (GWASs), identifying inherited genetic markers that influence retinal layer thicknesses and replicated our associations among the LIFE-Adult Study (N = 6313). Last, we performed a comparative analysis of phenome- and genome-wide associations to identify putative causal links between retinal layer thicknesses and both ocular and systemic conditions. Independent associations with incident mortality were detected for thinner photoreceptor segments (PSs) and, separately, ganglion cell complex layers. Phenotypic associations were detected between thinner retinal layers and ocular, neuropsychiatric, cardiometabolic, and pulmonary conditions. A GWAS of retinal layer thicknesses yielded 259 unique loci. Consistency between epidemiologic and genetic associations suggested links between a thinner retinal nerve fiber layer with glaucoma, thinner PS with age-related macular degeneration, and poor cardiometabolic and pulmonary function with a thinner PS. In conclusion, we identified multiple inherited genetic loci and acquired systemic cardio-metabolic-pulmonary conditions associated with thinner retinal layers and identify retinal layers wherein thinning is predictive of future ocular and systemic conditions.
Collapse
Affiliation(s)
- Seyedeh Maryam Zekavat
- Department of Ophthalmology, Massachusetts Eye and Ear, Harvard Medical School, Boston, MA 02114, USA
- Cardiovascular Research Center, Massachusetts General Hospital, Harvard Medical School, Boston, MA 02114, USA
- Program in Medical and Population Genetics, Broad Institute of MIT and Harvard, Cambridge, MA 02142, USA
| | - Saman Doroodgar Jorshery
- Program in Medical and Population Genetics, Broad Institute of MIT and Harvard, Cambridge, MA 02142, USA
- Departments of Computer Science/Medicine, University of Toronto, Toronto, ON M5S 1A1, Canada
- Vector Institute for Artificial Intelligence, Toronto, ON M5G 1M1, Canada
- Department of Computer Science and Electrical Engineering, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
- Cardiovascular Imaging Research Center, Massachusetts General Hospital, Harvard Medical School, Boston, MA 02114, USA
| | - Franziska G Rauscher
- Institute for Medical Informatics, Statistics, and Epidemiology (IMISE), Leipzig University, Leipzig 04107, Germany
- Leipzig Research Centre for Civilization Diseases (LIFE), Leipzig University, Leipzig 04103, Germany
| | - Katrin Horn
- Institute for Medical Informatics, Statistics, and Epidemiology (IMISE), Leipzig University, Leipzig 04107, Germany
| | | | - Satoshi Koyama
- Cardiovascular Research Center, Massachusetts General Hospital, Harvard Medical School, Boston, MA 02114, USA
- Program in Medical and Population Genetics, Broad Institute of MIT and Harvard, Cambridge, MA 02142, USA
| | - Trang T Nguyen
- Program in Medical and Population Genetics, Broad Institute of MIT and Harvard, Cambridge, MA 02142, USA
| | - Maria C Costanzo
- Program in Medical and Population Genetics, Broad Institute of MIT and Harvard, Cambridge, MA 02142, USA
| | - Dongkeun Jang
- Program in Medical and Population Genetics, Broad Institute of MIT and Harvard, Cambridge, MA 02142, USA
| | - Noël P Burtt
- Program in Medical and Population Genetics, Broad Institute of MIT and Harvard, Cambridge, MA 02142, USA
| | - Andreas Kühnapfel
- Institute for Medical Informatics, Statistics, and Epidemiology (IMISE), Leipzig University, Leipzig 04107, Germany
| | - Yusrah Shweikh
- Department of Ophthalmology, Massachusetts Eye and Ear, Harvard Medical School, Boston, MA 02114, USA
| | - Yixuan Ye
- Computational Biology and Bioinformatics Program, Yale School of Medicine, New Haven, CT 06511, USA
| | - Vineet Raghu
- Cardiovascular Research Center, Massachusetts General Hospital, Harvard Medical School, Boston, MA 02114, USA
- Program in Medical and Population Genetics, Broad Institute of MIT and Harvard, Cambridge, MA 02142, USA
- Cardiovascular Imaging Research Center, Massachusetts General Hospital, Harvard Medical School, Boston, MA 02114, USA
| | - Hongyu Zhao
- Computational Biology and Bioinformatics Program, Yale School of Medicine, New Haven, CT 06511, USA
- School of Public Health, Yale University, New Haven, CT 06510, USA
| | - Marzyeh Ghassemi
- Departments of Computer Science/Medicine, University of Toronto, Toronto, ON M5S 1A1, Canada
- Vector Institute for Artificial Intelligence, Toronto, ON M5G 1M1, Canada
- Department of Computer Science and Electrical Engineering, Massachusetts Institute of Technology, Cambridge, MA 02139, USA
| | - Tobias Elze
- Department of Ophthalmology, Massachusetts Eye and Ear, Harvard Medical School, Boston, MA 02114, USA
| | - Ayellet V Segrè
- Department of Ophthalmology, Massachusetts Eye and Ear, Harvard Medical School, Boston, MA 02114, USA
- Program in Medical and Population Genetics, Broad Institute of MIT and Harvard, Cambridge, MA 02142, USA
| | - Janey L Wiggs
- Department of Ophthalmology, Massachusetts Eye and Ear, Harvard Medical School, Boston, MA 02114, USA
- Program in Medical and Population Genetics, Broad Institute of MIT and Harvard, Cambridge, MA 02142, USA
| | - Lucian Del Priore
- Department of Ophthalmology and Visual Science, Yale School of Medicine, New Haven, CT 06510, USA
| | - Markus Scholz
- Institute for Medical Informatics, Statistics, and Epidemiology (IMISE), Leipzig University, Leipzig 04107, Germany
- Leipzig Research Centre for Civilization Diseases (LIFE), Leipzig University, Leipzig 04103, Germany
| | - Jay C Wang
- Department of Ophthalmology and Visual Science, Yale School of Medicine, New Haven, CT 06510, USA
- Northern California Retina Vitreous Associates, Mountain View, CA 94040, USA
| | - Pradeep Natarajan
- Cardiovascular Research Center, Massachusetts General Hospital, Harvard Medical School, Boston, MA 02114, USA
- Program in Medical and Population Genetics, Broad Institute of MIT and Harvard, Cambridge, MA 02142, USA
| | - Nazlee Zebardast
- Department of Ophthalmology, Massachusetts Eye and Ear, Harvard Medical School, Boston, MA 02114, USA
- Program in Medical and Population Genetics, Broad Institute of MIT and Harvard, Cambridge, MA 02142, USA
| |
Collapse
|
8
|
Wang YZ, Juroch K, Birch DG. Deep Learning-Assisted Measurements of Photoreceptor Ellipsoid Zone Area and Outer Segment Volume as Biomarkers for Retinitis Pigmentosa. Bioengineering (Basel) 2023; 10:1394. [PMID: 38135984 PMCID: PMC10740805 DOI: 10.3390/bioengineering10121394] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2023] [Revised: 11/13/2023] [Accepted: 11/29/2023] [Indexed: 12/24/2023] Open
Abstract
The manual segmentation of retinal layers from OCT scan images is time-consuming and costly. The deep learning approach has potential for the automatic delineation of retinal layers to significantly reduce the burden of human graders. In this study, we compared deep learning model (DLM) segmentation with manual correction (DLM-MC) to conventional manual grading (MG) for the measurements of the photoreceptor ellipsoid zone (EZ) area and outer segment (OS) volume in retinitis pigmentosa (RP) to assess whether DLM-MC can be a new gold standard for retinal layer segmentation and for the measurement of retinal layer metrics. Ninety-six high-speed 9 mm 31-line volume scans obtained from 48 patients with RPGR-associated XLRP were selected based on the following criteria: the presence of an EZ band within the scan limit and a detectable EZ in at least three B-scans in a volume scan. All the B-scan images in each volume scan were manually segmented for the EZ and proximal retinal pigment epithelium (pRPE) by two experienced human graders to serve as the ground truth for comparison. The test volume scans were also segmented by a DLM and then manually corrected for EZ and pRPE by the same two graders to obtain DLM-MC segmentation. The EZ area and OS volume were determined by interpolating the discrete two-dimensional B-scan EZ-pRPE layer over the scan area. Dice similarity, Bland-Altman analysis, correlation, and linear regression analyses were conducted to assess the agreement between DLM-MC and MG for the EZ area and OS volume measurements. For the EZ area, the overall mean dice score (SD) between DLM-MC and MG was 0.8524 (0.0821), which was comparable to 0.8417 (0.1111) between two MGs. For the EZ area > 1 mm2, the average dice score increased to 0.8799 (0.0614). When comparing DLM-MC to MG, the Bland-Altman plots revealed a mean difference (SE) of 0.0132 (0.0953) mm2 and a coefficient of repeatability (CoR) of 1.8303 mm2 for the EZ area and a mean difference (SE) of 0.0080 (0.0020) mm3 and a CoR of 0.0381 mm3 for the OS volume. The correlation coefficients (95% CI) were 0.9928 (0.9892-0.9952) and 0.9938 (0.9906-0.9958) for the EZ area and OS volume, respectively. The linear regression slopes (95% CI) were 0.9598 (0.9399-0.9797) and 1.0104 (0.9909-1.0298), respectively. The results from this study suggest that the manual correction of deep learning model segmentation can generate EZ area and OS volume measurements in excellent agreement with those of conventional manual grading in RP. Because DLM-MC is more efficient for retinal layer segmentation from OCT scan images, it has the potential to reduce the burden of human graders in obtaining quantitative measurements of biomarkers for assessing disease progression and treatment outcomes in RP.
Collapse
Affiliation(s)
- Yi-Zhong Wang
- Retina Foundation of the Southwest, 9600 North Central Expressway, Suite 200, Dallas, TX 75231, USA; (K.J.); (D.G.B.)
- Department of Ophthalmology, University of Texas Southwestern Medical Center, 5323 Harry Hines Blvd, Dallas, TX 75390, USA
| | - Katherine Juroch
- Retina Foundation of the Southwest, 9600 North Central Expressway, Suite 200, Dallas, TX 75231, USA; (K.J.); (D.G.B.)
| | - David Geoffrey Birch
- Retina Foundation of the Southwest, 9600 North Central Expressway, Suite 200, Dallas, TX 75231, USA; (K.J.); (D.G.B.)
- Department of Ophthalmology, University of Texas Southwestern Medical Center, 5323 Harry Hines Blvd, Dallas, TX 75390, USA
| |
Collapse
|
9
|
Zheng T, Li J, Tian H, Wu Q. The Process Analysis Method of SAR Target Recognition in Pre-Trained CNN Models. SENSORS (BASEL, SWITZERLAND) 2023; 23:6461. [PMID: 37514755 PMCID: PMC10384806 DOI: 10.3390/s23146461] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/18/2023] [Revised: 07/12/2023] [Accepted: 07/15/2023] [Indexed: 07/30/2023]
Abstract
Recently, attention has been paid to the convolutional neural network (CNN) based synthetic aperture radar (SAR) target recognition method. Because of its advantages of automatic feature extraction and the preservation of translation invariance, the recognition accuracies are stronger than traditional methods. However, similar to other deep learning models, CNN is a "black-box" model, whose working process is vague. It is difficult to locate the decision reasons. Because of this, we focus on the process analysis of a pre-trained CNN model. The role of the processing to feature extraction and final recognition decision is discussed. The discussed components of CNN models are convolution, activation function, and full connection. Here, the convolution processing can be deemed as image filtering. The activation function provides a nonlinear element of processing. Moreover, the fully connected layers can also further extract features. In the experiment, four classical CNN models, i.e., AlexNet, VGG16, GoogLeNet, and ResNet-50, are trained by public MSTAR data, which can realize ten-category SAR target recognition. These pre-trained CNN models are processing objects of the proposed process analysis method. After the analysis, the content of the SAR image target features concerned by these pre-trained CNN models is further clarified. In summary, we provide a paradigm to process the analysis of pre-trained CNN models used for SAR target recognition in this paper. To some degree, the adaptability of these models to SAR images is verified.
Collapse
Affiliation(s)
- Tong Zheng
- School of Artificial Intelligence, Beijing Technology and Business University, Beijing 100048, China
| | - Jin Li
- School of Artificial Intelligence, Beijing Technology and Business University, Beijing 100048, China
| | - Hao Tian
- School of Artificial Intelligence, Beijing Technology and Business University, Beijing 100048, China
| | - Qing Wu
- Heilongjiang Province Key Laboratory of Laser Spectroscopy Technology and Application, Harbin University of Science and Technology, Harbin 150006, China
| |
Collapse
|
10
|
Zekavat SM, Jorshery SD, Shweikh Y, Horn K, Rauscher FG, Sekimitsu S, Kayoma S, Ye Y, Raghu V, Zhao H, Ghassemi M, Elze T, Segrè AV, Wiggs JL, Scholz M, Priore LD, Wang JC, Natarajan P, Zebardast N. Insights into human health from phenome- and genome-wide analyses of UK Biobank retinal optical coherence tomography phenotypes. MEDRXIV : THE PREPRINT SERVER FOR HEALTH SCIENCES 2023:2023.05.16.23290063. [PMID: 37292770 PMCID: PMC10246137 DOI: 10.1101/2023.05.16.23290063] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
The human retina is a complex multi-layered tissue which offers a unique window into systemic health and disease. Optical coherence tomography (OCT) is widely used in eye care and allows the non-invasive, rapid capture of retinal measurements in exquisite detail. We conducted genome- and phenome-wide analyses of retinal layer thicknesses using macular OCT images from 44,823 UK Biobank participants. We performed phenome-wide association analyses, associating retinal thicknesses with 1,866 incident ICD-based conditions (median 10-year follow-up) and 88 quantitative traits and blood biomarkers. We performed genome-wide association analyses, identifying inherited genetic markers which influence the retina, and replicated our associations among 6,313 individuals from the LIFE-Adult Study. And lastly, we performed comparative association of phenome- and genome- wide associations to identify putative causal links between systemic conditions, retinal layer thicknesses, and ocular disease. Independent associations with incident mortality were detected for photoreceptor thinning and ganglion cell complex thinning. Significant phenotypic associations were detected between retinal layer thinning and ocular, neuropsychiatric, cardiometabolic and pulmonary conditions. Genome-wide association of retinal layer thicknesses yielded 259 loci. Consistency between epidemiologic and genetic associations suggested putative causal links between thinning of the retinal nerve fiber layer with glaucoma, photoreceptor segment with AMD, as well as poor cardiometabolic and pulmonary function with PS thinning, among other findings. In conclusion, retinal layer thinning predicts risk of future ocular and systemic disease. Furthermore, systemic cardio-metabolic-pulmonary conditions promote retinal thinning. Retinal imaging biomarkers, integrated into electronic health records, may inform risk prediction and potential therapeutic strategies.
Collapse
Affiliation(s)
- Seyedeh Maryam Zekavat
- Department of Ophthalmology, Massachusetts Eye and Ear, Harvard Medical School, Boston, MA, USA
- Cardiovascular Research Center, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
- Program in Medical and Population Genetics and Cardiovascular Disease Initiative, Broad Institute of MIT and Harvard, Cambridge, MA, USA
| | - Saman Doroodgar Jorshery
- Program in Medical and Population Genetics and Cardiovascular Disease Initiative, Broad Institute of MIT and Harvard, Cambridge, MA, USA
- Departments of Computer Science/Medicine, University of Toronto, Toronto, Canada
- Vector Institute for Artificial Intelligence, Toronto, ON, Canada
- Department of Computer Science and Electrical Engineering, Massachusetts Institute of Technology, Cambridge, MA, USA
- Cardiovascular Imaging Research Center, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
| | - Yusrah Shweikh
- Department of Ophthalmology, Massachusetts Eye and Ear, Harvard Medical School, Boston, MA, USA
| | - Katrin Horn
- Institute for Medical Informatics, Statistics and Epidemiology University of Leipzig, Germany and Leipzig Research Centre for Civilization Diseases (LIFE), Leipzig University, Leipzig, Germany
| | - Franziska G. Rauscher
- Institute for Medical Informatics, Statistics and Epidemiology University of Leipzig, Germany and Leipzig Research Centre for Civilization Diseases (LIFE), Leipzig University, Leipzig, Germany
| | | | - Satoshi Kayoma
- Cardiovascular Research Center, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
- Program in Medical and Population Genetics and Cardiovascular Disease Initiative, Broad Institute of MIT and Harvard, Cambridge, MA, USA
| | - Yixuan Ye
- Computational Biology and Bioinformatics Program, Yale University, New Haven, CT, USA
| | - Vineet Raghu
- Cardiovascular Research Center, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
- Program in Medical and Population Genetics and Cardiovascular Disease Initiative, Broad Institute of MIT and Harvard, Cambridge, MA, USA
- Cardiovascular Imaging Research Center, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
| | - Hongyu Zhao
- Computational Biology and Bioinformatics Program, Yale University, New Haven, CT, USA
- School of Public Health, Yale University, New Haven, CT, USA
| | - Marzyeh Ghassemi
- Departments of Computer Science/Medicine, University of Toronto, Toronto, Canada
- Vector Institute for Artificial Intelligence, Toronto, ON, Canada
- Department of Computer Science and Electrical Engineering, Massachusetts Institute of Technology, Cambridge, MA, USA
| | - Tobias Elze
- Department of Ophthalmology, Massachusetts Eye and Ear, Harvard Medical School, Boston, MA, USA
| | - Ayellet V. Segrè
- Department of Ophthalmology, Massachusetts Eye and Ear, Harvard Medical School, Boston, MA, USA
| | - Janey L. Wiggs
- Department of Ophthalmology, Massachusetts Eye and Ear, Harvard Medical School, Boston, MA, USA
| | - Markus Scholz
- Institute for Medical Informatics, Statistics and Epidemiology University of Leipzig, Germany and Leipzig Research Centre for Civilization Diseases (LIFE), Leipzig University, Leipzig, Germany
| | - Lucian Del Priore
- Department of Ophthalmology and Visual Science, Yale School of Medicine, New Haven, CT, USA
| | - Jay C. Wang
- Department of Ophthalmology and Visual Science, Yale School of Medicine, New Haven, CT, USA
- Northern California Retina Vitreous Associates, Mountain View, CA
| | - Pradeep Natarajan
- Cardiovascular Research Center, Massachusetts General Hospital, Harvard Medical School, Boston, MA, USA
- Program in Medical and Population Genetics and Cardiovascular Disease Initiative, Broad Institute of MIT and Harvard, Cambridge, MA, USA
| | - Nazlee Zebardast
- Department of Ophthalmology, Massachusetts Eye and Ear, Harvard Medical School, Boston, MA, USA
- Program in Medical and Population Genetics and Cardiovascular Disease Initiative, Broad Institute of MIT and Harvard, Cambridge, MA, USA
| |
Collapse
|
11
|
Ramajayam S, Rajavel S, Samidurai R, Cao Y. Finite-Time Synchronization for T–S Fuzzy Complex-Valued Inertial Delayed Neural Networks Via Decomposition Approach. Neural Process Lett 2023. [DOI: 10.1007/s11063-022-11117-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/10/2023]
|
12
|
Bao D, Wang L, Zhou X, Yang S, He K, Xu M. Automated detection and growth tracking of 3D bio-printed organoid clusters using optical coherence tomography with deep convolutional neural networks. Front Bioeng Biotechnol 2023; 11:1133090. [PMID: 37122853 PMCID: PMC10130530 DOI: 10.3389/fbioe.2023.1133090] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/28/2022] [Accepted: 03/31/2023] [Indexed: 05/02/2023] Open
Abstract
Organoids are advancing the development of accurate prediction of drug efficacy and toxicity in vitro. These advancements are attributed to the ability of organoids to recapitulate key structural and functional features of organs and parent tumor. Specifically, organoids are self-organized assembly with a multi-scale structure of 30-800 μm, which exacerbates the difficulty of non-destructive three-dimensional (3D) imaging, tracking and classification analysis for organoid clusters by traditional microscopy techniques. Here, we devise a 3D imaging, segmentation and analysis method based on Optical coherence tomography (OCT) technology and deep convolutional neural networks (CNNs) for printed organoid clusters (Organoid Printing and optical coherence tomography-based analysis, OPO). The results demonstrate that the organoid scale influences the segmentation effect of the neural network. The multi-scale information-guided optimized EGO-Net we designed achieves the best results, especially showing better recognition workout for the biologically significant organoid with diameter ≥50 μm than other neural networks. Moreover, OPO achieves to reconstruct the multiscale structure of organoid clusters within printed microbeads and calibrate the printing errors by segmenting the printed microbeads edges. Overall, the classification, tracking and quantitative analysis based on image reveal that the growth process of organoid undergoes morphological changes such as volume growth, cavity creation and fusion, and quantitative calculation of the volume demonstrates that the growth rate of organoid is associated with the initial scale. The new method we proposed enable the study of growth, structural evolution and heterogeneity for the organoid cluster, which is valuable for drug screening and tumor drug sensitivity detection based on organoids.
Collapse
Affiliation(s)
- Di Bao
- School of Automation, Hangzhou Dianzi University, Hangzhou, China
| | - Ling Wang
- School of Automation, Hangzhou Dianzi University, Hangzhou, China
- Key Laboratory of Medical Information and 3D Bioprinting of Zhejiang Province, Hangzhou, China
- *Correspondence: Ling Wang, ; Mingen Xu,
| | - Xiaofei Zhou
- School of Automation, Hangzhou Dianzi University, Hangzhou, China
- Key Laboratory of Medical Information and 3D Bioprinting of Zhejiang Province, Hangzhou, China
| | - Shanshan Yang
- School of Automation, Hangzhou Dianzi University, Hangzhou, China
- Key Laboratory of Medical Information and 3D Bioprinting of Zhejiang Province, Hangzhou, China
| | - Kangxin He
- Key Laboratory of Medical Information and 3D Bioprinting of Zhejiang Province, Hangzhou, China
| | - Mingen Xu
- School of Automation, Hangzhou Dianzi University, Hangzhou, China
- Key Laboratory of Medical Information and 3D Bioprinting of Zhejiang Province, Hangzhou, China
- *Correspondence: Ling Wang, ; Mingen Xu,
| |
Collapse
|
13
|
Computation and memory optimized spectral domain convolutional neural network for throughput and energy-efficient inference. APPL INTELL 2023; 53:4499-4523. [PMID: 35730044 PMCID: PMC9188280 DOI: 10.1007/s10489-022-03756-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 05/10/2022] [Indexed: 02/04/2023]
Abstract
Conventional convolutional neural networks (CNNs) present a high computational workload and memory access cost (CMC). Spectral domain CNNs (SpCNNs) offer a computationally efficient approach to compute CNN training and inference. This paper investigates CMC of SpCNNs and its contributing components analytically and then proposes a methodology to optimize CMC, under three strategies, to enhance inference performance. In this methodology, output feature map (OFM) size, OFM depth or both are progressively reduced under an accuracy constraint to compute performance-optimized CNN inference. Before conducting training or testing, it can provide designers guidelines and preliminary insights regarding techniques for optimum performance, least degradation in accuracy and a balanced performance-accuracy trade-off. This methodology was evaluated on MNIST and Fashion MNIST datasets using LeNet-5 and AlexNet architectures. When compared to state-of-the-art SpCNN models, LeNet-5 achieves up to 4.2× (batch inference) and 4.1× (single-image inference) higher throughputs and 10.5× (batch inference) and 4.2× (single-image inference) greater energy efficiency at a maximum loss of 3% in test accuracy. When compared to the baseline model used in this study, AlexNet delivers 11.6× (batch inference) and 5× (single-image inference) increased throughput and 25× (batch inference) and 8.8× (single-image inference) more energy-efficient inference with just 4.4% reduction in accuracy.
Collapse
|
14
|
Vachmanus S, Noraset T, Piyanonpong W, Rattananukrom T, Tuarob S. DeepMetaForge: A Deep Vision-Transformer Metadata-Fusion Network for Automatic Skin Lesion Classification. IEEE ACCESS 2023; 11:145467-145484. [DOI: 10.1109/access.2023.3345225] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/05/2025]
Affiliation(s)
- Sirawich Vachmanus
- Faculty of Information and Communication Technology, Mahidol University, Nakhon Pathom, Thailand
| | - Thanapon Noraset
- Faculty of Information and Communication Technology, Mahidol University, Nakhon Pathom, Thailand
| | - Waritsara Piyanonpong
- Division of Dermatology, Faculty of Medicine, Ramathibodi Hospital, Mahidol University, Bangkok, Thailand
| | - Teerapong Rattananukrom
- Division of Dermatology, Faculty of Medicine, Ramathibodi Hospital, Mahidol University, Bangkok, Thailand
| | - Suppawong Tuarob
- Faculty of Information and Communication Technology, Mahidol University, Nakhon Pathom, Thailand
| |
Collapse
|
15
|
He X, Zhong Z, Fang L, He M, Sebe N. Structure-Guided Cross-Attention Network for Cross-Domain OCT Fluid Segmentation. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2022; 32:309-320. [PMID: 37015552 DOI: 10.1109/tip.2022.3228163] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Accurate retinal fluid segmentation on Optical Coherence Tomography (OCT) images plays an important role in diagnosing and treating various eye diseases. The art deep models have shown promising performance on OCT image segmentation given pixel-wise annotated training data. However, the learned model will achieve poor performance on OCT images that are obtained from different devices (domains) due to the domain shift issue. This problem largely limits the real-world application of OCT image segmentation since the types of devices usually are different in each hospital. In this paper, we study the task of cross-domain OCT fluid segmentation, where we are given a labeled dataset of the source device (domain) and an unlabeled dataset of the target device (domain). The goal is to learn a model that can perform well on the target domain. To solve this problem, in this paper, we propose a novel Structure-guided Cross-Attention Network (SCAN), which leverages the retinal layer structure to facilitate domain alignment. Our SCAN is inspired by the fact that the retinal layer structure is robust to domains and can reflect regions that are important to fluid segmentation. In light of this, we build our SCAN in a multi-task manner by jointly learning the retinal structure prediction and fluid segmentation. To exploit the mutual benefit between layer structure and fluid segmentation, we further introduce a cross-attention module to measure the correlation between the layer-specific feature and the fluid-specific feature encouraging the model to concentrate on highly relative regions during domain alignment. Moreover, an adaptation difficulty map is evaluated based on the retinal structure predictions from different domains, which enforces the model focus on hard regions during structure-aware adversarial learning. Extensive experiments on the three domains of the RETOUCH dataset demonstrate the effectiveness of the proposed method and show that our approach produces state-of-the-art performance on cross-domain OCT fluid segmentation.
Collapse
|
16
|
An Intuitionistic Fuzzy Random Vector Functional Link Classifier. Neural Process Lett 2022. [DOI: 10.1007/s11063-022-11043-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|
17
|
Self-supervised patient-specific features learning for OCT image classification. Med Biol Eng Comput 2022; 60:2851-2863. [DOI: 10.1007/s11517-022-02627-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/13/2021] [Accepted: 04/28/2022] [Indexed: 11/26/2022]
|
18
|
|
19
|
Artificial Intelligence in Eye Disease: Recent Developments, Applications, and Surveys. Diagnostics (Basel) 2022; 12:diagnostics12081927. [PMID: 36010277 PMCID: PMC9406878 DOI: 10.3390/diagnostics12081927] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2022] [Revised: 08/05/2022] [Accepted: 08/09/2022] [Indexed: 11/16/2022] Open
Abstract
Artificial intelligence (AI) has expanded by finding applications in medical diagnosis for clinical support systems [...]
Collapse
|
20
|
Cao G, Zhang S, Mao H, Wu Y, Wang D, Dai C. A single-step regression method based on transformer for retinal layer segmentation. Phys Med Biol 2022; 67. [DOI: 10.1088/1361-6560/ac799a] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/28/2022] [Accepted: 06/16/2022] [Indexed: 11/11/2022]
Abstract
Abstract
The shape and structure of retinal layers are basic characteristics for the diagnosis of many ophthalmological diseases. Based on B-Scans of optical coherence tomography, most of retinal layer segmentation methods are composed of two-steps: classifying pixels and extracting retinal layers, in which the optimization of two independent steps decreases the accuracy. Although the methods based on deep learning are highly accurate, they require a large amount of labeled data. This paper proposes a single-step method based on transformer for retinal layer segmentation, which is trained by axial data (A-Scans), to obtain the boundary of each layer. The proposed method was evaluated on two public data sets. The first one contains eight retinal layer boundaries for diabetic macular edema, and the second one contains nine retinal layer boundaries for healthy controls and subjects with multiple sclerosis. Its absolute average distance errors are 0.99 pixels and 3.67 pixels, respectively, for the two sets, and its root mean square error is 1.29 pixels for the latter set. In addition, its accuracy is acceptable even if the training data is reduced to 0.3. The proposed method achieves state-of-the-art performance while maintaining the correct topology and requires less labeled data.
Collapse
|
21
|
ACP: Automatic Channel Pruning Method by Introducing Additional Loss for Deep Neural Networks. Neural Process Lett 2022. [DOI: 10.1007/s11063-022-10926-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
22
|
Liu H, Ding J, Xie X, Jiang X, Zhao Y, Wang X. Scalable multi-task Gaussian processes with neural embedding of coregionalization. Knowl Based Syst 2022. [DOI: 10.1016/j.knosys.2022.108775] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|
23
|
A Novel Focal Ordinal Loss for Assessment of Knee Osteoarthritis Severity. Neural Process Lett 2022. [DOI: 10.1007/s11063-022-10857-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
|
24
|
Yadav SK, Kafieh R, Zimmermann HG, Kauer-Bonin J, Nouri-Mahdavi K, Mohammadzadeh V, Shi L, Kadas EM, Paul F, Motamedi S, Brandt AU. Intraretinal Layer Segmentation Using Cascaded Compressed U-Nets. J Imaging 2022; 8:139. [PMID: 35621903 PMCID: PMC9146486 DOI: 10.3390/jimaging8050139] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/31/2022] [Revised: 04/23/2022] [Accepted: 05/03/2022] [Indexed: 12/24/2022] Open
Abstract
Reliable biomarkers quantifying neurodegeneration and neuroinflammation in central nervous system disorders such as Multiple Sclerosis, Alzheimer's dementia or Parkinson's disease are an unmet clinical need. Intraretinal layer thicknesses on macular optical coherence tomography (OCT) images are promising noninvasive biomarkers querying neuroretinal structures with near cellular resolution. However, changes are typically subtle, while tissue gradients can be weak, making intraretinal segmentation a challenging task. A robust and efficient method that requires no or minimal manual correction is an unmet need to foster reliable and reproducible research as well as clinical application. Here, we propose and validate a cascaded two-stage network for intraretinal layer segmentation, with both networks being compressed versions of U-Net (CCU-INSEG). The first network is responsible for retinal tissue segmentation from OCT B-scans. The second network segments eight intraretinal layers with high fidelity. At the post-processing stage, we introduce Laplacian-based outlier detection with layer surface hole filling by adaptive non-linear interpolation. Additionally, we propose a weighted version of focal loss to minimize the foreground-background pixel imbalance in the training data. We train our method using 17,458 B-scans from patients with autoimmune optic neuropathies, i.e., multiple sclerosis, and healthy controls. Voxel-wise comparison against manual segmentation produces a mean absolute error of 2.3 μm, outperforming current state-of-the-art methods on the same data set. Voxel-wise comparison against external glaucoma data leads to a mean absolute error of 2.6 μm when using the same gold standard segmentation approach, and 3.7 μm mean absolute error in an externally segmented data set. In scans from patients with severe optic atrophy, 3.5% of B-scan segmentation results were rejected by an experienced grader, whereas this was the case in 41.4% of B-scans segmented with a graph-based reference method. The validation results suggest that the proposed method can robustly segment macular scans from eyes with even severe neuroretinal changes.
Collapse
Affiliation(s)
- Sunil Kumar Yadav
- Experimental and Clinical Research Center, Max Delbrück Center for Molecular Medicine and Charité-Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin and Humboldt-Universität zu Berlin, 13125 Berlin, Germany; (S.K.Y.); (R.K.); (H.G.Z.); (J.K.-B.); (F.P.); (S.M.)
- Nocturne GmbH, 10119 Berlin, Germany;
| | - Rahele Kafieh
- Experimental and Clinical Research Center, Max Delbrück Center for Molecular Medicine and Charité-Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin and Humboldt-Universität zu Berlin, 13125 Berlin, Germany; (S.K.Y.); (R.K.); (H.G.Z.); (J.K.-B.); (F.P.); (S.M.)
| | - Hanna Gwendolyn Zimmermann
- Experimental and Clinical Research Center, Max Delbrück Center for Molecular Medicine and Charité-Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin and Humboldt-Universität zu Berlin, 13125 Berlin, Germany; (S.K.Y.); (R.K.); (H.G.Z.); (J.K.-B.); (F.P.); (S.M.)
| | - Josef Kauer-Bonin
- Experimental and Clinical Research Center, Max Delbrück Center for Molecular Medicine and Charité-Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin and Humboldt-Universität zu Berlin, 13125 Berlin, Germany; (S.K.Y.); (R.K.); (H.G.Z.); (J.K.-B.); (F.P.); (S.M.)
- Nocturne GmbH, 10119 Berlin, Germany;
| | - Kouros Nouri-Mahdavi
- Glaucoma Division, Stein Eye Institute, David Geffen School of Medicine, University of California Los Angeles, Los Angeles, CA 90095, USA; (K.N.-M.); (V.M.); (L.S.)
| | - Vahid Mohammadzadeh
- Glaucoma Division, Stein Eye Institute, David Geffen School of Medicine, University of California Los Angeles, Los Angeles, CA 90095, USA; (K.N.-M.); (V.M.); (L.S.)
| | - Lynn Shi
- Glaucoma Division, Stein Eye Institute, David Geffen School of Medicine, University of California Los Angeles, Los Angeles, CA 90095, USA; (K.N.-M.); (V.M.); (L.S.)
| | | | - Friedemann Paul
- Experimental and Clinical Research Center, Max Delbrück Center for Molecular Medicine and Charité-Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin and Humboldt-Universität zu Berlin, 13125 Berlin, Germany; (S.K.Y.); (R.K.); (H.G.Z.); (J.K.-B.); (F.P.); (S.M.)
- Department of Neurology, Charité-Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin and Humboldt-Universität zu Berlin, 10098 Berlin, Germany
| | - Seyedamirhosein Motamedi
- Experimental and Clinical Research Center, Max Delbrück Center for Molecular Medicine and Charité-Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin and Humboldt-Universität zu Berlin, 13125 Berlin, Germany; (S.K.Y.); (R.K.); (H.G.Z.); (J.K.-B.); (F.P.); (S.M.)
| | - Alexander Ulrich Brandt
- Experimental and Clinical Research Center, Max Delbrück Center for Molecular Medicine and Charité-Universitätsmedizin Berlin, Corporate Member of Freie Universität Berlin and Humboldt-Universität zu Berlin, 13125 Berlin, Germany; (S.K.Y.); (R.K.); (H.G.Z.); (J.K.-B.); (F.P.); (S.M.)
- Department of Neurology, University of California Irvine, Irvine, CA 92697, USA
| |
Collapse
|
25
|
Zhang J, Liu Y, Guo C, Zhan J. Optimized segmentation with image inpainting for semantic mapping in dynamic scenes. APPL INTELL 2022. [DOI: 10.1007/s10489-022-03487-3] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/02/2022]
|
26
|
PERSIST: Improving micro-expression spotting using better feature encodings and multi-scale Gaussian TCN. APPL INTELL 2022. [DOI: 10.1007/s10489-022-03553-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 09/29/2022]
|
27
|
Yang J, Tao Y, Xu Q, Zhang Y, Ma X, Yuan S, Chen Q. Self-Supervised Sequence Recovery for Semi-Supervised Retinal Layer Segmentation. IEEE J Biomed Health Inform 2022; 26:3872-3883. [PMID: 35412994 DOI: 10.1109/jbhi.2022.3166778] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
Automated layer segmentation plays an important role for retinal disease diagnosis in optical coherence tomography (OCT) images. However, the severe retinal diseases result in the performance degeneration of automated layer segmentation approaches. In this paper, we present a robust semi-supervised retinal layer segmentation network to relieve the model failures on abnormal retinas, in which we obtain the lesion features from the labeled images with disease-balanced distribution, and utilize the unlabeled images to supplement the layer structure information. Specifically, in our proposed method, the cross-consistency training is utilized over the predictions of the different decoders, and we enforce a consistency between different decoder predictions to improve the encoders representation. Then, we proposed a sequence prediction branch based on self-supervised manner, which is designed to predict the position of each jigsaw puzzle to obtain sensory perception of the retinal layer structure. To this task, a layer spatial pyramid pooling (LSPP) module is designed to extract multi-scale layer spatial features. Furthermore, we use the optical coherence tomography angiography (OCTA) to supplement the information damaged by diseases. The experimental results validate that our method achieves more robust results compared with current supervised segmentation methods. Meanwhile, advanced segmentation performance can be obtained compared with state-of-the-art semi-supervised segmentation methods.
Collapse
|
28
|
Gour N, Tanveer M, Khanna P. Challenges for ocular disease identification in the era of artificial intelligence. Neural Comput Appl 2022. [DOI: 10.1007/s00521-021-06770-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/19/2022]
|
29
|
Nazir A, Cheema MN, Sheng B, Li P, Li H, Xue G, Qin J, Kim J, Feng DD. ECSU-Net: An Embedded Clustering Sliced U-Net Coupled With Fusing Strategy for Efficient Intervertebral Disc Segmentation and Classification. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2022; 31:880-893. [PMID: 34951844 DOI: 10.1109/tip.2021.3136619] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Automatic vertebra segmentation from computed tomography (CT) image is the very first and a decisive stage in vertebra analysis for computer-based spinal diagnosis and therapy support system. However, automatic segmentation of vertebra remains challenging due to several reasons, including anatomic complexity of spine, unclear boundaries of the vertebrae associated with spongy and soft bones. Based on 2D U-Net, we have proposed an Embedded Clustering Sliced U-Net (ECSU-Net). ECSU-Net comprises of three modules named segmentation, intervertebral disc extraction (IDE) and fusion. The segmentation module follows an instance embedding clustering approach, where our three sliced sub-nets use axis of CT images to generate a coarse 2D segmentation along with embedding space with the same size of the input slices. Our IDE module is designed to classify vertebra and find the inter-space between two slices of segmented spine. Our fusion module takes the coarse segmentation (2D) and outputs the refined 3D results of vertebra. A novel adaptive discriminative loss (ADL) function is introduced to train the embedding space for clustering. In the fusion strategy, three modules are integrated via a learnable weight control component, which adaptively sets their contribution. We have evaluated classical and deep learning methods on Spineweb dataset-2. ECSU-Net has provided comparable performance to previous neural network based algorithms achieving the best segmentation dice score of 95.60% and classification accuracy of 96.20%, while taking less time and computation resources.
Collapse
|
30
|
Xiao X, Xu Y. Multi-target regression via self-parameterized Lasso and refactored target space. APPL INTELL 2021. [DOI: 10.1007/s10489-021-02238-0] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/24/2022]
|
31
|
Detection of Copy-Move Forgery in Digital Image Using Multi-scale, Multi-stage Deep Learning Model. Neural Process Lett 2021. [DOI: 10.1007/s11063-021-10620-9] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/20/2022]
|
32
|
Wang B, Wei W, Qiu S, Wang S, Li D, He H. Boundary Aware U-Net for Retinal Layers Segmentation in Optical Coherence Tomography Images. IEEE J Biomed Health Inform 2021; 25:3029-3040. [PMID: 33729959 DOI: 10.1109/jbhi.2021.3066208] [Citation(s) in RCA: 26] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Abstract
Retinal layers segmentation in optical coherence tomography (OCT) images is a critical step in the diagnosis of numerous ocular diseases. Automatic layers segmentation requires separating each individual layer instance with accurate boundary detection, but remains a challenging task since it suffers from speckle noise, intensity inhomogeneity, and the low contrast around boundary. In this work, we proposed a boundary aware U-Net (BAU-Net) for retinal layers segmentation by detecting accurate boundary. Based on encoder-decoder architecture, we design a dual tasks framework with low-level outputs for boundary detection and high-level outputs for layers segmentation. Specifically, we first use the multi-scale input strategy to enrich the spatial information in the deep features of encoder. For low-level features from encoder, we design an edge aware (EA) module in skip connection to extract the pure edge features. Then, a U-structure feature enhanced (UFE) module is designed in all skip connections to enlarge the features receptive fields from the encoder. Besides, a canny edge fusion (CEF) module is introduced to aforementioned architecture, which can fuse the priory edge information from segmentation task to boundary detection branch for a better predication. Furthermore, we model each boundary as a vertical coordinates distribution for boundary detection. Based on this distribution, a topology guarantee loss with combined A-scan regression loss and structure loss is proposed to make an accurate and guaranteed topological boundary set. The method is evaluated on two public datasets and the results demonstrate that the BAU-Net achieves promising performance than other state-of-the-art methods.
Collapse
|
33
|
Qi H, Zhang G, Jia H, Xing Z. A hybrid equilibrium optimizer algorithm for multi-level image segmentation. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2021; 18:4648-4678. [PMID: 34198458 DOI: 10.3934/mbe.2021236] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Threshlod image segmentation is a classic method of color image segmentation. In this paper, we propose a hybrid equilibrium optimizer algorithm for multi-level image segmentation. When multi-level threshold method calculates the neighborhood mean and median of color image, it takes huge challenge to find the optimal threshold. We use the proposed method to optimize the multi-level threshold method and get the optimal threshold from the color image. In order to test the performance of the proposed method, we select the CEC2015 dataset as the benchmark function. The result shows that the proposed method improves the optimization ability of the original algorithm. And then, the classic images and wood fiber images are taken as experimental objects to analyze the segmentation result. The experimental results show that the proposed method has a good performance in Uniformity measure, Peak Signal-to-Noise Ratio and Feature Similarity Index and CPU time.
Collapse
Affiliation(s)
- Hong Qi
- School of Information and Computer Engineering, Northeast Forestry University, China
| | - Guanglei Zhang
- School of Information and Computer Engineering, Northeast Forestry University, China
| | - Heming Jia
- School of Information Engineering, Sanming Universiy, China
| | - Zhikai Xing
- School of Electrical Engineering and Automation, Wuhan University, China
| |
Collapse
|
34
|
Tan B, Sim R, Chua J, Wong DWK, Yao X, Garhöfer G, Schmidl D, Werkmeister RM, Schmetterer L. Approaches to quantify optical coherence tomography angiography metrics. ANNALS OF TRANSLATIONAL MEDICINE 2020; 8:1205. [PMID: 33241054 PMCID: PMC7576021 DOI: 10.21037/atm-20-3246] [Citation(s) in RCA: 26] [Impact Index Per Article: 5.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/10/2020] [Accepted: 06/16/2020] [Indexed: 12/13/2022]
Abstract
Optical coherence tomography (OCT) has revolutionized the field of ophthalmology in the last three decades. As an OCT extension, OCT angiography (OCTA) utilizes a fast OCT system to detect motion contrast in ocular tissue and provides a three-dimensional representation of the ocular vasculature in a non-invasive, dye-free manner. The first OCT machine equipped with OCTA function was approved by U.S. Food and Drug Administration in 2016 and now it is widely applied in clinics. To date, numerous methods have been developed to aid OCTA interpretation and quantification. In this review, we focused on the workflow of OCTA-based interpretation, beginning from the generation of the OCTA images using signal decorrelation, which we divided into intensity-based, phase-based and phasor-based methods. We further discussed methods used to address image artifacts that are commonly observed in clinical settings, to the algorithms for image enhancement, binarization, and OCTA metrics extraction. We believe a better grasp of these technical aspects of OCTA will enhance the understanding of the technology and its potential application in disease diagnosis and management. Moreover, future studies will also explore the use of ocular OCTA as a window to link ocular vasculature to the function of other organs such as the kidney and brain.
Collapse
Affiliation(s)
- Bingyao Tan
- Institute for Health Technologies, Nanyang Technological University, Singapore, Singapore
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore, Singapore
- SERI-NTU Advanced Ocular Engineering (STANCE) Program, Nanyang Technological University, Singapore, Singapore
| | - Ralene Sim
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore, Singapore
| | - Jacqueline Chua
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore, Singapore
- Academic Clinical Program, Duke-NUS Medical School, Singapore, Singapore
| | - Damon W. K. Wong
- Institute for Health Technologies, Nanyang Technological University, Singapore, Singapore
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore, Singapore
- SERI-NTU Advanced Ocular Engineering (STANCE) Program, Nanyang Technological University, Singapore, Singapore
| | - Xinwen Yao
- Institute for Health Technologies, Nanyang Technological University, Singapore, Singapore
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore, Singapore
- SERI-NTU Advanced Ocular Engineering (STANCE) Program, Nanyang Technological University, Singapore, Singapore
| | - Gerhard Garhöfer
- Department of Clinical Pharmacology, Medical University of Vienna, Vienna, Austria
| | - Doreen Schmidl
- Department of Clinical Pharmacology, Medical University of Vienna, Vienna, Austria
| | - René M. Werkmeister
- Center for Medical Physics and Biomedical Engineering, Medical University of Vienna, Vienna, Austria
| | - Leopold Schmetterer
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore, Singapore
- SERI-NTU Advanced Ocular Engineering (STANCE) Program, Nanyang Technological University, Singapore, Singapore
- Academic Clinical Program, Duke-NUS Medical School, Singapore, Singapore
- Department of Clinical Pharmacology, Medical University of Vienna, Vienna, Austria
- Center for Medical Physics and Biomedical Engineering, Medical University of Vienna, Vienna, Austria
- Yong Loo Lin School of Medicine, National University of Singapore and National University Health System, Singapore, Singapore
- Department of Ophthalmology, Lee Kong Chian School of Medicine, Nanyang Technological University, Singapore, Singapore
- Institute of Molecular and Clinical Ophthalmology, Basel, Switzerland
| |
Collapse
|
35
|
Lu Z, Bai Y, Chen Y, Su C, Lu S, Zhan T, Hong X, Wang S. The classification of gliomas based on a Pyramid dilated convolution resnet model. Pattern Recognit Lett 2020. [DOI: 10.1016/j.patrec.2020.03.007] [Citation(s) in RCA: 47] [Impact Index Per Article: 9.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/28/2022]
|