1
|
Hu Y, Gong M, Qiu Z, Liu J, Shen H, Yuan M, Zhang X, Li H, Lu H, Liu J. COph100: A comprehensive fundus image registration dataset from infants constituting the "RIDIRP" database. Sci Data 2025; 12:99. [PMID: 39824846 PMCID: PMC11742693 DOI: 10.1038/s41597-025-04426-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2024] [Accepted: 01/03/2025] [Indexed: 01/20/2025] Open
Abstract
Retinal image registration is vital for diagnostic therapeutic applications within the field of ophthalmology. Existing public datasets, focusing on adult retinal pathologies with high-quality images, have limited number of image pairs and neglect clinical challenges. To address this gap, we introduce COph100, a novel and challenging dataset known as the Comprehensive Ophthalmology Retinal Image Registration dataset for infants with a wide range of image quality issues constituting the public "RIDIRP" database. COph100 consists of 100 eyes, each with 2 to 9 examination sessions, amounting to a total of 491 image pairs carefully selected from the publicly available dataset. We manually labeled the corresponding ground truth image points and provided automatic vessel segmentation masks for each image. We have assessed COph100 in terms of image quality and registration outcomes using state-of-the-art algorithms. This resource enables a robust comparison of retinal registration methodologies and aids in the analysis of disease progression in infants, thereby deepening our understanding of pediatric ophthalmic conditions.
Collapse
Affiliation(s)
- Yan Hu
- Research Institute of Trustworthy Autonomous Systems and Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China.
| | - Mingdao Gong
- Research Institute of Trustworthy Autonomous Systems and Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China
| | - Zhongxi Qiu
- Research Institute of Trustworthy Autonomous Systems and Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China
| | - Jiabao Liu
- Research Institute of Trustworthy Autonomous Systems and Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China
| | - Hongli Shen
- Research Institute of Trustworthy Autonomous Systems and Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China
| | - Mingzhen Yuan
- Beijing Tongren Eye Center, Beijing Tongren Hospital, Capital Medical University, Beijing Ophthalmology and Visual Sciences Key Laboratory, Beijing, China
| | - Xiaoqing Zhang
- Center for High Performance Computing and Shenzhen Key Laboratory of Intelligent Bioinformatics, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Heng Li
- Research Institute of Trustworthy Autonomous Systems and Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China
| | - Hai Lu
- Beijing Tongren Eye Center, Beijing Tongren Hospital, Capital Medical University, Beijing Ophthalmology and Visual Sciences Key Laboratory, Beijing, China
| | - Jiang Liu
- Research Institute of Trustworthy Autonomous Systems and Department of Computer Science and Engineering, Southern University of Science and Technology, Shenzhen, China.
| |
Collapse
|
2
|
Wei J, Guo X, Wang Y, Zhang Y, Zhao W, Han S, Liu C, Yang X, Liang W. Realizing real-time optical molecular imaging in peripheral nerve tissue via Rhodamine B. Front Med (Lausanne) 2024; 11:1461520. [PMID: 39659623 PMCID: PMC11628308 DOI: 10.3389/fmed.2024.1461520] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/08/2024] [Accepted: 11/12/2024] [Indexed: 12/12/2024] Open
Abstract
Background Iatrogenic nerve injury is a consequential complication during surgery. Thus, real-time imaging of peripheral nerve (PN) possesses significant clinical implications. In recent years, the rapid advancements in optical molecular imaging (OMI) technology have provided essential technical foundations for the implementation of PN fluorescence imaging. This study aimed to realize real-time OMI of PNs via Rhodamine B. Methods Phosphate buffered saline (PBS), normal saline (NS), 5% glucose solution (GS), and fetal bovine serum (FBS) were selected for measuring the fluorescence spectra of Rhodamine B solutions prepared in each formulation. Rhodamine B solutions, with varying doses dissolved in 100 μL of each formulation, were prepared and applied to the exposed PNs of the mice for incubation later. To ascertain the optimal formulation and dose of Rhodamine B, an analysis was performed on the signal-to-background ratio (SBR) of the nerves. Based on the experimental results, we proceeded to incubate Rhodamine B solution on the PN tissue of mice and human subjects, as well as on neuronal cells, to verify the binding sites of Rhodamine B with nerve. Subsequently, histological studies were conducted to validate the binding site between Rhodamine B and the nerves. Finally, we injected the optimal combination of Rhodamine B solution into mice via the tail vein and collected the SBR of mouse nerve tissues at different time intervals to determine the optimal pre-injection time. Fluorescence images of various tissues were collected, and Hematoxylin and Eosin (H&E) staining results were observed to determine the metabolism of Rhodamine B in mice and its toxicity. Results The excitation peak of Rhodamine B in PBS, NS, 5% GS, and FBS formulations was 554 nm, and the emission peak was 576 nm. In PBS group, the maximum SBR was 15.37 ± 0.68 while the dose of Rhodamine B was 8 nmol. Through ex-vivo validation on fresh human nerve tissue and verification using mouse and human tissue sections, we observed fluorescent signals of Rhodamine Bin the regions of nerve tissue and the fluorescence signals were all concentrated on the neuronal cell membranes. After injection, the fluorescent signal in nerve tissue reached its peak at 24 hours (h), coinciding with the highest SBR (5.93 ± 0.92) in mouse nerve tissues at this time point. Additionally, the fluorescence signal could be maintained for at least 48 h. Within 24 h, lung dilation and fusion of alveoli occurred. Then these pathological manifestations gradually diminished, returning to normal at 2 weeks (w), with no significant acute or chronic adverse reactions observed in other tissues. Conclusion Rhodamine B enables fluorescence imaging of PNs and has the potential for clinical translation.
Collapse
Affiliation(s)
- Jinzheng Wei
- Department of Orthopaedics, First Hospital of Shanxi Medical University, Taiyuan, China
- Biomedical Engineering Research Center, Shanxi Medical University, Taiyuan, China
| | - Xinyu Guo
- Biomedical Engineering Research Center, Shanxi Medical University, Taiyuan, China
- Department of Urology, First Hospital of Shanxi Medical University, Taiyuan, China
| | - Yixi Wang
- School of Electronic Engineering, Xidian University, Xi’an, China
| | - Yunmeng Zhang
- Biomedical Engineering Research Center, Shanxi Medical University, Taiyuan, China
- Department of Urology, First Hospital of Shanxi Medical University, Taiyuan, China
| | - Wei Zhao
- Department of Orthopaedics, First Hospital of Shanxi Medical University, Taiyuan, China
| | - Shufeng Han
- Department of Orthopaedics, First Hospital of Shanxi Medical University, Taiyuan, China
| | - Chao Liu
- Biomedical Engineering Research Center, Shanxi Medical University, Taiyuan, China
- Department of Urology, First Hospital of Shanxi Medical University, Taiyuan, China
| | - Xiaofeng Yang
- Biomedical Engineering Research Center, Shanxi Medical University, Taiyuan, China
- Department of Urology, First Hospital of Shanxi Medical University, Taiyuan, China
| | - Wenkai Liang
- Department of Orthopaedics, First Hospital of Shanxi Medical University, Taiyuan, China
| |
Collapse
|
3
|
Kamran SA, Hossain KF, Ong J, Waisberg E, Zaman N, Baker SA, Lee AG, Tavakkoli A. FA4SANS-GAN: A Novel Machine Learning Generative Adversarial Network to Further Understand Ophthalmic Changes in Spaceflight Associated Neuro-Ocular Syndrome (SANS). OPHTHALMOLOGY SCIENCE 2024; 4:100493. [PMID: 38682031 PMCID: PMC11046204 DOI: 10.1016/j.xops.2024.100493] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/10/2022] [Revised: 01/11/2024] [Accepted: 02/05/2024] [Indexed: 05/01/2024]
Abstract
Purpose To provide an automated system for synthesizing fluorescein angiography (FA) images from color fundus photographs for averting risks associated with fluorescein dye and extend its future application to spaceflight associated neuro-ocular syndrome (SANS) detection in spaceflight where resources are limited. Design Development and validation of a novel conditional generative adversarial network (GAN) trained on limited amount of FA and color fundus images with diabetic retinopathy and control cases. Participants Color fundus and FA paired images for unique patients were collected from a publicly available study. Methods FA4SANS-GAN was trained to generate FA images from color fundus photographs using 2 multiscale generators coupled with 2 patch-GAN discriminators. Eight hundred fifty color fundus and FA images were utilized for training by augmenting images from 17 unique patients. The model was evaluated on 56 fluorescein images collected from 14 unique patients. In addition, it was compared with 3 other GAN architectures trained on the same data set. Furthermore, we test the robustness of the models against acquisition noise and retaining structural information when introduced to artificially created biological markers. Main Outcome Measures For GAN synthesis, metric Fréchet Inception Distance (FID) and Kernel Inception Distance (KID). Also, two 1-sided tests (TOST) based on Welch's t test for measuring statistical significance. Results On test FA images, mean FID for FA4SANS-GAN was 39.8 (standard deviation, 9.9), which is better than GANgio model's mean of 43.2 (standard deviation, 13.7), Pix2PixHD's mean of 57.3 (standard deviation, 11.5) and Pix2Pix's mean of 67.5 (standard deviation, 11.7). Similarly for KID, FA4SANS-GAN achieved mean of 0.00278 (standard deviation, 0.00167) which is better than other 3 model's mean KID of 0.00303 (standard deviation, 0.00216), 0.00609 (standard deviation, 0.00238), 0.00784 (standard deviation, 0.00218). For TOST measurement, FA4SANS-GAN was proven to be statistically significant versus GANgio (P = 0.006); versus Pix2PixHD (P < 0.00001); and versus Pix2Pix (P < 0.00001). Conclusions Our study has shown FA4SANS-GAN to be statistically significant for 2 GAN synthesis metrics. Moreover, it is robust against acquisition noise, and can retain clear biological markers compared with the other 3 GAN architectures. This deployment of this model can be crucial in the International Space Station for detecting SANS. Financial Disclosures The authors have no proprietary or commercial interest in any materials discussed in this article.
Collapse
Affiliation(s)
- Sharif Amit Kamran
- Human-Machine Perception Laboratory, Department of Computer Science and Engineering, University of Nevada, Reno, Reno, Nevada
| | - Khondker Fariha Hossain
- Human-Machine Perception Laboratory, Department of Computer Science and Engineering, University of Nevada, Reno, Reno, Nevada
| | - Joshua Ong
- Department of Ophthalmology and Visual Sciences, University of Michigan Kellogg Eye Center, Ann Arbor, Michigan
| | - Ethan Waisberg
- Department of Ophthalmology, University College Dublin School of Medicine, Belfield, Dublin, Ireland
| | - Nasif Zaman
- Human-Machine Perception Laboratory, Department of Computer Science and Engineering, University of Nevada, Reno, Reno, Nevada
| | - Salah A. Baker
- Department of Physiology and Cell Biology, University of Nevada School of Medicine, Reno, Nevada
| | - Andrew G. Lee
- Center for Space Medicine, Baylor College of Medicine, Houston, Texas
- Department of Ophthalmology, Blanton Eye Institute, Houston Methodist Hospital, Houston, Texas
- Houston Methodist Research Institute, Houston Methodist Hospital, Houston, Texas
- Departments of Ophthalmology, Neurology, and Neurosurgery, Weill Cornell Medicine, New York, New York
- Department of Ophthalmology, University of Texas Medical Branch, Galveston, Texas
- Department of Ophthalmology, University of Texas MD Anderson Cancer Center, Houston, Texas
- Department of Ophthalmology, Texas A&M College of Medicine, Texas
- Department of Ophthalmology, The University of Iowa Hospitals and Clinics, Iowa City, Iowa
| | - Alireza Tavakkoli
- Human-Machine Perception Laboratory, Department of Computer Science and Engineering, University of Nevada, Reno, Reno, Nevada
| |
Collapse
|
4
|
Zhang J, Qing C, Li Y, Wang Y. BCSwinReg: A cross-modal attention network for CBCT-to-CT multimodal image registration. Comput Biol Med 2024; 171:107990. [PMID: 38377717 DOI: 10.1016/j.compbiomed.2024.107990] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/29/2023] [Revised: 12/26/2023] [Accepted: 01/13/2024] [Indexed: 02/22/2024]
Abstract
Computed tomography (CT) and cone beam computed tomography (CBCT) registration plays an important role in radiotherapy. However, the poor quality of CBCT makes CBCT-CT multimodal registration challenging. Effective feature fusion and mapping often lead to better registration results for multimodal registration. Therefore, we proposed a new backbone network BCSwinReg and a cross-modal attention module CrossSwin. Specifically, a cross-modal attention CrossSwin is designed to promote multi-modal feature fusion, map the multi-modal domain to the common domain, and thus helping the network learn the correspondence between images better. Furthermore, a new network, BCSwinReg, is proposed to discover correspondence through cross-attention exchange information, obtain multi-level semantic information through a multi-resolution strategy, and finally integrate the deformation of multi-resolutions by the divide-conquer cascade method. We performed experiments on the publicly available 4D-Lung dataset to demonstrate the effectiveness of CrossSwin and BCSwinReg. Compared with VoxelMorph, the BCSwinReg has obtained performance improvements of 3.3% in Dice Similarity Coefficient (DSC) and 0.19 in the average 95% Hausdorff distance (HD95).
Collapse
Affiliation(s)
- Jieming Zhang
- The East China University of Science and Technology, Shanghai, 200237, China
| | - Chang Qing
- The East China University of Science and Technology, Shanghai, 200237, China.
| | - Yu Li
- The East China University of Science and Technology, Shanghai, 200237, China
| | - Yaqi Wang
- The East China University of Science and Technology, Shanghai, 200237, China
| |
Collapse
|
5
|
Nadana Ravishankar T, Ramprasath M, Daniel A, Selvarajan S, Subbiah P, Balusamy B. White shark optimizer with optimal deep learning based effective unmanned aerial vehicles communication and scene classification. Sci Rep 2023; 13:23041. [PMID: 38155207 PMCID: PMC10754923 DOI: 10.1038/s41598-023-50064-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2023] [Accepted: 12/14/2023] [Indexed: 12/30/2023] Open
Abstract
Unmanned aerial vehicles (UAVs) become a promising enabler for the next generation of wireless networks with the tremendous growth in electronics and communications. The application of UAV communications comprises messages relying on coverage extension for transmission networks after disasters, Internet of Things (IoT) devices, and dispatching distress messages from the device positioned within the coverage hole to the emergency centre. But there are some problems in enhancing UAV clustering and scene classification using deep learning approaches for enhancing performance. This article presents a new White Shark Optimizer with Optimal Deep Learning based Effective Unmanned Aerial Vehicles Communication and Scene Classification (WSOODL-UAVCSC) technique. UAV clustering and scene categorization present many deep learning challenges in disaster management: scene understanding complexity, data variability and abundance, visual data feature extraction, nonlinear and high-dimensional data, adaptability and generalization, real-time decision making, UAV clustering optimization, sparse and incomplete data. the need to handle complex, high-dimensional data, adapt to changing environments, and make quick, correct decisions in critical situations drives deep learning in UAV clustering and scene categorization. The purpose of the WSOODL-UAVCSC technique is to cluster the UAVs for effective communication and scene classification. The WSO algorithm is utilized for the optimization of the UAV clustering process and enables to accomplish effective communication and interaction in the network. With dynamic adjustment of the clustering, the WSO algorithm improves the performance and robustness of the UAV system. For the scene classification process, the WSOODL-UAVCSC technique involves capsule network (CapsNet) feature extraction, marine predators algorithm (MPA) based hyperparameter tuning, and echo state network (ESN) classification. A wide-ranging simulation analysis was conducted to validate the enriched performance of the WSOODL-UAVCSC approach. Extensive result analysis pointed out the enhanced performance of the WSOODL-UAVCSC method over other existing techniques. The WSOODL-UAVCSC method achieved an accuracy of 99.12%, precision of 97.45%, recall of 98.90%, and F1-score of 98.10% when compared to other existing techniques.
Collapse
Affiliation(s)
- T Nadana Ravishankar
- Department of Data Science and Business Systems, SRM Institute of Science and Technology, Kattankulathur, Chennai, Tamil Nadu, India
| | - M Ramprasath
- Department of Data Science and Business Systems, SRM Institute of Science and Technology, Kattankulathur, Chennai, India
| | - A Daniel
- Computer Science & Engineering. Amity School of Engineering and Technology (ASET), Amity University, Gwalior, Madhya Pradesh, India
| | - Shitharth Selvarajan
- Department of Computer Science, Kebri Dehar University, Kebri Dehar, Ethiopia.
- School of Built Environment, Engineering and Computing, Leeds Beckett University, Leeds, LS1 3HE, UK.
| | - Priyanga Subbiah
- Department of Networking and Communications, Faculty of Engineering and Technology, SRM Institute of Science and Technology, Kattankulathur, Chengalpattu District, Tamil Nadu, 603203, India
| | | |
Collapse
|
6
|
Kapsala Z, Pallikaris A, Tsilimbaris MK. Assessment of a Novel Semi-Automated Algorithm for the Quantification of the Parafoveal Capillary Network. Clin Ophthalmol 2023; 17:1661-1674. [PMID: 37313218 PMCID: PMC10259575 DOI: 10.2147/opth.s407695] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/12/2023] [Accepted: 05/01/2023] [Indexed: 06/15/2023] Open
Abstract
Introduction We present a novel semi-automated computerized method for the detection and quantification of parafoveal capillary network (PCN) in fluorescein angiography (FA) images. Material and Methods An algorithm detecting the superficial parafoveal capillary bed in high-resolution grayscale FA images and creating a one-pixel-wide PCN skeleton was developed using MatLab software. In addition to PCN detection, capillary density and branch point density in two circular areas centered on the center of the foveal avascular zone of 500μm and 750μm radius was calculated by the algorithm. Three consecutive FA images with distinguishable PCN from 56 eyes from 56 subjects were used for analysis. Both manual and semi-automated detection of the PCN and branch points was performed and compared. Three different intensity thresholds were used for the PCN detection to optimize the method defined as mean(I)+0.05*SD(I), mean(I) and mean(I)-0.05*SD(I), where I is the grayscale intensity of each image and SD the standard deviation. Limits of agreement (LoA), intraclass correlation coefficient (ICC) and Pearson's correlation coefficient (r) were calculated. Results Using mean(I)-0.05*SD(I) as threshold the average difference in PCN density between semi-automated and manual method was 0.197 (0.316) deg-1 at 500μm radius and 0.409 (0.562) deg-1 at 750μm radius. The LoA were -0.421 to 0.817 and -0.693 to 1.510 deg-1, respectively. The average difference of branch point density between semi-automated and manual method was zero for both areas; LoA were -0.001 to 0.002 and -0.001 to 0.001 branch points/degrees2, respectively. The other two intensity thresholds provided wider LoA for both metrics. The semi-automated algorithm showed great repeatability (ICC>0.91 in the 500μm radius and ICC>0.84 in the 750μm radius) for both metrics. Conclusion This semi-automated algorithm seems to provide readings in agreement with those of manual capillary tracing in FA. Larger prospective studies are needed to confirm the utility of the algorithm in clinical practice.
Collapse
Affiliation(s)
- Zoi Kapsala
- Department of Neurology and Sensory Organs, Medical School, University of Crete, Heraklion, Greece
| | - Aristofanis Pallikaris
- Department of Neurology and Sensory Organs, Medical School, University of Crete, Heraklion, Greece
- Vardinoyiannion Eye Institute of Crete, Medical School, University of Crete, Heraklion, Greece
| | - Miltiadis K Tsilimbaris
- Department of Neurology and Sensory Organs, Medical School, University of Crete, Heraklion, Greece
- Vardinoyiannion Eye Institute of Crete, Medical School, University of Crete, Heraklion, Greece
| |
Collapse
|
7
|
Martínez-Río J, Carmona EJ, Cancelas D, Novo J, Ortega M. Deformable registration of multimodal retinal images using a weakly supervised deep learning approach. Neural Comput Appl 2023. [DOI: 10.1007/s00521-023-08454-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/30/2023]
Abstract
AbstractThere are different retinal vascular imaging modalities widely used in clinical practice to diagnose different retinal pathologies. The joint analysis of these multimodal images is of increasing interest since each of them provides common and complementary visual information. However, if we want to facilitate the comparison of two images, obtained with different techniques and containing the same retinal region of interest, it will be necessary to make a previous registration of both images. Here, we present a weakly supervised deep learning methodology for robust deformable registration of multimodal retinal images, which is applied to implement a method for the registration of fluorescein angiography (FA) and optical coherence tomography angiography (OCTA) images. This methodology is strongly inspired by VoxelMorph, a general unsupervised deep learning framework of the state of the art for deformable registration of unimodal medical images. The method was evaluated in a public dataset with 172 pairs of FA and superficial plexus OCTA images. The degree of alignment of the common information (blood vessels) and preservation of the non-common information (image background) in the transformed image were measured using the Dice coefficient (DC) and zero-normalized cross-correlation (ZNCC), respectively. The average values of the mentioned metrics, including the standard deviations, were DC = 0.72 ± 0.10 and ZNCC = 0.82 ± 0.04. The time required to obtain each pair of registered images was 0.12 s. These results outperform rigid and deformable registration methods with which our method was compared.
Collapse
|
8
|
Cocianu CL, Uscatu CR, Stan AD. Evolutionary Image Registration: A Review. SENSORS (BASEL, SWITZERLAND) 2023; 23:967. [PMID: 36679771 PMCID: PMC9865935 DOI: 10.3390/s23020967] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/05/2022] [Revised: 01/10/2023] [Accepted: 01/11/2023] [Indexed: 06/17/2023]
Abstract
Image registration is one of the most important image processing tools enabling recognition, classification, detection and other analysis tasks. Registration methods are used to solve a large variety of real-world problems, including remote sensing, computer vision, geophysics, medical image analysis, surveillance, and so on. In the last few years, nature-inspired algorithms and metaheuristics have been successfully used to address the image registration problem, becoming a solid alternative for direct optimization methods. The aim of this paper is to investigate and summarize a series of state-of-the-art works reporting evolutionary-based registration methods. The papers were selected using the PRISMA 2020 method. The reported algorithms are reviewed and compared in terms of evolutionary components, fitness function, image similarity measures and algorithm accuracy indexes used in the alignment process.
Collapse
|