1
|
Pan L, Cai Z, Hu D, Zhu W, Shi F, Tao W, Wu Q, Xiao S, Chen X. Research on registration method for enface image using multi-feature fusion. Phys Med Biol 2024; 69:215037. [PMID: 39413811 DOI: 10.1088/1361-6560/ad87a5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/04/2024] [Accepted: 10/16/2024] [Indexed: 10/18/2024]
Abstract
Objective.The purpose of this work is to accurately and quickly register the Optical coherence tomography (OCT) projection (enface) images at adjacent time points, and to solve the problem of interference caused by CNV lesions on the registration features.Approach.In this work, a multi-feature registration strategy was proposed, in which a combined feature (com-feature) containing 3D information, intersection information and SURF feature was designed. Firstly, the coordinates of all feature points were extracted as combined features, and then these feature coordinates were added to the initial vascular coordinate set simplified by the Douglas-Peucker algorithm as the point set for registration. Finally, the coherent point drift registration algorithm was used to register the enface coordinate point sets of adjacent time series.Main results.The newly designed features significantly improve the success rate of global registration of vascular networks in enface images, while the simplification step greatly improves the registration speed on the basis of preserving vascular features. The MSE, DSC and time complexity of the proposed method are 0.07993, 0.9693 and 42.7016 s, respectively.Significance.CNV is a serious retinal disease in ophthalmology. The registration of OCT enface images at adjacent time points can timely monitor the progress of the disease and assist doctors in making diagnoses. The proposed method not only improves the accuracy of OCT enface image registration, but also significantly reduces the time complexity. It has good registration results in clinical routine and provides a more efficient method for clinical diagnosis and treatment.
Collapse
Affiliation(s)
- Lingjiao Pan
- Department of Electrical Information Engineering, Jiangsu University of Technology, Changzhou, People's Republic of China
| | - Zhongwang Cai
- Department of Mechanical Engineering, Jiangsu University of Technology, Changzhou, People's Republic of China
| | - Derong Hu
- Department of Mechanical Engineering, Jiangsu University of Technology, Changzhou, People's Republic of China
| | - Weifang Zhu
- Department of Information Engineering, Suzhou University, Suzhou, People's Republic of China
| | - Fei Shi
- Department of Information Engineering, Suzhou University, Suzhou, People's Republic of China
| | - Weige Tao
- Department of Electrical Information Engineering, Jiangsu University of Technology, Changzhou, People's Republic of China
| | - Quanyu Wu
- Department of Electrical Information Engineering, Jiangsu University of Technology, Changzhou, People's Republic of China
| | - Shuyan Xiao
- Department of Electrical Information Engineering, Jiangsu University of Technology, Changzhou, People's Republic of China
| | - Xinjian Chen
- Department of Information Engineering, Suzhou University, Suzhou, People's Republic of China
| |
Collapse
|
2
|
Wang CY, Sadrieh FK, Shen YT, Chen SE, Kim S, Chen V, Raghavendra A, Wang D, Saeedi O, Tao Y. MEMO: dataset and methods for robust multimodal retinal image registration with large or small vessel density differences. BIOMEDICAL OPTICS EXPRESS 2024; 15:3457-3479. [PMID: 38855695 PMCID: PMC11161385 DOI: 10.1364/boe.516481] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 01/18/2024] [Revised: 03/20/2024] [Accepted: 04/18/2024] [Indexed: 06/11/2024]
Abstract
The measurement of retinal blood flow (RBF) in capillaries can provide a powerful biomarker for the early diagnosis and treatment of ocular diseases. However, no single modality can determine capillary flowrates with high precision. Combining erythrocyte-mediated angiography (EMA) with optical coherence tomography angiography (OCTA) has the potential to achieve this goal, as EMA can measure the absolute RBF of retinal microvasculature and OCTA can provide the structural images of capillaries. However, multimodal retinal image registration between these two modalities remains largely unexplored. To fill this gap, we establish MEMO, the first public multimodal EMA and OCTA retinal image dataset. A unique challenge in multimodal retinal image registration between these modalities is the relatively large difference in vessel density (VD). To address this challenge, we propose a segmentation-based deep-learning framework (VDD-Reg), which provides robust results despite differences in vessel density. VDD-Reg consists of a vessel segmentation module and a registration module. To train the vessel segmentation module, we further designed a two-stage semi-supervised learning framework (LVD-Seg) combining supervised and unsupervised losses. We demonstrate that VDD-Reg outperforms existing methods quantitatively and qualitatively for cases of both small VD differences (using the CF-FA dataset) and large VD differences (using our MEMO dataset). Moreover, VDD-Reg requires as few as three annotated vessel segmentation masks to maintain its accuracy, demonstrating its feasibility.
Collapse
Affiliation(s)
- Chiao-Yi Wang
- Department of Bioengineering, University of Maryland, College Park, MD 20742, USA
| | | | - Yi-Ting Shen
- Department of Electrical and Computer Engineering, University of Maryland, College Park, MD 20742, USA
| | - Shih-En Chen
- Department of Ophthalmology and Visual Sciences, University of Maryland School of Medicine, Baltimore, MD 21201, USA
| | - Sarah Kim
- Department of Ophthalmology and Visual Sciences, University of Maryland School of Medicine, Baltimore, MD 21201, USA
| | - Victoria Chen
- Department of Ophthalmology and Visual Sciences, University of Maryland School of Medicine, Baltimore, MD 21201, USA
| | - Achyut Raghavendra
- Department of Ophthalmology and Visual Sciences, University of Maryland School of Medicine, Baltimore, MD 21201, USA
| | - Dongyi Wang
- Department of Biological and Agricultural Engineering, University of Arkansas, Fayetteville, AR 72701, USA
| | - Osamah Saeedi
- Department of Ophthalmology and Visual Sciences, University of Maryland School of Medicine, Baltimore, MD 21201, USA
| | - Yang Tao
- Department of Bioengineering, University of Maryland, College Park, MD 20742, USA
| |
Collapse
|
3
|
Rivas-Villar D, Motschi AR, Pircher M, Hitzenberger CK, Schranz M, Roberts PK, Schmidt-Erfurth U, Bogunović H. Automated inter-device 3D OCT image registration using deep learning and retinal layer segmentation. BIOMEDICAL OPTICS EXPRESS 2023; 14:3726-3747. [PMID: 37497506 PMCID: PMC10368062 DOI: 10.1364/boe.493047] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 04/12/2023] [Revised: 05/18/2023] [Accepted: 05/26/2023] [Indexed: 07/28/2023]
Abstract
Optical coherence tomography (OCT) is the most widely used imaging modality in ophthalmology. There are multiple variations of OCT imaging capable of producing complementary information. Thus, registering these complementary volumes is desirable in order to combine their information. In this work, we propose a novel automated pipeline to register OCT images produced by different devices. This pipeline is based on two steps: a multi-modal 2D en-face registration based on deep learning, and a Z-axis (axial axis) registration based on the retinal layer segmentation. We evaluate our method using data from a Heidelberg Spectralis and an experimental PS-OCT device. The empirical results demonstrated high-quality registrations, with mean errors of approximately 46 µm for the 2D registration and 9.59 µm for the Z-axis registration. These registrations may help in multiple clinical applications such as the validation of layer segmentations among others.
Collapse
Affiliation(s)
- David Rivas-Villar
- Centro de investigacion CITIC, Universidade da Coruña, 15071 A Coruña, Spain
- Grupo VARPA, Instituto de Investigacion Biomédica de A Coruña (INIBIC), Universidade da Coruña, 15006 A Coruña, Spain
| | - Alice R Motschi
- Medical University of Vienna, Center for Medical Physics and Biomedical Engineering, Vienna, Austria
| | - Michael Pircher
- Medical University of Vienna, Center for Medical Physics and Biomedical Engineering, Vienna, Austria
| | - Christoph K Hitzenberger
- Medical University of Vienna, Center for Medical Physics and Biomedical Engineering, Vienna, Austria
| | - Markus Schranz
- Medical University of Vienna, Center for Medical Physics and Biomedical Engineering, Vienna, Austria
| | - Philipp K Roberts
- Medical University of Vienna, Department of Ophthalmology and Optometry, Vienna, Austria
| | - Ursula Schmidt-Erfurth
- Medical University of Vienna, Department of Ophthalmology and Optometry, Vienna, Austria
| | - Hrvoje Bogunović
- Medical University of Vienna, Department of Ophthalmology and Optometry, Christian Doppler Lab for Artificial Intelligence in Retina, Vienna, Austria
| |
Collapse
|
4
|
Cavichini M, Bartsch DUG, Warter A, Singh S, An C, Wang Y, Zhang J, Nguyen T, Freeman WR. Accuracy and Time Comparison Between Side-by-Side and Artificial Intelligence Overlayed Images. Ophthalmic Surg Lasers Imaging Retina 2023; 54:108-113. [PMID: 36780638 DOI: 10.3928/23258160-20230130-03] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/15/2023]
Abstract
BACKGROUND AND OBJECTIVE The purpose of this study was to evaluate the accuracy and the time to find a lesion, taken in different platforms, color fundus photographs and infrared scanning laser ophthalmoscope images, using the traditional side-by-side (SBS) colocalization technique to an artificial intelligence (AI)-assisted technique. PATIENTS AND METHODS Fifty-three pathological lesions were studied in 11 eyes. Images were aligned using SBS and AI overlaid methods. The location of each color fundus lesion on the corresponding infrared scanning laser ophthalmoscope image was analyzed twice, one time for each method, on different days, for two specialists, in random order. The outcomes for each method were measured and recorded by an independent observer. RESULTS The colocalization AI method was superior to the conventional in accuracy and time (P < .001), with a mean time to colocalize 37% faster. The error rate using AI was 0% compared with 18% in SBS measurements. CONCLUSIONS AI permitted a more accurate and faster colocalization of pathologic lesions than the conventional method. [Ophthalmic Surg Lasers Imaging Retina 2023;54:108-113.].
Collapse
|
5
|
An C, Wang Y, Zhang J, Nguyen TQ. Self-Supervised Rigid Registration for Multimodal Retinal Images. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2022; 31:5733-5747. [PMID: 36040946 PMCID: PMC11211857 DOI: 10.1109/tip.2022.3201476] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
The ability to accurately overlay one modality retinal image to another is critical in ophthalmology. Our previous framework achieved the state-of-the-art results for multimodal retinal image registration. However, it requires human-annotated labels due to the supervised approach of the previous work. In this paper, we propose a self-supervised multimodal retina registration method to alleviate the burdens of time and expense to prepare for training data, that is, aiming to automatically register multimodal retinal images without any human annotations. Specially, we focus on registering color fundus images with infrared reflectance and fluorescein angiography images, and compare registration results with several conventional and supervised and unsupervised deep learning methods. From the experimental results, the proposed self-supervised framework achieves a comparable accuracy comparing to the state-of-the-art supervised learning method in terms of registration accuracy and Dice coefficient.
Collapse
|
6
|
Abdelmotaal H, Sharaf M, Soliman W, Wasfi E, Kedwany SM. Bridging the resources gap: deep learning for fluorescein angiography and optical coherence tomography macular thickness map image translation. BMC Ophthalmol 2022; 22:355. [PMID: 36050661 PMCID: PMC9434904 DOI: 10.1186/s12886-022-02577-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/15/2021] [Accepted: 08/23/2022] [Indexed: 11/29/2022] Open
Abstract
Background To assess the ability of the pix2pix generative adversarial network (pix2pix GAN) to synthesize clinically useful optical coherence tomography (OCT) color-coded macular thickness maps based on a modest-sized original fluorescein angiography (FA) dataset and the reverse, to be used as a plausible alternative to either imaging technique in patients with diabetic macular edema (DME). Methods Original images of 1,195 eyes of 708 nonconsecutive diabetic patients with or without DME were retrospectively analyzed. OCT macular thickness maps and corresponding FA images were preprocessed for use in training and testing the proposed pix2pix GAN. The best quality synthesized images using the test set were selected based on the Fréchet inception distance score, and their quality was studied subjectively by image readers and objectively by calculating the peak signal-to-noise ratio, structural similarity index, and Hamming distance. We also used original and synthesized images in a trained deep convolutional neural network (DCNN) to plot the difference between synthesized images and their ground-truth analogues and calculate the learned perceptual image patch similarity metric. Results The pix2pix GAN-synthesized images showed plausible subjectively and objectively assessed quality, which can provide a clinically useful alternative to either image modality. Conclusion Using the pix2pix GAN to synthesize mutually dependent OCT color-coded macular thickness maps or FA images can overcome issues related to machine unavailability or clinical situations that preclude the performance of either imaging technique. Trial registration ClinicalTrials.gov Identifier: NCT05105620, November 2021. “Retrospectively registered”.
Collapse
Affiliation(s)
- Hazem Abdelmotaal
- Department of Ophthalmology, Faculty of Medicine, Assiut University, Assiut, 71515, Egypt.
| | - Mohamed Sharaf
- Department of Ophthalmology, Faculty of Medicine, Assiut University, Assiut, 71515, Egypt
| | - Wael Soliman
- Department of Ophthalmology, Faculty of Medicine, Assiut University, Assiut, 71515, Egypt
| | - Ehab Wasfi
- Department of Ophthalmology, Faculty of Medicine, Assiut University, Assiut, 71515, Egypt
| | - Salma M Kedwany
- Department of Ophthalmology, Faculty of Medicine, Assiut University, Assiut, 71515, Egypt
| |
Collapse
|
7
|
Zhang J, Wang Y, Dai J, Cavichini M, Bartsch DUG, Freeman WR, Nguyen TQ, An C. Two-Step Registration on Multi-Modal Retinal Images via Deep Neural Networks. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2022; 31:823-838. [PMID: 34932479 PMCID: PMC8912939 DOI: 10.1109/tip.2021.3135708] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Multi-modal retinal image registration plays an important role in the ophthalmological diagnosis process. The conventional methods lack robustness in aligning multi-modal images of various imaging qualities. Deep-learning methods have not been widely developed for this task, especially for the coarse-to-fine registration pipeline. To handle this task, we propose a two-step method based on deep convolutional networks, including a coarse alignment step and a fine alignment step. In the coarse alignment step, a global registration matrix is estimated by three sequentially connected networks for vessel segmentation, feature detection and description, and outlier rejection, respectively. In the fine alignment step, a deformable registration network is set up to find pixel-wise correspondence between a target image and a coarsely aligned image from the previous step to further improve the alignment accuracy. Particularly, an unsupervised learning framework is proposed to handle the difficulties of inconsistent modalities and lack of labeled training data for the fine alignment step. The proposed framework first changes multi-modal images into a same modality through modality transformers, and then adopts photometric consistency loss and smoothness loss to train the deformable registration network. The experimental results show that the proposed method achieves state-of-the-art results in Dice metrics and is more robust in challenging cases.
Collapse
|
8
|
Pancreatic cancer segmentation in unregistered multi-parametric MRI with adversarial learning and multi-scale supervision. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2021.09.058] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/19/2022]
|
9
|
Zhang J, Wang Y, Bartsch DUG, Freeman WR, Nguyen TQ, An C. Perspective Distortion Correction for Multi-Modal Registration between Ultra-Widefield and Narrow-Angle Retinal Images. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2021; 2021:4086-4091. [PMID: 34892126 PMCID: PMC9359414 DOI: 10.1109/embc46164.2021.9631084] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Multi-modal retinal image registration between 2D Ultra-Widefield (UWF) and narrow-angle (NA) images has not been well-studied, since most existing methods mainly focus on NA image alignment. The stereographic projection model used in UWF imaging causes strong distortions in peripheral areas, which leads to inferior alignment quality. We propose a distortion correction method that remaps the UWF images based on estimated camera view points of NA images. In addition, we set up a CNN-based registration pipeline for UWF and NA images, which consists of the distortion correction method and three networks for vessel segmentation, feature detection and matching, and outlier rejection. Experimental results on our collected dataset shows the effectiveness of the proposed pipeline and the distortion correction method.
Collapse
|
10
|
Ho CJ, Wang Y, Zhang J, Nguyen T, An C. A Convolutional Neural Network Pipeline For Multi-Temporal Retinal Image Registration. INTERNATIONAL SOC DESIGN CONFERENCE. INTERNATIONAL SOC DESIGN CONFERENCE 2021; 2021:27-28. [PMID: 35949978 PMCID: PMC9359415 DOI: 10.1109/isocc53507.2021.9613906] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
A sequence of images is usually captured to observe the change of health status in medical diagnosis. However, an image sequence taken over year usually suffers from severe deformation, making it time-consuming for physicians to match corresponding patterns. In this paper, we propose a coarse-to-fine pipeline for retinal image registration based on convolutional neural network. By leveraging the three components of the pipeline: feature matching, outlier rejection, and local registration, we recover the deformation and accurately align multi-temporal image sequences. Experimental results show that the proposed network is robust to severe deformation as well as illumination and contrast variations. With the proposed registration pipeline, the change of image patterns over time can be identified through visual analysis.
Collapse
Affiliation(s)
- Chi-Jui Ho
- Department of Electrical and Computer Engineering, UC San Diego, La Jolla, USA
| | - Yiqian Wang
- Department of Electrical and Computer Engineering, UC San Diego, La Jolla, USA
| | - Junkang Zhang
- Department of Electrical and Computer Engineering, UC San Diego, La Jolla, USA
| | - Truong Nguyen
- Department of Electrical and Computer Engineering, UC San Diego, La Jolla, USA
| | - Cheolhong An
- Department of Electrical and Computer Engineering, UC San Diego, La Jolla, USA
| |
Collapse
|
11
|
Huang W, Yang H, Liu X, Li C, Zhang I, Wang R, Zheng H, Wang S. A Coarse-to-Fine Deformable Transformation Framework for Unsupervised Multi-Contrast MR Image Registration with Dual Consistency Constraint. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:2589-2599. [PMID: 33577451 DOI: 10.1109/tmi.2021.3059282] [Citation(s) in RCA: 22] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/10/2023]
Abstract
Multi-contrast magnetic resonance (MR) image registration is useful in the clinic to achieve fast and accurate imaging-based disease diagnosis and treatment planning. Nevertheless, the efficiency and performance of the existing registration algorithms can still be improved. In this paper, we propose a novel unsupervised learning-based framework to achieve accurate and efficient multi-contrast MR image registration. Specifically, an end-to-end coarse-to-fine network architecture consisting of affine and deformable transformations is designed to improve the robustness and achieve end-to-end registration. Furthermore, a dual consistency constraint and a new prior knowledge-based loss function are developed to enhance the registration performances. The proposed method has been evaluated on a clinical dataset containing 555 cases, and encouraging performances have been achieved. Compared to the commonly utilized registration methods, including VoxelMorph, SyN, and LT-Net, the proposed method achieves better registration performance with a Dice score of 0.8397± 0.0756 in identifying stroke lesions. With regards to the registration speed, our method is about 10 times faster than the most competitive method of SyN (Affine) when testing on a CPU. Moreover, we prove that our method can still perform well on more challenging tasks with lacking scanning information data, showing the high robustness for the clinical application.
Collapse
|
12
|
Li J, Feng C, Lin X, Qian X. Utilizing GCN and Meta-Learning Strategy in Unsupervised Domain Adaptation for Pancreatic Cancer Segmentation. IEEE J Biomed Health Inform 2021; 26:79-89. [PMID: 34057903 DOI: 10.1109/jbhi.2021.3085092] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Automated pancreatic cancer segmentation is highly crucial for computer-assisted diagnosis. The general practice is to label images from selected modalities since it is expensive to label all modalities. This practice brought about a significant interest in learning the knowledge transfer from the labeled modalities to unlabeled ones. However, the imaging parameter inconsistency between modalities leads to a domain shift, limiting the transfer learning performance. Therefore, we propose an unsupervised domain adaptation segmentation framework for pancreatic cancer based on GCN and meta-learning strategy. Our model first transforms the source image into a target-like visual appearance through the synergistic collaboration between image and feature adaptation. Specifically, we employ encoders incorporating adversarial learning to separate domain-invariant features from domain-specific ones to achieve visual appearance translation. Then, the meta-learning strategy with good generalization capabilities is exploited to strike a reasonable balance in the training of the source and transformed images. Thus, the model acquires more correlated features and improve the adaptability to the target images. Moreover, a GCN is introduced to supervise the high-dimensional abstract features directly related to the segmentation outcomes, and hence ensure the integrity of key structural features. Extensive experiments on four multi-parameter pancreatic-cancer magnetic resonance imaging datasets demonstrate improved performance in all adaptation directions, confirming our model's effectiveness for unlabeled pancreatic cancer images. The results are promising for reducing the burden of annotation and improving the performance of computer-aided diagnosis of pancreatic cancer. Our source codes will be released at https://github.com/SJTUBME-QianLab/UDAseg, once this manuscript is accepted for publication.
Collapse
|
13
|
Golkar E, Rabbani H, Dehghani A. Hybrid registration of retinal fluorescein angiography and optical coherence tomography images of patients with diabetic retinopathy. BIOMEDICAL OPTICS EXPRESS 2021; 12:1707-1724. [PMID: 33796382 PMCID: PMC7984788 DOI: 10.1364/boe.415939] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/27/2020] [Revised: 01/26/2021] [Accepted: 02/21/2021] [Indexed: 05/10/2023]
Abstract
Diabetic retinopathy (DR) is a common ophthalmic disease among diabetic patients. It is essential to diagnose DR in the early stages of treatment. Various imaging systems have been proposed to detect and visualize retina diseases. The fluorescein angiography (FA) imaging technique is now widely used as a gold standard technique to evaluate the clinical manifestations of DR. Optical coherence tomography (OCT) imaging is another technique that provides 3D information of the retinal structure. The FA and OCT images are captured in two different phases and field of views and image fusion of these modalities are of interest to clinicians. This paper proposes a hybrid registration framework based on the extraction and refinement of segmented major blood vessels of retinal images. The newly extracted features significantly improve the success rate of global registration results in the complex blood vessel network of retinal images. Afterward, intensity-based and deformable transformations are utilized to further compensate the motion magnitude between the FA and OCT images. Experimental results of 26 images of the various stages of DR patients indicate that this algorithm yields promising registration and fusion results for clinical routine.
Collapse
Affiliation(s)
- Ehsan Golkar
- Medical Image and Signal Processing Research Center, School of Advanced Technologies in Medicine, Isfahan University of Medical Sciences, Isfahan, Iran
| | - Hossein Rabbani
- Medical Image and Signal Processing Research Center, School of Advanced Technologies in Medicine, Isfahan University of Medical Sciences, Isfahan, Iran
| | - Alireza Dehghani
- Eye Research Center, Isfahan University of Medical Sciences, Isfahan, Iran and Didavaran Eye Clinic, Isfahan, Iran
| |
Collapse
|
14
|
Wang Y, Zhang J, Cavichini M, Bartsch DUG, Freeman WR, Nguyen TQ, An C. Robust Content-Adaptive Global Registration for Multimodal Retinal Images Using Weakly Supervised Deep-Learning Framework. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2021; 30:3167-3178. [PMID: 33600314 DOI: 10.1109/tip.2021.3058570] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Multimodal retinal imaging plays an important role in ophthalmology. We propose a content-adaptive multimodal retinal image registration method in this paper that focuses on the globally coarse alignment and includes three weakly supervised neural networks for vessel segmentation, feature detection and description, and outlier rejection. We apply the proposed framework to register color fundus images with infrared reflectance and fluorescein angiography images, and compare it with several conventional and deep learning methods. Our proposed framework demonstrates a significant improvement in robustness and accuracy reflected by a higher success rate and Dice coefficient compared with other methods.
Collapse
|
15
|
De Silva T, Chew EY, Hotaling N, Cukras CA. Deep-learning based multi-modal retinal image registration for the longitudinal analysis of patients with age-related macular degeneration. BIOMEDICAL OPTICS EXPRESS 2021; 12:619-636. [PMID: 33520392 PMCID: PMC7818952 DOI: 10.1364/boe.408573] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/28/2020] [Revised: 10/29/2020] [Accepted: 10/30/2020] [Indexed: 05/23/2023]
Abstract
This work reports a deep-learning based registration algorithm that aligns multi-modal retinal images collected from longitudinal clinical studies to achieve accuracy and robustness required for analysis of structural changes in large-scale clinical data. Deep-learning networks that mirror the architecture of conventional feature-point-based registration were evaluated with different networks that solved for registration affine parameters, image patch displacements, and patch displacements within the region of overlap. The ground truth images for deep learning-based approaches were derived from successful conventional feature-based registration. Cross-sectional and longitudinal affine registrations were performed across color fundus photography (CFP), fundus autofluorescence (FAF), and infrared reflectance (IR) image modalities. For mono-modality longitudinal registration, the conventional feature-based registration method achieved mean errors in the range of 39-53 µm (depending on the modality) whereas the deep learning method with region overlap prediction exhibited mean errors in the range 54-59 µm. For cross-sectional multi-modality registration, the conventional method exhibited gross failures with large errors in more than 50% of the cases while the proposed deep-learning method achieved robust performance with no gross failures and mean errors in the range 66-69 µm. Thus, the deep learning-based method achieved superior overall performance across all modalities. The accuracy and robustness reported in this work provide important advances that will facilitate clinical research and enable a detailed study of the progression of retinal diseases such as age-related macular degeneration.
Collapse
Affiliation(s)
- Tharindu De Silva
- National Eye Institute, National Institutes of Health, Bethesda, MD 20892, USA
| | - Emily Y Chew
- National Eye Institute, National Institutes of Health, Bethesda, MD 20892, USA
| | - Nathan Hotaling
- National Center for Advancing Translational Science, National Institutes of Health, Bethesda, MD 20892, USA
| | - Catherine A Cukras
- National Eye Institute, National Institutes of Health, Bethesda, MD 20892, USA
| |
Collapse
|
16
|
Cavichini M, An C, Bartsch DUG, Jhingan M, Amador-Patarroyo MJ, Long CP, Zhang J, Wang Y, Chan AX, Madala S, Nguyen T, Freeman WR. Artificial Intelligence for Automated Overlay of Fundus Camera and Scanning Laser Ophthalmoscope Images. Transl Vis Sci Technol 2020; 9:56. [PMID: 33173612 PMCID: PMC7594596 DOI: 10.1167/tvst.9.2.56] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/12/2020] [Accepted: 09/23/2020] [Indexed: 02/06/2023] Open
Abstract
Purpose The purpose of this study was to evaluate the ability to align two types of retinal images taken on different platforms; color fundus (CF) photographs and infrared scanning laser ophthalmoscope (IR SLO) images using mathematical warping and artificial intelligence (AI). Methods We collected 109 matched pairs of CF and IR SLO images. An AI algorithm utilizing two separate networks was developed. A style transfer network (STN) was used to segment vessel structures. A registration network was used to align the segmented images to each. Neither network used a ground truth dataset. A conventional image warping algorithm was used as a control. Software displayed image pairs as a 5 × 5 checkerboard grid composed of alternating subimages. This technique permitted vessel alignment determination by human observers and 5 masked graders evaluated alignment by the AI and conventional warping in 25 fields for each image. Results Our new AI method was superior to conventional warping at generating vessel alignment as judged by masked human graders (P < 0.0001). The average number of good/excellent matches increased from 90.5% to 94.4% with AI method. Conclusions AI permitted a more accurate overlay of CF and IR SLO images than conventional mathematical warping. This is a first step toward developing an AI that could allow overlay of all types of fundus images by utilizing vascular landmarks. Translational Relevance The ability to align and overlay imaging data from multiple instruments and manufacturers will permit better analysis of this complex data helping understand disease and predict treatment.
Collapse
Affiliation(s)
- Melina Cavichini
- Jacobs Retina Center, Shiley Eye Institute, University of California San Diego, La Jolla, CA, USA.,Departamento de Oftalmologia, Faculdade de Medicina do ABC, Santo Andre, Brazil
| | - Cheolhong An
- Department of Electrical and Computer Engineering, University of California San Diego, La Jolla, CA, USA
| | - Dirk-Uwe G Bartsch
- Jacobs Retina Center, Shiley Eye Institute, University of California San Diego, La Jolla, CA, USA
| | - Mahima Jhingan
- Jacobs Retina Center, Shiley Eye Institute, University of California San Diego, La Jolla, CA, USA.,Aravind Eye Hospital, Madurai, India
| | - Manuel J Amador-Patarroyo
- Jacobs Retina Center, Shiley Eye Institute, University of California San Diego, La Jolla, CA, USA.,Escuela Superior de Oftalmologia, Instituto Barraquer de America, Bogota, Colombia
| | - Christopher P Long
- University of California San Diego School of Medicine, La Jolla, CA, USA
| | - Junkang Zhang
- Department of Electrical and Computer Engineering, University of California San Diego, La Jolla, CA, USA
| | - Yiqian Wang
- Department of Electrical and Computer Engineering, University of California San Diego, La Jolla, CA, USA
| | - Alison X Chan
- University of California San Diego School of Medicine, La Jolla, CA, USA
| | - Samantha Madala
- University of California San Diego School of Medicine, La Jolla, CA, USA
| | - Truong Nguyen
- Department of Electrical and Computer Engineering, University of California San Diego, La Jolla, CA, USA
| | - William R Freeman
- Jacobs Retina Center, Shiley Eye Institute, University of California San Diego, La Jolla, CA, USA
| |
Collapse
|
17
|
Luo G, Chen X, Shi F, Peng Y, Xiang D, Chen Q, Xu X, Zhu W, Fan Y. Multimodal affine registration for ICGA and MCSL fundus images of high myopia. BIOMEDICAL OPTICS EXPRESS 2020; 11:4443-4457. [PMID: 32923055 PMCID: PMC7449720 DOI: 10.1364/boe.393178] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/19/2020] [Revised: 06/29/2020] [Accepted: 07/06/2020] [Indexed: 06/11/2023]
Abstract
The registration between indocyanine green angiography (ICGA) and multi-color scanning laser (MCSL) imaging fundus images is vital for the joint linear lesion segmentation in ICGA and MCSL and the evaluation whether MCSL can replace ICGA as a non-invasive diagnosis for linear lesion. To our best knowledge, there are no studies focusing on the image registration between these two modalities. In this paper, we propose a framework based on convolutional neural networks for the multimodal affine registration between ICGA and MCSL images, which contains two parts: coarse registration stage and fine registration stage. In the coarse registration stage, the optic disc is segmented and its centroid is used as a matching point to perform coarse registration. The fine registration stage regresses affine parameters directly using jointly supervised and weakly-supervised loss function. Experimental results show the effectiveness of the proposed method, which lays a sound foundation for further evaluation of non-invasive diagnosis of linear lesion based on MCSL.
Collapse
Affiliation(s)
- Gaohui Luo
- School of Electronic and Information Engineering, Soochow University, Suzhou 215006, China
- contributed equally
| | - Xinjian Chen
- School of Electronic and Information Engineering, Soochow University, Suzhou 215006, China
- State Key Laboratory of Radiation Medicine and Protection, Soochow University, Suzhou 215123, China
- contributed equally
| | - Fei Shi
- School of Electronic and Information Engineering, Soochow University, Suzhou 215006, China
| | - Yunzhen Peng
- School of Electronic and Information Engineering, Soochow University, Suzhou 215006, China
| | - Dehui Xiang
- School of Electronic and Information Engineering, Soochow University, Suzhou 215006, China
| | - Qiuying Chen
- Shanghai General Hospital, Shanghai 200080, China
| | - Xun Xu
- Shanghai General Hospital, Shanghai 200080, China
| | - Weifang Zhu
- School of Electronic and Information Engineering, Soochow University, Suzhou 215006, China
| | - Ying Fan
- Shanghai General Hospital, Shanghai 200080, China
| |
Collapse
|
18
|
Almasi R, Vafaei A, Ghasemi Z, Ommani MR, Dehghani AR, Rabbani H. Registration of fluorescein angiography and optical coherence tomography images of curved retina via scanning laser ophthalmoscopy photographs. BIOMEDICAL OPTICS EXPRESS 2020; 11:3455-3476. [PMID: 33014544 PMCID: PMC7510895 DOI: 10.1364/boe.395784] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/20/2020] [Revised: 05/27/2020] [Accepted: 05/27/2020] [Indexed: 05/18/2023]
Abstract
Accurate and automatic registration of multimodal retinal images such as fluorescein angiography (FA) and optical coherence tomography (OCT) enables utilization of supplementary information. FA is a gold standard imaging modality that depicts neurovascular structure of retina and is used for diagnosing neurovascular-related diseases such as diabetic retinopathy (DR). Unlike FA, OCT is non-invasive retinal imaging modality that provides cross-sectional data of retina. Due to differences in contrast, resolution and brightness of multimodal retinal images, the images resulted from vessel extraction of image pairs are not exactly the same. Also, prevalent feature detection, extraction and matching schemes do not result in perfect matches. In addition, the relationships between retinal image pairs are usually modeled by affine transformation, which cannot generate accurate alignments due to the non-planar retina surface. In this paper, a precise registration scheme is proposed to align FA and OCT images via scanning laser ophthalmoscopy (SLO) photographs as intermediate images. For this purpose, first a retinal vessel segmentation is applied to extract main blood vessels from the FA and SLO images. Next, a novel global registration is proposed based on the Gaussian model for curved surface of retina. For doing so, first a global rigid transformation is applied to FA vessel-map image using a new feature-based method to align it with SLO vessel-map photograph, in a way that outlier matched features resulted from not-perfect vessel segmentation are completely eliminated. After that, the transformed image is globally registered again considering Gaussian model for curved surface of retina to improve the precision of the previous step. Eventually a local non-rigid transformation is exploited to register two images perfectly. The experimental results indicate the presented scheme is more precise compared to other registration methods.
Collapse
Affiliation(s)
- Ramin Almasi
- Department of Computer Engineering, Faculty of Engineering, University of Isfahan, Isfahan, Iran
| | - Abbas Vafaei
- Department of Computer Engineering, Faculty of Engineering, University of Isfahan, Isfahan, Iran
| | - Zeinab Ghasemi
- Department of Electrical and Computer Engineering, University of Detroit Mercy, Detroit, MI 48202, USA
| | | | - Ali Reza Dehghani
- Didavaran Eye Clinic, Isfahan, Iran
- Department of Ophthalmology, School of Medicine, Isfahan University of Medical Sciences, Isfahan, Iran
| | - Hossein Rabbani
- Medical Image & Signal Processing Research Center, Isfahan University of Medical Sciences, Isfahan, Iran
| |
Collapse
|
19
|
Grey-Wolf-Based Wang's Demons for Retinal Image Registration. ENTROPY 2020; 22:e22060659. [PMID: 33286433 PMCID: PMC7517193 DOI: 10.3390/e22060659] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/07/2020] [Revised: 06/04/2020] [Accepted: 06/06/2020] [Indexed: 11/28/2022]
Abstract
Image registration has an imperative role in medical imaging. In this work, a grey-wolf optimizer (GWO)-based non-rigid demons registration is proposed to support the retinal image registration process. A comparative study of the proposed GWO-based demons registration framework with cuckoo search, firefly algorithm, and particle swarm optimization-based demons registration is conducted. In addition, a comparative analysis of different demons registration methods, such as Wang’s demons, Tang’s demons, and Thirion’s demons which are optimized using the proposed GWO is carried out. The results established the superiority of the GWO-based framework which achieved 0.9977 correlation, and fast processing compared to the use of the other optimization algorithms. Moreover, GWO-based Wang’s demons performed better accuracy compared to the Tang’s demons and Thirion’s demons framework. It also achieved the best less registration error of 8.36 × 10−5.
Collapse
|
20
|
Pan L, Shi F, Xiang D, Yu K, Duan L, Zheng J, Chen X. OCTRexpert:A Feature-based 3D Registration Method for Retinal OCT Images. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2020; 29:3885-3897. [PMID: 31995490 DOI: 10.1109/tip.2020.2967589] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
Medical image registration can be used for studying longitudinal and cross-sectional data, quantitatively monitoring disease progression and guiding computer assisted diagnosis and treatments. However, deformable registration which enables more precise and quantitative comparison has not been well developed for retinal optical coherence tomography (OCT) images. This paper proposes a new 3D registration approach for retinal OCT data called OCTRexpert. To the best of our knowledge, the proposed algorithm is the first full 3D registration approach for retinal OCT images which can be applied to longitudinal OCT images for both normal and serious pathological subjects. In this approach, a pre-processing method is first performed to remove eye motion artifact and then a novel design-detection-deformation strategy is applied for the registration. In the design step, a couple of features are designed for each voxel in the image. In the detection step, active voxels are selected and the point-to-point correspondences between the subject and template images are established. In the deformation step, the image is hierarchically deformed according to the detected correspondences in multi-resolution. The proposed method is evaluated on a dataset with longitudinal OCT images from 20 healthy subjects and 4 subjects diagnosed with serious Choroidal Neovascularization (CNV). Experimental results show that the proposed registration algorithm consistently yields statistically significant improvements in both Dice similarity coefficient and the average unsigned surface error compared with the other registration methods.
Collapse
|
21
|
Motta D, Casaca W, Paiva A. Vessel Optimal Transport for Automated Alignment of Retinal Fundus Images. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2019; 28:6154-6168. [PMID: 31283507 DOI: 10.1109/tip.2019.2925287] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/09/2023]
Abstract
Optimal transport has emerged as a promising and useful tool for supporting modern image processing applications such as medical imaging and scientific visualization. Indeed, the optimal transport theory enables great flexibility in modeling problems related to image registration, as different optimization resources can be successfully used as well as the choice of suitable matching models to align the images. In this paper, we introduce an automated framework for fundus image registration which unifies optimal transport theory, image processing tools, and graph matching schemes into a functional and concise methodology. Given two ocular fundus images, we construct representative graphs which embed in their structures spatial and topological information from the eye's blood vessels. The graphs produced are then used as input by our optimal transport model in order to establish a correspondence between their sets of nodes. Finally, geometric transformations are performed between the images so as to accomplish the registration task properly. Our formulation relies on the solid mathematical foundation of optimal transport as a constrained optimization problem, being also robust when dealing with outliers created during the matching stage. We demonstrate the accuracy and effectiveness of the present framework throughout a comprehensive set of qualitative and quantitative comparisons against several influential state-of-the-art methods on various fundus image databases.
Collapse
|
22
|
Bashiri FS, Baghaie A, Rostami R, Yu Z, D’Souza RM. Multi-Modal Medical Image Registration with Full or Partial Data: A Manifold Learning Approach. J Imaging 2018; 5:5. [PMID: 34470183 PMCID: PMC8320870 DOI: 10.3390/jimaging5010005] [Citation(s) in RCA: 16] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/20/2018] [Revised: 12/23/2018] [Accepted: 12/25/2018] [Indexed: 11/16/2022] Open
Abstract
Multi-modal image registration is the primary step in integrating information stored in two or more images, which are captured using multiple imaging modalities. In addition to intensity variations and structural differences between images, they may have partial or full overlap, which adds an extra hurdle to the success of registration process. In this contribution, we propose a multi-modal to mono-modal transformation method that facilitates direct application of well-founded mono-modal registration methods in order to obtain accurate alignment of multi-modal images in both cases, with complete (full) and incomplete (partial) overlap. The proposed transformation facilitates recovering strong scales, rotations, and translations. We explain the method thoroughly and discuss the choice of parameters. For evaluation purposes, the effectiveness of the proposed method is examined and compared with widely used information theory-based techniques using simulated and clinical human brain images with full data. Using RIRE dataset, mean absolute error of 1.37, 1.00, and 1.41 mm are obtained for registering CT images with PD-, T1-, and T2-MRIs, respectively. In the end, we empirically investigate the efficacy of the proposed transformation in registering multi-modal partially overlapped images.
Collapse
Affiliation(s)
- Fereshteh S. Bashiri
- Department of Electrical Engineering, University of Wisconsin-Milwaukee, Milwaukee, WI 53211, USA
| | - Ahmadreza Baghaie
- Department of Electrical Engineering, University of Wisconsin-Milwaukee, Milwaukee, WI 53211, USA
| | - Reihaneh Rostami
- Department of Computer Science, University of Wisconsin-Milwaukee, Milwaukee, WI 53211, USA
| | - Zeyun Yu
- Department of Electrical Engineering, University of Wisconsin-Milwaukee, Milwaukee, WI 53211, USA
- Department of Computer Science, University of Wisconsin-Milwaukee, Milwaukee, WI 53211, USA
| | - Roshan M. D’Souza
- Department of Mechanical Engineering, University of Wisconsin-Milwaukee, Milwaukee, WI 53211, USA
| |
Collapse
|
23
|
A-RANSAC: Adaptive random sample consensus method in multimodal retinal image registration. Biomed Signal Process Control 2018. [DOI: 10.1016/j.bspc.2018.06.002] [Citation(s) in RCA: 18] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/20/2022]
|