1
|
Galazis C, Shepperd S, Brouwer EJP, Queiros S, Alskaf E, Anjari M, Chiribiri A, Lee J, Bharath AA, Varela M. High-Resolution Maps of Left Atrial Displacements and Strains Estimated With 3D Cine MRI Using Online Learning Neural Networks. IEEE TRANSACTIONS ON MEDICAL IMAGING 2025; 44:2056-2067. [PMID: 40030862 DOI: 10.1109/tmi.2025.3526364] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/05/2025]
Abstract
The functional analysis of the left atrium (LA) is important for evaluating cardiac health and understanding diseases like atrial fibrillation. Cine MRI is ideally placed for the detailed 3D characterization of LA motion and deformation but is lacking appropriate acquisition and analysis tools. Here, we propose tools for the Analysis of Left Atrial Displacements and DeformatIons using online learning neural Networks (Aladdin) and present a technical feasibility study on how Aladdin can characterize 3D LA function globally and regionally. Aladdin includes an online segmentation and image registration network, and a strain calculation pipeline tailored to the LA. We create maps of LA Displacement Vector Field (DVF) magnitude and LA principal strain values from images of 10 healthy volunteers and 8 patients with cardiovascular disease (CVD), of which 2 had large left ventricular ejection fraction (LVEF) impairment. We additionally create an atlas of these biomarkers using the data from the healthy volunteers. Results showed that Aladdin can accurately track the LA wall across the cardiac cycle and characterize its motion and deformation. Global LA function markers assessed with Aladdin agree well with estimates from 2D Cine MRI. A more marked active contraction phase was observed in the healthy cohort, while the CVD $\text {LVEF}_{\downarrow } $ group showed overall reduced LA function. Aladdin is uniquely able to identify LA regions with abnormal deformation metrics that may indicate focal pathology. We expect Aladdin to have important clinical applications as it can non-invasively characterize atrial pathophysiology. All source code and data are available at: https://github.com/cgalaz01/aladdin_cmr_la.
Collapse
|
2
|
Hossain T, Zhang M. MGAug: Multimodal Geometric Augmentation in Latent Spaces of Image Deformations. Med Image Anal 2025; 102:103540. [PMID: 40117987 DOI: 10.1016/j.media.2025.103540] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/16/2024] [Revised: 01/26/2025] [Accepted: 03/02/2025] [Indexed: 03/23/2025]
Abstract
Geometric transformations have been widely used to augment the size of training images. Existing methods often assume a unimodal distribution of the underlying transformations between images, which limits their power when data with multimodal distributions occur. In this paper, we propose a novel model, Multimodal Geometric Augmentation (MGAug), that for the first time generates augmenting transformations in a multimodal latent space of geometric deformations. To achieve this, we first develop a deep network that embeds the learning of latent geometric spaces of diffeomorphic transformations (a.k.a. diffeomorphisms) in a variational autoencoder (VAE). A mixture of multivariate Gaussians is formulated in the tangent space of diffeomorphisms and serves as a prior to approximate the hidden distribution of image transformations. We then augment the original training dataset by deforming images using randomly sampled transformations from the learned multimodal latent space of VAE. To validate the efficiency of our model, we jointly learn the augmentation strategy with two distinct domain-specific tasks: multi-class classification on both synthetic 2D and real 3D brain MRIs, and segmentation on real 3D brain MRIs dataset. We also compare MGAug with state-of-the-art transformation-based image augmentation algorithms. Experimental results show that our proposed approach outperforms all baselines by significantly improved prediction accuracy. Our code is publicly available at GitHub.
Collapse
Affiliation(s)
- Tonmoy Hossain
- Department of Computer Science, Department of Electrical and Computer Engineering, School of Engineering & Applied Science, University of Virginia, Charlottesville, 22903, VA, United States
| | - Miaomiao Zhang
- Department of Computer Science, Department of Electrical and Computer Engineering, School of Engineering & Applied Science, University of Virginia, Charlottesville, 22903, VA, United States.
| |
Collapse
|
3
|
Chen J, Liu Y, Wei S, Bian Z, Subramanian S, Carass A, Prince JL, Du Y. A survey on deep learning in medical image registration: New technologies, uncertainty, evaluation metrics, and beyond. Med Image Anal 2025; 100:103385. [PMID: 39612808 PMCID: PMC11730935 DOI: 10.1016/j.media.2024.103385] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/12/2023] [Revised: 10/27/2024] [Accepted: 11/01/2024] [Indexed: 12/01/2024]
Abstract
Deep learning technologies have dramatically reshaped the field of medical image registration over the past decade. The initial developments, such as regression-based and U-Net-based networks, established the foundation for deep learning in image registration. Subsequent progress has been made in various aspects of deep learning-based registration, including similarity measures, deformation regularizations, network architectures, and uncertainty estimation. These advancements have not only enriched the field of image registration but have also facilitated its application in a wide range of tasks, including atlas construction, multi-atlas segmentation, motion estimation, and 2D-3D registration. In this paper, we present a comprehensive overview of the most recent advancements in deep learning-based image registration. We begin with a concise introduction to the core concepts of deep learning-based image registration. Then, we delve into innovative network architectures, loss functions specific to registration, and methods for estimating registration uncertainty. Additionally, this paper explores appropriate evaluation metrics for assessing the performance of deep learning models in registration tasks. Finally, we highlight the practical applications of these novel techniques in medical imaging and discuss the future prospects of deep learning-based image registration.
Collapse
Affiliation(s)
- Junyu Chen
- Department of Radiology and Radiological Science, Johns Hopkins School of Medicine, MD, USA.
| | - Yihao Liu
- Department of Electrical and Computer Engineering, Johns Hopkins University, MD, USA
| | - Shuwen Wei
- Department of Electrical and Computer Engineering, Johns Hopkins University, MD, USA
| | - Zhangxing Bian
- Department of Electrical and Computer Engineering, Johns Hopkins University, MD, USA
| | - Shalini Subramanian
- Department of Radiology and Radiological Science, Johns Hopkins School of Medicine, MD, USA
| | - Aaron Carass
- Department of Electrical and Computer Engineering, Johns Hopkins University, MD, USA
| | - Jerry L Prince
- Department of Electrical and Computer Engineering, Johns Hopkins University, MD, USA
| | - Yong Du
- Department of Radiology and Radiological Science, Johns Hopkins School of Medicine, MD, USA
| |
Collapse
|
4
|
Comte V, Alenya M, Urru A, Recober J, Nakaki A, Crovetto F, Camara O, Gratacós E, Eixarch E, Crispi F, Piella G, Ceresa M, González Ballester MA. Deep cascaded registration and weakly-supervised segmentation of fetal brain MRI. Heliyon 2025; 11:e40148. [PMID: 39816514 PMCID: PMC11732682 DOI: 10.1016/j.heliyon.2024.e40148] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2024] [Revised: 09/28/2024] [Accepted: 11/04/2024] [Indexed: 01/18/2025] Open
Abstract
Deformable image registration is a cornerstone of many medical image analysis applications, particularly in the context of fetal brain magnetic resonance imaging (MRI), where precise registration is essential for studying the rapidly evolving fetal brain during pregnancy and potentially identifying neurodevelopmental abnormalities. While deep learning has become the leading approach for medical image registration, traditional convolutional neural networks (CNNs) often fall short in capturing fine image details due to their bias toward low spatial frequencies. To address this challenge, we introduce a deep learning registration framework comprising multiple cascaded convolutional networks. These networks predict a series of incremental deformation fields that transform the moving image at various spatial frequency levels, ensuring accurate alignment with the fixed image. This multi-resolution approach allows for a more accurate and detailed registration process, capturing both coarse and fine image structures. Our method outperforms existing state-of-the-art techniques, including other multi-resolution strategies, by a substantial margin. Furthermore, we integrate our registration method into a multi-atlas segmentation pipeline and showcase its competitive performance compared to nnU-Net, achieved using only a small subset of annotated images as atlases. This approach is particularly valuable in the context of fetal brain MRI, where annotated datasets are limited. Our pipeline for registration and multi-atlas segmentation is publicly available at https://github.com/ValBcn/CasReg.
Collapse
Affiliation(s)
- Valentin Comte
- BCN MedTech, Department of Information and Communication Technologies, Universitat Pompeu Fabra, Barcelona, Spain
- European Commission, Joint Research Centre (JRC), Geel, Belgium
| | - Mireia Alenya
- BCN MedTech, Department of Information and Communication Technologies, Universitat Pompeu Fabra, Barcelona, Spain
| | - Andrea Urru
- BCN MedTech, Department of Information and Communication Technologies, Universitat Pompeu Fabra, Barcelona, Spain
| | - Judith Recober
- BCN MedTech, Department of Information and Communication Technologies, Universitat Pompeu Fabra, Barcelona, Spain
| | - Ayako Nakaki
- BCNatal, Fetal Medicine Research Center (Hospital Clínic and Hospital Sant Joan de Déu), University of Barcelona, Barcelona, Spain
- Institut d’Investigacions Biomédiques August Pi i Sunyer (IDIBAPS), Barcelona, Spain
| | - Francesca Crovetto
- BCNatal, Fetal Medicine Research Center (Hospital Clínic and Hospital Sant Joan de Déu), University of Barcelona, Barcelona, Spain
| | - Oscar Camara
- BCN MedTech, Department of Information and Communication Technologies, Universitat Pompeu Fabra, Barcelona, Spain
| | - Eduard Gratacós
- BCNatal, Fetal Medicine Research Center (Hospital Clínic and Hospital Sant Joan de Déu), University of Barcelona, Barcelona, Spain
- Institut d’Investigacions Biomédiques August Pi i Sunyer (IDIBAPS), Barcelona, Spain
- Centre for Biomedical Research on Rare Diseases (CIBERER), Barcelona, Spain
| | - Elisenda Eixarch
- BCNatal, Fetal Medicine Research Center (Hospital Clínic and Hospital Sant Joan de Déu), University of Barcelona, Barcelona, Spain
- Institut d’Investigacions Biomédiques August Pi i Sunyer (IDIBAPS), Barcelona, Spain
- Centre for Biomedical Research on Rare Diseases (CIBERER), Barcelona, Spain
| | - Fatima Crispi
- BCNatal, Fetal Medicine Research Center (Hospital Clínic and Hospital Sant Joan de Déu), University of Barcelona, Barcelona, Spain
- Institut d’Investigacions Biomédiques August Pi i Sunyer (IDIBAPS), Barcelona, Spain
- Centre for Biomedical Research on Rare Diseases (CIBERER), Barcelona, Spain
| | - Gemma Piella
- BCN MedTech, Department of Information and Communication Technologies, Universitat Pompeu Fabra, Barcelona, Spain
| | - Mario Ceresa
- BCN MedTech, Department of Information and Communication Technologies, Universitat Pompeu Fabra, Barcelona, Spain
- European Commission, Joint Research Centre (JRC), Ispra, Italy
| | - Miguel A. González Ballester
- BCN MedTech, Department of Information and Communication Technologies, Universitat Pompeu Fabra, Barcelona, Spain
- ICREA, Barcelona, Spain
| |
Collapse
|
5
|
Duan T, Chen W, Ruan M, Zhang X, Shen S, Gu W. Unsupervised deep learning-based medical image registration: a survey. Phys Med Biol 2025; 70:02TR01. [PMID: 39667278 DOI: 10.1088/1361-6560/ad9e69] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/02/2024] [Accepted: 12/12/2024] [Indexed: 12/14/2024]
Abstract
In recent decades, medical image registration technology has undergone significant development, becoming one of the core technologies in medical image analysis. With the rise of deep learning, deep learning-based medical image registration methods have achieved revolutionary improvements in processing speed and automation, showing great potential, especially in unsupervised learning. This paper briefly introduces the core concepts of deep learning-based unsupervised image registration, followed by an in-depth discussion of innovative network architectures and a detailed review of these studies, highlighting their unique contributions. Additionally, this paper explores commonly used loss functions, datasets, and evaluation metrics. Finally, we discuss the main challenges faced by various categories and propose potential future research topics. This paper surveys the latest advancements in unsupervised deep neural network-based medical image registration methods, aiming to help active readers interested in this field gain a deep understanding of this exciting area.
Collapse
Affiliation(s)
- Taisen Duan
- School of Computer, Electronics and Information, Guangxi University, Nanning 530004, People's Republic of China
- Guangxi Key Laboratory of Multimedia Communications and Network Technology, Guangxi University, Nanning 530004, People's Republic of China
| | - Wenkang Chen
- School of Computer, Electronics and Information, Guangxi University, Nanning 530004, People's Republic of China
- Guangxi Key Laboratory of Multimedia Communications and Network Technology, Guangxi University, Nanning 530004, People's Republic of China
| | - Meilin Ruan
- School of Computer, Electronics and Information, Guangxi University, Nanning 530004, People's Republic of China
| | - Xuejun Zhang
- School of Computer, Electronics and Information, Guangxi University, Nanning 530004, People's Republic of China
- Guangxi Key Laboratory of Multimedia Communications and Network Technology, Guangxi University, Nanning 530004, People's Republic of China
| | - Shaofei Shen
- School of Computer, Electronics and Information, Guangxi University, Nanning 530004, People's Republic of China
- Guangxi Key Laboratory of Multimedia Communications and Network Technology, Guangxi University, Nanning 530004, People's Republic of China
| | - Weiyu Gu
- School of Computer, Electronics and Information, Guangxi University, Nanning 530004, People's Republic of China
- Guangxi Key Laboratory of Multimedia Communications and Network Technology, Guangxi University, Nanning 530004, People's Republic of China
| |
Collapse
|
6
|
Liu L, Aviles-Rivero AI, Schonlieb CB. Contrastive Registration for Unsupervised Medical Image Segmentation. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2025; 36:147-159. [PMID: 37983143 DOI: 10.1109/tnnls.2023.3332003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/22/2023]
Abstract
Medical image segmentation is an important task in medical imaging, as it serves as the first step for clinical diagnosis and treatment planning. While major success has been reported using deep learning supervised techniques, they assume a large and well-representative labeled set. This is a strong assumption in the medical domain where annotations are expensive, time-consuming, and inherent to human bias. To address this problem, unsupervised segmentation techniques have been proposed in the literature. Yet, none of the existing unsupervised segmentation techniques reach accuracies that come even near to the state-of-the-art of supervised segmentation methods. In this work, we present a novel optimization model framed in a new convolutional neural network (CNN)-based contrastive registration architecture for unsupervised medical image segmentation called CLMorph. The core idea of our approach is to exploit image-level registration and feature-level contrastive learning, to perform registration-based segmentation. First, we propose an architecture to capture the image-to-image transformation mapping via registration for unsupervised medical image segmentation. Second, we embed a contrastive learning mechanism in the registration architecture to enhance the discriminative capacity of the network at the feature level. We show that our proposed CLMorph technique mitigates the major drawbacks of existing unsupervised techniques. We demonstrate, through numerical and visual experiments, that our technique substantially outperforms the current state-of-the-art unsupervised segmentation methods on two major medical image datasets.
Collapse
|
7
|
Shen C, Zhu H, Zhou Y, Liu Y, Yi S, Dong L, Zhao W, Brady DJ, Cao X, Ma Z, Lin Y. Continuous 3D Myocardial Motion Tracking via Echocardiography. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:4236-4252. [PMID: 38935475 DOI: 10.1109/tmi.2024.3419780] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/29/2024]
Abstract
Myocardial motion tracking stands as an essential clinical tool in the prevention and detection of cardiovascular diseases (CVDs), the foremost cause of death globally. However, current techniques suffer from incomplete and inaccurate motion estimation of the myocardium in both spatial and temporal dimensions, hindering the early identification of myocardial dysfunction. To address these challenges, this paper introduces the Neural Cardiac Motion Field (NeuralCMF). NeuralCMF leverages implicit neural representation (INR) to model the 3D structure and the comprehensive 6D forward/backward motion of the heart. This method surpasses pixel-wise limitations by offering the capability to continuously query the precise shape and motion of the myocardium at any specific point throughout the cardiac cycle, enhancing the detailed analysis of cardiac dynamics beyond traditional speckle tracking. Notably, NeuralCMF operates without the need for paired datasets, and its optimization is self-supervised through the physics knowledge priors in both space and time dimensions, ensuring compatibility with both 2D and 3D echocardiogram video inputs. Experimental validations across three representative datasets support the robustness and innovative nature of the NeuralCMF, marking significant advantages over existing state-of-the-art methods in cardiac imaging and motion tracking. Code is available at: https://njuvision.github.io/NeuralCMF.
Collapse
|
8
|
Li J, Tuckute G, Fedorenko E, Edlow BL, Dalca AV, Fischl B. JOSA: Joint surface-based registration and atlas construction of brain geometry and function. Med Image Anal 2024; 98:103292. [PMID: 39173411 DOI: 10.1016/j.media.2024.103292] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/27/2023] [Revised: 06/21/2024] [Accepted: 07/30/2024] [Indexed: 08/24/2024]
Abstract
Surface-based cortical registration is an important topic in medical image analysis and facilitates many downstream applications. Current approaches for cortical registration are mainly driven by geometric features, such as sulcal depth and curvature, and often assume that registration of folding patterns leads to alignment of brain function. However, functional variability of anatomically corresponding areas across subjects has been widely reported, particularly in higher-order cognitive areas. In this work, we present JOSA, a novel cortical registration framework that jointly models the mismatch between geometry and function while simultaneously learning an unbiased population-specific atlas. Using a semi-supervised training strategy, JOSA achieves superior registration performance in both geometry and function to the state-of-the-art methods but without requiring functional data at inference. This learning framework can be extended to any auxiliary data to guide spherical registration that is available during training but is difficult or impossible to obtain during inference, such as parcellations, architectonic identity, transcriptomic information, and molecular profiles. By recognizing the mismatch between geometry and function, JOSA provides new insights into the future development of registration methods using joint analysis of brain structure and function.
Collapse
Affiliation(s)
- Jian Li
- A. A. Martinos Center for Biomedical Imaging, Department of Radiology, Massachusetts General Hospital and Harvard Medical School, United States of America; Center for Neurotechnology and Neurorecovery, Department of Neurology, Massachusetts General Hospital and Harvard Medical School, United States of America.
| | - Greta Tuckute
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, United States of America; McGovern Institute for Brain Research, Massachusetts Institute of Technology, United States of America
| | - Evelina Fedorenko
- Department of Brain and Cognitive Sciences, Massachusetts Institute of Technology, United States of America; McGovern Institute for Brain Research, Massachusetts Institute of Technology, United States of America; Program in Speech Hearing Bioscience and Technology, Harvard University, United States of America
| | - Brian L Edlow
- A. A. Martinos Center for Biomedical Imaging, Department of Radiology, Massachusetts General Hospital and Harvard Medical School, United States of America; Center for Neurotechnology and Neurorecovery, Department of Neurology, Massachusetts General Hospital and Harvard Medical School, United States of America
| | - Adrian V Dalca
- A. A. Martinos Center for Biomedical Imaging, Department of Radiology, Massachusetts General Hospital and Harvard Medical School, United States of America; Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology, United States of America
| | - Bruce Fischl
- A. A. Martinos Center for Biomedical Imaging, Department of Radiology, Massachusetts General Hospital and Harvard Medical School, United States of America; Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology, United States of America
| |
Collapse
|
9
|
Hua Y, Xu K, Yang X. Variational image registration with learned prior using multi-stage VAEs. Comput Biol Med 2024; 178:108785. [PMID: 38925089 DOI: 10.1016/j.compbiomed.2024.108785] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/23/2023] [Revised: 05/16/2024] [Accepted: 06/15/2024] [Indexed: 06/28/2024]
Abstract
Variational Autoencoders (VAEs) are an efficient variational inference technique coupled with the generated network. Due to the uncertainty provided by variational inference, VAEs have been applied in medical image registration. However, a critical problem in VAEs is that the simple prior cannot provide suitable regularization, which leads to the mismatch between the variational posterior and prior. An optimal prior can close the gap between the evidence's real and variational posterior. In this paper, we propose a multi-stage VAE to learn the optimal prior, which is the aggregated posterior. A lightweight VAE is used to generate the aggregated posterior as a whole. It is an effective way to estimate the distribution of the high-dimensional aggregated posterior that commonly exists in medical image registration based on VAEs. A factorized telescoping classifier is trained to estimate the density ratio of a simple given prior and aggregated posterior, aiming to calculate the KL divergence between the variational and aggregated posterior more accurately. We analyze the KL divergence and find that the finer the factorization, the smaller the KL divergence is. However, too fine a partition is not conducive to registration accuracy. Moreover, the diagonal hypothesis of the variational posterior's covariance ignores the relationship between latent variables in image registration. To address this issue, we learn a covariance matrix with low-rank information to enable correlations with each dimension of the variational posterior. The covariance matrix is further used as a measure to reduce the uncertainty of deformation fields. Experimental results on four public medical image datasets demonstrate that our proposed method outperforms other methods in negative log-likelihood (NLL) and achieves better registration accuracy.
Collapse
Affiliation(s)
- Yong Hua
- College of Computer Science and Software Engineering, Shenzhen University, Shenzhen, 518060, Guangdong, China
| | - Kangrong Xu
- College of Computer Science and Software Engineering, Shenzhen University, Shenzhen, 518060, Guangdong, China
| | - Xuan Yang
- College of Computer Science and Software Engineering, Shenzhen University, Shenzhen, 518060, Guangdong, China.
| |
Collapse
|
10
|
Hernandez M, Ramon Julvez U. Insights into traditional Large Deformation Diffeomorphic Metric Mapping and unsupervised deep-learning for diffeomorphic registration and their evaluation. Comput Biol Med 2024; 178:108761. [PMID: 38908357 DOI: 10.1016/j.compbiomed.2024.108761] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/29/2023] [Revised: 06/04/2024] [Accepted: 06/13/2024] [Indexed: 06/24/2024]
Abstract
This paper explores the connections between traditional Large Deformation Diffeomorphic Metric Mapping methods and unsupervised deep-learning approaches for non-rigid registration, particularly emphasizing diffeomorphic registration. The study provides useful insights and establishes connections between the methods, thereby facilitating a profound understanding of the methodological landscape. The methods considered in our study are extensively evaluated in T1w MRI images using traditional NIREP and Learn2Reg OASIS evaluation protocols with a focus on fairness, to establish equitable benchmarks and facilitate informed comparisons. Through a comprehensive analysis of the results, we address key questions, including the intricate relationship between accuracy and transformation quality in performance, the disentanglement of the influence of registration ingredients on performance, and the determination of benchmark methods and baselines. We offer valuable insights into the strengths and limitations of both traditional and deep-learning methods, shedding light on their comparative performance and guiding future advancements in the field.
Collapse
Affiliation(s)
- Monica Hernandez
- Computer Science Department, University of Zaragoza, Spain; Aragon Institute on Engineering Research, Spain.
| | - Ubaldo Ramon Julvez
- Computer Science Department, University of Zaragoza, Spain; Aragon Institute on Engineering Research, Spain
| |
Collapse
|
11
|
Liu Y, Hu Z. An unsupervised learning-based guidewire shape registration for vascular intervention surgery robot. Comput Methods Biomech Biomed Engin 2024:1-17. [PMID: 39037421 DOI: 10.1080/10255842.2024.2372637] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/16/2023] [Accepted: 06/20/2024] [Indexed: 07/23/2024]
Abstract
In a vascular interventional surgery robot(VISR), a high transparency master-slave system can aid physicians in the more precise manipulation of guidewires for navigation and operation within blood vessels. However, deformations arising from the movement of the guidewire can affect the accuracy of the registration, thus reducing the transparency of the master-slave system. In this study, the degree of the guidewire's deformation is analyzed based on the Kirchhoff model. An unsupervised learning-based guidewire shape registration method(UL-GSR) is proposed to estimate geometric transformations by learning displacement field functions. It can effectively achieve precise registration of flexible bodies. This method not only demonstrates high registration accuracy but also performs robustly under different complexity degrees of guidewire shapes. The experiments have demonstrated that the UL-GSR method significantly improves the accuracy of shape point set registration between the master and slave sides, thus enhancing the transparency and operational reliability of the VISR system.
Collapse
Affiliation(s)
- Yueling Liu
- School of Electronics and Electrical Engineering, Shanghai University of Engineering Science, Shanghai, China
| | - Zhi Hu
- Laboratory of Intelligent Control and Robotics, Shanghai University of Engineering Science, Shanghai, China
| |
Collapse
|
12
|
Alae Eddine EB, Scheiber C, Grenier T, Janier M, Flaus A. CT-guided spatial normalization of nuclear hybrid imaging adapted to enlarged ventricles: Impact on striatal uptake quantification. Neuroimage 2024; 294:120631. [PMID: 38701993 DOI: 10.1016/j.neuroimage.2024.120631] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/09/2023] [Revised: 04/25/2024] [Accepted: 04/30/2024] [Indexed: 05/06/2024] Open
Abstract
INTRODUCTION Spatial normalization is a prerequisite step for the quantitative analysis of SPECT or PET brain images using volume-of-interest (VOI) template or voxel-based analysis. MRI-guided spatial normalization is the gold standard, but the wide use of PET/CT or SPECT/CT in routine clinical practice makes CT-guided spatial normalization a necessary alternative. Ventricular enlargement is observed with aging, and it hampers the spatial normalization of the lateral ventricles and striatal regions, limiting their analysis. The aim of the present study was to propose a robust spatial normalization method based on CT scans that takes into account features of the aging brain to reduce bias in the CT-guided striatal analysis of SPECT images. METHODS We propose an enhanced CT-guided spatial normalization pipeline based on SPM12. Performance of the proposed pipeline was assessed on visually normal [123I]-FP-CIT SPECT/CT images. SPM12 default CT-guided spatial normalization was used as reference method. The metrics assessed were the overlap between the spatially normalized lateral ventricles and caudate/putamen VOIs, and the computation of caudate and putamen specific binding ratios (SBR). RESULTS In total 231 subjects (mean age ± SD = 61.9 ± 15.5 years) were included in the statistical analysis. The mean overlap between the spatially normalized lateral ventricles of subjects and the caudate VOI and the mean SBR of caudate were respectively 38.40 % (± SD = 19.48 %) of the VOI and 1.77 (± 0.79) when performing SPM12 default spatial normalization. The mean overlap decreased to 9.13 % (± SD = 1.41 %, P < 0.001) of the VOI and the SBR of caudate increased to 2.38 (± 0.51, P < 0.0001) when performing the proposed pipeline. Spatially normalized lateral ventricles did not overlap with putamen VOI using either method. The mean putamen SBR value derived from the proposed spatial normalization (2.75 ± 0.54) was not significantly different from that derived from the default SPM12 spatial normalization (2.83 ± 0.52, P > 0.05). CONCLUSION The automatic CT-guided spatial normalization used herein led to a less biased spatial normalization of SPECT images, hence an improved semi-quantitative analysis. The proposed pipeline could be implemented in clinical routine to perform a more robust SBR computation using hybrid imaging.
Collapse
Affiliation(s)
- El Barkaoui Alae Eddine
- Département de médecine nucléaire, Groupement Hospitalier Est, Hospices Civils de Lyon, Bron, France; INSA-Lyon, Universite Claude Bernard Lyon 1, CNRS, Inserm, CREATIS UMR 5220, U1294, F-69100, LYON, France
| | - Christian Scheiber
- Département de médecine nucléaire, Groupement Hospitalier Est, Hospices Civils de Lyon, Bron, France; Institut des Sciences Cognitives Marc Jeannerod, UMR 5229, CNRS, CRNL, Université Claude Bernard Lyon 1, Lyon, France
| | - Thomas Grenier
- INSA-Lyon, Universite Claude Bernard Lyon 1, CNRS, Inserm, CREATIS UMR 5220, U1294, F-69100, LYON, France
| | - Marc Janier
- Département de médecine nucléaire, Groupement Hospitalier Est, Hospices Civils de Lyon, Bron, France; Faculté de Médecine Lyon Est, Université Claude Bernard Lyon 1, Lyon, France; Laboratoire d'Automatique, de génie des procédés et de génie pharmaceutique, LAGEPP, UMR 5007 UCBL1 - CNRS, Lyon, France
| | - Anthime Flaus
- Département de médecine nucléaire, Groupement Hospitalier Est, Hospices Civils de Lyon, Bron, France; Faculté de Médecine Lyon Est, Université Claude Bernard Lyon 1, Lyon, France; Centre de Recherche en Neurosciences de Lyon, INSERM U1028/CNRS UMR5292, Lyon, France.
| |
Collapse
|
13
|
Zou J, Song Y, Liu L, Aviles-Rivero AI, Qin J. Unsupervised lung CT image registration via stochastic decomposition of deformation fields. Comput Med Imaging Graph 2024; 115:102397. [PMID: 38735104 DOI: 10.1016/j.compmedimag.2024.102397] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/08/2023] [Revised: 01/30/2024] [Accepted: 05/01/2024] [Indexed: 05/14/2024]
Abstract
We address the problem of lung CT image registration, which underpins various diagnoses and treatments for lung diseases. The main crux of the problem is the large deformation that the lungs undergo during respiration. This physiological process imposes several challenges from a learning point of view. In this paper, we propose a novel training scheme, called stochastic decomposition, which enables deep networks to effectively learn such a difficult deformation field during lung CT image registration. The key idea is to stochastically decompose the deformation field, and supervise the registration by synthetic data that have the corresponding appearance discrepancy. The stochastic decomposition allows for revealing all possible decompositions of the deformation field. At the learning level, these decompositions can be seen as a prior to reduce the ill-posedness of the registration yielding to boost the performance. We demonstrate the effectiveness of our framework on Lung CT data. We show, through extensive numerical and visual results, that our technique outperforms existing methods.
Collapse
Affiliation(s)
- Jing Zou
- Center for Smart Health, School of Nursing, The Hong Kong Polytechnic University, Hung Hom, Hong Kong, China
| | - Youyi Song
- Department of Data Science, School of Science, China Pharmaceutical University, Nan Jing, 210009, China
| | - Lihao Liu
- Department of Applied Mathematics and Theoretical Physics, University of Cambridge, Cambridge, CB30WA, UK
| | - Angelica I Aviles-Rivero
- Department of Applied Mathematics and Theoretical Physics, University of Cambridge, Cambridge, CB30WA, UK
| | - Jing Qin
- Center for Smart Health, School of Nursing, The Hong Kong Polytechnic University, Hung Hom, Hong Kong, China.
| |
Collapse
|
14
|
Huang Z, Zhao R, Leung FHF, Banerjee S, Lam KM, Zheng YP, Ling SH. Landmark Localization From Medical Images With Generative Distribution Prior. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:2679-2692. [PMID: 38421850 DOI: 10.1109/tmi.2024.3371948] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/02/2024]
Abstract
In medical image analysis, anatomical landmarks usually contain strong prior knowledge of their structural information. In this paper, we propose to promote medical landmark localization by modeling the underlying landmark distribution via normalizing flows. Specifically, we introduce the flow-based landmark distribution prior as a learnable objective function into a regression-based landmark localization framework. Moreover, we employ an integral operation to make the mapping from heatmaps to coordinates differentiable to further enhance heatmap-based localization with the learned distribution prior. Our proposed Normalizing Flow-based Distribution Prior (NFDP) employs a straightforward backbone and non-problem-tailored architecture (i.e., ResNet18), which delivers high-fidelity outputs across three X-ray-based landmark localization datasets. Remarkably, the proposed NFDP can do the job with minimal additional computational burden as the normalizing flows module is detached from the framework on inferencing. As compared to existing techniques, our proposed NFDP provides a superior balance between prediction accuracy and inference speed, making it a highly efficient and effective approach. The source code of this paper is available at https://github.com/jacksonhzx95/NFDP.
Collapse
|
15
|
Hoffmann M, Hoopes A, Greve DN, Fischl B, Dalca AV. Anatomy-aware and acquisition-agnostic joint registration with SynthMorph. IMAGING NEUROSCIENCE (CAMBRIDGE, MASS.) 2024; 2:1-33. [PMID: 39015335 PMCID: PMC11247402 DOI: 10.1162/imag_a_00197] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 01/25/2023] [Revised: 04/27/2024] [Accepted: 05/21/2024] [Indexed: 07/18/2024]
Abstract
Affine image registration is a cornerstone of medical-image analysis. While classical algorithms can achieve excellent accuracy, they solve a time-consuming optimization for every image pair. Deep-learning (DL) methods learn a function that maps an image pair to an output transform. Evaluating the function is fast, but capturing large transforms can be challenging, and networks tend to struggle if a test-image characteristic shifts from the training domain, such as the resolution. Most affine methods are agnostic to the anatomy the user wishes to align, meaning the registration will be inaccurate if algorithms consider all structures in the image. We address these shortcomings with SynthMorph, a fast, symmetric, diffeomorphic, and easy-to-use DL tool for joint affine-deformable registration of any brain image without preprocessing. First, we leverage a strategy that trains networks with widely varying images synthesized from label maps, yielding robust performance across acquisition specifics unseen at training. Second, we optimize the spatial overlap of select anatomical labels. This enables networks to distinguish anatomy of interest from irrelevant structures, removing the need for preprocessing that excludes content which would impinge on anatomy-specific registration. Third, we combine the affine model with a deformable hypernetwork that lets users choose the optimal deformation-field regularity for their specific data, at registration time, in a fraction of the time required by classical methods. This framework is applicable to learning anatomy-aware, acquisition-agnostic registration of any anatomy with any architecture, as long as label maps are available for training. We analyze how competing architectures learn affine transforms and compare state-of-the-art registration tools across an extremely diverse set of neuroimaging data, aiming to truly capture the behavior of methods in the real world. SynthMorph demonstrates high accuracy and is available at https://w3id.org/synthmorph, as a single complete end-to-end solution for registration of brain magnetic resonance imaging (MRI) data.
Collapse
Affiliation(s)
- Malte Hoffmann
- Athinoula A. Martinos Center for Biomedical Imaging, Charlestown, MA, United States
- Department of Radiology, Massachusetts General Hospital, Boston, MA, United States
- Department of Radiology, Harvard Medical School, Boston, MA, United States
| | - Andrew Hoopes
- Athinoula A. Martinos Center for Biomedical Imaging, Charlestown, MA, United States
- Department of Radiology, Massachusetts General Hospital, Boston, MA, United States
- Computer Science & Artificial Intelligence Laboratory, Massachusetts Institute of Technology, Cambridge, MA, United States
| | - Douglas N. Greve
- Athinoula A. Martinos Center for Biomedical Imaging, Charlestown, MA, United States
- Department of Radiology, Massachusetts General Hospital, Boston, MA, United States
- Department of Radiology, Harvard Medical School, Boston, MA, United States
| | - Bruce Fischl
- Athinoula A. Martinos Center for Biomedical Imaging, Charlestown, MA, United States
- Department of Radiology, Massachusetts General Hospital, Boston, MA, United States
- Department of Radiology, Harvard Medical School, Boston, MA, United States
- Computer Science & Artificial Intelligence Laboratory, Massachusetts Institute of Technology, Cambridge, MA, United States
| | - Adrian V. Dalca
- Athinoula A. Martinos Center for Biomedical Imaging, Charlestown, MA, United States
- Department of Radiology, Massachusetts General Hospital, Boston, MA, United States
- Department of Radiology, Harvard Medical School, Boston, MA, United States
- Computer Science & Artificial Intelligence Laboratory, Massachusetts Institute of Technology, Cambridge, MA, United States
| |
Collapse
|
16
|
Chang Q, Wang Y. Structure-aware independently trained multi-scale registration network for cardiac images. Med Biol Eng Comput 2024; 62:1795-1808. [PMID: 38381202 DOI: 10.1007/s11517-024-03039-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/08/2023] [Accepted: 01/31/2024] [Indexed: 02/22/2024]
Abstract
Image registration is a primary task in various medical image analysis applications. However, cardiac image registration is difficult due to the large non-rigid deformation of the heart and the complex anatomical structure. This paper proposes a structure-aware independently trained multi-scale registration network (SIMReg) to address this challenge. Using image pairs of different resolutions, independently train each registration network to extract image features of large deformation image pairs at different resolutions. In the testing stage, the large deformation registration is decomposed into a multi-scale registration process, and the deformation fields of different resolutions are fused by a step-by-step deformation method, thus solving the difficulty of directly processing large deformation. Meanwhile, the targeted introduction of MIND (modality independent neighborhood descriptor) structural features to guide network training enhances the registration of cardiac structural contours and improves the registration effect of local details. Experiments were carried out on the open cardiac dataset ACDC (automated cardiac diagnosis challenge), and the average Dice value of the experimental results of the proposed method was 0.833. Comparative experiments showed that the proposed SIMReg could better solve the problem of heart image registration and achieve a better registration effect on cardiac images.
Collapse
Affiliation(s)
- Qing Chang
- School of Information Science and Engineering, East China University of Science and Technology, Shanghai, China
| | - Yaqi Wang
- School of Information Science and Engineering, East China University of Science and Technology, Shanghai, China.
| |
Collapse
|
17
|
Yao Y, Zhong J, Zhang L, Khan S, Chen W. CartiMorph: A framework for automated knee articular cartilage morphometrics. Med Image Anal 2024; 91:103035. [PMID: 37992496 DOI: 10.1016/j.media.2023.103035] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2022] [Revised: 08/25/2023] [Accepted: 11/13/2023] [Indexed: 11/24/2023]
Abstract
We introduce CartiMorph, a framework for automated knee articular cartilage morphometrics. It takes an image as input and generates quantitative metrics for cartilage subregions, including the percentage of full-thickness cartilage loss (FCL), mean thickness, surface area, and volume. CartiMorph leverages the power of deep learning models for hierarchical image feature representation. Deep learning models were trained and validated for tissue segmentation, template construction, and template-to-image registration. We established methods for surface-normal-based cartilage thickness mapping, FCL estimation, and rule-based cartilage parcellation. Our cartilage thickness map showed less error in thin and peripheral regions. We evaluated the effectiveness of the adopted segmentation model by comparing the quantitative metrics obtained from model segmentation and those from manual segmentation. The root-mean-squared deviation of the FCL measurements was less than 8%, and strong correlations were observed for the mean thickness (Pearson's correlation coefficient ρ∈[0.82,0.97]), surface area (ρ∈[0.82,0.98]) and volume (ρ∈[0.89,0.98]) measurements. We compared our FCL measurements with those from a previous study and found that our measurements deviated less from the ground truths. We observed superior performance of the proposed rule-based cartilage parcellation method compared with the atlas-based approach. CartiMorph has the potential to promote imaging biomarkers discovery for knee osteoarthritis.
Collapse
Affiliation(s)
- Yongcheng Yao
- CU Lab of AI in Radiology (CLAIR), Department of Imaging and Interventional Radiology, The Chinese University of Hong Kong, Hong Kong, China.
| | - Junru Zhong
- CU Lab of AI in Radiology (CLAIR), Department of Imaging and Interventional Radiology, The Chinese University of Hong Kong, Hong Kong, China
| | - Liping Zhang
- CU Lab of AI in Radiology (CLAIR), Department of Imaging and Interventional Radiology, The Chinese University of Hong Kong, Hong Kong, China
| | - Sheheryar Khan
- School of Professional Education and Executive Development, The Hong Kong Polytechnic University, Hong Kong, China
| | - Weitian Chen
- CU Lab of AI in Radiology (CLAIR), Department of Imaging and Interventional Radiology, The Chinese University of Hong Kong, Hong Kong, China.
| |
Collapse
|
18
|
Tzitzimpasis P, Zachiu C, Raaymakers BW, Ries M. SOLID: a novel similarity metric for mono-modal and multi-modal deformable image registration. Phys Med Biol 2023; 69:015020. [PMID: 38048629 DOI: 10.1088/1361-6560/ad120e] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/17/2023] [Accepted: 12/04/2023] [Indexed: 12/06/2023]
Abstract
Medical image registration is an integral part of various clinical applications including image guidance, motion tracking, therapy assessment and diagnosis. We present a robust approach for mono-modal and multi-modal medical image registration. To this end, we propose the novel shape operator based local image distance (SOLID) which estimates the similarity of images by comparing their second-order curvature information. Our similarity metric is rigorously tailored to be suitable for comparing images from different medical imaging modalities or image contrasts. A critical element of our method is the extraction of local features using higher-order shape information, enabling the accurate identification and registration of smaller structures. In order to assess the efficacy of the proposed similarity metric, we have implemented a variational image registration algorithm that relies on the principle of matching the curvature information of the given images. The performance of the proposed algorithm has been evaluated against various alternative state-of-the-art variational registration algorithms. Our experiments involve mono-modal as well as multi-modal and cross-contrast co-registration tasks in a broad variety of anatomical regions. Compared to the evaluated alternative registration methods, the results indicate a very favorable accuracy, precision and robustness of the proposed SOLID method in various highly challenging registration tasks.
Collapse
Affiliation(s)
- Paris Tzitzimpasis
- Department of Radiotherapy, UMC Utrecht, Heidelberglaan 100, 3508 GA, Utrecht, The Netherlands
| | - Cornel Zachiu
- Department of Radiotherapy, UMC Utrecht, Heidelberglaan 100, 3508 GA, Utrecht, The Netherlands
| | - Bas W Raaymakers
- Department of Radiotherapy, UMC Utrecht, Heidelberglaan 100, 3508 GA, Utrecht, The Netherlands
| | - Mario Ries
- Imaging Division, UMC Utrecht, Heidelberglaan 100, 3508 GA, Utrecht, The Netherlands
| |
Collapse
|
19
|
Wang AQ, Yu EM, Dalca AV, Sabuncu MR. A robust and interpretable deep learning framework for multi-modal registration via keypoints. Med Image Anal 2023; 90:102962. [PMID: 37769550 PMCID: PMC10591968 DOI: 10.1016/j.media.2023.102962] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/04/2022] [Revised: 08/24/2023] [Accepted: 09/07/2023] [Indexed: 10/03/2023]
Abstract
We present KeyMorph, a deep learning-based image registration framework that relies on automatically detecting corresponding keypoints. State-of-the-art deep learning methods for registration often are not robust to large misalignments, are not interpretable, and do not incorporate the symmetries of the problem. In addition, most models produce only a single prediction at test-time. Our core insight which addresses these shortcomings is that corresponding keypoints between images can be used to obtain the optimal transformation via a differentiable closed-form expression. We use this observation to drive the end-to-end learning of keypoints tailored for the registration task, and without knowledge of ground-truth keypoints. This framework not only leads to substantially more robust registration but also yields better interpretability, since the keypoints reveal which parts of the image are driving the final alignment. Moreover, KeyMorph can be designed to be equivariant under image translations and/or symmetric with respect to the input image ordering. Finally, we show how multiple deformation fields can be computed efficiently and in closed-form at test time corresponding to different transformation variants. We demonstrate the proposed framework in solving 3D affine and spline-based registration of multi-modal brain MRI scans. In particular, we show registration accuracy that surpasses current state-of-the-art methods, especially in the context of large displacements. Our code is available at https://github.com/alanqrwang/keymorph.
Collapse
Affiliation(s)
- Alan Q Wang
- School of Electrical and Computer Engineering, Cornell University and Cornell Tech, New York, NY 10044, USA; Department of Radiology, Weill Cornell Medical School, New York, NY 10065, USA.
| | - Evan M Yu
- Iterative Scopes, Cambridge, MA 02139, USA
| | - Adrian V Dalca
- Computer Science and Artificial Intelligence Lab at the Massachusetts Institute of Technology, Cambridge, MA 02139, USA; A.A. Martinos Center for Biomedical Imaging at the Massachusetts General Hospital, Charlestown, MA 02129, USA
| | - Mert R Sabuncu
- School of Electrical and Computer Engineering, Cornell University and Cornell Tech, New York, NY 10044, USA; Department of Radiology, Weill Cornell Medical School, New York, NY 10065, USA
| |
Collapse
|
20
|
Abbasi S, Mehdizadeh A, Boveiri HR, Mosleh Shirazi MA, Javidan R, Khayami R, Tavakoli M. Unsupervised deep learning registration model for multimodal brain images. J Appl Clin Med Phys 2023; 24:e14177. [PMID: 37823748 PMCID: PMC10647957 DOI: 10.1002/acm2.14177] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/02/2023] [Revised: 05/29/2023] [Accepted: 09/14/2023] [Indexed: 10/13/2023] Open
Abstract
Multimodal image registration is a key for many clinical image-guided interventions. However, it is a challenging task because of complicated and unknown relationships between different modalities. Currently, deep supervised learning is the state-of-theart method at which the registration is conducted in end-to-end manner and one-shot. Therefore, a huge ground-truth data is required to improve the results of deep neural networks for registration. Moreover, supervised methods may yield models that bias towards annotated structures. Here, to deal with above challenges, an alternative approach is using unsupervised learning models. In this study, we have designed a novel deep unsupervised Convolutional Neural Network (CNN)-based model based on computer tomography/magnetic resonance (CT/MR) co-registration of brain images in an affine manner. For this purpose, we created a dataset consisting of 1100 pairs of CT/MR slices from the brain of 110 neuropsychic patients with/without tumor. At the next step, 12 landmarks were selected by a well-experienced radiologist and annotated on each slice resulting in the computation of series of metrics evaluation, target registration error (TRE), Dice similarity, Hausdorff, and Jaccard coefficients. The proposed method could register the multimodal images with TRE 9.89, Dice similarity 0.79, Hausdorff 7.15, and Jaccard 0.75 that are appreciable for clinical applications. Moreover, the approach registered the images in an acceptable time 203 ms and can be appreciable for clinical usage due to the short registration time and high accuracy. Here, the results illustrated that our proposed method achieved competitive performance against other related approaches from both reasonable computation time and the metrics evaluation.
Collapse
Affiliation(s)
- Samaneh Abbasi
- Department of Medical Physics and EngineeringSchool of MedicineShiraz University of Medical SciencesShirazIran
| | - Alireza Mehdizadeh
- Research Center for Neuromodulation and PainShiraz University of Medical SciencesShirazIran
| | - Hamid Reza Boveiri
- Department of Computer Engineering and ITShiraz University of TechnologyShirazIran
| | - Mohammad Amin Mosleh Shirazi
- Ionizing and Non‐Ionizing Radiation Protection Research Center, School of Paramedical SciencesShiraz University of Medical SciencesShirazIran
| | - Reza Javidan
- Department of Computer Engineering and ITShiraz University of TechnologyShirazIran
| | - Raouf Khayami
- Department of Computer Engineering and ITShiraz University of TechnologyShirazIran
| | - Meysam Tavakoli
- Department of Radiation Oncologyand Winship Cancer InstituteEmory UniversityAtlantaGeorgiaUSA
| |
Collapse
|
21
|
Joshi A, Hong Y. R2Net: Efficient and flexible diffeomorphic image registration using Lipschitz continuous residual networks. Med Image Anal 2023; 89:102917. [PMID: 37598607 DOI: 10.1016/j.media.2023.102917] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/13/2022] [Revised: 06/26/2023] [Accepted: 07/25/2023] [Indexed: 08/22/2023]
Abstract
Classical diffeomorphic image registration methods, while being accurate, face the challenges of high computational costs. Deep learning based approaches provide a fast alternative to address these issues; however, most existing deep solutions either lose the good property of diffeomorphism or have limited flexibility to capture large deformations, under the assumption that deformations are driven by stationary velocity fields (SVFs). Also, the adopted squaring and scaling technique for integrating SVFs is time- and memory-consuming, hindering deep methods from handling large image volumes. In this paper, we present an unsupervised diffeomorphic image registration framework, which uses deep residual networks (ResNets) as numerical approximations of the underlying continuous diffeomorphic setting governed by ordinary differential equations, which is parameterized by either SVFs or time-varying (non-stationary) velocity fields. This flexible parameterization in our Residual Registration Network (R2Net) not only provides the model's ability to capture large deformation but also reduces the time and memory cost when integrating velocity fields for deformation generation. Also, we introduce a Lipschitz continuity constraint into the ResNet block to help achieve diffeomorphic deformations. To enhance the ability of our model for handling images with large volume sizes, we employ a hierarchical extension with a multi-phase learning strategy to solve the image registration task in a coarse-to-fine fashion. We demonstrate our models on four 3D image registration tasks with a wide range of anatomies, including brain MRIs, cine cardiac MRIs, and lung CT scans. Compared to classical methods SyN and diffeomorphic VoxelMorph, our models achieve comparable or better registration accuracy with much smoother deformations. Our source code is available online at https://github.com/ankitajoshi15/R2Net.
Collapse
Affiliation(s)
- Ankita Joshi
- School of Computing, University of Georgia, Athens, 30602, USA
| | - Yi Hong
- Department of Computer Science and Engineering, Shanghai Jiao Tong University, Shanghai, 200240, China.
| |
Collapse
|
22
|
Gan Z, Sun W, Liao K, Yang X. Probabilistic Modeling for Image Registration Using Radial Basis Functions: Application to Cardiac Motion Estimation. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2023; 34:7324-7338. [PMID: 35073271 DOI: 10.1109/tnnls.2022.3141119] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Cardiovascular diseases (CVDs) are the leading cause of death, affecting the cardiac dynamics over the cardiac cycle. Estimation of cardiac motion plays an essential role in many medical clinical tasks. This article proposes a probabilistic framework for image registration using compact support radial basis functions (CSRBFs) to estimate cardiac motion. A variational inference-based generative model with convolutional neural networks (CNNs) is proposed to learn the probabilistic coefficients of CSRBFs used in image deformation. We designed two networks to estimate the deformation coefficients of CSRBFs: the first one solves the spatial transformation using given control points, and the second one models the transformation using drifting control points. The given-point-based network estimates the probabilistic coefficients of control points. In contrast, the drifting-point-based model predicts the probabilistic coefficients and spatial distribution of control points simultaneously. To regularize these coefficients, we derive the bending energy (BE) in the variational bound by defining the covariance of coefficients. The proposed framework has been evaluated on the cardiac motion estimation and the calculation of the myocardial strain. In the experiments, 1409 slice pairs of end-diastolic (ED) and end-systolic (ES) phase in 4-D cardiac magnetic resonance (MR) images selected from three public datasets are employed to evaluate our networks. The experimental results show that our framework outperforms the state-of-the-art registration methods concerning the deformation smoothness and registration accuracy.
Collapse
|
23
|
Sheikhjafari A, Krishnaswamy D, Noga M, Ray N, Punithakumar K. Deep Learning Based Parameterization of Diffeomorphic Image Registration for Cardiac Image Segmentation. IEEE Trans Nanobioscience 2023; 22:800-807. [PMID: 37220045 DOI: 10.1109/tnb.2023.3276867] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
Cardiac segmentation from magnetic resonance imaging (MRI) is one of the essential tasks in analyzing the anatomy and function of the heart for the assessment and diagnosis of cardiac diseases. However, cardiac MRI generates hundreds of images per scan, and manual annotation of them is challenging and time-consuming, and therefore processing these images automatically is of interest. This study proposes a novel end-to-end supervised cardiac MRI segmentation framework based on a diffeomorphic deformable registration that can segment cardiac chambers from 2D and 3D images or volumes. To represent actual cardiac deformation, the method parameterizes the transformation using radial and rotational components computed via deep learning, with a set of paired images and segmentation masks used for training. The formulation guarantees transformations that are invertible and prevents mesh folding, which is essential for preserving the topology of the segmentation results. A physically plausible transformation is achieved by employing diffeomorphism in computing the transformations and activation functions that constrain the range of the radial and rotational components. The method was evaluated over three different data sets and showed significant improvements compared to exacting learning and non-learning based methods in terms of the Dice score and Hausdorff distance metrics.
Collapse
|
24
|
Ye M, Yang D, Huang Q, Kanski M, Axel L, Metaxas DN. SequenceMorph: A Unified Unsupervised Learning Framework for Motion Tracking on Cardiac Image Sequences. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2023; 45:10409-10426. [PMID: 37022840 DOI: 10.1109/tpami.2023.3243040] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Modern medical imaging techniques, such as ultrasound (US) and cardiac magnetic resonance (MR) imaging, have enabled the evaluation of myocardial deformation directly from an image sequence. While many traditional cardiac motion tracking methods have been developed for the automated estimation of the myocardial wall deformation, they are not widely used in clinical diagnosis, due to their lack of accuracy and efficiency. In this paper, we propose a novel deep learning-based fully unsupervised method, SequenceMorph, for in vivo motion tracking in cardiac image sequences. In our method, we introduce the concept of motion decomposition and recomposition. We first estimate the inter-frame (INF) motion field between any two consecutive frames, by a bi-directional generative diffeomorphic registration neural network. Using this result, we then estimate the Lagrangian motion field between the reference frame and any other frame, through a differentiable composition layer. Our framework can be extended to incorporate another registration network, to further reduce the accumulated errors introduced in the INF motion tracking step, and to refine the Lagrangian motion estimation. By utilizing temporal information to perform reasonable estimations of spatio-temporal motion fields, this novel method provides a useful solution for image sequence motion tracking. Our method has been applied to US (echocardiographic) and cardiac MR (untagged and tagged cine) image sequences; the results show that SequenceMorph is significantly superior to conventional motion tracking methods, in terms of the cardiac motion tracking accuracy and inference efficiency.
Collapse
|
25
|
Yang G, Xu M, Chen W, Qiao X, Shi H, Hu Y. A brain CT-based approach for predicting and analyzing stroke-associated pneumonia from intracerebral hemorrhage. Front Neurol 2023; 14:1139048. [PMID: 37332986 PMCID: PMC10272424 DOI: 10.3389/fneur.2023.1139048] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/06/2023] [Accepted: 05/08/2023] [Indexed: 06/20/2023] Open
Abstract
Introduction Stroke-associated pneumonia (SAP) is a common complication of stroke that can increase the mortality rate of patients and the burden on their families. In contrast to prior clinical scoring models that rely on baseline data, we propose constructing models based on brain CT scans due to their accessibility and clinical universality. Methods Our study aims to explore the mechanism behind the distribution and lesion areas of intracerebral hemorrhage (ICH) in relation to pneumonia, we utilized an MRI atlas that could present brain structures and a registration method in our program to extract features that may represent this relationship. We developed three machine learning models to predict the occurrence of SAP using these features. Ten-fold cross-validation was applied to evaluate the performance of models. Additionally, we constructed a probability map through statistical analysis that could display which brain regions are more frequently impacted by hematoma in patients with SAP based on four types of pneumonia. Results Our study included a cohort of 244 patients, and we extracted 35 features that captured the invasion of ICH to different brain regions for model development. We evaluated the performance of three machine learning models, namely, logistic regression, support vector machine, and random forest, in predicting SAP, and the AUCs for these models ranged from 0.77 to 0.82. The probability map revealed that the distribution of ICH varied between the left and right brain hemispheres in patients with moderate and severe SAP, and we identified several brain structures, including the left-choroid-plexus, right-choroid-plexus, right-hippocampus, and left-hippocampus, that were more closely related to SAP based on feature selection. Additionally, we observed that some statistical indicators of ICH volume, such as mean and maximum values, were proportional to the severity of SAP. Discussion Our findings suggest that our method is effective in classifying the development of pneumonia based on brain CT scans. Furthermore, we identified distinct characteristics, such as volume and distribution, of ICH in four different types of SAP.
Collapse
Affiliation(s)
- Guangtong Yang
- School of Control Science and Engineering, Shandong University, Jinan, China
| | - Min Xu
- Neurointensive Care Unit, Shengli Oilfield Central Hospital, Dongying, China
| | - Wei Chen
- Department of Radiology, Shandong First Medical University and Shandong Academy of Medical Sciences, Jinan, China
| | - Xu Qiao
- School of Control Science and Engineering, Shandong University, Jinan, China
| | - Hongfeng Shi
- Neurointensive Care Unit, Shengli Oilfield Central Hospital, Dongying, China
| | - Yongmei Hu
- School of Control Science and Engineering, Shandong University, Jinan, China
| |
Collapse
|
26
|
Lu J, Jin R, Wang M, Song E, Ma G. A bidirectional registration neural network for cardiac motion tracking using cine MRI images. Comput Biol Med 2023; 160:107001. [PMID: 37187138 DOI: 10.1016/j.compbiomed.2023.107001] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2022] [Revised: 03/15/2023] [Accepted: 05/02/2023] [Indexed: 05/17/2023]
Abstract
Using cine magnetic resonance imaging (cine MRI) images to track cardiac motion helps users to analyze the myocardial strain, and is of great importance in clinical applications. At present, most of the automatic deep learning-based motion tracking methods compare two images without considering temporal information between MRI frames, which easily leads to the lack of consistency of the generated motion fields. Even though a small number of works take into account the temporal factor, they are usually computationally intensive or have limitations on image length. To solve this problem, we propose a bidirectional convolution neural network for motion tracking of cardiac cine MRI images. This network leverages convolutional blocks to extract spatial features from three-dimensional (3D) image registration pairs, and models the temporal relations through a bidirectional recurrent neural network to obtain the Lagrange motion field between the reference image and other images. Compared with previous pairwise registration methods, the proposed method can automatically learn spatiotemporal information from multiple images with fewer parameters. We evaluated our model on three public cardiac cine MRI datasets. The experimental results demonstrated that the proposed method can significantly improve the motion tracking accuracy. The average Dice coefficient between estimated segmentation and manual segmentation has reached almost 0.85 on the widely used Automatic Cardiac Diagnostic Challenge (ACDC) dataset.
Collapse
Affiliation(s)
- Jiayi Lu
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, Hubei, 430074, China
| | - Renchao Jin
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, Hubei, 430074, China.
| | - Manyang Wang
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, Hubei, 430074, China
| | - Enmin Song
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, Hubei, 430074, China
| | - Guangzhi Ma
- School of Computer Science and Technology, Huazhong University of Science and Technology, Wuhan, Hubei, 430074, China
| |
Collapse
|
27
|
Iglesias JE. A ready-to-use machine learning tool for symmetric multi-modality registration of brain MRI. Sci Rep 2023; 13:6657. [PMID: 37095168 PMCID: PMC10126156 DOI: 10.1038/s41598-023-33781-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/05/2023] [Accepted: 04/19/2023] [Indexed: 04/26/2023] Open
Abstract
Volumetric registration of brain MRI is routinely used in human neuroimaging, e.g., to align different MRI modalities, to measure change in longitudinal analysis, to map an individual to a template, or in registration-based segmentation. Classical registration techniques based on numerical optimization have been very successful in this domain, and are implemented in widespread software suites like ANTs, Elastix, NiftyReg, or DARTEL. Over the last 7-8 years, learning-based techniques have emerged, which have a number of advantages like high computational efficiency, potential for higher accuracy, easy integration of supervision, and the ability to be part of a meta-architectures. However, their adoption in neuroimaging pipelines has so far been almost inexistent. Reasons include: lack of robustness to changes in MRI modality and resolution; lack of robust affine registration modules; lack of (guaranteed) symmetry; and, at a more practical level, the requirement of deep learning expertise that may be lacking at neuroimaging research sites. Here, we present EasyReg, an open-source, learning-based registration tool that can be easily used from the command line without any deep learning expertise or specific hardware. EasyReg combines the features of classical registration tools, the capabilities of modern deep learning methods, and the robustness to changes in MRI modality and resolution provided by our recent work in domain randomization. As a result, EasyReg is: fast; symmetric; diffeomorphic (and thus invertible); agnostic to MRI modality and resolution; compatible with affine and nonlinear registration; and does not require any preprocessing or parameter tuning. We present results on challenging registration tasks, showing that EasyReg is as accurate as classical methods when registering 1 mm isotropic scans within MRI modality, but much more accurate across modalities and resolutions. EasyReg is publicly available as part of FreeSurfer; see https://surfer.nmr.mgh.harvard.edu/fswiki/EasyReg .
Collapse
Affiliation(s)
- Juan Eugenio Iglesias
- Athinoula A. Martinos Center for Biomedical Imaging, Massachusetts General Hospital and Harvard Medical School, Boston, 02129, USA.
- Department of Medical Physics and Biomedical Engineering, University College London, London, WC1V 6LJ, UK.
- Computer Science and Artificial Intelligence Laboratory, Massachusetts Institute of Technology, Boston, 02139, USA.
| |
Collapse
|
28
|
Pastor-Serrano O, Habraken S, Hoogeman M, Lathouwers D, Schaart D, Nomura Y, Xing L, Perkó Z. A probabilistic deep learning model of inter-fraction anatomical variations in radiotherapy. Phys Med Biol 2023; 68:085018. [PMID: 36958058 PMCID: PMC10481950 DOI: 10.1088/1361-6560/acc71d] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/20/2022] [Revised: 02/20/2023] [Accepted: 03/23/2023] [Indexed: 03/25/2023]
Abstract
Objective. In radiotherapy, the internal movement of organs between treatment sessions causes errors in the final radiation dose delivery. To assess the need for adaptation, motion models can be used to simulate dominant motion patterns and assess anatomical robustness before delivery. Traditionally, such models are based on principal component analysis (PCA) and are either patient-specific (requiring several scans per patient) or population-based, applying the same set of deformations to all patients. We present a hybrid approach which, based on population data, allows to predict patient-specific inter-fraction variations for an individual patient.Approach. We propose a deep learning probabilistic framework that generates deformation vector fields warping a patient's planning computed tomography (CT) into possible patient-specific anatomies. This daily anatomy model (DAM) uses few random variables capturing groups of correlated movements. Given a new planning CT, DAM estimates the joint distribution over the variables, with each sample from the distribution corresponding to a different deformation. We train our model using dataset of 312 CT pairs with prostate, bladder, and rectum delineations from 38 prostate cancer patients. For 2 additional patients (22 CTs), we compute the contour overlap between real and generated images, and compare the sampled and 'ground truth' distributions of volume and center of mass changes.Results. With a DICE score of 0.86 ± 0.05 and a distance between prostate contours of 1.09 ± 0.93 mm, DAM matches and improves upon previously published PCA-based models, using as few as 8 latent variables. The overlap between distributions further indicates that DAM's sampled movements match the range and frequency of clinically observed daily changes on repeat CTs.Significance. Conditioned only on planning CT values and organ contours of a new patient without any pre-processing, DAM can accurately deformations seen during following treatment sessions, enabling anatomically robust treatment planning and robustness evaluation against inter-fraction anatomical changes.
Collapse
Affiliation(s)
- Oscar Pastor-Serrano
- Delft University of Technology,
Department of Radiation Science & Technology, Delft, The
Netherlands
- Stanford University, Department of
Radiation Oncology, Stanford, CA, United States of America
| | - Steven Habraken
- Erasmus University Medical Center,
Department of Radiotherapy, Rotterdam, The Netherlands
- HollandPTC, Department of Medical
Physics and Informatics, Delft, The Netherlands
| | - Mischa Hoogeman
- Erasmus University Medical Center,
Department of Radiotherapy, Rotterdam, The Netherlands
- HollandPTC, Department of Medical
Physics and Informatics, Delft, The Netherlands
| | - Danny Lathouwers
- Delft University of Technology,
Department of Radiation Science & Technology, Delft, The
Netherlands
| | - Dennis Schaart
- Delft University of Technology,
Department of Radiation Science & Technology, Delft, The
Netherlands
- HollandPTC, Department of Medical
Physics and Informatics, Delft, The Netherlands
| | - Yusuke Nomura
- Stanford University, Department of
Radiation Oncology, Stanford, CA, United States of America
| | - Lei Xing
- Stanford University, Department of
Radiation Oncology, Stanford, CA, United States of America
| | - Zoltán Perkó
- Delft University of Technology,
Department of Radiation Science & Technology, Delft, The
Netherlands
| |
Collapse
|
29
|
Aganj I, Fischl B. Intermediate Deformable Image Registration via Windowed Cross-Correlation. PROCEEDINGS. IEEE INTERNATIONAL SYMPOSIUM ON BIOMEDICAL IMAGING 2023; 2023:10.1109/isbi53787.2023.10230715. [PMID: 37691967 PMCID: PMC10485808 DOI: 10.1109/isbi53787.2023.10230715] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/12/2023]
Abstract
In population and longitudinal imaging studies that employ deformable image registration, more accurate results can be achieved by initializing deformable registration with the results of affine registration where global misalignments have been considerably reduced. Such affine registration, however, is limited to linear transformations and it cannot account for large nonlinear anatomical variations, such as those between pre- and post-operative images or across different subject anatomies. In this work, we introduce a new intermediate deformable image registration (IDIR) technique that recovers large deformations via windowed cross-correlation, and provide an efficient implementation based on the fast Fourier transform. We evaluate our method on 2D X-ray and 3D magnetic resonance images, demonstrating its ability to align substantial nonlinear anatomical variations within a few iterations.
Collapse
Affiliation(s)
- Iman Aganj
- Martinos Center for Biomedical Imaging, Massachusetts General Hospital, Harvard Medical School
| | - Bruce Fischl
- Martinos Center for Biomedical Imaging, Massachusetts General Hospital, Harvard Medical School
| |
Collapse
|
30
|
Xiao X, Dong S, Yu Y, Li Y, Yang G, Qiu Z. MAE-TransRNet: An improved transformer-ConvNet architecture with masked autoencoder for cardiac MRI registration. Front Med (Lausanne) 2023; 10:1114571. [PMID: 36968818 PMCID: PMC10033952 DOI: 10.3389/fmed.2023.1114571] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/02/2022] [Accepted: 02/14/2023] [Indexed: 03/11/2023] Open
Abstract
The heart is a relatively complex non-rigid motion organ in the human body. Quantitative motion analysis of the heart takes on a critical significance to help doctors with accurate diagnosis and treatment. Moreover, cardiovascular magnetic resonance imaging (CMRI) can be used to perform a more detailed quantitative analysis evaluation for cardiac diagnosis. Deformable image registration (DIR) has become a vital task in biomedical image analysis since tissue structures have variability in medical images. Recently, the model based on masked autoencoder (MAE) has recently been shown to be effective in computer vision tasks. Vision Transformer has the context aggregation ability to restore the semantic information in the original image regions by using a low proportion of visible image patches to predict the masked image patches. A novel Transformer-ConvNet architecture is proposed in this study based on MAE for medical image registration. The core of the Transformer is designed as a masked autoencoder (MAE) and a lightweight decoder structure, and feature extraction before the downstream registration task is transformed into the self-supervised learning task. This study also rethinks the calculation method of the multi-head self-attention mechanism in the Transformer encoder. We improve the query-key-value-based dot product attention by introducing both depthwise separable convolution (DWSC) and squeeze and excitation (SE) modules into the self-attention module to reduce the amount of parameter computation to highlight image details and maintain high spatial resolution image features. In addition, concurrent spatial and channel squeeze and excitation (scSE) module is embedded into the CNN structure, which also proves to be effective for extracting robust feature representations. The proposed method, called MAE-TransRNet, has better generalization. The proposed model is evaluated on the cardiac short-axis public dataset (with images and labels) at the 2017 Automated Cardiac Diagnosis Challenge (ACDC). The relevant qualitative and quantitative results (e.g., dice performance and Hausdorff distance) suggest that the proposed model can achieve superior results over those achieved by the state-of-the-art methods, thus proving that MAE and improved self-attention are more effective and promising for medical image registration tasks. Codes and models are available at https://github.com/XinXiao101/MAE-TransRNet.
Collapse
Affiliation(s)
- Xin Xiao
- College of Information and Computer Engineering, Northeast Forestry University, Harbin, China
| | - Suyu Dong
- College of Information and Computer Engineering, Northeast Forestry University, Harbin, China
- *Correspondence: Suyu Dong
| | - Yang Yu
- Department of Cardiovascular Surgery, Beijing Anzhen Hospital, Capital Medical University, Beijing, China
| | - Yan Li
- College of Information and Computer Engineering, Northeast Forestry University, Harbin, China
- Yan Li
| | - Guangyuan Yang
- First Affiliated Hospital, Jiamusi University, Jiamusi, China
- Guangyuan Yang
| | - Zhaowen Qiu
- College of Information and Computer Engineering, Northeast Forestry University, Harbin, China
- Zhaowen Qiu
| |
Collapse
|
31
|
Fu J, Tzortzakakis A, Barroso J, Westman E, Ferreira D, Moreno R. Fast three-dimensional image generation for healthy brain aging using diffeomorphic registration. Hum Brain Mapp 2023; 44:1289-1308. [PMID: 36468536 PMCID: PMC9921328 DOI: 10.1002/hbm.26165] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/16/2022] [Revised: 11/15/2022] [Accepted: 11/16/2022] [Indexed: 12/12/2022] Open
Abstract
Predicting brain aging can help in the early detection and prognosis of neurodegenerative diseases. Longitudinal cohorts of healthy subjects scanned through magnetic resonance imaging (MRI) have been essential to understand the structural brain changes due to aging. However, these cohorts suffer from missing data due to logistic issues in the recruitment of subjects. This paper proposes a methodology for filling up missing data in longitudinal cohorts with anatomically plausible images that capture the subject-specific aging process. The proposed methodology is developed within the framework of diffeomorphic registration. First, two novel modules are introduced within Synthmorph, a fast, state-of-the-art deep learning-based diffeomorphic registration method, to simulate the aging process between the first and last available MRI scan for each subject in three-dimensional (3D). The use of image registration also makes the generated images plausible by construction. Second, we used six image similarity measurements to rearrange the generated images to the specific age range. Finally, we estimated the age of every generated image by using the assumption of linear brain decay in healthy subjects. The methodology was evaluated on 2662 T1-weighted MRI scans from 796 healthy participants from 3 different longitudinal cohorts: Alzheimer's Disease Neuroimaging Initiative, Open Access Series of Imaging Studies-3, and Group of Neuropsychological Studies of the Canary Islands (GENIC). In total, we generated 7548 images to simulate the access of a scan per subject every 6 months in these cohorts. We evaluated the quality of the synthetic images using six quantitative measurements and a qualitative assessment by an experienced neuroradiologist with state-of-the-art results. The assumption of linear brain decay was accurate in these cohorts (R2 ∈ [.924, .940]). The experimental results show that the proposed methodology can produce anatomically plausible aging predictions that can be used to enhance longitudinal datasets. Compared to deep learning-based generative methods, diffeomorphic registration is more likely to preserve the anatomy of the different structures of the brain, which makes it more appropriate for its use in clinical applications. The proposed methodology is able to efficiently simulate anatomically plausible 3D MRI scans of brain aging of healthy subjects from two images scanned at two different time points.
Collapse
Affiliation(s)
- Jingru Fu
- Division of Biomedical ImagingDepartment of Biomedical Engineering and Health Systems, KTH Royal Institute of TechnologyStockholmSweden
| | - Antonios Tzortzakakis
- Division of RadiologyDepartment for Clinical Science, Intervention and Technology (CLINTEC), Karolinska InstitutetStockholmSweden
- Medical Radiation Physics and Nuclear MedicineFunctional Unit of Nuclear Medicine, Karolinska University HospitalHuddingeStockholmSweden
| | - José Barroso
- Department of PsychologyFaculty of Health Sciences, University Fernando Pessoa CanariasLas PalmasSpain
| | - Eric Westman
- Division of Clinical GeriatricsCentre for Alzheimer Research, Department of Neurobiology, Care Sciences, and Society (NVS), Karolinska InstitutetStockholmSweden
- Department of NeuroimagingCentre for Neuroimaging Sciences, Institute of Psychiatry, Psychology and Neuroscience, King's College LondonLondonUnited Kingdom
| | - Daniel Ferreira
- Division of Clinical GeriatricsCentre for Alzheimer Research, Department of Neurobiology, Care Sciences, and Society (NVS), Karolinska InstitutetStockholmSweden
| | - Rodrigo Moreno
- Division of Biomedical ImagingDepartment of Biomedical Engineering and Health Systems, KTH Royal Institute of TechnologyStockholmSweden
| | | |
Collapse
|
32
|
Ruthven M, Miquel ME, King AP. A segmentation-informed deep learning framework to register dynamic two-dimensional magnetic resonance images of the vocal tract during speech. Biomed Signal Process Control 2023; 80:104290. [PMID: 36743699 PMCID: PMC9746295 DOI: 10.1016/j.bspc.2022.104290] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/04/2022] [Revised: 09/29/2022] [Accepted: 10/08/2022] [Indexed: 11/06/2022]
Abstract
Objective Dynamic magnetic resonance (MR) imaging enables visualisation of articulators during speech. There is growing interest in quantifying articulator motion in two-dimensional MR images of the vocal tract, to better understand speech production and potentially inform patient management decisions. Image registration is an established way to achieve this quantification. Recently, segmentation-informed deformable registration frameworks have been developed and have achieved state-of-the-art accuracy. This work aims to adapt such a framework and optimise it for estimating displacement fields between dynamic two-dimensional MR images of the vocal tract during speech. Methods A deep-learning-based registration framework was developed and compared with current state-of-the-art registration methods and frameworks (two traditional methods and three deep-learning-based frameworks, two of which are segmentation informed). The accuracy of the methods and frameworks was evaluated using the Dice coefficient (DSC), average surface distance (ASD) and a metric based on velopharyngeal closure. The metric evaluated if the fields captured a clinically relevant and quantifiable aspect of articulator motion. Results The segmentation-informed frameworks achieved higher DSCs and lower ASDs and captured more velopharyngeal closures than the traditional methods and the framework that was not segmentation informed. All segmentation-informed frameworks achieved similar DSCs and ASDs. However, the proposed framework captured the most velopharyngeal closures. Conclusions A framework was successfully developed and found to more accurately estimate articulator motion than five current state-of-the-art methods and frameworks. Significance The first deep-learning-based framework specifically for registering dynamic two-dimensional MR images of the vocal tract during speech has been developed and evaluated.
Collapse
Affiliation(s)
- Matthieu Ruthven
- Clinical Physics, Barts Health NHS Trust, West Smithfield, London EC1A 7BE, United Kingdom,School of Biomedical Engineering & Imaging Sciences, King’s College London, King’s Health Partners, St Thomas’ Hospital, London SE1 7EH, United Kingdom,Corresponding author at: Clinical Physics, Barts Health NHS Trust, West Smithfield, London EC1A 7BE, United Kingdom.
| | - Marc E. Miquel
- Clinical Physics, Barts Health NHS Trust, West Smithfield, London EC1A 7BE, United Kingdom,Digital Environment Research Institute (DERI), Empire House, 67-75 New Road, Queen Mary University of London, London E1 1HH, United Kingdom,Advanced Cardiovascular Imaging, Barts NIHR BRC, Queen Mary University of London, London EC1M 6BQ, United Kingdom
| | - Andrew P. King
- School of Biomedical Engineering & Imaging Sciences, King’s College London, King’s Health Partners, St Thomas’ Hospital, London SE1 7EH, United Kingdom
| |
Collapse
|
33
|
Generative myocardial motion tracking via latent space exploration with biomechanics-informed prior. Med Image Anal 2023; 83:102682. [PMID: 36403311 DOI: 10.1016/j.media.2022.102682] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/29/2022] [Revised: 08/15/2022] [Accepted: 11/02/2022] [Indexed: 11/09/2022]
Abstract
Myocardial motion and deformation are rich descriptors that characterize cardiac function. Image registration, as the most commonly used technique for myocardial motion tracking, is an ill-posed inverse problem which often requires prior assumptions on the solution space. In contrast to most existing approaches which impose explicit generic regularization such as smoothness, in this work we propose a novel method that can implicitly learn an application-specific biomechanics-informed prior and embed it into a neural network-parameterized transformation model. Particularly, the proposed method leverages a variational autoencoder-based generative model to learn a manifold for biomechanically plausible deformations. The motion tracking then can be performed via traversing the learnt manifold to search for the optimal transformations while considering the sequence information. The proposed method is validated on three public cardiac cine MRI datasets with comprehensive evaluations. The results demonstrate that the proposed method can outperform other approaches, yielding higher motion tracking accuracy with reasonable volume preservation and better generalizability to varying data distributions. It also enables better estimates of myocardial strains, which indicates the potential of the method in characterizing spatiotemporal signatures for understanding cardiovascular diseases.
Collapse
|
34
|
Zakeri A, Hokmabadi A, Bi N, Wijesinghe I, Nix MG, Petersen SE, Frangi AF, Taylor ZA, Gooya A. DragNet: Learning-based deformable registration for realistic cardiac MR sequence generation from a single frame. Med Image Anal 2023; 83:102678. [PMID: 36403308 DOI: 10.1016/j.media.2022.102678] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/08/2022] [Revised: 08/24/2022] [Accepted: 10/27/2022] [Indexed: 11/06/2022]
Abstract
Deformable image registration (DIR) can be used to track cardiac motion. Conventional DIR algorithms aim to establish a dense and non-linear correspondence between independent pairs of images. They are, nevertheless, computationally intensive and do not consider temporal dependencies to regulate the estimated motion in a cardiac cycle. In this paper, leveraging deep learning methods, we formulate a novel hierarchical probabilistic model, termed DragNet, for fast and reliable spatio-temporal registration in cine cardiac magnetic resonance (CMR) images and for generating synthetic heart motion sequences. DragNet is a variational inference framework, which takes an image from the sequence in combination with the hidden states of a recurrent neural network (RNN) as inputs to an inference network per time step. As part of this framework, we condition the prior probability of the latent variables on the hidden states of the RNN utilised to capture temporal dependencies. We further condition the posterior of the motion field on a latent variable from hierarchy and features from the moving image. Subsequently, the RNN updates the hidden state variables based on the feature maps of the fixed image and the latent variables. Different from traditional methods, DragNet performs registration on unseen sequences in a forward pass, which significantly expedites the registration process. Besides, DragNet enables generating a large number of realistic synthetic image sequences given only one frame, where the corresponding deformations are also retrieved. The probabilistic framework allows for computing spatio-temporal uncertainties in the estimated motion fields. Our results show that DragNet performance is comparable with state-of-the-art methods in terms of registration accuracy, with the advantage of offering analytical pixel-wise motion uncertainty estimation across a cardiac cycle and being a motion generator. We will make our code publicly available.
Collapse
Affiliation(s)
- Arezoo Zakeri
- Centre for Computational Imaging and Simulation Technologies in Biomedicine, School of Computing, University of Leeds, UK.
| | - Alireza Hokmabadi
- Centre for Computational Imaging and Simulation Technologies in Biomedicine, School of Computing, University of Leeds, UK
| | - Ning Bi
- Centre for Computational Imaging and Simulation Technologies in Biomedicine, School of Computing, University of Leeds, UK
| | - Isuru Wijesinghe
- Centre for Computational Imaging and Simulation Technologies in Biomedicine, School of Mechanical Engineering, University of Leeds, UK
| | - Michael G Nix
- Leeds Cancer Centre, Leeds Teaching Hospitals NHS Trust, UK
| | - Steffen E Petersen
- William Harvey Research Institute, NIHR Barts Biomedical Research Centre, Queen Mary University of London, UK; Barts Heart Centre, St Bartholomew's Hospital, Barts Health NHS Trust, London, UK; Health Data Research UK, London, UK; Alan Turing Institute, London, UK
| | - Alejandro F Frangi
- Centre for Computational Imaging and Simulation Technologies in Biomedicine, School of Computing, University of Leeds, UK
| | - Zeike A Taylor
- Centre for Computational Imaging and Simulation Technologies in Biomedicine, School of Mechanical Engineering, University of Leeds, UK
| | - Ali Gooya
- Alan Turing Institute, London, UK; School of Computing Science, University of Glasgow, Glasgow, UK.
| |
Collapse
|
35
|
Jiang X, Isogai T, Chi J, Danuser G. Fine-grained, nonlinear registration of live cell movies reveals spatiotemporal organization of diffuse molecular processes. PLoS Comput Biol 2022; 18:e1009667. [PMID: 36584219 PMCID: PMC9870159 DOI: 10.1371/journal.pcbi.1009667] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/23/2021] [Revised: 01/23/2023] [Accepted: 11/28/2022] [Indexed: 01/01/2023] Open
Abstract
We present an application of nonlinear image registration to align in microscopy time lapse sequences for every frame the cell outline and interior with the outline and interior of the same cell in a reference frame. The registration relies on a subcellular fiducial marker, a cell motion mask, and a topological regularization that enforces diffeomorphism on the registration without significant loss of granularity. This allows spatiotemporal analysis of extremely noisy and diffuse molecular processes across the entire cell. We validate the registration method for different fiducial markers by measuring the intensity differences between predicted and original time lapse sequences of Actin cytoskeleton images and by uncovering zones of spatially organized GEF- and GTPase signaling dynamics visualized by FRET-based activity biosensors in MDA-MB-231 cells. We then demonstrate applications of the registration method in conjunction with stochastic time-series analysis. We describe distinct zones of locally coherent dynamics of the cytoplasmic protein Profilin in U2OS cells. Further analysis of the Profilin dynamics revealed strong relationships with Actin cytoskeleton reorganization during cell symmetry-breaking and polarization. This study thus provides a framework for extracting information to explore functional interactions between cell morphodynamics, protein distributions, and signaling in cells undergoing continuous shape changes. Matlab code implementing the proposed registration method is available at https://github.com/DanuserLab/Mask-Regularized-Diffeomorphic-Cell-Registration.
Collapse
Affiliation(s)
- Xuexia Jiang
- Lyda Hill Department of Bioinformatics, UT Southwestern Medical Center, Dallas, Texas, United States of America
| | - Tadamoto Isogai
- Lyda Hill Department of Bioinformatics, UT Southwestern Medical Center, Dallas, Texas, United States of America
| | - Joseph Chi
- Lyda Hill Department of Bioinformatics, UT Southwestern Medical Center, Dallas, Texas, United States of America
| | - Gaudenz Danuser
- Lyda Hill Department of Bioinformatics, UT Southwestern Medical Center, Dallas, Texas, United States of America
- * E-mail:
| |
Collapse
|
36
|
Nakao M, Nakamura M, Matsuda T. Image-to-Graph Convolutional Network for 2D/3D Deformable Model Registration of Low-Contrast Organs. IEEE TRANSACTIONS ON MEDICAL IMAGING 2022; 41:3747-3761. [PMID: 35901001 DOI: 10.1109/tmi.2022.3194517] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Organ shape reconstruction based on a single-projection image during treatment has wide clinical scope, e.g., in image-guided radiotherapy and surgical guidance. We propose an image-to-graph convolutional network that achieves deformable registration of a three-dimensional (3D) organ mesh for a low-contrast two-dimensional (2D) projection image. This framework enables simultaneous training of two types of transformation: from the 2D projection image to a displacement map, and from the sampled per-vertex feature to a 3D displacement that satisfies the geometrical constraint of the mesh structure. Assuming application to radiation therapy, the 2D/3D deformable registration performance is verified for multiple abdominal organs that have not been targeted to date, i.e., the liver, stomach, duodenum, and kidney, and for pancreatic cancer. The experimental results show shape prediction considering relationships among multiple organs can be used to predict respiratory motion and deformation from digitally reconstructed radiographs with clinically acceptable accuracy.
Collapse
|
37
|
Chen P, Guo Y, Wang D, Chen C. Dlung: Unsupervised Few-Shot Diffeomorphic Respiratory Motion Modeling. JOURNAL OF SHANGHAI JIAOTONG UNIVERSITY (SCIENCE) 2022; 28:1-10. [PMID: 36406811 PMCID: PMC9660014 DOI: 10.1007/s12204-022-2525-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Subscribe] [Scholar Register] [Received: 01/27/2021] [Accepted: 09/17/2021] [Indexed: 11/15/2022]
Abstract
Lung image registration plays an important role in lung analysis applications, such as respiratory motion modeling. Unsupervised learning-based image registration methods that can compute the deformation without the requirement of supervision attract much attention. However, it is noteworthy that they have two drawbacks: they do not handle the problem of limited data and do not guarantee diffeomorphic (topology-preserving) properties, especially when large deformation exists in lung scans. In this paper, we present an unsupervised few-shot learning-based diffeomorphic lung image registration, namely Dlung. We employ fine-tuning techniques to solve the problem of limited data and apply the scaling and squaring method to accomplish the diffeomorphic registration. Furthermore, atlas-based registration on spatio-temporal (4D) images is performed and thoroughly compared with baseline methods. Dlung achieves the highest accuracy with diffeomorphic properties. It constructs accurate and fast respiratory motion models with limited data. This research extends our knowledge of respiratory motion modeling.
Collapse
Affiliation(s)
- Peizhi Chen
- College of Computer and Information Engineering, Xiamen University of Technology, Xiamen, Fujian, 361024 China
- Fujian Key Laboratory of Pattern Recognition and Image Understanding, Xiamen, Fujian, 361024 China
| | - Yifan Guo
- College of Computer and Information Engineering, Xiamen University of Technology, Xiamen, Fujian, 361024 China
| | - Dahan Wang
- College of Computer and Information Engineering, Xiamen University of Technology, Xiamen, Fujian, 361024 China
- Fujian Key Laboratory of Pattern Recognition and Image Understanding, Xiamen, Fujian, 361024 China
| | - Chinling Chen
- College of Computer and Information Engineering, Xiamen University of Technology, Xiamen, Fujian, 361024 China
- School of Information Engineering, Changchun Sci-Tech University, Changchun, 130600 China
- Department of Computer Science and Information Engineering, Chaoyang University of Technology, Taichung, Taiwan, 41349 China
| |
Collapse
|
38
|
Chen J, Frey EC, He Y, Segars WP, Li Y, Du Y. TransMorph: Transformer for unsupervised medical image registration. Med Image Anal 2022; 82:102615. [PMID: 36156420 PMCID: PMC9999483 DOI: 10.1016/j.media.2022.102615] [Citation(s) in RCA: 153] [Impact Index Per Article: 51.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/31/2021] [Revised: 06/29/2022] [Accepted: 09/02/2022] [Indexed: 01/04/2023]
Abstract
In the last decade, convolutional neural networks (ConvNets) have been a major focus of research in medical image analysis. However, the performances of ConvNets may be limited by a lack of explicit consideration of the long-range spatial relationships in an image. Recently, Vision Transformer architectures have been proposed to address the shortcomings of ConvNets and have produced state-of-the-art performances in many medical imaging applications. Transformers may be a strong candidate for image registration because their substantially larger receptive field enables a more precise comprehension of the spatial correspondence between moving and fixed images. Here, we present TransMorph, a hybrid Transformer-ConvNet model for volumetric medical image registration. This paper also presents diffeomorphic and Bayesian variants of TransMorph: the diffeomorphic variants ensure the topology-preserving deformations, and the Bayesian variant produces a well-calibrated registration uncertainty estimate. We extensively validated the proposed models using 3D medical images from three applications: inter-patient and atlas-to-patient brain MRI registration and phantom-to-CT registration. The proposed models are evaluated in comparison to a variety of existing registration methods and Transformer architectures. Qualitative and quantitative results demonstrate that the proposed Transformer-based model leads to a substantial performance improvement over the baseline methods, confirming the effectiveness of Transformers for medical image registration.
Collapse
Affiliation(s)
- Junyu Chen
- Russell H. Morgan Department of Radiology and Radiological Science, Johns Hopkins Medical Institutes, Baltimore, MD, USA; Department of Electrical and Computer Engineering, Johns Hopkins University, Baltimore, MD, USA.
| | - Eric C Frey
- Russell H. Morgan Department of Radiology and Radiological Science, Johns Hopkins Medical Institutes, Baltimore, MD, USA; Department of Electrical and Computer Engineering, Johns Hopkins University, Baltimore, MD, USA.
| | - Yufan He
- NVIDIA Corporation, Bethesda, MD, USA.
| | - William P Segars
- Carl E. Ravin Advanced Imaging Laboratories, Department of Radiology, Duke University Medical Center, Durham, NC, USA.
| | - Ye Li
- Center for Advanced Medical Computing and Analysis, Massachusetts General Hospital and Harvard Medical School, Boston, MA, USA.
| | - Yong Du
- Russell H. Morgan Department of Radiology and Radiological Science, Johns Hopkins Medical Institutes, Baltimore, MD, USA.
| |
Collapse
|
39
|
Liu X, Yuan D, Xue K, Li JB, Zhao H, Liu H, Wang T. Diffeomorphic matching with multiscale kernels based on sparse parameterization for cross-view target detection. APPL INTELL 2022. [DOI: 10.1007/s10489-022-03668-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/27/2022]
|
40
|
Meng Q, Qin C, Bai W, Liu T, de Marvao A, O’Regan DP, Rueckert D. MulViMotion: Shape-Aware 3D Myocardial Motion Tracking From Multi-View Cardiac MRI. IEEE TRANSACTIONS ON MEDICAL IMAGING 2022; 41:1961-1974. [PMID: 35201985 PMCID: PMC7613225 DOI: 10.1109/tmi.2022.3154599] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/22/2021] [Revised: 02/07/2022] [Accepted: 02/11/2022] [Indexed: 06/14/2023]
Abstract
Recovering the 3D motion of the heart from cine cardiac magnetic resonance (CMR) imaging enables the assessment of regional myocardial function and is important for understanding and analyzing cardiovascular disease. However, 3D cardiac motion estimation is challenging because the acquired cine CMR images are usually 2D slices which limit the accurate estimation of through-plane motion. To address this problem, we propose a novel multi-view motion estimation network (MulViMotion), which integrates 2D cine CMR images acquired in short-axis and long-axis planes to learn a consistent 3D motion field of the heart. In the proposed method, a hybrid 2D/3D network is built to generate dense 3D motion fields by learning fused representations from multi-view images. To ensure that the motion estimation is consistent in 3D, a shape regularization module is introduced during training, where shape information from multi-view images is exploited to provide weak supervision to 3D motion estimation. We extensively evaluate the proposed method on 2D cine CMR images from 580 subjects of the UK Biobank study for 3D motion tracking of the left ventricular myocardium. Experimental results show that the proposed method quantitatively and qualitatively outperforms competing methods.
Collapse
Affiliation(s)
- Qingjie Meng
- Biomedical Image Analysis GroupDepartment of ComputingImperial College LondonLondonSW7 2AZU.K.
| | - Chen Qin
- School of EngineeringInstitute for Digital Communications, The University of EdinburghEdinburghEH9 9JLU.K.
| | - Wenjia Bai
- Biomedical Image Analysis GroupDepartment of ComputingImperial College LondonLondonSW7 2AZU.K.
- Department of Brain SciencesImperial College LondonLondonSW7 2AZU.K.
| | - Tianrui Liu
- Biomedical Image Analysis GroupDepartment of ComputingImperial College LondonLondonSW7 2AZU.K.
| | - Antonio de Marvao
- MRC London Institute of Medical SciencesImperial College LondonLondonW12 0HSU.K.
| | - Declan P O’Regan
- MRC London Institute of Medical SciencesImperial College LondonLondonW12 0HSU.K.
| | - Daniel Rueckert
- Biomedical Image Analysis GroupDepartment of ComputingImperial College LondonLondonSW7 2AZU.K.
- Faculty of Informatics and MedicineTechnical University of Munich85748MunichGermany
| |
Collapse
|
41
|
Dense correspondence of deformable volumetric images via deep spectral embedding and descriptor learning. Med Image Anal 2022; 82:102604. [DOI: 10.1016/j.media.2022.102604] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/23/2022] [Revised: 07/14/2022] [Accepted: 08/24/2022] [Indexed: 11/17/2022]
|
42
|
Hernandez M, Ramon-Julvez U, Sierra-Tome D. Partial Differential Equation-Constrained Diffeomorphic Registration from Sum of Squared Differences to Normalized Cross-Correlation, Normalized Gradient Fields, and Mutual Information: A Unifying Framework. SENSORS (BASEL, SWITZERLAND) 2022; 22:3735. [PMID: 35632143 PMCID: PMC9146848 DOI: 10.3390/s22103735] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 04/20/2022] [Revised: 05/10/2022] [Accepted: 05/11/2022] [Indexed: 06/15/2023]
Abstract
This work proposes a unifying framework for extending PDE-constrained Large Deformation Diffeomorphic Metric Mapping (PDE-LDDMM) with the sum of squared differences (SSD) to PDE-LDDMM with different image similarity metrics. We focused on the two best-performing variants of PDE-LDDMM with the spatial and band-limited parameterizations of diffeomorphisms. We derived the equations for gradient-descent and Gauss-Newton-Krylov (GNK) optimization with Normalized Cross-Correlation (NCC), its local version (lNCC), Normalized Gradient Fields (NGFs), and Mutual Information (MI). PDE-LDDMM with GNK was successfully implemented for NCC and lNCC, substantially improving the registration results of SSD. For these metrics, GNK optimization outperformed gradient-descent. However, for NGFs, GNK optimization was not able to overpass the performance of gradient-descent. For MI, GNK optimization involved the product of huge dense matrices, requesting an unaffordable memory load. The extensive evaluation reported the band-limited version of PDE-LDDMM based on the deformation state equation with NCC and lNCC image similarities among the best performing PDE-LDDMM methods. In comparison with benchmark deep learning-based methods, our proposal reached or surpassed the accuracy of the best-performing models. In NIREP16, several configurations of PDE-LDDMM outperformed ANTS-lNCC, the best benchmark method. Although NGFs and MI usually underperformed the other metrics in our evaluation, these metrics showed potentially competitive results in a multimodal deformable experiment. We believe that our proposed image similarity extension over PDE-LDDMM will promote the use of physically meaningful diffeomorphisms in a wide variety of clinical applications depending on deformable image registration.
Collapse
Affiliation(s)
- Monica Hernandez
- Aragon Institute of Engineering Research (I3A), 50018 Zaragoza, Spain
- Department of Computer Sciences, University of Zaragoza (UZ), 50018 Zaragoza, Spain; (U.R.-J.); (D.S.-T.)
| | - Ubaldo Ramon-Julvez
- Department of Computer Sciences, University of Zaragoza (UZ), 50018 Zaragoza, Spain; (U.R.-J.); (D.S.-T.)
| | - Daniel Sierra-Tome
- Department of Computer Sciences, University of Zaragoza (UZ), 50018 Zaragoza, Spain; (U.R.-J.); (D.S.-T.)
| |
Collapse
|
43
|
Li H, Fan Y. MDReg-Net: Multi-resolution diffeomorphic image registration using fully convolutional networks with deep self-supervision. Hum Brain Mapp 2022; 43:2218-2231. [PMID: 35072327 PMCID: PMC8996355 DOI: 10.1002/hbm.25782] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/14/2021] [Revised: 12/15/2021] [Accepted: 01/08/2022] [Indexed: 11/09/2022] Open
Abstract
We present a diffeomorphic image registration algorithm to learn spatial transformations between pairs of images to be registered using fully convolutional networks (FCNs) under a self-supervised learning setting. Particularly, a deep neural network is trained to estimate diffeomorphic spatial transformations between pairs of images by maximizing an image-wise similarity metric between fixed and warped moving images, similar to those adopted in conventional image registration algorithms. The network is implemented in a multi-resolution image registration framework to optimize and learn spatial transformations at different image resolutions jointly and incrementally with deep self-supervision in order to better handle large deformation between images. A spatial Gaussian smoothing kernel is integrated with the FCNs to yield sufficiently smooth deformation fields for diffeomorphic image registration. The spatial transformations learned at coarser resolutions are utilized to warp the moving image, which is subsequently used as input to the network for learning incremental transformations at finer resolutions. This procedure proceeds recursively to the full image resolution and the accumulated transformations serve as the final transformation to warp the moving image at the finest resolution. Experimental results for registering high-resolution 3D structural brain magnetic resonance (MR) images have demonstrated that image registration networks trained by our method obtain robust, diffeomorphic image registration results within seconds with improved accuracy compared with state-of-the-art image registration algorithms.
Collapse
Affiliation(s)
- Hongming Li
- Center for Biomedical Image Computing and Analytics (CBICA), Department of Radiology, Perelman School of MedicineUniversity of PennsylvaniaPhiladelphiaPennsylvaniaUSA
| | - Yong Fan
- Center for Biomedical Image Computing and Analytics (CBICA), Department of Radiology, Perelman School of MedicineUniversity of PennsylvaniaPhiladelphiaPennsylvaniaUSA
| | | |
Collapse
|
44
|
Wang H, Li Q, Yuan Y, Zhang Z, Wang K, Zhang H. Inter-subject registration-based one-shot segmentation with alternating union network for cardiac MRI images. Med Image Anal 2022; 79:102455. [PMID: 35453066 DOI: 10.1016/j.media.2022.102455] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/12/2021] [Revised: 04/04/2022] [Accepted: 04/08/2022] [Indexed: 11/30/2022]
Abstract
Medical image segmentation based on deep-learning networks makes great progress in assisting disease diagnosis. However, currently, the training of most networks still requires a large amount of data with labels. In reality, labeling a considerable number of medical images is challenging and time-consuming. In order to tackle this challenge, a new one-shot segmentation framework for cardiac MRI images based on an inter-subject registration model called Alternating Union Network (AUN) is proposed in this study. The label of the source image is warped with deformation fields discovered from AUN to segment target images directly. Initially, the volumes are pre-processed by aligning affinely and adjusting the global intensity to simplify subsequent deformation registration. AUN consists of two kinds of subnetworks trained alternately to optimize segmentation gradually. The first kind of subnetwork takes a pair of volumes as inputs and registers them using global intensity similarity. The second kind of subnetwork, which takes the predicted labels generated from the previous subnetwork and the labels refined using the information of intrinsic anatomical structures of interest as inputs, is intensity-independent and focuses attention on registering structures of interest. Specifically, the input of AUN is a pair of a labeled image with the texture in regions of interest removed and a target image. Additionally, a new similarity measurement more appropriate for registering such image pair is defined as Local Squared Error (LSE). The proposed registration-based one-shot segmentation pays attention to the problem of the lack of labeled medical images. In AUN, only one labeled volume is required and a large number of unlabeled ones can be leveraged to improve segmentation performance, which has great advantages in clinical application. In addition, the intensity-independent subnetwork and LSE proposed in this study empower the framework to segment medical images with complicated intensity distribution.
Collapse
Affiliation(s)
- Heying Wang
- School of Computer Science and Technology, Harbin Institute of Technology, Nangang District, Harbin 150000, China
| | - Qince Li
- School of Computer Science and Technology, Harbin Institute of Technology, Nangang District, Harbin 150000, China; Peng Cheng Laboratory, Nanshan District, Shenzhen 518000, China.
| | - Yongfeng Yuan
- School of Computer Science and Technology, Harbin Institute of Technology, Nangang District, Harbin 150000, China
| | - Ze Zhang
- School of Computer Science and Technology, Harbin Institute of Technology, Nangang District, Harbin 150000, China
| | - Kuanquan Wang
- School of Computer Science and Technology, Harbin Institute of Technology, Nangang District, Harbin 150000, China
| | - Henggui Zhang
- Peng Cheng Laboratory, Nanshan District, Shenzhen 518000, China; School of Physics and Astronomy, The University of Manchester, Manchester M139PL, UK; Key Laboratory of Medical Electrophysiology of Ministry of Education and Medical Electrophysiological Key Laboratory of Sichuan Province, Institute of Cardiovascular Research, Southwest Medical University, Luzhou 646000, China
| |
Collapse
|
45
|
Abbasi S, Tavakoli M, Boveiri HR, Mosleh Shirazi MA, Khayami R, Khorasani H, Javidan R, Mehdizadeh A. Medical image registration using unsupervised deep neural network: A scoping literature review. Biomed Signal Process Control 2022. [DOI: 10.1016/j.bspc.2021.103444] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/26/2023]
|
46
|
Hoffmann M, Billot B, Greve DN, Iglesias JE, Fischl B, Dalca AV. SynthMorph: Learning Contrast-Invariant Registration Without Acquired Images. IEEE TRANSACTIONS ON MEDICAL IMAGING 2022; 41:543-558. [PMID: 34587005 PMCID: PMC8891043 DOI: 10.1109/tmi.2021.3116879] [Citation(s) in RCA: 48] [Impact Index Per Article: 16.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/10/2023]
Abstract
We introduce a strategy for learning image registration without acquired imaging data, producing powerful networks agnostic to contrast introduced by magnetic resonance imaging (MRI). While classical registration methods accurately estimate the spatial correspondence between images, they solve an optimization problem for every new image pair. Learning-based techniques are fast at test time but limited to registering images with contrasts and geometric content similar to those seen during training. We propose to remove this dependency on training data by leveraging a generative strategy for diverse synthetic label maps and images that exposes networks to a wide range of variability, forcing them to learn more invariant features. This approach results in powerful networks that accurately generalize to a broad array of MRI contrasts. We present extensive experiments with a focus on 3D neuroimaging, showing that this strategy enables robust and accurate registration of arbitrary MRI contrasts even if the target contrast is not seen by the networks during training. We demonstrate registration accuracy surpassing the state of the art both within and across contrasts, using a single model. Critically, training on arbitrary shapes synthesized from noise distributions results in competitive performance, removing the dependency on acquired data of any kind. Additionally, since anatomical label maps are often available for the anatomy of interest, we show that synthesizing images from these dramatically boosts performance, while still avoiding the need for real intensity images. Our code is available at doic https://w3id.org/synthmorph.
Collapse
|
47
|
A training-free recursive multiresolution framework for diffeomorphic deformable image registration. APPL INTELL 2022. [DOI: 10.1007/s10489-021-03062-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/02/2022]
|
48
|
Kang M, Hu X, Huang W, Scott MR, Reyes M. Dual-stream Pyramid Registration Network. Med Image Anal 2022; 78:102379. [DOI: 10.1016/j.media.2022.102379] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/22/2020] [Revised: 01/25/2022] [Accepted: 01/27/2022] [Indexed: 10/19/2022]
|
49
|
Atlas-ISTN: Joint Segmentation, Registration and Atlas Construction with Image-and-Spatial Transformer Networks. Med Image Anal 2022; 78:102383. [DOI: 10.1016/j.media.2022.102383] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2021] [Revised: 11/24/2021] [Accepted: 02/01/2022] [Indexed: 11/16/2022]
|
50
|
Sanchez-Martinez S, Camara O, Piella G, Cikes M, González-Ballester MÁ, Miron M, Vellido A, Gómez E, Fraser AG, Bijnens B. Machine Learning for Clinical Decision-Making: Challenges and Opportunities in Cardiovascular Imaging. Front Cardiovasc Med 2022; 8:765693. [PMID: 35059445 PMCID: PMC8764455 DOI: 10.3389/fcvm.2021.765693] [Citation(s) in RCA: 14] [Impact Index Per Article: 4.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/27/2021] [Accepted: 12/07/2021] [Indexed: 11/30/2022] Open
Abstract
The use of machine learning (ML) approaches to target clinical problems is called to revolutionize clinical decision-making in cardiology. The success of these tools is dependent on the understanding of the intrinsic processes being used during the conventional pathway by which clinicians make decisions. In a parallelism with this pathway, ML can have an impact at four levels: for data acquisition, predominantly by extracting standardized, high-quality information with the smallest possible learning curve; for feature extraction, by discharging healthcare practitioners from performing tedious measurements on raw data; for interpretation, by digesting complex, heterogeneous data in order to augment the understanding of the patient status; and for decision support, by leveraging the previous steps to predict clinical outcomes, response to treatment or to recommend a specific intervention. This paper discusses the state-of-the-art, as well as the current clinical status and challenges associated with the two later tasks of interpretation and decision support, together with the challenges related to the learning process, the auditability/traceability, the system infrastructure and the integration within clinical processes in cardiovascular imaging.
Collapse
Affiliation(s)
| | - Oscar Camara
- Department of Information and Communication Technologies, University Pompeu Fabra, Barcelona, Spain
| | - Gemma Piella
- Department of Information and Communication Technologies, University Pompeu Fabra, Barcelona, Spain
| | - Maja Cikes
- Department of Cardiovascular Diseases, University of Zagreb School of Medicine, University Hospital Centre Zagreb, Zagreb, Croatia
| | | | - Marius Miron
- Joint Research Centre, European Commission, Seville, Spain
| | - Alfredo Vellido
- Computer Science Department, Intelligent Data Science and Artificial Intelligence (IDEAI-UPC) Research Center, Universitat Politècnica de Catalunya, Barcelona, Spain
| | - Emilia Gómez
- Department of Information and Communication Technologies, University Pompeu Fabra, Barcelona, Spain
- Joint Research Centre, European Commission, Seville, Spain
| | - Alan G. Fraser
- School of Medicine, Cardiff University, Cardiff, United Kingdom
| | - Bart Bijnens
- August Pi i Sunyer Biomedical Research Institute (IDIBAPS), Barcelona, Spain
- ICREA, Barcelona, Spain
- Department of Cardiovascular Sciences, KU Leuven, Leuven, Belgium
| |
Collapse
|