1
|
Huang J, Wu Y, Wang F, Fang Y, Nan Y, Alkan C, Abraham D, Liao C, Xu L, Gao Z, Wu W, Zhu L, Chen Z, Lally P, Bangerter N, Setsompop K, Guo Y, Rueckert D, Wang G, Yang G. Data- and Physics-Driven Deep Learning Based Reconstruction for Fast MRI: Fundamentals and Methodologies. IEEE Rev Biomed Eng 2025; 18:152-171. [PMID: 39437302 DOI: 10.1109/rbme.2024.3485022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/25/2024]
Abstract
Magnetic Resonance Imaging (MRI) is a pivotal clinical diagnostic tool, yet its extended scanning times often compromise patient comfort and image quality, especially in volumetric, temporal and quantitative scans. This review elucidates recent advances in MRI acceleration via data and physics-driven models, leveraging techniques from algorithm unrolling models, enhancement-based methods, and plug-and-play models to the emerging full spectrum of generative model-based methods. We also explore the synergistic integration of data models with physics-based insights, encompassing the advancements in multi-coil hardware accelerations like parallel imaging and simultaneous multi-slice imaging, and the optimization of sampling patterns. We then focus on domain-specific challenges and opportunities, including image redundancy exploitation, image integrity, evaluation metrics, data heterogeneity, and model generalization. This work also discusses potential solutions and future research directions, with an emphasis on the role of data harmonization and federated learning for further improving the general applicability and performance of these methods in MRI reconstruction.
Collapse
|
2
|
Zhou W, Ji J, Cui W, Wang Y, Yi Y. Unsupervised Domain Adaptation Fundus Image Segmentation via Multi-Scale Adaptive Adversarial Learning. IEEE J Biomed Health Inform 2024; 28:5792-5803. [PMID: 38090822 DOI: 10.1109/jbhi.2023.3342422] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/19/2023]
Abstract
Segmentation of the Optic Disc (OD) and Optic Cup (OC) is crucial for the early detection and treatment of glaucoma. Despite the strides made in deep neural networks, incorporating trained segmentation models for clinical application remains challenging due to domain shifts arising from disparities in fundus images across different healthcare institutions. To tackle this challenge, this study introduces an innovative unsupervised domain adaptation technique called Multi-scale Adaptive Adversarial Learning (MAAL), which consists of three key components. The Multi-scale Wasserstein Patch Discriminator (MWPD) module is designed to extract domain-specific features at multiple scales, enhancing domain classification performance and offering valuable guidance for the segmentation network. To further enhance model generalizability and explore domain-invariant features, we introduce the Adaptive Weighted Domain Constraint (AWDC) module. During training, this module dynamically assigns varying weights to different scales, allowing the model to adaptively focus on informative features. Furthermore, the Pixel-level Feature Enhancement (PFE) module enhances low-level features extracted at shallow network layers by incorporating refined high-level features. This integration ensures the preservation of domain-invariant information, effectively addressing domain variation and mitigating the loss of global features. Two publicly accessible fundus image databases are employed to demonstrate the effectiveness of our MAAL method in mitigating model degradation and improving segmentation performance. The achieved results outperform current state-of-the-art (SOTA) methods in both OD and OC segmentation.
Collapse
|
3
|
Nowak S, Bischoff LM, Pennig L, Kaya K, Isaak A, Theis M, Block W, Pieper CC, Kuetting D, Zimmer S, Nickenig G, Attenberger UI, Sprinkart AM, Luetkens JA. Deep Learning Virtual Contrast-Enhanced T1 Mapping for Contrast-Free Myocardial Extracellular Volume Assessment. J Am Heart Assoc 2024; 13:e035599. [PMID: 39344639 DOI: 10.1161/jaha.124.035599] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/25/2024] [Accepted: 08/19/2024] [Indexed: 10/01/2024]
Abstract
BACKGROUND The acquisition of contrast-enhanced T1 maps to calculate extracellular volume (ECV) requires contrast agent administration and is time consuming. This study investigates generative adversarial networks for contrast-free, virtual extracellular volume (vECV) by generating virtual contrast-enhanced T1 maps. METHODS AND RESULTS This retrospective study includes 2518 registered native and contrast-enhanced T1 maps from 1000 patients who underwent cardiovascular magnetic resonance at 1.5 Tesla. Recent hematocrit values of 123 patients (hold-out test) and 96 patients from a different institution (external evaluation) allowed for calculation of conventional ECV. A generative adversarial network was trained to generate virtual contrast-enhanced T1 maps from native T1 maps for vECV creation. Mean and SD of the difference per patient (ΔECV) were calculated and compared by permutation of the 2-sided t test with 10 000 resamples. For ECV and vECV, differences in area under the receiver operating characteristic curve (AUC) for discriminating hold-out test patients with normal cardiovascular magnetic resonance versus myocarditis or amyloidosis were tested with Delong's test. ECV and vECV showed a high agreement in patients with myocarditis (ΔECV: hold-out test, 2.0%±1.5%; external evaluation, 1.9%±1.7%) and normal cardiovascular magnetic resonance (ΔECV: hold-out test, 1.9%±1.4%; external evaluation, 1.5%±1.2%), but variations in amyloidosis were higher (ΔECV: hold-out test, 6.2%±6.0%; external evaluation, 15.5%±6.4%). In the hold-out test, ECV and vECV had a comparable AUC for the diagnosis of myocarditis (ECV AUC, 0.77 versus vECV AUC, 0.76; P=0.76) and amyloidosis (ECV AUC, 0.99 versus vECV AUC, 0.96; P=0.52). CONCLUSIONS Generation of vECV on the basis of native T1 maps is feasible. Multicenter training data are required to further enhance generalizability of vECV in amyloidosis.
Collapse
Affiliation(s)
- Sebastian Nowak
- Department of Diagnostic and Interventional Radiology University Hospital Bonn Bonn Germany
- Quantitative Imaging Laboratory Bonn (QILaB) University Hospital Bonn Bonn Germany
| | - Leon M Bischoff
- Department of Diagnostic and Interventional Radiology University Hospital Bonn Bonn Germany
- Quantitative Imaging Laboratory Bonn (QILaB) University Hospital Bonn Bonn Germany
| | - Lenhard Pennig
- Department of Diagnostic and Interventional Radiology University Hospital Cologne Cologne Germany
| | - Kenan Kaya
- Department of Diagnostic and Interventional Radiology University Hospital Cologne Cologne Germany
| | - Alexander Isaak
- Department of Diagnostic and Interventional Radiology University Hospital Bonn Bonn Germany
- Quantitative Imaging Laboratory Bonn (QILaB) University Hospital Bonn Bonn Germany
| | - Maike Theis
- Department of Diagnostic and Interventional Radiology University Hospital Bonn Bonn Germany
- Quantitative Imaging Laboratory Bonn (QILaB) University Hospital Bonn Bonn Germany
| | - Wolfgang Block
- Department of Diagnostic and Interventional Radiology University Hospital Bonn Bonn Germany
- Quantitative Imaging Laboratory Bonn (QILaB) University Hospital Bonn Bonn Germany
| | - Claus C Pieper
- Department of Diagnostic and Interventional Radiology University Hospital Bonn Bonn Germany
| | - Daniel Kuetting
- Department of Diagnostic and Interventional Radiology University Hospital Bonn Bonn Germany
- Quantitative Imaging Laboratory Bonn (QILaB) University Hospital Bonn Bonn Germany
| | - Sebastian Zimmer
- Department of Internal Medicine II, Heart Center University Hospital Bonn Bonn Germany
| | - Georg Nickenig
- Department of Internal Medicine II, Heart Center University Hospital Bonn Bonn Germany
| | - Ulrike I Attenberger
- Department of Diagnostic and Interventional Radiology University Hospital Bonn Bonn Germany
| | - Alois M Sprinkart
- Department of Diagnostic and Interventional Radiology University Hospital Bonn Bonn Germany
- Quantitative Imaging Laboratory Bonn (QILaB) University Hospital Bonn Bonn Germany
| | - Julian A Luetkens
- Department of Diagnostic and Interventional Radiology University Hospital Bonn Bonn Germany
- Quantitative Imaging Laboratory Bonn (QILaB) University Hospital Bonn Bonn Germany
| |
Collapse
|
4
|
Wang S, Wu R, Jia S, Diakite A, Li C, Liu Q, Zheng H, Ying L. Knowledge-driven deep learning for fast MR imaging: Undersampled MR image reconstruction from supervised to un-supervised learning. Magn Reson Med 2024; 92:496-518. [PMID: 38624162 DOI: 10.1002/mrm.30105] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/03/2023] [Revised: 03/19/2024] [Accepted: 03/20/2024] [Indexed: 04/17/2024]
Abstract
Deep learning (DL) has emerged as a leading approach in accelerating MRI. It employs deep neural networks to extract knowledge from available datasets and then applies the trained networks to reconstruct accurate images from limited measurements. Unlike natural image restoration problems, MRI involves physics-based imaging processes, unique data properties, and diverse imaging tasks. This domain knowledge needs to be integrated with data-driven approaches. Our review will introduce the significant challenges faced by such knowledge-driven DL approaches in the context of fast MRI along with several notable solutions, which include learning neural networks and addressing different imaging application scenarios. The traits and trends of these techniques have also been given which have shifted from supervised learning to semi-supervised learning, and finally, to unsupervised learning methods. In addition, MR vendors' choices of DL reconstruction have been provided along with some discussions on open questions and future directions, which are critical for the reliable imaging systems.
Collapse
Affiliation(s)
- Shanshan Wang
- Paul C Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Ruoyou Wu
- Paul C Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Sen Jia
- Paul C Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Alou Diakite
- Paul C Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
- University of Chinese Academy of Sciences, Beijing, China
| | - Cheng Li
- Paul C Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Qiegen Liu
- Department of Electronic Information Engineering, Nanchang University, Nanchang, China
| | - Hairong Zheng
- Paul C Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Leslie Ying
- Department of Biomedical Engineering and Department of Electrical Engineering, The State University of New York, Buffalo, New York, USA
| |
Collapse
|
5
|
Gao Y, Xiong Z, Shan S, Liu Y, Rong P, Li M, Wilman AH, Pike GB, Liu F, Sun H. Plug-and-Play latent feature editing for orientation-adaptive quantitative susceptibility mapping neural networks. Med Image Anal 2024; 94:103160. [PMID: 38552528 DOI: 10.1016/j.media.2024.103160] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/18/2023] [Revised: 03/09/2024] [Accepted: 03/23/2024] [Indexed: 04/16/2024]
Abstract
Quantitative susceptibility mapping (QSM) is a post-processing technique for deriving tissue magnetic susceptibility distribution from MRI phase measurements. Deep learning (DL) algorithms hold great potential for solving the ill-posed QSM reconstruction problem. However, a significant challenge facing current DL-QSM approaches is their limited adaptability to magnetic dipole field orientation variations during training and testing. In this work, we propose a novel Orientation-Adaptive Latent Feature Editing (OA-LFE) module to learn the encoding of acquisition orientation vectors and seamlessly integrate them into the latent features of deep networks. Importantly, it can be directly Plug-and-Play (PnP) into various existing DL-QSM architectures, enabling reconstructions of QSM from arbitrary magnetic dipole orientations. Its effectiveness is demonstrated by combining the OA-LFE module into our previously proposed phase-to-susceptibility single-step instant QSM (iQSM) network, which was initially tailored for pure-axial acquisitions. The proposed OA-LFE-empowered iQSM, which we refer to as iQSM+, is trained in a simulated-supervised manner on a specially-designed simulation brain dataset. Comprehensive experiments are conducted on simulated and in vivo human brain datasets, encompassing subjects ranging from healthy individuals to those with pathological conditions. These experiments involve various MRI platforms (3T and 7T) and aim to compare our proposed iQSM+ against several established QSM reconstruction frameworks, including the original iQSM. The iQSM+ yields QSM images with significantly improved accuracies and mitigates artifacts, surpassing other state-of-the-art DL-QSM algorithms. The PnP OA-LFE module's versatility was further demonstrated by its successful application to xQSM, a distinct DL-QSM network for dipole inversion. In conclusion, this work introduces a new DL paradigm, allowing researchers to develop innovative QSM methods without requiring a complete overhaul of their existing architectures.
Collapse
Affiliation(s)
- Yang Gao
- School of Computer Science and Engineering, Central South University, Changsha, China.
| | - Zhuang Xiong
- School of Electrical Engineering and Computer Science, University of Queensland, Brisbane, Australia
| | - Shanshan Shan
- State Key Laboratory of Radiation, Medicine and Protection, School for Radiological and Interdisciplinary Sciences (RAD-X), Collaborative Innovation Center of Radiation Medicine of Jiangsu Higher Education Institutions, Soochow University, Suzhou, China
| | - Yin Liu
- Department of Radiology, The Third Xiangya Hospital, Central South University, Changsha, China
| | - Pengfei Rong
- Department of Radiology, The Third Xiangya Hospital, Central South University, Changsha, China
| | - Min Li
- School of Computer Science and Engineering, Central South University, Changsha, China
| | - Alan H Wilman
- Department of Radiology and Diagnostic Imaging, University of Alberta, Edmonton, Canada
| | - G Bruce Pike
- Departments of Radiology and Clinical Neurosciences, Hotchkiss Brain Institute, University of Calgary, Calgary, Canada
| | - Feng Liu
- School of Electrical Engineering and Computer Science, University of Queensland, Brisbane, Australia
| | - Hongfu Sun
- School of Electrical Engineering and Computer Science, University of Queensland, Brisbane, Australia; School of Engineering, University of Newcastle, Newcastle, Australia
| |
Collapse
|
6
|
Huang W, Xu W, Wan R, Zhang P, Zha Y, Pang M. Auto Diagnosis of Parkinson's Disease Via a Deep Learning Model Based on Mixed Emotional Facial Expressions. IEEE J Biomed Health Inform 2024; 28:2547-2557. [PMID: 37022035 DOI: 10.1109/jbhi.2023.3239780] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/27/2023]
Abstract
Parkinson's disease (PD) is a common degenerative disease of the nervous system in the elderly. The early diagnosis of PD is very important for potential patients to receive prompt treatment and avoid the aggravation of the disease. Recent studies have found that PD patients always suffer from emotional expression disorder, thus forming the characteristics of "masked faces". Based on this, we thus propose an auto PD diagnosis method based on mixed emotional facial expressions in the paper. Specifically, the proposed method is cast into four steps: Firstly, we synthesize virtual face images containing six basic expressions (i.e., anger, disgust, fear, happiness, sadness, and surprise) via generative adversarial learning, in order to approximate the premorbid expressions of PD patients; Secondly, we design an effective screening scheme to assess the quality of the above synthesized facial expression images and then shortlist the high-quality ones; Thirdly, we train a deep feature extractor accompanied with a facial expression classifier based on the mixture of the original facial expression images of the PD patients, the high-quality synthesized facial expression images of PD patients, and the normal facial expression images from other public face datasets; Finally, with the well-trained deep feature extractor, we thus adopt it to extract the latent expression features for six facial expression images of a potential PD patient to conduct PD/non-PD prediction. To show real-world impacts, we also collected a new facial expression dataset of PD patients in collaboration with a hospital. Extensive experiments are conducted to validate the effectiveness of the proposed method for PD diagnosis and facial expression recognition.
Collapse
|
7
|
Huang P, Zhang C, Zhang X, Li X, Dong L, Ying L. Self-Supervised Deep Unrolled Reconstruction Using Regularization by Denoising. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:1203-1213. [PMID: 37962993 PMCID: PMC11056277 DOI: 10.1109/tmi.2023.3332614] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2023]
Abstract
Deep learning methods have been successfully used in various computer vision tasks. Inspired by that success, deep learning has been explored in magnetic resonance imaging (MRI) reconstruction. In particular, integrating deep learning and model-based optimization methods has shown considerable advantages. However, a large amount of labeled training data is typically needed for high reconstruction quality, which is challenging for some MRI applications. In this paper, we propose a novel reconstruction method, named DURED-Net, that enables interpretable self-supervised learning for MR image reconstruction by combining a self-supervised denoising network and a plug-and-play method. We aim to boost the reconstruction performance of Noise2Noise in MR reconstruction by adding an explicit prior that utilizes imaging physics. Specifically, the leverage of a denoising network for MRI reconstruction is achieved using Regularization by Denoising (RED). Experiment results demonstrate that the proposed method requires a reduced amount of training data to achieve high reconstruction quality among the state-of-the-art approaches utilizing Noise2Noise.
Collapse
|
8
|
Dar SUH, Öztürk Ş, Özbey M, Oguz KK, Çukur T. Parallel-stream fusion of scan-specific and scan-general priors for learning deep MRI reconstruction in low-data regimes. Comput Biol Med 2023; 167:107610. [PMID: 37883853 DOI: 10.1016/j.compbiomed.2023.107610] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/10/2023] [Revised: 09/20/2023] [Accepted: 10/17/2023] [Indexed: 10/28/2023]
Abstract
Magnetic resonance imaging (MRI) is an essential diagnostic tool that suffers from prolonged scan times. Reconstruction methods can alleviate this limitation by recovering clinically usable images from accelerated acquisitions. In particular, learning-based methods promise performance leaps by employing deep neural networks as data-driven priors. A powerful approach uses scan-specific (SS) priors that leverage information regarding the underlying physical signal model for reconstruction. SS priors are learned on each individual test scan without the need for a training dataset, albeit they suffer from computationally burdening inference with nonlinear networks. An alternative approach uses scan-general (SG) priors that instead leverage information regarding the latent features of MRI images for reconstruction. SG priors are frozen at test time for efficiency, albeit they require learning from a large training dataset. Here, we introduce a novel parallel-stream fusion model (PSFNet) that synergistically fuses SS and SG priors for performant MRI reconstruction in low-data regimes, while maintaining competitive inference times to SG methods. PSFNet implements its SG prior based on a nonlinear network, yet it forms its SS prior based on a linear network to maintain efficiency. A pervasive framework for combining multiple priors in MRI reconstruction is algorithmic unrolling that uses serially alternated projections, causing error propagation under low-data regimes. To alleviate error propagation, PSFNet combines its SS and SG priors via a novel parallel-stream architecture with learnable fusion parameters. Demonstrations are performed on multi-coil brain MRI for varying amounts of training data. PSFNet outperforms SG methods in low-data regimes, and surpasses SS methods with few tens of training samples. On average across tasks, PSFNet achieves 3.1 dB higher PSNR, 2.8% higher SSIM, and 0.3 × lower RMSE than baselines. Furthermore, in both supervised and unsupervised setups, PSFNet requires an order of magnitude lower samples compared to SG methods, and enables an order of magnitude faster inference compared to SS methods. Thus, the proposed model improves deep MRI reconstruction with elevated learning and computational efficiency.
Collapse
Affiliation(s)
- Salman Ul Hassan Dar
- Department of Internal Medicine III, Heidelberg University Hospital, 69120, Heidelberg, Germany; AI Health Innovation Cluster, Heidelberg, Germany
| | - Şaban Öztürk
- Department of Electrical and Electronics Engineering, Bilkent University, Ankara 06800, Turkey; Department of Electrical-Electronics Engineering, Amasya University, Amasya 05100, Turkey
| | - Muzaffer Özbey
- Department of Electrical and Computer Engineering, University of Illinois Urbana-Champaign, IL 61820, United States
| | - Kader Karli Oguz
- Department of Radiology, University of California, Davis, CA 95616, United States; Department of Radiology, Hacettepe University, Ankara, Turkey
| | - Tolga Çukur
- Department of Electrical and Electronics Engineering, Bilkent University, Ankara 06800, Turkey; Department of Radiology, Hacettepe University, Ankara, Turkey; National Magnetic Resonance Research Center (UMRAM), Bilkent University, Ankara 06800, Turkey; Neuroscience Graduate Program, Bilkent University, Ankara 06800, Turkey.
| |
Collapse
|
9
|
Desai AD, Ozturkler BM, Sandino CM, Boutin R, Willis M, Vasanawala S, Hargreaves BA, Ré C, Pauly JM, Chaudhari AS. Noise2Recon: Enabling SNR-robust MRI reconstruction with semi-supervised and self-supervised learning. Magn Reson Med 2023; 90:2052-2070. [PMID: 37427449 DOI: 10.1002/mrm.29759] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/05/2022] [Revised: 05/23/2023] [Accepted: 05/24/2023] [Indexed: 07/11/2023]
Abstract
PURPOSE To develop a method for building MRI reconstruction neural networks robust to changes in signal-to-noise ratio (SNR) and trainable with a limited number of fully sampled scans. METHODS We propose Noise2Recon, a consistency training method for SNR-robust accelerated MRI reconstruction that can use both fully sampled (labeled) and undersampled (unlabeled) scans. Noise2Recon uses unlabeled data by enforcing consistency between model reconstructions of undersampled scans and their noise-augmented counterparts. Noise2Recon was compared to compressed sensing and both supervised and self-supervised deep learning baselines. Experiments were conducted using retrospectively accelerated data from the mridata three-dimensional fast-spin-echo knee and two-dimensional fastMRI brain datasets. All methods were evaluated in label-limited settings and among out-of-distribution (OOD) shifts, including changes in SNR, acceleration factors, and datasets. An extensive ablation study was conducted to characterize the sensitivity of Noise2Recon to hyperparameter choices. RESULTS In label-limited settings, Noise2Recon achieved better structural similarity, peak signal-to-noise ratio, and normalized-RMS error than all baselines and matched performance of supervised models, which were trained with14 × $$ 14\times $$ more fully sampled scans. Noise2Recon outperformed all baselines, including state-of-the-art fine-tuning and augmentation techniques, among low-SNR scans and when generalizing to OOD acceleration factors. Augmentation extent and loss weighting hyperparameters had negligible impact on Noise2Recon compared to supervised methods, which may indicate increased training stability. CONCLUSION Noise2Recon is a label-efficient reconstruction method that is robust to distribution shifts, such as changes in SNR, acceleration factors, and others, with limited or no fully sampled training data.
Collapse
Affiliation(s)
- Arjun D Desai
- Department of Electrical Engineering, Stanford University, Stanford, California, USA
- Department of Radiology, Stanford University, Stanford, California, USA
| | - Batu M Ozturkler
- Department of Electrical Engineering, Stanford University, Stanford, California, USA
| | - Christopher M Sandino
- Department of Electrical Engineering, Stanford University, Stanford, California, USA
| | - Robert Boutin
- Department of Radiology, Stanford University, Stanford, California, USA
| | - Marc Willis
- Department of Radiology, Stanford University, Stanford, California, USA
| | | | - Brian A Hargreaves
- Department of Electrical Engineering, Stanford University, Stanford, California, USA
- Department of Radiology, Stanford University, Stanford, California, USA
| | - Christopher Ré
- Department of Computer Science, Stanford University, Stanford, California, USA
| | - John M Pauly
- Department of Electrical Engineering, Stanford University, Stanford, California, USA
| | - Akshay S Chaudhari
- Department of Radiology, Stanford University, Stanford, California, USA
- Department of Biomedical Data Science, Stanford University, Stanford, California, USA
| |
Collapse
|
10
|
Zhang Q, Shu J, Chen C, Teng Z, Gu Z, Li F, Kan J. Optimization of pneumonia CT classification model using RepVGG and spatial attention features. Front Med (Lausanne) 2023; 10:1233724. [PMID: 37795420 PMCID: PMC10546926 DOI: 10.3389/fmed.2023.1233724] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/12/2023] [Accepted: 09/05/2023] [Indexed: 10/06/2023] Open
Abstract
Introduction Pneumonia is a common and widespread infectious disease that seriously affects the life and health of patients. Especially in recent years, the outbreak of COVID-19 has caused a sharp rise in the number of confirmed cases of epidemic spread. Therefore, early detection and treatment of pneumonia are very important. However, the uneven gray distribution and structural intricacy of pneumonia images substantially impair the classification accuracy of pneumonia. In this classification task of COVID-19 and other pneumonia, because there are some commonalities between this pneumonia, even a small gap will lead to the risk of prediction deviation, it is difficult to achieve high classification accuracy by directly using the current network model to optimize the classification model. Methods Consequently, an optimization method for the CT classification model of COVID-19 based on RepVGG was proposed. In detail, it is made up of two essential modules, feature extraction backbone and spatial attention block, which allows it to extract spatial attention features while retaining the benefits of RepVGG. Results The model's inference time is significantly reduced, and it shows better learning ability than RepVGG on both the training and validation sets. Compared with the existing advanced network models VGG-16, ResNet-50, GoogleNet, ViT, AlexNet, MobileViT, ConvNeXt, ShuffleNet, and RepVGG_b0, our model has demonstrated the best performance in a lot of indicators. In testing, it achieved an accuracy of 0.951, an F1 score of 0.952, and a Youden index of 0.902. Discussion Overall, multiple experiments on the large dataset of SARS-CoV-2 CT-scan dataset reveal that this method outperforms most basic models in terms of classification and screening of COVID-19 CT, and has a significant reference value. Simultaneously, in the inspection experiment, this method outperformed other networks with residual structures.
Collapse
Affiliation(s)
| | - Jianhua Shu
- School of Medical Information Engineering, Anhui University of Chinese Medicine, Hefei, China
| | | | | | | | | | - Junling Kan
- School of Medical Information Engineering, Anhui University of Chinese Medicine, Hefei, China
| |
Collapse
|
11
|
Güngör A, Dar SU, Öztürk Ş, Korkmaz Y, Bedel HA, Elmas G, Ozbey M, Çukur T. Adaptive diffusion priors for accelerated MRI reconstruction. Med Image Anal 2023; 88:102872. [PMID: 37384951 DOI: 10.1016/j.media.2023.102872] [Citation(s) in RCA: 41] [Impact Index Per Article: 20.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/31/2022] [Revised: 04/13/2023] [Accepted: 06/12/2023] [Indexed: 07/01/2023]
Abstract
Deep MRI reconstruction is commonly performed with conditional models that de-alias undersampled acquisitions to recover images consistent with fully-sampled data. Since conditional models are trained with knowledge of the imaging operator, they can show poor generalization across variable operators. Unconditional models instead learn generative image priors decoupled from the operator to improve reliability against domain shifts related to the imaging operator. Recent diffusion models are particularly promising given their high sample fidelity. Nevertheless, inference with a static image prior can perform suboptimally. Here we propose the first adaptive diffusion prior for MRI reconstruction, AdaDiff, to improve performance and reliability against domain shifts. AdaDiff leverages an efficient diffusion prior trained via adversarial mapping over large reverse diffusion steps. A two-phase reconstruction is executed following training: a rapid-diffusion phase that produces an initial reconstruction with the trained prior, and an adaptation phase that further refines the result by updating the prior to minimize data-consistency loss. Demonstrations on multi-contrast brain MRI clearly indicate that AdaDiff outperforms competing conditional and unconditional methods under domain shifts, and achieves superior or on par within-domain performance.
Collapse
Affiliation(s)
- Alper Güngör
- Department of Electrical and Electronics Engineering, Bilkent University, Ankara 06800, Turkey; National Magnetic Resonance Research Center (UMRAM), Bilkent University, Ankara 06800, Turkey; ASELSAN Research Center, Ankara 06200, Turkey
| | - Salman Uh Dar
- Department of Electrical and Electronics Engineering, Bilkent University, Ankara 06800, Turkey; National Magnetic Resonance Research Center (UMRAM), Bilkent University, Ankara 06800, Turkey; Department of Internal Medicine III, Heidelberg University Hospital, Heidelberg 69120, Germany
| | - Şaban Öztürk
- Department of Electrical and Electronics Engineering, Bilkent University, Ankara 06800, Turkey; National Magnetic Resonance Research Center (UMRAM), Bilkent University, Ankara 06800, Turkey; Department of Electrical and Electronics Engineering, Amasya University, Amasya 05100, Turkey
| | - Yilmaz Korkmaz
- Department of Electrical and Electronics Engineering, Bilkent University, Ankara 06800, Turkey; National Magnetic Resonance Research Center (UMRAM), Bilkent University, Ankara 06800, Turkey
| | - Hasan A Bedel
- Department of Electrical and Electronics Engineering, Bilkent University, Ankara 06800, Turkey; National Magnetic Resonance Research Center (UMRAM), Bilkent University, Ankara 06800, Turkey
| | - Gokberk Elmas
- Department of Electrical and Electronics Engineering, Bilkent University, Ankara 06800, Turkey; National Magnetic Resonance Research Center (UMRAM), Bilkent University, Ankara 06800, Turkey
| | - Muzaffer Ozbey
- Department of Electrical and Electronics Engineering, Bilkent University, Ankara 06800, Turkey; National Magnetic Resonance Research Center (UMRAM), Bilkent University, Ankara 06800, Turkey
| | - Tolga Çukur
- Department of Electrical and Electronics Engineering, Bilkent University, Ankara 06800, Turkey; National Magnetic Resonance Research Center (UMRAM), Bilkent University, Ankara 06800, Turkey; Neuroscience Program, Bilkent University, Ankara 06800, Turkey.
| |
Collapse
|
12
|
Huang P, Li H, Liu R, Zhang X, Li X, Liang D, Ying L. Accelerating MRI Using Vision Transformer with Unpaired Unsupervised Training. PROCEEDINGS OF THE INTERNATIONAL SOCIETY FOR MAGNETIC RESONANCE IN MEDICINE ... SCIENTIFIC MEETING AND EXHIBITION. INTERNATIONAL SOCIETY FOR MAGNETIC RESONANCE IN MEDICINE. SCIENTIFIC MEETING AND EXHIBITION 2023; 31:2933. [PMID: 37600538 PMCID: PMC10440071] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Key Words] [Grants] [Subscribe] [Scholar Register] [Indexed: 08/22/2023]
Affiliation(s)
- Peizhou Huang
- Biomedical Engineering, State University of New York at Buffalo, Buffalo, NY, United States
| | - Hongyu Li
- Electrical Engineering, State University of New York at Buffalo, Buffalo, NY, United States
| | - Ruiying Liu
- Electrical Engineering, State University of New York at Buffalo, Buffalo, NY, United States
| | - Xiaoliang Zhang
- Biomedical Engineering, State University of New York at Buffalo, Buffalo, NY, United States
| | - Xiaojuan Li
- Program of Advanced Musculoskeletal Imaging (PAMI), Cleveland Clinic, Cleveland, OH, United States
| | - Dong Liang
- Paul C. Lauterbur Research Center for Biomedical Imaging, SIAT CAS, Shenzhen, China
| | - Leslie Ying
- Biomedical Engineering, State University of New York at Buffalo, Buffalo, NY, United States
- Electrical Engineering, State University of New York at Buffalo, Buffalo, NY, United States
| |
Collapse
|
13
|
Qiao Z, Li L, Zhao X, Liu L, Zhang Q, Hechmi S, Atri M, Li X. An enhanced Runge Kutta boosted machine learning framework for medical diagnosis. Comput Biol Med 2023; 160:106949. [PMID: 37159961 DOI: 10.1016/j.compbiomed.2023.106949] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/24/2022] [Revised: 03/27/2023] [Accepted: 04/15/2023] [Indexed: 05/11/2023]
Abstract
With the development and maturity of machine learning methods, medical diagnosis aided with machine learning methods has become a popular method to assist doctors in diagnosing and treating patients. However, machine learning methods are greatly affected by their hyperparameters, for instance, the kernel parameter in kernel extreme learning machine (KELM) and the learning rate in residual neural networks (ResNet). If the hyperparameters are appropriately set, the performance of the classifier can be significantly improved. To boost the performance of the machine learning methods, this paper proposes to improve the Runge Kutta optimizer (RUN) to adaptively adjust the hyperparameters of the machine learning methods for medical diagnosis purposes. Although RUN has a solid mathematical theoretical foundation, there are still some performance defects when dealing with complex optimization problems. To remedy these defects, this paper proposes a new enhanced RUN method with a grey wolf mechanism and an orthogonal learning mechanism called GORUN. The superior performance of the GORUN was validated against other well-established optimizers on IEEE CEC 2017 benchmark functions. Then, the proposed GORUN is employed to optimize the machine learning models, including the KELM and ResNet, to construct robust models for medical diagnosis. The performance of the proposed machine learning framework was validated on several medical data sets, and the experimental results have demonstrated its superiority.
Collapse
Affiliation(s)
- Zenglin Qiao
- School of Science, Beijing University of Posts and Telecommunications, Beijing, 100876, China.
| | - Lynn Li
- China Telecom Stocks Co.,Ltd., Hangzhou Branch, Hangzhou, 310000, China.
| | - Xinchao Zhao
- School of Science, Beijing University of Posts and Telecommunications, Beijing, 100876, China.
| | - Lei Liu
- College of Computer Science, Sichuan University, Chengdu, Sichuan, 610065, China.
| | - Qian Zhang
- School of Data Science and Artificial Intelligence, Wenzhou University of Technology, Wenzhou, Zhejiang, 325035, China.
| | - Shili Hechmi
- Dept. Computer Sciences, Tabuk University, Tabuk, Saudi Arabia.
| | - Mohamed Atri
- College of Computer Science, King Khalid University, Abha, Saudi Arabia.
| | - Xiaohua Li
- Library, Wenzhou University, Wenzhou, Zhejiang, 325035, China.
| |
Collapse
|
14
|
Yi K, Li H, Xu C, Zhong G, Ding Z, Zhang G, Guan X, Zhong M, Li G, Jiang N, Zhang Y. Morphological feature recognition of different differentiation stages of induced ADSCs based on deep learning. Comput Biol Med 2023; 159:106906. [PMID: 37084638 DOI: 10.1016/j.compbiomed.2023.106906] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/29/2022] [Revised: 03/14/2023] [Accepted: 04/10/2023] [Indexed: 04/23/2023]
Abstract
In order to accurately identify the morphological features of different differentiation stages of induced Adipose Derived Stem Cells (ADSCs) and judge the differentiation types of induced ADSCs, a morphological feature recognition method of different differentiation stages of induced ADSCs based on deep learning is proposed. Using the super-resolution image acquisition method of ADSCs differentiation based on stimulated emission depletion imaging, after obtaining the super-resolution images at different stages of inducing ADSCs differentiation, the noise of the obtained image is removed and the image quality is optimized through the ADSCs differentiation image denoising model based on low rank nonlocal sparse representation; The denoised image is taken as the recognition target of the morphological feature recognition method for ADSCs differentiation image based on the improved Visual Geometry Group (VGG-19) convolutional neural network. Through the improved VGG-19 convolutional neural network and class activation mapping method, the morphological feature recognition and visual display of the recognition results at different stages of inducing ADSCs differentiation are realized. After testing, this method can accurately identify the morphological features of different differentiation stages of induced ADSCs, and is available.
Collapse
Affiliation(s)
- Ke Yi
- School of Information Engineering, East China Jiaotong University, 330013 Nanchang, Jiangxi, China
| | - Han Li
- Meta Platforms, Inc., Menlo Park, CA 94025, USA
| | - Cheng Xu
- School of Information Engineering, East China Jiaotong University, 330013 Nanchang, Jiangxi, China
| | - Guoqing Zhong
- School of Information Engineering, East China Jiaotong University, 330013 Nanchang, Jiangxi, China
| | - Zhiquan Ding
- School of Information Engineering, East China Jiaotong University, 330013 Nanchang, Jiangxi, China
| | - Guolong Zhang
- School of Information Engineering, East China Jiaotong University, 330013 Nanchang, Jiangxi, China
| | - Xiaohui Guan
- The National Engineering Research Center for Bioengineering Drugs and the Technologies, Nanchang University, Nanchang, China
| | - Meiling Zhong
- School of Materials Science and Engineering, East China Jiaotong University, Nanchang, China
| | - Guanghui Li
- School of Information Engineering, East China Jiaotong University, 330013 Nanchang, Jiangxi, China
| | - Nan Jiang
- School of Information Engineering, East China Jiaotong University, 330013 Nanchang, Jiangxi, China
| | - Yuejin Zhang
- School of Information Engineering, East China Jiaotong University, 330013 Nanchang, Jiangxi, China.
| |
Collapse
|
15
|
Zhang K, Lin P, Pan J, Xu P, Qiu X, Crookes D, Hua L, Wang L. End to End Multitask Joint Learning Model for Osteoporosis Classification in CT Images. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2023; 2023:3018320. [PMID: 36970245 PMCID: PMC10036193 DOI: 10.1155/2023/3018320] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/10/2022] [Revised: 02/23/2023] [Accepted: 03/01/2023] [Indexed: 03/18/2023]
Abstract
Osteoporosis is a significant global health concern that can be difficult to detect early due to a lack of symptoms. At present, the examination of osteoporosis depends mainly on methods containing dual-energyX-ray, quantitative CT, etc., which are high costs in terms of equipment and human time. Therefore, a more efficient and economical method is urgently needed for diagnosing osteoporosis. With the development of deep learning, automatic diagnosis models for various diseases have been proposed. However, the establishment of these models generally requires images with only lesion areas, and annotating the lesion areas is time-consuming. To address this challenge, we propose a joint learning framework for osteoporosis diagnosis that combines localization, segmentation, and classification to enhance diagnostic accuracy. Our method includes a boundary heat map regression branch for thinning segmentation and a gated convolution module for adjusting context features in the classification module. We also integrate segmentation and classification features and propose a feature fusion module to adjust the weight of different levels of vertebrae. We trained our model on a self-built dataset and achieved an overall accuracy rate of 93.3% for the three label categories (normal, osteopenia, and osteoporosis) in the testing datasets. The area under the curve for the normal category is 0.973; for the osteopenia category, it is 0.965; and for the osteoporosis category, it is 0.985. Our method provides a promising alternative for the diagnosis of osteoporosis at present.
Collapse
Affiliation(s)
- Kun Zhang
- School of Electrical Engineering, Nantong University, Nantong, Jiangsu 226001, China
- Nantong Key Laboratory of Intelligent Control and Intelligent Computing, Nantong, Jiangsu 226001, China
- Nantong Key Laboratory of Intelligent Medicine Innovation and Transformation, Nantong, Jiangsu 226001, China
| | - Pengcheng Lin
- School of Electrical Engineering, Nantong University, Nantong, Jiangsu 226001, China
| | - Jing Pan
- Department of Radiology, Affiliated Hospital 2 of Nantong University, Nantong, Jiangsu 226001, China
| | - Peixia Xu
- School of Electrical Engineering, Nantong University, Nantong, Jiangsu 226001, China
| | - Xuechen Qiu
- College of Mechanical Engineering, Donghua University, Shanghai 201620, China
| | - Danny Crookes
- School of Electronics, Electrical Engineering and Computer Science, Queen's University Belfast, Belfast BT7 1NN, UK
| | - Liang Hua
- School of Electrical Engineering, Nantong University, Nantong, Jiangsu 226001, China
| | - Lin Wang
- Department of Radiology, Affiliated Hospital 2 of Nantong University, Nantong, Jiangsu 226001, China
| |
Collapse
|
16
|
Gu Z, Li Y, Wang Z, Kan J, Shu J, Wang Q. Classification of Diabetic Retinopathy Severity in Fundus Images Using the Vision Transformer and Residual Attention. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2023; 2023:1305583. [PMID: 36636467 PMCID: PMC9831706 DOI: 10.1155/2023/1305583] [Citation(s) in RCA: 7] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 05/26/2022] [Revised: 12/04/2022] [Accepted: 12/05/2022] [Indexed: 01/05/2023]
Abstract
Diabetic retinopathy (DR) is a common retinal vascular disease, which can cause severe visual impairment. It is of great clinical significance to use fundus images for intelligent diagnosis of DR. In this paper, an intelligent DR classification model of fundus images is proposed. This method can detect all the five stages of DR, including of no DR, mild, moderate, severe, and proliferative. This model is composed of two key modules. FEB, feature extraction block, is mainly used for feature extraction of fundus images, and GPB, grading prediction block, is used to classify the five stages of DR. The transformer in the FEB has more fine-grained attention that can pay more attention to retinal hemorrhage and exudate areas. The residual attention in the GPB can effectively capture different spatial regions occupied by different classes of objects. Comprehensive experiments on DDR datasets well demonstrate the superiority of our method, and compared with the benchmark method, our method has achieved competitive performance.
Collapse
Affiliation(s)
- Zongyun Gu
- College of Medical Information Engineering, Anhui University of Traditional Chinese Medicine, Hefei 230012, China
- Artificial Intelligence Research Institute of Hefei Comprehensive National Science Center (Anhui Artificial Intelligence Laboratory), Hefei 230026, China
| | - Yan Li
- Joint Surgery Department, Hefei First People's Hospital, Hefei 230061, China
| | - Zijian Wang
- College of Medical Information Engineering, Anhui University of Traditional Chinese Medicine, Hefei 230012, China
| | - Junling Kan
- College of Medical Information Engineering, Anhui University of Traditional Chinese Medicine, Hefei 230012, China
| | - Jianhua Shu
- College of Medical Information Engineering, Anhui University of Traditional Chinese Medicine, Hefei 230012, China
| | - Qing Wang
- College of Medical Information Engineering, Anhui University of Traditional Chinese Medicine, Hefei 230012, China
| |
Collapse
|
17
|
Yaman B, Gu H, Hosseini SAH, Demirel OB, Moeller S, Ellermann J, Uğurbil K, Akçakaya M. Multi-mask self-supervised learning for physics-guided neural networks in highly accelerated magnetic resonance imaging. NMR IN BIOMEDICINE 2022; 35:e4798. [PMID: 35789133 PMCID: PMC9669191 DOI: 10.1002/nbm.4798] [Citation(s) in RCA: 18] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/10/2021] [Revised: 06/30/2022] [Accepted: 07/02/2022] [Indexed: 06/15/2023]
Abstract
Self-supervised learning has shown great promise because of its ability to train deep learning (DL) magnetic resonance imaging (MRI) reconstruction methods without fully sampled data. Current self-supervised learning methods for physics-guided reconstruction networks split acquired undersampled data into two disjoint sets, where one is used for data consistency (DC) in the unrolled network, while the other is used to define the training loss. In this study, we propose an improved self-supervised learning strategy that more efficiently uses the acquired data to train a physics-guided reconstruction network without a database of fully sampled data. The proposed multi-mask self-supervised learning via data undersampling (SSDU) applies a holdout masking operation on the acquired measurements to split them into multiple pairs of disjoint sets for each training sample, while using one of these pairs for DC units and the other for defining loss, thereby more efficiently using the undersampled data. Multi-mask SSDU is applied on fully sampled 3D knee and prospectively undersampled 3D brain MRI datasets, for various acceleration rates and patterns, and compared with the parallel imaging method, CG-SENSE, and single-mask SSDU DL-MRI, as well as supervised DL-MRI when fully sampled data are available. The results on knee MRI show that the proposed multi-mask SSDU outperforms SSDU and performs as well as supervised DL-MRI. A clinical reader study further ranks the multi-mask SSDU higher than supervised DL-MRI in terms of signal-to-noise ratio and aliasing artifacts. Results on brain MRI show that multi-mask SSDU achieves better reconstruction quality compared with SSDU. The reader study demonstrates that multi-mask SSDU at R = 8 significantly improves reconstruction compared with single-mask SSDU at R = 8, as well as CG-SENSE at R = 2.
Collapse
Affiliation(s)
- Burhaneddin Yaman
- Department of Electrical and Computer EngineeringUniversity of MinnesotaMinneapolisMinnesotaUSA
- Center for Magnetic Resonance ResearchUniversity of MinnesotaMinneapolisMinnesotaUSA
| | - Hongyi Gu
- Department of Electrical and Computer EngineeringUniversity of MinnesotaMinneapolisMinnesotaUSA
- Center for Magnetic Resonance ResearchUniversity of MinnesotaMinneapolisMinnesotaUSA
| | - Seyed Amir Hossein Hosseini
- Department of Electrical and Computer EngineeringUniversity of MinnesotaMinneapolisMinnesotaUSA
- Center for Magnetic Resonance ResearchUniversity of MinnesotaMinneapolisMinnesotaUSA
| | - Omer Burak Demirel
- Department of Electrical and Computer EngineeringUniversity of MinnesotaMinneapolisMinnesotaUSA
- Center for Magnetic Resonance ResearchUniversity of MinnesotaMinneapolisMinnesotaUSA
| | - Steen Moeller
- Center for Magnetic Resonance ResearchUniversity of MinnesotaMinneapolisMinnesotaUSA
| | - Jutta Ellermann
- Center for Magnetic Resonance ResearchUniversity of MinnesotaMinneapolisMinnesotaUSA
| | - Kâmil Uğurbil
- Center for Magnetic Resonance ResearchUniversity of MinnesotaMinneapolisMinnesotaUSA
| | - Mehmet Akçakaya
- Department of Electrical and Computer EngineeringUniversity of MinnesotaMinneapolisMinnesotaUSA
- Center for Magnetic Resonance ResearchUniversity of MinnesotaMinneapolisMinnesotaUSA
| |
Collapse
|
18
|
Zhang F, Zhang Y, Zhu X, Chen X, Du H, Zhang X. PregGAN: A prognosis prediction model for breast cancer based on conditional generative adversarial networks. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 224:107026. [PMID: 35872384 DOI: 10.1016/j.cmpb.2022.107026] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/01/2021] [Revised: 07/13/2022] [Accepted: 07/13/2022] [Indexed: 06/15/2023]
Abstract
BACKGROUND AND OBJECTIVE Generative adversarial network (GAN) is able to learn from a set of training data and generate new data with the same characteristics as the training data. Based on the characteristics of GAN, this paper developed its capability as a tool of disease prognosis prediction, and proposed a prognostic model PregGAN based on conditional generative adversarial network (CGAN). METHODS The idea of PregGAN is to generate the prognosis prediction results based on the clinical data of patients. PregGAN added the clinical data as conditions to the training process. Conditions were used as the input to the generator along with noises. The generator synthesized new samples using the noises vectors and the conditions. In order to solve the mode collapse problem during PregGAN training, Wasserstein distance and gradient penalty strategy were used to make the training process more stable. RESULTS In the prognosis prediction experiments using the METABRIC breast cancer dataset, PregGAN achieved good results, with the average accurate (ACC) of 90.6% and the average AUC (area under curve) of 0.946. CONCLUSIONS Experimental results show that PregGAN is a reliable prognosis predictive model for breast cancer. Due to the strong ability of probability distribution learning, PregGAN can also be used for the prognosis prediction of other diseases.
Collapse
Affiliation(s)
- Fan Zhang
- Henan Key Laboratory of Big Data Analysis and Processing, Henan University, Kaifeng 475004, China; Henan Engineering Laboratory of Spatial Information Processing, Henan University, Kaifeng 475004, China
| | - Yingqi Zhang
- Henan Key Laboratory of Big Data Analysis and Processing, Henan University, Kaifeng 475004, China
| | - Xiaoke Zhu
- Henan Key Laboratory of Big Data Analysis and Processing, Henan University, Kaifeng 475004, China
| | - Xiaopan Chen
- Henan Key Laboratory of Big Data Analysis and Processing, Henan University, Kaifeng 475004, China
| | - Haishun Du
- School of Artificial Intelligence, Henan University, Kaifeng 475004, China
| | - Xinhong Zhang
- School of Software, Henan University, Kaifeng 475004, China.
| |
Collapse
|
19
|
Korkmaz Y, Dar SUH, Yurt M, Ozbey M, Cukur T. Unsupervised MRI Reconstruction via Zero-Shot Learned Adversarial Transformers. IEEE TRANSACTIONS ON MEDICAL IMAGING 2022; 41:1747-1763. [PMID: 35085076 DOI: 10.1109/tmi.2022.3147426] [Citation(s) in RCA: 88] [Impact Index Per Article: 29.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Supervised reconstruction models are characteristically trained on matched pairs of undersampled and fully-sampled data to capture an MRI prior, along with supervision regarding the imaging operator to enforce data consistency. To reduce supervision requirements, the recent deep image prior framework instead conjoins untrained MRI priors with the imaging operator during inference. Yet, canonical convolutional architectures are suboptimal in capturing long-range relationships, and priors based on randomly initialized networks may yield suboptimal performance. To address these limitations, here we introduce a novel unsupervised MRI reconstruction method based on zero-Shot Learned Adversarial TransformERs (SLATER). SLATER embodies a deep adversarial network with cross-attention transformers to map noise and latent variables onto coil-combined MR images. During pre-training, this unconditional network learns a high-quality MRI prior in an unsupervised generative modeling task. During inference, a zero-shot reconstruction is then performed by incorporating the imaging operator and optimizing the prior to maximize consistency to undersampled data. Comprehensive experiments on brain MRI datasets clearly demonstrate the superior performance of SLATER against state-of-the-art unsupervised methods.
Collapse
|
20
|
Ovalle-Magallanes E, Avina-Cervantes JG, Cruz-Aceves I, Ruiz-Pinales J. Improving convolutional neural network learning based on a hierarchical bezier generative model for stenosis detection in X-ray images. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 219:106767. [PMID: 35364481 DOI: 10.1016/j.cmpb.2022.106767] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/17/2021] [Revised: 03/09/2022] [Accepted: 03/19/2022] [Indexed: 06/14/2023]
Abstract
BACKGROUND AND OBJECTIVE Automatic detection of stenosis on X-ray Coronary Angiography (XCA) images may help diagnose early coronary artery disease. Stenosis is manifested by a buildup of plaque in the arteries, decreasing the blood flow to the heart, increasing the risk of a heart attack. Convolutional Neural Networks (CNNs) have been successfully applied to identify pathological, regular, and featured tissues on rich and diverse medical image datasets. Nevertheless, CNNs find operative and performing limitations while working with small and poorly diversified databases. Transfer learning from large natural image datasets (such as ImageNet) has become a de-facto method to improve neural networks performance in the medical image domain. METHODS This paper proposes a novel Hierarchical Bezier-based Generative Model (HBGM) to improve the CNNs training process to detect stenosis. Herein, artificial image patches are generated to enlarge the original database, speeding up network convergence. The artificial dataset consists of 10,000 images containing 50% stenosis and 50% non-stenosis cases. Besides, a reliable Fréchet Inception Distance (FID) is used to evaluate the generated data quantitatively. Therefore, by using the proposed framework, the network is pre-trained with the artificial datasets and subsequently fine-tuned using the real XCA training dataset. The real dataset consists of 250 XCA image patches, selecting 125 images for stenosis and the remainder for non-stenosis cases. Furthermore, a Convolutional Block Attention Module (CBAM) was included in the network architecture as a self-attention mechanism to improve the efficiency of the network. RESULTS The results showed that the pre-trained networks using the proposed generative model outperformed the results concerning training from scratch. Particularly, an accuracy, precision, sensitivity, and F1-score of 0.8934, 0.9031, 0.8746, 0.8880, 0.9111, respectively, were achieved. The generated artificial dataset obtains a mean FID of 84.0886, with more realistic visual XCA images. CONCLUSIONS Different ResNet architectures for stenosis detection have been evaluated, including attention modules into the network. Numerical results demonstrated that by using the HBGM is obtained a higher performance than training from scratch, even outperforming the ImageNet pre-trained models.
Collapse
Affiliation(s)
- Emmanuel Ovalle-Magallanes
- Telematics and Digital Signal Processing Research groups (CAs), Engineering Division, Campus Irapuato-Salamanca, University of Guanajuato, Carretera Salamanca-Valle de Santiago km 3.5 + 1.8km, Comunidad de Palo Blanco, Salamanca, 36885 Guanajuato, Mexico.
| | - Juan Gabriel Avina-Cervantes
- Telematics and Digital Signal Processing Research groups (CAs), Engineering Division, Campus Irapuato-Salamanca, University of Guanajuato, Carretera Salamanca-Valle de Santiago km 3.5 + 1.8km, Comunidad de Palo Blanco, Salamanca, 36885 Guanajuato, Mexico.
| | - Ivan Cruz-Aceves
- CONACYT, Center for Research in Mathematics (CIMAT), A.C., Jalisco S/N, Col. Valenciana, Guanajuato, 36000 Guanajuato, Mexico.
| | - Jose Ruiz-Pinales
- Telematics and Digital Signal Processing Research groups (CAs), Engineering Division, Campus Irapuato-Salamanca, University of Guanajuato, Carretera Salamanca-Valle de Santiago km 3.5 + 1.8km, Comunidad de Palo Blanco, Salamanca, 36885 Guanajuato, Mexico.
| |
Collapse
|
21
|
Eberhardt B, Poser BA, Shah NJ, Felder J. B1 field map synthesis with generative deep learning used in the design of parallel-transmit RF pulses for ultra-high field MRI. Z Med Phys 2022; 32:334-345. [PMID: 35144850 PMCID: PMC9948838 DOI: 10.1016/j.zemedi.2021.12.003] [Citation(s) in RCA: 7] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2021] [Revised: 11/27/2021] [Accepted: 12/23/2021] [Indexed: 10/19/2022]
Abstract
Spoke trajectory parallel transmit (pTX) excitation in ultra-high field MRI enables B1+ inhomogeneities arising from the shortened RF wavelength in biological tissue to be mitigated. To this end, current RF excitation pulse design algorithms either employ the acquisition of field maps with subsequent non-linear optimization or a universal approach applying robust pre-computed pulses. We suggest and evaluate an intermediate method that uses a subset of acquired field maps combined with generative machine learning models to reduce the pulse calibration time while offering more tailored excitation than robust pulses (RP). The possibility of employing image-to-image translation and semantic image synthesis machine learning models based on generative adversarial networks (GANs) to deduce the missing field maps is examined. Additionally, an RF pulse design that employs a predictive machine learning model to find solutions for the non-linear (two-spokes) pulse design problem is investigated. As a proof of concept, we present simulation results obtained with the suggested machine learning approaches that were trained on a limited data-set, acquired in vivo. The achieved excitation homogeneity based on a subset of half of the B1+ maps acquired in the calibration scans and half of the B1+ maps synthesized with GANs is comparable with state of the art pulse design methods when using the full set of calibration data while halving the total calibration time. By employing RP dictionaries or machine-learning RF pulse predictions, the total calibration time can be reduced significantly as these methods take only seconds or milliseconds per slice, respectively.
Collapse
Affiliation(s)
- Boris Eberhardt
- Institute of Neuroscience and Medicine 4, Forschungszentrum Jülich, Jüich, Germany; RWTH Aachen University, Aachen, Germany.
| | - Benedikt A. Poser
- Department of Cognitive Neuroscience, Faculty of Psychology and Neuroscience, Maastricht University, Maastricht, Netherlands
| | - N. Jon Shah
- Institute of Neuroscience and Medicine 4, Forschungszentrum Jülich, Jüich, Germany,Institute of Neuroscience and Medicine 11, Forschungszentrum Jülich, Jülich, Germany,Department of Neurology, RWTH Aachen University, Aachen, Germany,JARA-BRAIN, Translational Medicine, Aachen, Germany
| | - Jörg Felder
- Institute of Neuroscience and Medicine 4, Forschungszentrum Jülich, Jüich, Germany; RWTH Aachen University, Aachen, Germany.
| |
Collapse
|
22
|
Lei K, Syed AB, Zhu X, Pauly JM, Vasanawala SS. Artifact- and content-specific quality assessment for MRI with image rulers. Med Image Anal 2022; 77:102344. [PMID: 35091278 PMCID: PMC8901552 DOI: 10.1016/j.media.2021.102344] [Citation(s) in RCA: 12] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/25/2021] [Revised: 12/22/2021] [Accepted: 12/27/2021] [Indexed: 11/27/2022]
Abstract
In clinical practice MR images are often first seen by radiologists long after the scan. If image quality is inadequate either patients have to return for an additional scan, or a suboptimal interpretation is rendered. An automatic image quality assessment (IQA) would enable real-time remediation. Existing IQA works for MRI give only a general quality score, agnostic to the cause of and solution to low-quality scans. Furthermore, radiologists' image quality requirements vary with the scan type and diagnostic task. Therefore, the same score may have different implications for different scans. We propose a framework with multi-task CNN model trained with calibrated labels and inferenced with image rulers. Labels calibrated by human inputs follow a well-defined and efficient labeling task. Image rulers address varying quality standards and provide a concrete way of interpreting raw scores from the CNN. The model supports assessments of two of the most common artifacts in MRI: noise and motion. It achieves accuracies of around 90%, 6% better than the best previous method examined, and 3% better than human experts on noise assessment. Our experiments show that label calibration, image rulers, and multi-task training improve the model's performance and generalizability.
Collapse
|
23
|
Zeng G, Guo Y, Zhan J, Wang Z, Lai Z, Du X, Qu X, Guo D. A review on deep learning MRI reconstruction without fully sampled k-space. BMC Med Imaging 2021; 21:195. [PMID: 34952572 PMCID: PMC8710001 DOI: 10.1186/s12880-021-00727-9] [Citation(s) in RCA: 33] [Impact Index Per Article: 8.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2021] [Accepted: 12/08/2021] [Indexed: 11/10/2022] Open
Abstract
BACKGROUND Magnetic resonance imaging (MRI) is an effective auxiliary diagnostic method in clinical medicine, but it has always suffered from the problem of long acquisition time. Compressed sensing and parallel imaging are two common techniques to accelerate MRI reconstruction. Recently, deep learning provides a new direction for MRI, while most of them require a large number of data pairs for training. However, there are many scenarios where fully sampled k-space data cannot be obtained, which will seriously hinder the application of supervised learning. Therefore, deep learning without fully sampled data is indispensable. MAIN TEXT In this review, we first introduce the forward model of MRI as a classic inverse problem, and briefly discuss the connection of traditional iterative methods to deep learning. Next, we will explain how to train reconstruction network without fully sampled data from the perspective of obtaining prior information. CONCLUSION Although the reviewed methods are used for MRI reconstruction, they can also be extended to other areas where ground-truth is not available. Furthermore, we may anticipate that the combination of traditional methods and deep learning will produce better reconstruction results.
Collapse
Affiliation(s)
- Gushan Zeng
- School of Computer and Information Engineering, Fujian Engineering Research Center for Medical Data Mining and Application, Xiamen University of Technology, Xiamen, China
| | - Yi Guo
- School of Computer and Information Engineering, Fujian Engineering Research Center for Medical Data Mining and Application, Xiamen University of Technology, Xiamen, China
| | - Jiaying Zhan
- School of Computer and Information Engineering, Fujian Engineering Research Center for Medical Data Mining and Application, Xiamen University of Technology, Xiamen, China
| | - Zi Wang
- Department of Electronic Science, National Institute for Data Science in Health and Medicine, Xiamen University, Xiamen, China
| | - Zongying Lai
- School of Information Engineering, Jimei University, Xiamen, China
| | - Xiaofeng Du
- School of Computer and Information Engineering, Fujian Engineering Research Center for Medical Data Mining and Application, Xiamen University of Technology, Xiamen, China
| | - Xiaobo Qu
- Department of Electronic Science, National Institute for Data Science in Health and Medicine, Xiamen University, Xiamen, China
| | - Di Guo
- School of Computer and Information Engineering, Fujian Engineering Research Center for Medical Data Mining and Application, Xiamen University of Technology, Xiamen, China.
| |
Collapse
|
24
|
Sandino CM, Cole EK, Alkan C, Chaudhari AS, Loening AM, Hyun D, Dahl J, Imran AAZ, Wang AS, Vasanawala SS. Upstream Machine Learning in Radiology. Radiol Clin North Am 2021; 59:967-985. [PMID: 34689881 PMCID: PMC8549864 DOI: 10.1016/j.rcl.2021.07.009] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022]
Abstract
Machine learning (ML) and Artificial intelligence (AI) has the potential to dramatically improve radiology practice at multiple stages of the imaging pipeline. Most of the attention has been garnered by applications focused on improving the end of the pipeline: image interpretation. However, this article reviews how AI/ML can be applied to improve upstream components of the imaging pipeline, including exam modality selection, hardware design, exam protocol selection, data acquisition, image reconstruction, and image processing. A breadth of applications and their potential for impact is shown across multiple imaging modalities, including ultrasound, computed tomography, and MRI.
Collapse
Affiliation(s)
- Christopher M Sandino
- Department of Electrical Engineering, Stanford University, 350 Serra Mall, Stanford, CA 94305, USA
| | - Elizabeth K Cole
- Department of Electrical Engineering, Stanford University, 350 Serra Mall, Stanford, CA 94305, USA
| | - Cagan Alkan
- Department of Electrical Engineering, Stanford University, 350 Serra Mall, Stanford, CA 94305, USA
| | - Akshay S Chaudhari
- Department of Biomedical Data Science, 1201 Welch Road, Stanford, CA 94305, USA; Department of Radiology, Stanford University, 1201 Welch Road, Stanford, CA 94305, USA
| | - Andreas M Loening
- Department of Radiology, Stanford University, 1201 Welch Road, Stanford, CA 94305, USA
| | - Dongwoon Hyun
- Department of Radiology, Stanford University, 1201 Welch Road, Stanford, CA 94305, USA
| | - Jeremy Dahl
- Department of Radiology, Stanford University, 1201 Welch Road, Stanford, CA 94305, USA
| | | | - Adam S Wang
- Department of Radiology, Stanford University, 1201 Welch Road, Stanford, CA 94305, USA
| | - Shreyas S Vasanawala
- Department of Radiology, Stanford University, 1201 Welch Road, Stanford, CA 94305, USA.
| |
Collapse
|
25
|
Lahiri A, Wang G, Ravishankar S, Fessler JA. Blind Primed Supervised (BLIPS) Learning for MR Image Reconstruction. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:3113-3124. [PMID: 34191725 PMCID: PMC8672324 DOI: 10.1109/tmi.2021.3093770] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
This paper examines a combined supervised-unsupervised framework involving dictionary-based blind learning and deep supervised learning for MR image reconstruction from under-sampled k-space data. A major focus of the work is to investigate the possible synergy of learned features in traditional shallow reconstruction using adaptive sparsity-based priors and deep prior-based reconstruction. Specifically, we propose a framework that uses an unrolled network to refine a blind dictionary learning-based reconstruction. We compare the proposed method with strictly supervised deep learning-based reconstruction approaches on several datasets of varying sizes and anatomies. We also compare the proposed method to alternative approaches for combining dictionary-based methods with supervised learning in MR image reconstruction. The improvements yielded by the proposed framework suggest that the blind dictionary-based approach preserves fine image details that the supervised approach can iteratively refine, suggesting that the features learned using the two methods are complementary.
Collapse
|
26
|
Huang B, Xiao H, Liu W, Zhang Y, Wu H, Wang W, Yang Y, Yang Y, Miller GW, Li T, Cai J. MRI super-resolution via realistic downsampling with adversarial learning. Phys Med Biol 2021; 66. [PMID: 34474407 DOI: 10.1088/1361-6560/ac232e] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/06/2021] [Accepted: 09/02/2021] [Indexed: 11/12/2022]
Abstract
Many deep learning (DL) frameworks have demonstrated state-of-the-art performance in the super-resolution (SR) task of magnetic resonance imaging, but most performances have been achieved with simulated low-resolution (LR) images rather than LR images from real acquisition. Due to the limited generalizability of the SR network, enhancement is not guaranteed for real LR images because of the unreality of the training LR images. In this study, we proposed a DL-based SR framework with an emphasis on data construction to achieve better performance on real LR MR images. The framework comprised two steps: (a) downsampling training using a generative adversarial network (GAN) to construct more realistic and perfectly matched LR/high-resolution (HR) pairs. The downsampling GAN input was real LR and HR images. The generator translated the HR images to LR images and the discriminator distinguished the patch-level difference between the synthetic and real LR images. (b) SR training was performed using an enhance4d deep super-resolution network (EDSR). In the controlled experiments, three EDSRs were trained using our proposed method, Gaussian blur, and k-space zero-filling. As for the data, liver MR images were obtained from 24 patients using breath-hold serial LR and HR scans (only HR images were used in the conventional methods). The k-space zero-filling group delivered almost zero enhancement on the real LR images and the Gaussian group produced a considerable number of artifacts. The proposed method exhibited significantly better resolution enhancement and fewer artifacts compared with the other two networks. Our method outperformed the Gaussian method by an improvement of 0.111 ± 0.016 in the structural similarity index and 2.76 ± 0.98 dB in the peak signal-to-noise ratio. The blind/reference-less image spatial quality evaluator metric of the conventional Gaussian method and proposed method were 46.6 ± 4.2 and 34.1 ± 2.4, respectively.
Collapse
Affiliation(s)
- Bangyan Huang
- Department of Engineering and Applied Physics, University of Science and Technology of China, Hefei, People's Republic of China
| | - Haonan Xiao
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Hong Kong SAR, People's Republic of China
| | - Weiwei Liu
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education/Beijing), Department of Radiation Oncology, Beijing Cancer Hospital and Institute, Peking University Cancer Hospital and Institute, Beijing, People's Republic of China
| | - Yibao Zhang
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education/Beijing), Department of Radiation Oncology, Beijing Cancer Hospital and Institute, Peking University Cancer Hospital and Institute, Beijing, People's Republic of China
| | - Hao Wu
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education/Beijing), Department of Radiation Oncology, Beijing Cancer Hospital and Institute, Peking University Cancer Hospital and Institute, Beijing, People's Republic of China
| | - Weihu Wang
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education/Beijing), Department of Radiation Oncology, Beijing Cancer Hospital and Institute, Peking University Cancer Hospital and Institute, Beijing, People's Republic of China
| | - Yunhuan Yang
- Department of Engineering and Applied Physics, University of Science and Technology of China, Hefei, People's Republic of China
| | - Yidong Yang
- Department of Engineering and Applied Physics, University of Science and Technology of China, Hefei, People's Republic of China
| | - G Wilson Miller
- Department of Radiology and Medical Imaging, The University of Virginia, Charlottesville, VA, United States of America
| | - Tian Li
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Hong Kong SAR, People's Republic of China
| | - Jing Cai
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Hong Kong SAR, People's Republic of China
| |
Collapse
|
27
|
Yaman B, Hosseini SAH, Moeller S, Ellermann J, Uğurbil K, Akçakaya M. Self-supervised learning of physics-guided reconstruction neural networks without fully sampled reference data. Magn Reson Med 2020; 84:3172-3191. [PMID: 32614100 PMCID: PMC7811359 DOI: 10.1002/mrm.28378] [Citation(s) in RCA: 127] [Impact Index Per Article: 25.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/16/2019] [Revised: 05/21/2020] [Accepted: 05/22/2020] [Indexed: 12/25/2022]
Abstract
PURPOSE To develop a strategy for training a physics-guided MRI reconstruction neural network without a database of fully sampled data sets. METHODS Self-supervised learning via data undersampling (SSDU) for physics-guided deep learning reconstruction partitions available measurements into two disjoint sets, one of which is used in the data consistency (DC) units in the unrolled network and the other is used to define the loss for training. The proposed training without fully sampled data is compared with fully supervised training with ground-truth data, as well as conventional compressed-sensing and parallel imaging methods using the publicly available fastMRI knee database. The same physics-guided neural network is used for both proposed SSDU and supervised training. The SSDU training is also applied to prospectively two-fold accelerated high-resolution brain data sets at different acceleration rates, and compared with parallel imaging. RESULTS Results on five different knee sequences at an acceleration rate of 4 shows that the proposed self-supervised approach performs closely with supervised learning, while significantly outperforming conventional compressed-sensing and parallel imaging, as characterized by quantitative metrics and a clinical reader study. The results on prospectively subsampled brain data sets, in which supervised learning cannot be used due to lack of ground-truth reference, show that the proposed self-supervised approach successfully performs reconstruction at high acceleration rates (4, 6, and 8). Image readings indicate improved visual reconstruction quality with the proposed approach compared with parallel imaging at acquisition acceleration. CONCLUSION The proposed SSDU approach allows training of physics-guided deep learning MRI reconstruction without fully sampled data, while achieving comparable results with supervised deep learning MRI trained on fully sampled data.
Collapse
Affiliation(s)
- Burhaneddin Yaman
- Department of Electrical and Computer Engineering, University of Minnesota, Minneapolis, MN
- Center for Magnetic Resonance Research, University of Minnesota, Minneapolis, MN
| | - Seyed Amir Hossein Hosseini
- Department of Electrical and Computer Engineering, University of Minnesota, Minneapolis, MN
- Center for Magnetic Resonance Research, University of Minnesota, Minneapolis, MN
| | - Steen Moeller
- Center for Magnetic Resonance Research, University of Minnesota, Minneapolis, MN
| | - Jutta Ellermann
- Center for Magnetic Resonance Research, University of Minnesota, Minneapolis, MN
| | - Kâmil Uğurbil
- Center for Magnetic Resonance Research, University of Minnesota, Minneapolis, MN
| | - Mehmet Akçakaya
- Department of Electrical and Computer Engineering, University of Minnesota, Minneapolis, MN
- Center for Magnetic Resonance Research, University of Minnesota, Minneapolis, MN
| |
Collapse
|