1
|
Kumar PA, Gunasundari R. A lightweight adaptive spatial channel attention efficient net B3 based generative adversarial network approach for MR image reconstruction from under sampled data. Magn Reson Imaging 2025; 117:110281. [PMID: 39672285 DOI: 10.1016/j.mri.2024.110281] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/21/2024] [Revised: 11/15/2024] [Accepted: 11/23/2024] [Indexed: 12/15/2024]
Abstract
Magnetic Resonance Imaging (MRI) stands out as a notable non-invasive method for medical imaging assessments, widely employed in early medical diagnoses due to its exceptional resolution in portraying soft tissue structures. However, the MRI method faces challenges with its inherently slow acquisition process, stemming from the sequential sampling in k-space and limitations in traversal speed due to physiological and hardware constraints. Compressed Sensing in MRI (CS-MRI) accelerates image acquisition by utilizing greatly under-sampled k-space information. Despite its advantages, conventional CS-MRI encounters issues such as sluggish iterations and artefacts at higher acceleration factors. Recent advancements integrate deep learning models into CS-MRI, inspired by successes in various computer vision domains. It has drawn significant attention from the MRI community because of its great potential for image reconstruction from undersampled k-space data in fast MRI. This paper proposes a lightweight Adaptive Spatial-Channel Attention EfficientNet B3-based Generative Adversarial Network (ASCA-EffNet GAN) for fast, high-quality MR image reconstruction from greatly under-sampled k-space information in CS-MRI. The proposed GAN employs a U-net generator with ASCA-based EfficientNet B3 for encoder blocks and a ResNet decoder. The discriminator is a binary classifier with ASCA-based EfficientNet B3, a fully connected layer and a sigmoid layer. The EfficientNet B3 utilizes a compound scaling strategy that achieves a balance amongst model depth, width, and resolution, resulting in optimal performance with a reduced number of parameters. Furthermore, the adaptive attention mechanisms in the proposed ASCA-EffNet GAN effectively capture spatial and channel-wise features, contributing to detailed anatomical structure reconstruction. Experimental evaluations on the dataset demonstrate ASCA-EffNet GAN's superior performance across various metrics, surpassing conventional reconstruction methods. Hence, ASCA-EffNet GAN showcases remarkable reconstruction capabilities even under high under-sampling rates, making it suitable for clinical applications.
Collapse
Affiliation(s)
- Penta Anil Kumar
- Dept. of Electronics and Communication Engineering, Puducherry Technological University, Puducherry 605014, India.
| | - Ramalingam Gunasundari
- Dept. of Electronics and Communication Engineering, Puducherry Technological University, Puducherry 605014, India
| |
Collapse
|
2
|
Wang S, Wu R, Jia S, Diakite A, Li C, Liu Q, Zheng H, Ying L. Knowledge-driven deep learning for fast MR imaging: Undersampled MR image reconstruction from supervised to un-supervised learning. Magn Reson Med 2024; 92:496-518. [PMID: 38624162 DOI: 10.1002/mrm.30105] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/03/2023] [Revised: 03/19/2024] [Accepted: 03/20/2024] [Indexed: 04/17/2024]
Abstract
Deep learning (DL) has emerged as a leading approach in accelerating MRI. It employs deep neural networks to extract knowledge from available datasets and then applies the trained networks to reconstruct accurate images from limited measurements. Unlike natural image restoration problems, MRI involves physics-based imaging processes, unique data properties, and diverse imaging tasks. This domain knowledge needs to be integrated with data-driven approaches. Our review will introduce the significant challenges faced by such knowledge-driven DL approaches in the context of fast MRI along with several notable solutions, which include learning neural networks and addressing different imaging application scenarios. The traits and trends of these techniques have also been given which have shifted from supervised learning to semi-supervised learning, and finally, to unsupervised learning methods. In addition, MR vendors' choices of DL reconstruction have been provided along with some discussions on open questions and future directions, which are critical for the reliable imaging systems.
Collapse
Affiliation(s)
- Shanshan Wang
- Paul C Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Ruoyou Wu
- Paul C Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Sen Jia
- Paul C Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Alou Diakite
- Paul C Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
- University of Chinese Academy of Sciences, Beijing, China
| | - Cheng Li
- Paul C Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Qiegen Liu
- Department of Electronic Information Engineering, Nanchang University, Nanchang, China
| | - Hairong Zheng
- Paul C Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Leslie Ying
- Department of Biomedical Engineering and Department of Electrical Engineering, The State University of New York, Buffalo, New York, USA
| |
Collapse
|
3
|
Yuan T, Yang J, Chi J, Yu T, Liu F. A cross-domain complex convolution neural network for undersampled magnetic resonance image reconstruction. Magn Reson Imaging 2024; 108:86-97. [PMID: 38331053 DOI: 10.1016/j.mri.2024.02.004] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/06/2023] [Revised: 02/01/2024] [Accepted: 02/05/2024] [Indexed: 02/10/2024]
Abstract
To introduce a new cross-domain complex convolution neural network for accurate MR image reconstruction from undersampled k-space data. Most reconstruction methods utilize neural networks or cascade neural networks in either the image domain and/or the k-space domain. However, these methods encounter several challenges: 1) Applying neural networks directly in the k-space domain is suboptimal for feature extraction; 2) Classic image-domain networks have difficulty in fully extracting texture features; and 3) Existing cross-domain methods still face challenges in extracting and fusing features from both image and k-space domains simultaneously. In this work, we propose a novel deep-learning-based 2-D single-coil complex-valued MR reconstruction network termed TEID-Net. TEID-Net integrates three modules: 1) TE-Net, an image-domain-based sub-network designed to enhance contrast in input features by incorporating a Texture Enhancement Module; 2) ID-Net, an intermediate-domain sub-network tailored to operate in the image-Fourier space, with the specific goal of reducing aliasing artifacts realized by leveraging the superior incoherence property of the decoupled one-dimensional signals; and 3) TEID-Net, a cross-domain reconstruction network in which ID-Nets and TE-Nets are combined and cascaded to boost the quality of image reconstruction further. Extensive experiments have been conducted on the fastMRI and Calgary-Campinas datasets. Results demonstrate the effectiveness of the proposed TEID-Net in mitigating undersampling-induced artifacts and producing high-quality image reconstructions, outperforming several state-of-the-art methods while utilizing fewer network parameters. The cross-domain TEID-Net excels in restoring tissue structures and intricate texture details. The results illustrate that TEID-Net is particularly well-suited for regular Cartesian undersampling scenarios.
Collapse
Affiliation(s)
- Tengfei Yuan
- College of Electronics and Information, Qingdao University, Qingdao, Shandong, China
| | - Jie Yang
- College of Mechanical and Electrical Engineering, Qingdao University, Qingdao, Shandong, China
| | - Jieru Chi
- College of Electronics and Information, Qingdao University, Qingdao, Shandong, China.
| | - Teng Yu
- College of Electronics and Information, Qingdao University, Qingdao, Shandong, China
| | - Feng Liu
- School of Electrical Engineering and Computer Science, University of Queensland, Brisbane, Brisbane, Australia
| |
Collapse
|
4
|
Papanastasiou G, Dikaios N, Huang J, Wang C, Yang G. Is Attention all You Need in Medical Image Analysis? A Review. IEEE J Biomed Health Inform 2024; 28:1398-1411. [PMID: 38157463 DOI: 10.1109/jbhi.2023.3348436] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/03/2024]
Abstract
Medical imaging is a key component in clinical diagnosis, treatment planning and clinical trial design, accounting for almost 90% of all healthcare data. CNNs achieved performance gains in medical image analysis (MIA) over the last years. CNNs can efficiently model local pixel interactions and be trained on small-scale MI data. Despite their important advances, typical CNN have relatively limited capabilities in modelling "global" pixel interactions, which restricts their generalisation ability to understand out-of-distribution data with different "global" information. The recent progress of Artificial Intelligence gave rise to Transformers, which can learn global relationships from data. However, full Transformer models need to be trained on large-scale data and involve tremendous computational complexity. Attention and Transformer compartments ("Transf/Attention") which can well maintain properties for modelling global relationships, have been proposed as lighter alternatives of full Transformers. Recently, there is an increasing trend to co-pollinate complementary local-global properties from CNN and Transf/Attention architectures, which led to a new era of hybrid models. The past years have witnessed substantial growth in hybrid CNN-Transf/Attention models across diverse MIA problems. In this systematic review, we survey existing hybrid CNN-Transf/Attention models, review and unravel key architectural designs, analyse breakthroughs, and evaluate current and future opportunities as well as challenges. We also introduced an analysis framework on generalisation opportunities of scientific and clinical impact, based on which new data-driven domain generalisation and adaptation methods can be stimulated.
Collapse
|
5
|
Shao L, Chen B, Zhang Z, Zhang Z, Chen X. Artificial intelligence generated content (AIGC) in medicine: A narrative review. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2024; 21:1672-1711. [PMID: 38303483 DOI: 10.3934/mbe.2024073] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/03/2024]
Abstract
Recently, artificial intelligence generated content (AIGC) has been receiving increased attention and is growing exponentially. AIGC is generated based on the intentional information extracted from human-provided instructions by generative artificial intelligence (AI) models. AIGC quickly and automatically generates large amounts of high-quality content. Currently, there is a shortage of medical resources and complex medical procedures in medicine. Due to its characteristics, AIGC can help alleviate these problems. As a result, the application of AIGC in medicine has gained increased attention in recent years. Therefore, this paper provides a comprehensive review on the recent state of studies involving AIGC in medicine. First, we present an overview of AIGC. Furthermore, based on recent studies, the application of AIGC in medicine is reviewed from two aspects: medical image processing and medical text generation. The basic generative AI models, tasks, target organs, datasets and contribution of studies are considered and summarized. Finally, we also discuss the limitations and challenges faced by AIGC and propose possible solutions with relevant studies. We hope this review can help readers understand the potential of AIGC in medicine and obtain some innovative ideas in this field.
Collapse
Affiliation(s)
- Liangjing Shao
- Academy for Engineering & Technology, Fudan University, Shanghai 200433, China
- Shanghai Key Laboratory of Medical Image Computing and Computer Assisted Intervention, Fudan University, Shanghai 200032, China
| | - Benshuang Chen
- Academy for Engineering & Technology, Fudan University, Shanghai 200433, China
- Shanghai Key Laboratory of Medical Image Computing and Computer Assisted Intervention, Fudan University, Shanghai 200032, China
| | - Ziqun Zhang
- Information office, Fudan University, Shanghai 200032, China
| | - Zhen Zhang
- Baoshan Branch of Ren Ji Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai 200444, China
| | - Xinrong Chen
- Academy for Engineering & Technology, Fudan University, Shanghai 200433, China
- Shanghai Key Laboratory of Medical Image Computing and Computer Assisted Intervention, Fudan University, Shanghai 200032, China
| |
Collapse
|
6
|
Ekanayake M, Pawar K, Harandi M, Egan G, Chen Z. McSTRA: A multi-branch cascaded swin transformer for point spread function-guided robust MRI reconstruction. Comput Biol Med 2024; 168:107775. [PMID: 38061154 DOI: 10.1016/j.compbiomed.2023.107775] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2023] [Revised: 11/23/2023] [Accepted: 11/27/2023] [Indexed: 01/10/2024]
Abstract
Deep learning MRI reconstruction methods are often based on Convolutional neural network (CNN) models; however, they are limited in capturing global correlations among image features due to the intrinsic locality of the convolution operation. Conversely, the recent vision transformer models (ViT) are capable of capturing global correlations by applying self-attention operations on image patches. Nevertheless, the existing transformer models for MRI reconstruction rarely leverage the physics of MRI. In this paper, we propose a novel physics-based transformer model titled, the Multi-branch Cascaded Swin Transformers (McSTRA) for robust MRI reconstruction. McSTRA combines several interconnected MRI physics-related concepts with the Swin transformers: it exploits global MRI features via the shifted window self-attention mechanism; it extracts MRI features belonging to different spectral components via a multi-branch setup; it iterates between intermediate de-aliasing and data consistency via a cascaded network with intermediate loss computations; furthermore, we propose a point spread function-guided positional embedding generation mechanism for the Swin transformers which exploit the spread of the aliasing artifacts for effective reconstruction. With the combination of all these components, McSTRA outperforms the state-of-the-art methods while demonstrating robustness in adversarial conditions such as higher accelerations, noisy data, different undersampling protocols, out-of-distribution data, and abnormalities in anatomy.
Collapse
Affiliation(s)
- Mevan Ekanayake
- Monash Biomedical Imaging, Monash University, Australia; Department of Electrical and Computer Systems Engineering, Monash University, Australia.
| | - Kamlesh Pawar
- Monash Biomedical Imaging, Monash University, Australia
| | - Mehrtash Harandi
- Department of Electrical and Computer Systems Engineering, Monash University, Australia
| | - Gary Egan
- Monash Biomedical Imaging, Monash University, Australia; School of Psychological Sciences, Monash University, Australia
| | - Zhaolin Chen
- Monash Biomedical Imaging, Monash University, Australia; Department of Data Science and AI, Monash University, Australia
| |
Collapse
|
7
|
蔡 昕, 侯 学, 杨 光, 聂 生. [Application of generative adversarial network in magnetic resonance image reconstruction]. SHENG WU YI XUE GONG CHENG XUE ZA ZHI = JOURNAL OF BIOMEDICAL ENGINEERING = SHENGWU YIXUE GONGCHENGXUE ZAZHI 2023; 40:582-588. [PMID: 37380400 PMCID: PMC10307593 DOI: 10.7507/1001-5515.202204007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Subscribe] [Scholar Register] [Received: 04/02/2022] [Revised: 04/06/2023] [Indexed: 06/30/2023]
Abstract
Magnetic resonance imaging (MRI) is an important medical imaging method, whose major limitation is its long scan time due to the imaging mechanism, increasing patients' cost and waiting time for the examination. Currently, parallel imaging (PI) and compress sensing (CS) together with other reconstruction technologies have been proposed to accelerate image acquisition. However, the image quality of PI and CS depends on the image reconstruction algorithms, which is far from satisfying in respect to both the image quality and the reconstruction speed. In recent years, image reconstruction based on generative adversarial network (GAN) has become a research hotspot in the field of magnetic resonance imaging because of its excellent performance. In this review, we summarized the recent development of application of GAN in MRI reconstruction in both single- and multi-modality acceleration, hoping to provide a useful reference for interested researchers. In addition, we analyzed the characteristics and limitations of existing technologies and forecasted some development trends in this field.
Collapse
Affiliation(s)
- 昕 蔡
- 上海理工大学 健康科学与工程学院(上海 200093)School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai 200093, P. R. China
| | - 学文 侯
- 上海理工大学 健康科学与工程学院(上海 200093)School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai 200093, P. R. China
| | - 光 杨
- 上海理工大学 健康科学与工程学院(上海 200093)School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai 200093, P. R. China
| | - 生东 聂
- 上海理工大学 健康科学与工程学院(上海 200093)School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai 200093, P. R. China
| |
Collapse
|
8
|
Eisenmenger LB, Peret A, Roberts GS, Spahic A, Tang C, Kuner AD, Grayev AM, Field AS, Rowley HA, Kennedy TA. Focused Abbreviated Survey MRI Protocols for Brain and Spine Imaging. Radiographics 2023; 43:e220147. [PMID: 37167089 PMCID: PMC10262597 DOI: 10.1148/rg.220147] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/14/2022] [Revised: 10/12/2022] [Accepted: 10/18/2022] [Indexed: 05/13/2023]
Abstract
There has been extensive growth in both the technical development and the clinical applications of MRI, establishing this modality as one of the most powerful diagnostic imaging tools. However, long examination and image interpretation times still limit the application of MRI, especially in emergent clinical settings. Rapid and abbreviated MRI protocols have been developed as alternatives to standard MRI, with reduced imaging times, and in some cases limited numbers of sequences, to more efficiently answer specific clinical questions. A group of rapid MRI protocols used at the authors' institution, referred to as FAST (focused abbreviated survey techniques), are designed to include or exclude emergent or urgent conditions or screen for specific entities. These FAST protocols provide adequate diagnostic image quality with use of accelerated approaches to produce imaging studies faster than traditional methods. FAST protocols have become critical diagnostic screening tools at the authors' institution, allowing confident and efficient confirmation or exclusion of actionable findings. The techniques commonly used to reduce imaging times, the imaging protocols used at the authors' institution, and future directions in FAST imaging are reviewed to provide a practical and comprehensive overview of FAST MRI for practicing neuroradiologists. ©RSNA, 2023 Quiz questions for this article are available in the supplemental material.
Collapse
Affiliation(s)
| | | | - Grant S. Roberts
- From the Departments of Radiology (L.B.E., A.P., A.D.K., A.M.G.,
A.S.F., H.A.R., T.A.K.) and Medical Physics (G.S.R., A.S., C.T.), University of
Wisconsin School of Medicine and Public Health, 600 Highland Ave, Madison, WI
53792-3252
| | - Alma Spahic
- From the Departments of Radiology (L.B.E., A.P., A.D.K., A.M.G.,
A.S.F., H.A.R., T.A.K.) and Medical Physics (G.S.R., A.S., C.T.), University of
Wisconsin School of Medicine and Public Health, 600 Highland Ave, Madison, WI
53792-3252
| | - Chenwei Tang
- From the Departments of Radiology (L.B.E., A.P., A.D.K., A.M.G.,
A.S.F., H.A.R., T.A.K.) and Medical Physics (G.S.R., A.S., C.T.), University of
Wisconsin School of Medicine and Public Health, 600 Highland Ave, Madison, WI
53792-3252
| | - Anthony D. Kuner
- From the Departments of Radiology (L.B.E., A.P., A.D.K., A.M.G.,
A.S.F., H.A.R., T.A.K.) and Medical Physics (G.S.R., A.S., C.T.), University of
Wisconsin School of Medicine and Public Health, 600 Highland Ave, Madison, WI
53792-3252
| | - Allison M. Grayev
- From the Departments of Radiology (L.B.E., A.P., A.D.K., A.M.G.,
A.S.F., H.A.R., T.A.K.) and Medical Physics (G.S.R., A.S., C.T.), University of
Wisconsin School of Medicine and Public Health, 600 Highland Ave, Madison, WI
53792-3252
| | - Aaron S. Field
- From the Departments of Radiology (L.B.E., A.P., A.D.K., A.M.G.,
A.S.F., H.A.R., T.A.K.) and Medical Physics (G.S.R., A.S., C.T.), University of
Wisconsin School of Medicine and Public Health, 600 Highland Ave, Madison, WI
53792-3252
| | - Howard A. Rowley
- From the Departments of Radiology (L.B.E., A.P., A.D.K., A.M.G.,
A.S.F., H.A.R., T.A.K.) and Medical Physics (G.S.R., A.S., C.T.), University of
Wisconsin School of Medicine and Public Health, 600 Highland Ave, Madison, WI
53792-3252
| | - Tabassum A. Kennedy
- From the Departments of Radiology (L.B.E., A.P., A.D.K., A.M.G.,
A.S.F., H.A.R., T.A.K.) and Medical Physics (G.S.R., A.S., C.T.), University of
Wisconsin School of Medicine and Public Health, 600 Highland Ave, Madison, WI
53792-3252
| |
Collapse
|
9
|
Cai X, Hou X, Sun R, Chang X, Zhu H, Jia S, Nie S. Accelerating image reconstruction for multi-contrast MRI based on Y-Net3. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2023:XST230012. [PMID: 37248943 DOI: 10.3233/xst-230012] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/31/2023]
Abstract
BACKGROUND As one of the significant preoperative imaging modalities in medical diagnosis, Magnetic resonance imaging (MRI) takes a long scanning time due to its special imaging principle. OBJECTIVE We propose an innovative MRI reconstruction strategy and data consistency method based on deep learning to reconstruct high-quality brain MRIs from down-sampled data and accelerate the MR imaging process. METHODS Sixteen healthy subjects undergoing T1-weighted spin-echo (SE) and T2-weighted fast spin-echo (FSE) sequences by a 1.5T MRI scanner were recruited. A Y-Net3+ network was used to facilitate the high-quality MRI reconstruction through context information. In addition, the existing data consistency fidelity method was improved. The difference between the reconstructed K-space and the original K-space was shorten by the linear regression algorithm. Therefore, the redundant artifacts derived from under-sampling were avoided. The Structural Similarity (SSIM) and Peak Signal to Noise Ratio (PSNR) were applied to quantitatively evaluate image reconstruction performance of different down-sampling patterns. RESULTS Compared with the classical Y-Net, Y-Net3+ network improved SSIM and PSNR of MRI images from 0.9164±0.0178 and 33.2216±3.2919 to 0.9387±0.0363 and 35.1785±3.3105, respectively, under compressed sensing reconstruction with acceleration factor of 4. The improved network increases signal-to-noise ratio and adds more image texture information in the reconstructed images. Furthermore, in the process of data consistency, linear regression analysis was used to reduce the difference between the reconstructed K-space and the original K-space, so that the SSIM and PSNR were increased to 0.9808±0.0081 and 40.9254±1.1911, respectively. CONCLUSIONS The improved Y-Net combined with data consistency fidelity method elucidates its potential in reconstructing high-quality T2-weighted images from the down-sampled data by fully exploring the T1-weighted information. With the advantage of avoiding down-sampled artifacts, the improved network exhibits remarkable clinical promise for fast MRI applications.
Collapse
Affiliation(s)
- Xin Cai
- School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai, China
| | - Xuewen Hou
- Shanghai Kangda COLORFUL Healthcare Co., Ltd, Shanghai, China
| | - Rong Sun
- School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai, China
| | - Xiao Chang
- School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai, China
| | - Honglin Zhu
- School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai, China
| | - Shouqiang Jia
- Department of Imaging, Jinan People's Hospital affiliated to Shandong First Medical University, Shandong, China
| | - Shengdong Nie
- School of Health Science and Engineering, University of Shanghai for Science and Technology, Shanghai, China
| |
Collapse
|
10
|
Hossain MB, Kwon KC, Shinde RK, Imtiaz SM, Kim N. A Hybrid Residual Attention Convolutional Neural Network for Compressed Sensing Magnetic Resonance Image Reconstruction. Diagnostics (Basel) 2023; 13:diagnostics13071306. [PMID: 37046524 PMCID: PMC10093476 DOI: 10.3390/diagnostics13071306] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/20/2023] [Revised: 03/20/2023] [Accepted: 03/29/2023] [Indexed: 04/03/2023] Open
Abstract
We propose a dual-domain deep learning technique for accelerating compressed sensing magnetic resonance image reconstruction. An advanced convolutional neural network with residual connectivity and an attention mechanism was developed for frequency and image domains. First, the sensor domain subnetwork estimates the unmeasured frequencies of k-space to reduce aliasing artifacts. Second, the image domain subnetwork performs a pixel-wise operation to remove blur and noisy artifacts. The skip connections efficiently concatenate the feature maps to alleviate the vanishing gradient problem. An attention gate in each decoder layer enhances network generalizability and speeds up image reconstruction by eliminating irrelevant activations. The proposed technique reconstructs real-valued clinical images from sparsely sampled k-spaces that are identical to the reference images. The performance of this novel approach was compared with state-of-the-art direct mapping, single-domain, and multi-domain methods. With acceleration factors (AFs) of 4 and 5, our method improved the mean peak signal-to-noise ratio (PSNR) to 8.67 and 9.23, respectively, compared with the single-domain Unet model; similarly, our approach increased the average PSNR to 3.72 and 4.61, respectively, compared with the multi-domain W-net. Remarkably, using an AF of 6, it enhanced the PSNR by 9.87 ± 1.55 and 6.60 ± 0.38 compared with Unet and W-net, respectively.
Collapse
|
11
|
Satrya GB, Ramatryana INA, Shin SY. Compressive Sensing of Medical Images Based on HSV Color Space. SENSORS (BASEL, SWITZERLAND) 2023; 23:2616. [PMID: 36904821 PMCID: PMC10006955 DOI: 10.3390/s23052616] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 12/06/2022] [Revised: 02/06/2023] [Accepted: 02/21/2023] [Indexed: 06/18/2023]
Abstract
Recently, compressive sensing (CS) schemes have been studied as a new compression modality that exploits the sensing matrix in the measurement scheme and the reconstruction scheme to recover the compressed signal. In addition, CS is exploited in medical imaging (MI) to support efficient sampling, compression, transmission, and storage of a large amount of MI. Although CS of MI has been extensively investigated, the effect of color space in CS of MI has not yet been studied in the literature. To fulfill these requirements, this article proposes a novel CS of MI based on hue-saturation value (HSV), using spread spectrum Fourier sampling (SSFS) and sparsity averaging with reweighted analysis (SARA). An HSV loop that performs SSFS is proposed to obtain a compressed signal. Next, HSV-SARA is proposed to reconstruct MI from the compressed signal. A set of color MIs is investigated, such as colonoscopy, magnetic resonance imaging of the brain and eye, and wireless capsule endoscopy images. Experiments were performed to show the superiority of HSV-SARA over benchmark methods in terms of signal-to-noise ratio (SNR), structural similarity (SSIM) index, and measurement rate (MR). The experiments showed that a color MI, with a resolution of 256×256 pixels, could be compressed by the proposed CS at MR of 0.1, and could be improved in terms of SNR being 15.17% and SSIM being 2.53%. The proposed HSV-SARA can be a solution for color medical image compression and sampling to improve the image acquisition of medical devices.
Collapse
Affiliation(s)
| | - I Nyoman Apraz Ramatryana
- Department of IT Convergence Engineering, Kumoh National Institute of Technology, Gumi 39177, Republic of Korea
| | - Soo Young Shin
- Department of IT Convergence Engineering, Kumoh National Institute of Technology, Gumi 39177, Republic of Korea
| |
Collapse
|
12
|
Hossain MB, Kwon KC, Imtiaz SM, Nam OS, Jeon SH, Kim N. De-Aliasing and Accelerated Sparse Magnetic Resonance Image Reconstruction Using Fully Dense CNN with Attention Gates. Bioengineering (Basel) 2022; 10:22. [PMID: 36671594 PMCID: PMC9854709 DOI: 10.3390/bioengineering10010022] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/25/2022] [Revised: 12/19/2022] [Accepted: 12/19/2022] [Indexed: 12/24/2022] Open
Abstract
When sparsely sampled data are used to accelerate magnetic resonance imaging (MRI), conventional reconstruction approaches produce significant artifacts that obscure the content of the image. To remove aliasing artifacts, we propose an advanced convolutional neural network (CNN) called fully dense attention CNN (FDA-CNN). We updated the Unet model with the fully dense connectivity and attention mechanism for MRI reconstruction. The main benefit of FDA-CNN is that an attention gate in each decoder layer increases the learning process by focusing on the relevant image features and provides a better generalization of the network by reducing irrelevant activations. Moreover, densely interconnected convolutional layers reuse the feature maps and prevent the vanishing gradient problem. Additionally, we also implement a new, proficient under-sampling pattern in the phase direction that takes low and high frequencies from the k-space both randomly and non-randomly. The performance of FDA-CNN was evaluated quantitatively and qualitatively with three different sub-sampling masks and datasets. Compared with five current deep learning-based and two compressed sensing MRI reconstruction techniques, the proposed method performed better as it reconstructed smoother and brighter images. Furthermore, FDA-CNN improved the mean PSNR by 2 dB, SSIM by 0.35, and VIFP by 0.37 compared with Unet for the acceleration factor of 5.
Collapse
Affiliation(s)
- Md. Biddut Hossain
- School of Information and Communication Engineering, Chungbuk National University, Cheongju-si 28644, Chungcheongbuk-do, Republic of Korea
| | - Ki-Chul Kwon
- School of Information and Communication Engineering, Chungbuk National University, Cheongju-si 28644, Chungcheongbuk-do, Republic of Korea
| | - Shariar Md Imtiaz
- School of Information and Communication Engineering, Chungbuk National University, Cheongju-si 28644, Chungcheongbuk-do, Republic of Korea
| | - Oh-Seung Nam
- School of Information and Communication Engineering, Chungbuk National University, Cheongju-si 28644, Chungcheongbuk-do, Republic of Korea
| | - Seok-Hee Jeon
- Department of Electronics Engineering, Incheon National University, 119 Academy-ro, Yeonsu-gu, Incheon 22012, Gyeonggi-do, Republic of Korea
| | - Nam Kim
- School of Information and Communication Engineering, Chungbuk National University, Cheongju-si 28644, Chungcheongbuk-do, Republic of Korea
| |
Collapse
|
13
|
Dalmaz O, Yurt M, Cukur T. ResViT: Residual Vision Transformers for Multimodal Medical Image Synthesis. IEEE TRANSACTIONS ON MEDICAL IMAGING 2022; 41:2598-2614. [PMID: 35436184 DOI: 10.1109/tmi.2022.3167808] [Citation(s) in RCA: 122] [Impact Index Per Article: 40.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Generative adversarial models with convolutional neural network (CNN) backbones have recently been established as state-of-the-art in numerous medical image synthesis tasks. However, CNNs are designed to perform local processing with compact filters, and this inductive bias compromises learning of contextual features. Here, we propose a novel generative adversarial approach for medical image synthesis, ResViT, that leverages the contextual sensitivity of vision transformers along with the precision of convolution operators and realism of adversarial learning. ResViT's generator employs a central bottleneck comprising novel aggregated residual transformer (ART) blocks that synergistically combine residual convolutional and transformer modules. Residual connections in ART blocks promote diversity in captured representations, while a channel compression module distills task-relevant information. A weight sharing strategy is introduced among ART blocks to mitigate computational burden. A unified implementation is introduced to avoid the need to rebuild separate synthesis models for varying source-target modality configurations. Comprehensive demonstrations are performed for synthesizing missing sequences in multi-contrast MRI, and CT images from MRI. Our results indicate superiority of ResViT against competing CNN- and transformer-based methods in terms of qualitative observations and quantitative metrics.
Collapse
|
14
|
Huang J, Wu Y, Wu H, Yang G. Fast MRI Reconstruction: How Powerful Transformers Are? ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2022; 2022:2066-2070. [PMID: 36085682 DOI: 10.1109/embc48229.2022.9871475] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Magnetic resonance imaging (MRI) is a widely used non-radiative and non-invasive method for clinical interro-gation of organ structures and metabolism, with an inherently long scanning time. Methods by k-space undersampling and deep learning based reconstruction have been popularised to accelerate the scanning process. This work focuses on investigating how powerful transformers are for fast MRI by exploiting and comparing different novel network architectures. In particular, a generative adversarial network (GAN) based Swin transformer (ST-GAN) was introduced for the fast MRI reconstruction. To further preserve the edge and texture information, edge enhanced GAN based Swin transformer (EES-GAN) and texture enhanced GAN based Swin transformer (TES-GAN) were also developed, where a dual-discriminator GAN structure was applied. We compared our proposed GAN based transformers, standalone Swin transformer and other convolutional neural networks based GAN model in terms of the evaluation metrics PSNR, SSIM and FID. We showed that transformers work well for the MRI reconstruction from different undersampling conditions. The utilisation of GAN's adversarial structure improves the quality of images reconstructed when undersampled for 30% or higher. The code is publicly available at https://github.comJayanglab/SwinGANMR.
Collapse
|
15
|
Korkmaz Y, Dar SUH, Yurt M, Ozbey M, Cukur T. Unsupervised MRI Reconstruction via Zero-Shot Learned Adversarial Transformers. IEEE TRANSACTIONS ON MEDICAL IMAGING 2022; 41:1747-1763. [PMID: 35085076 DOI: 10.1109/tmi.2022.3147426] [Citation(s) in RCA: 88] [Impact Index Per Article: 29.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Supervised reconstruction models are characteristically trained on matched pairs of undersampled and fully-sampled data to capture an MRI prior, along with supervision regarding the imaging operator to enforce data consistency. To reduce supervision requirements, the recent deep image prior framework instead conjoins untrained MRI priors with the imaging operator during inference. Yet, canonical convolutional architectures are suboptimal in capturing long-range relationships, and priors based on randomly initialized networks may yield suboptimal performance. To address these limitations, here we introduce a novel unsupervised MRI reconstruction method based on zero-Shot Learned Adversarial TransformERs (SLATER). SLATER embodies a deep adversarial network with cross-attention transformers to map noise and latent variables onto coil-combined MR images. During pre-training, this unconditional network learns a high-quality MRI prior in an unsupervised generative modeling task. During inference, a zero-shot reconstruction is then performed by incorporating the imaging operator and optimizing the prior to maximize consistency to undersampled data. Comprehensive experiments on brain MRI datasets clearly demonstrate the superior performance of SLATER against state-of-the-art unsupervised methods.
Collapse
|
16
|
Xiong YT, Zeng W, Xu L, Guo JX, Liu C, Chen JT, Du XY, Tang W. Virtual reconstruction of midfacial bone defect based on generative adversarial network. Head Face Med 2022; 18:19. [PMID: 35761334 PMCID: PMC9235085 DOI: 10.1186/s13005-022-00325-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2022] [Accepted: 05/19/2022] [Indexed: 02/08/2023] Open
Abstract
BACKGROUND The study aims to evaluate the accuracy of the generative adversarial networks (GAN) for reconstructing bony midfacial defects. METHODS According to anatomy, the bony midface was divided into five subunit structural regions and artificial defects are manually created on the corresponding CT images. GAN is trained to reconstruct artificial defects to their previous normal shape and tested. The clinical defects are reconstructed by the trained GAN, where the midspan defects were used for qualitative evaluation and the unilateral defects were used for quantitative evaluation. The cosine similarity and the mean error are used to evaluate the accuracy of reconstruction. The Mann-Whitney U test is used to detect whether reconstruction errors were consistent in artificial and unilateral clinical defects. RESULTS This study included 518 normal CT data, with 415 in training set and 103 in testing set, and 17 real patient data, with 2 midspan defects and 15 unilateral defects. Reconstruction of midspan clinical defects assessed by experts is acceptable. The cosine similarity in the reconstruction of artificial defects and unilateral clinical defects is 0.97 ± 0.01 and 0.96 ± 0.01, P = 0.695. The mean error in the reconstruction of artificial defects and unilateral clinical defects is 0.59 ± 0.31 mm and 0.48 ± 0.08 mm, P = 0.09. CONCLUSION GAN-based virtual reconstruction technology has reached a high accuracy in testing set, and statistical tests suggest that it can achieve similar results in real patient data. This study has preliminarily solved the problem of bony midfacial defect without reference.
Collapse
Affiliation(s)
- Yu-Tao Xiong
- State Key Laboratory of Oral Diseases and National Clinical Research Centre for Oral Diseases and Department of Oral and Maxillofacial Surgery, West China Hospital of Stomatology, Sichuan University, No.14, 3rd section of Ren Min Nan Road, Chengdu, 610041, China
| | - Wei Zeng
- State Key Laboratory of Oral Diseases and National Clinical Research Centre for Oral Diseases and Department of Oral and Maxillofacial Surgery, West China Hospital of Stomatology, Sichuan University, No.14, 3rd section of Ren Min Nan Road, Chengdu, 610041, China
| | - Lei Xu
- Machine Intelligence Laboratory, College of Computer Science, Sichuan University, Chengdu, 610065, China
| | - Ji-Xiang Guo
- Machine Intelligence Laboratory, College of Computer Science, Sichuan University, Chengdu, 610065, China
| | - Chang Liu
- State Key Laboratory of Oral Diseases and National Clinical Research Centre for Oral Diseases and Department of Oral and Maxillofacial Surgery, West China Hospital of Stomatology, Sichuan University, No.14, 3rd section of Ren Min Nan Road, Chengdu, 610041, China
| | - Jun-Tian Chen
- State Key Laboratory of Oral Diseases and National Clinical Research Centre for Oral Diseases and Department of Oral and Maxillofacial Surgery, West China Hospital of Stomatology, Sichuan University, No.14, 3rd section of Ren Min Nan Road, Chengdu, 610041, China
| | - Xin-Ya Du
- Department of Stomatology, the People's Hospital of Longhua, Shenzhen, 518109, China
| | - Wei Tang
- State Key Laboratory of Oral Diseases and National Clinical Research Centre for Oral Diseases and Department of Oral and Maxillofacial Surgery, West China Hospital of Stomatology, Sichuan University, No.14, 3rd section of Ren Min Nan Road, Chengdu, 610041, China.
| |
Collapse
|
17
|
Wu W, Hu D, Cong W, Shan H, Wang S, Niu C, Yan P, Yu H, Vardhanabhuti V, Wang G. Stabilizing deep tomographic reconstruction: Part A. Hybrid framework and experimental results. PATTERNS (NEW YORK, N.Y.) 2022; 3:100474. [PMID: 35607623 PMCID: PMC9122961 DOI: 10.1016/j.patter.2022.100474] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/26/2021] [Revised: 12/24/2021] [Accepted: 03/01/2022] [Indexed: 12/16/2022]
Abstract
A recent PNAS paper reveals that several popular deep reconstruction networks are unstable. Specifically, three kinds of instabilities were reported: (1) strong image artefacts from tiny perturbations, (2) small features missed in a deeply reconstructed image, and (3) decreased imaging performance with increased input data. Here, we propose an analytic compressed iterative deep (ACID) framework to address this challenge. ACID synergizes a deep network trained on big data, kernel awareness from compressed sensing (CS)-inspired processing, and iterative refinement to minimize the data residual relative to real measurement. Our study demonstrates that the ACID reconstruction is accurate, is stable, and sheds light on the converging mechanism of the ACID iteration under a bounded relative error norm assumption. ACID not only stabilizes an unstable deep reconstruction network but also is resilient against adversarial attacks to the whole ACID workflow, being superior to classic sparsity-regularized reconstruction and eliminating the three kinds of instabilities.
Collapse
Affiliation(s)
- Weiwen Wu
- Biomedical Imaging Center, Center for Biotechnology and Interdisciplinary Studies, Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY, USA
- School of Biomedical Engineering, Sun Yat-sen University, Shenzhen, Guangdong, China
- Department of Diagnostic Radiology, Li Ka Shing Faculty of Medicine, University of Hong Kong, Hong Kong SAR, China
| | - Dianlin Hu
- The Laboratory of Image Science and Technology, Southeast University, Nanjing, China
| | - Wenxiang Cong
- Biomedical Imaging Center, Center for Biotechnology and Interdisciplinary Studies, Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY, USA
| | - Hongming Shan
- Biomedical Imaging Center, Center for Biotechnology and Interdisciplinary Studies, Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY, USA
- Institute of Science and Technology for Brain-inspired Intelligence, Fudan University, Shanghai, China
| | - Shaoyu Wang
- Department of Electrical & Computer Engineering, University of Massachusetts Lowell, Lowell, MA, USA
| | - Chuang Niu
- Biomedical Imaging Center, Center for Biotechnology and Interdisciplinary Studies, Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY, USA
| | - Pingkun Yan
- Biomedical Imaging Center, Center for Biotechnology and Interdisciplinary Studies, Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY, USA
| | - Hengyong Yu
- Department of Electrical & Computer Engineering, University of Massachusetts Lowell, Lowell, MA, USA
| | - Varut Vardhanabhuti
- Department of Diagnostic Radiology, Li Ka Shing Faculty of Medicine, University of Hong Kong, Hong Kong SAR, China
| | - Ge Wang
- Biomedical Imaging Center, Center for Biotechnology and Interdisciplinary Studies, Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY, USA
| |
Collapse
|
18
|
Pal A, Rathi Y. A review and experimental evaluation of deep learning methods for MRI reconstruction. THE JOURNAL OF MACHINE LEARNING FOR BIOMEDICAL IMAGING 2022; 1:001. [PMID: 35722657 PMCID: PMC9202830] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 10/25/2022]
Abstract
Following the success of deep learning in a wide range of applications, neural network-based machine-learning techniques have received significant interest for accelerating magnetic resonance imaging (MRI) acquisition and reconstruction strategies. A number of ideas inspired by deep learning techniques for computer vision and image processing have been successfully applied to nonlinear image reconstruction in the spirit of compressed sensing for accelerated MRI. Given the rapidly growing nature of the field, it is imperative to consolidate and summarize the large number of deep learning methods that have been reported in the literature, to obtain a better understanding of the field in general. This article provides an overview of the recent developments in neural-network based approaches that have been proposed specifically for improving parallel imaging. A general background and introduction to parallel MRI is also given from a classical view of k-space based reconstruction methods. Image domain based techniques that introduce improved regularizers are covered along with k-space based methods which focus on better interpolation strategies using neural networks. While the field is rapidly evolving with plenty of papers published each year, in this review, we attempt to cover broad categories of methods that have shown good performance on publicly available data sets. Limitations and open problems are also discussed and recent efforts for producing open data sets and benchmarks for the community are examined.
Collapse
Affiliation(s)
- Arghya Pal
- Department of Psychiatry and Radiology, Harvard Medical School, Boston, MA, USA
| | - Yogesh Rathi
- Department of Psychiatry and Radiology, Harvard Medical School, Boston, MA, USA
| |
Collapse
|
19
|
Liao Z, Liao K, Shen H, van Boxel MF, Prijs J, Jaarsma RL, Doornberg JN, Hengel AVD, Verjans JW. CNN Attention Guidance for Improved Orthopedics Radiographic Fracture Classification. IEEE J Biomed Health Inform 2022; 26:3139-3150. [PMID: 35192467 DOI: 10.1109/jbhi.2022.3152267] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
Convolutional neural networks (CNNs) have gained significant popularity in orthopedic imaging in recent years due to their ability to solve fracture classification problems. A common criticism of CNNs is their opaque learning and reasoning process, making it difficult to trust machine diagnosis and the subsequent adoption of such algorithms in clinical setting. This is especially true when the CNN is trained with limited amount of medical data, which is a common issue as curating sufficiently large amount of annotated medical image data is a long and costly process. While interest has been devoted to explaining CNN learnt knowledge by visualizing network attention, the utilization of the visualized attention to improve network learning has been rarely investigated. This paper explores the effectiveness of regularizing CNN network with human-provided attention guidance on where in the image the network should look for answering clues. On two orthopedics radiographic fracture classification datasets, through extensive experiments we demonstrate that explicit human-guided attention indeed can direct correct network attention and consequently significantly improve classification performance. The development code for the proposed attention guidance is publicly available on GitHub.
Collapse
|
20
|
Huang J, Ding W, Lv J, Yang J, Dong H, Del Ser J, Xia J, Ren T, Wong ST, Yang G. Edge-enhanced dual discriminator generative adversarial network for fast MRI with parallel imaging using multi-view information. APPL INTELL 2022; 52:14693-14710. [PMID: 36199853 PMCID: PMC9526695 DOI: 10.1007/s10489-021-03092-w] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Accepted: 12/09/2021] [Indexed: 12/24/2022]
Abstract
In clinical medicine, magnetic resonance imaging (MRI) is one of the most important tools for diagnosis, triage, prognosis, and treatment planning. However, MRI suffers from an inherent slow data acquisition process because data is collected sequentially in k-space. In recent years, most MRI reconstruction methods proposed in the literature focus on holistic image reconstruction rather than enhancing the edge information. This work steps aside this general trend by elaborating on the enhancement of edge information. Specifically, we introduce a novel parallel imaging coupled dual discriminator generative adversarial network (PIDD-GAN) for fast multi-channel MRI reconstruction by incorporating multi-view information. The dual discriminator design aims to improve the edge information in MRI reconstruction. One discriminator is used for holistic image reconstruction, whereas the other one is responsible for enhancing edge information. An improved U-Net with local and global residual learning is proposed for the generator. Frequency channel attention blocks (FCA Blocks) are embedded in the generator for incorporating attention mechanisms. Content loss is introduced to train the generator for better reconstruction quality. We performed comprehensive experiments on Calgary-Campinas public brain MR dataset and compared our method with state-of-the-art MRI reconstruction methods. Ablation studies of residual learning were conducted on the MICCAI13 dataset to validate the proposed modules. Results show that our PIDD-GAN provides high-quality reconstructed MR images, with well-preserved edge information. The time of single-image reconstruction is below 5ms, which meets the demand of faster processing.
Collapse
Affiliation(s)
- Jiahao Huang
- College of Information Science and Technology, Zhejiang Shuren University, 310015 Hangzhou, China
- National Heart and Lung Institute, Imperial College London, London, UK
| | - Weiping Ding
- School of Information Science and Technology, Nantong University, 226019 Nantong, China
| | - Jun Lv
- School of Computer and Control Engineering, Yantai University, 264005 Yantai, China
| | - Jingwen Yang
- Department of Prosthodontics, Peking University School and Hospital of Stomatology, Beijing, China
| | - Hao Dong
- Center on Frontiers of Computing Studies, Peking University, Beijing, China
| | - Javier Del Ser
- TECNALIA, Basque Research and Technology Alliance (BRTA), 48160 Derio, Spain
- University of the Basque Country (UPV/EHU), 48013 Bilbao, Spain
| | - Jun Xia
- Department of Radiology, Shenzhen Second People’s Hospital, The First Afliated Hospital of Shenzhen University Health Science Center, Shenzhen, China
| | - Tiaojuan Ren
- College of Information Science and Technology, Zhejiang Shuren University, 310015 Hangzhou, China
| | - Stephen T. Wong
- Systems Medicine and Bioengineering Department, Departments of Radiology and Pathology, Houston Methodist Cancer Center, Houston Methodist Hospital, Weill Cornell Medicine, 77030 Houston, TX USA
| | - Guang Yang
- National Heart and Lung Institute, Imperial College London, London, UK
- Cardiovascular Research Centre, Royal Brompton Hospital, London, UK
| |
Collapse
|
21
|
Wu X, Li C, Zeng X, Wei H, Deng HW, Zhang J, Xu M. CryoETGAN: Cryo-Electron Tomography Image Synthesis via Unpaired Image Translation. Front Physiol 2022; 13:760404. [PMID: 35370760 PMCID: PMC8970048 DOI: 10.3389/fphys.2022.760404] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/18/2021] [Accepted: 01/17/2022] [Indexed: 12/02/2022] Open
Abstract
Cryo-electron tomography (Cryo-ET) has been regarded as a revolution in structural biology and can reveal molecular sociology. Its unprecedented quality enables it to visualize cellular organelles and macromolecular complexes at nanometer resolution with native conformations. Motivated by developments in nanotechnology and machine learning, establishing machine learning approaches such as classification, detection and averaging for Cryo-ET image analysis has inspired broad interest. Yet, deep learning-based methods for biomedical imaging typically require large labeled datasets for good results, which can be a great challenge due to the expense of obtaining and labeling training data. To deal with this problem, we propose a generative model to simulate Cryo-ET images efficiently and reliably: CryoETGAN. This cycle-consistent and Wasserstein generative adversarial network (GAN) is able to generate images with an appearance similar to the original experimental data. Quantitative and visual grading results on generated images are provided to show that the results of our proposed method achieve better performance compared to the previous state-of-the-art simulation methods. Moreover, CryoETGAN is stable to train and capable of generating plausibly diverse image samples.
Collapse
Affiliation(s)
- Xindi Wu
- Computational Biology Department, Carnegie Mellon University, Pittsburgh, PA, United States
| | - Chengkun Li
- École Polytechnique Fédérale de Lausanne, Lausanne, Switzerland
| | - Xiangrui Zeng
- Computational Biology Department, Carnegie Mellon University, Pittsburgh, PA, United States
| | - Haocheng Wei
- Department of Electrical & Computer Engineering, University of Toronto, Toronto, ON, Canada
| | - Hong-Wen Deng
- Center for Biomedical Informatics & Genomics, Tulane University, New Orleans, LA, United States
| | - Jing Zhang
- Department of Computer Science, University of California, Irvine, Irvine, CA, United States
| | - Min Xu
- Computational Biology Department, Carnegie Mellon University, Pittsburgh, PA, United States
| |
Collapse
|
22
|
Mudeng V, Kim M, Choe SW. Objective Numerical Evaluation of Diffuse, Optically Reconstructed Images Using Structural Similarity Index. BIOSENSORS 2021; 11:504. [PMID: 34940261 PMCID: PMC8699273 DOI: 10.3390/bios11120504] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/31/2021] [Revised: 12/06/2021] [Accepted: 12/06/2021] [Indexed: 05/13/2023]
Abstract
Diffuse optical tomography is emerging as a non-invasive optical modality used to evaluate tissue information by obtaining the optical properties' distribution. Two procedures are performed to produce reconstructed absorption and reduced scattering images, which provide structural information that can be used to locate inclusions within tissues with the assistance of a known light intensity around the boundary. These methods are referred to as a forward problem and an inverse solution. Once the reconstructed image is obtained, a subjective measurement is used as the conventional way to assess the image. Hence, in this study, we developed an algorithm designed to numerically assess reconstructed images to identify inclusions using the structural similarity (SSIM) index. We compared four SSIM algorithms with 168 simulated reconstructed images involving the same inclusion position with different contrast ratios and inclusion sizes. A multiscale, improved SSIM containing a sharpness parameter (MS-ISSIM-S) was proposed to represent the potential evaluation compared with the human visible perception. The results indicated that the proposed MS-ISSIM-S is suitable for human visual perception by demonstrating a reduction of similarity score related to various contrasts with a similar size of inclusion; thus, this metric is promising for the objective numerical assessment of diffuse, optically reconstructed images.
Collapse
Affiliation(s)
- Vicky Mudeng
- Department of Medical IT Convergence Engineering, Kumoh National Institute of Technology, Gumi 39253, Korea;
- Department of Electrical Engineering, Institut Teknologi Kalimantan, Balikpapan 76127, Indonesia
| | - Minseok Kim
- Department of Mechanical System Engineering, Kumoh National Institute of Technology, Gumi 39177, Korea
- Department of Aeronautics, Mechanical and Electronic Convergence Engineering, Kumoh National Institute of Technology, Gumi 39177, Korea
| | - Se-woon Choe
- Department of Medical IT Convergence Engineering, Kumoh National Institute of Technology, Gumi 39253, Korea;
- Department of IT Convergence Engineering, Kumoh National Institute of Technology, Gumi 39253, Korea
| |
Collapse
|
23
|
Lv J, Li G, Tong X, Chen W, Huang J, Wang C, Yang G. Transfer learning enhanced generative adversarial networks for multi-channel MRI reconstruction. Comput Biol Med 2021; 134:104504. [PMID: 34062366 DOI: 10.1016/j.compbiomed.2021.104504] [Citation(s) in RCA: 33] [Impact Index Per Article: 8.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/24/2021] [Revised: 05/17/2021] [Accepted: 05/17/2021] [Indexed: 12/23/2022]
Abstract
Deep learning based generative adversarial networks (GAN) can effectively perform image reconstruction with under-sampled MR data. In general, a large number of training samples are required to improve the reconstruction performance of a certain model. However, in real clinical applications, it is difficult to obtain tens of thousands of raw patient data to train the model since saving k-space data is not in the routine clinical flow. Therefore, enhancing the generalizability of a network based on small samples is urgently needed. In this study, three novel applications were explored based on parallel imaging combined with the GAN model (PI-GAN) and transfer learning. The model was pre-trained with public Calgary brain images and then fine-tuned for use in (1) patients with tumors in our center; (2) different anatomies, including knee and liver; (3) different k-space sampling masks with acceleration factors (AFs) of 2 and 6. As for the brain tumor dataset, the transfer learning results could remove the artifacts found in PI-GAN and yield smoother brain edges. The transfer learning results for the knee and liver were superior to those of the PI-GAN model trained with its own dataset using a smaller number of training cases. However, the learning procedure converged more slowly in the knee datasets compared to the learning in the brain tumor datasets. The reconstruction performance was improved by transfer learning both in the models with AFs of 2 and 6. Of these two models, the one with AF = 2 showed better results. The results also showed that transfer learning with the pre-trained model could solve the problem of inconsistency between the training and test datasets and facilitate generalization to unseen data.
Collapse
Affiliation(s)
- Jun Lv
- School of Computer and Control Engineering, Yantai University, Yantai, China
| | - Guangyuan Li
- School of Computer and Control Engineering, Yantai University, Yantai, China
| | - Xiangrong Tong
- School of Computer and Control Engineering, Yantai University, Yantai, China
| | | | - Jiahao Huang
- School of Optics and Photonics, Beijing Institute of Technology, Beijing, China
| | - Chengyan Wang
- Human Phenome Institute, Fudan University, Shanghai, China.
| | - Guang Yang
- Cardiovascular Research Centre, Royal Brompton Hospital, London, SW3 6NP, UK; National Heart and Lung Institute, Imperial College London, London, SW7 2AZ, UK.
| |
Collapse
|
24
|
Yang D, Liu J, Wang Y, Xu B, Wang X. Application of a Generative Adversarial Network in Image Reconstruction of Magnetic Induction Tomography. SENSORS 2021; 21:s21113869. [PMID: 34205157 PMCID: PMC8199933 DOI: 10.3390/s21113869] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/28/2021] [Revised: 05/29/2021] [Accepted: 05/31/2021] [Indexed: 11/16/2022]
Abstract
Image reconstruction of Magnetic induction tomography (MIT) is an ill-posed problem. The non-linear characteristics lead many difficulties to its solution. In this paper, a method based on a Generative Adversarial Network (GAN) is presented to tackle these barriers. Firstly, the principle of MIT is analyzed. Then the process for finding the global optimum of conductivity distribution is described as a training process, and the GAN model is proposed. Finally, the image was reconstructed by a part of the model (the generator). All datasets are obtained from an eight-channel MIT model by COMSOL Multiphysics software. The voltage measurement samples are used as input to the trained network, and its output is an estimate for image reconstruction of the internal conductivity distribution. The results based on the proposed model and the traditional algorithms were compared, which have shown that average root mean squared error of reconstruction results obtained by the proposed method is 0.090, and the average correlation coefficient with original images is 0.940, better than corresponding indicators of BPNN and Tikhonov regularization algorithms. Accordingly, the GAN algorithm was able to fit the non-linear relationship between input and output, and visual images also show that it solved the usual problems of artifact in traditional algorithm and hot pixels in L2 regularization, which is of great significance for other ill-posed or non-linear problems.
Collapse
Affiliation(s)
- Dan Yang
- Key Laboratory of Data Analytics and Optimization for Smart Industry, Northeastern University, Shenyang 110819, China; (J.L.); (X.W.)
- Key Laboratory of Infrared Optoelectric Materials and Micro-Nano Devices, Shenyang 110819, China;
- College of Information Science and Engineering, Northeastern University, Shenyang 110819, China
- Correspondence: ; Tel.: +86-135-1428-6842
| | - Jiahua Liu
- Key Laboratory of Data Analytics and Optimization for Smart Industry, Northeastern University, Shenyang 110819, China; (J.L.); (X.W.)
- Key Laboratory of Infrared Optoelectric Materials and Micro-Nano Devices, Shenyang 110819, China;
| | - Yuchen Wang
- Key Laboratory of Infrared Optoelectric Materials and Micro-Nano Devices, Shenyang 110819, China;
- College of Information Science and Engineering, Northeastern University, Shenyang 110819, China
| | - Bin Xu
- College of Computer Science and Engineering, Northeastern University, Shenyang 110819, China;
| | - Xu Wang
- Key Laboratory of Data Analytics and Optimization for Smart Industry, Northeastern University, Shenyang 110819, China; (J.L.); (X.W.)
| |
Collapse
|
25
|
Lv J, Wang C, Yang G. PIC-GAN: A Parallel Imaging Coupled Generative Adversarial Network for Accelerated Multi-Channel MRI Reconstruction. Diagnostics (Basel) 2021; 11:61. [PMID: 33401777 PMCID: PMC7824530 DOI: 10.3390/diagnostics11010061] [Citation(s) in RCA: 27] [Impact Index Per Article: 6.8] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/05/2020] [Revised: 12/28/2020] [Accepted: 12/29/2020] [Indexed: 12/16/2022] Open
Abstract
In this study, we proposed a model combing parallel imaging (PI) with generative adversarial network (GAN) architecture (PIC-GAN) for accelerated multi-channel magnetic resonance imaging (MRI) reconstruction. This model integrated data fidelity and regularization terms into the generator to benefit from multi-coils information and provide an "end-to-end" reconstruction. Besides, to better preserve image details during reconstruction, we combined the adversarial loss with pixel-wise loss in both image and frequency domains. The proposed PIC-GAN framework was evaluated on abdominal and knee MRI images using 2, 4 and 6-fold accelerations with different undersampling patterns. The performance of the PIC-GAN was compared to the sparsity-based parallel imaging (L1-ESPIRiT), the variational network (VN), and conventional GAN with single-channel images as input (zero-filled (ZF)-GAN). Experimental results show that our PIC-GAN can effectively reconstruct multi-channel MR images at a low noise level and improved structure similarity of the reconstructed images. PIC-GAN has yielded the lowest Normalized Mean Square Error (in ×10-5) (PIC-GAN: 0.58 ± 0.37, ZF-GAN: 1.93 ± 1.41, VN: 1.87 ± 1.28, L1-ESPIRiT: 2.49 ± 1.04 for abdominal MRI data and PIC-GAN: 0.80 ± 0.26, ZF-GAN: 0.93 ± 0.29, VN:1.18 ± 0.31, L1-ESPIRiT: 1.28 ± 0.24 for knee MRI data) and the highest Peak Signal to Noise Ratio (PIC-GAN: 34.43 ± 1.92, ZF-GAN: 31.45 ± 4.0, VN: 29.26 ± 2.98, L1-ESPIRiT: 25.40 ± 1.88 for abdominal MRI data and PIC-GAN: 34.10 ± 1.09, ZF-GAN: 31.47 ± 1.05, VN: 30.01 ± 1.01, L1-ESPIRiT: 28.01 ± 0.98 for knee MRI data) compared to ZF-GAN, VN and L1-ESPIRiT with an under-sampling factor of 6. The proposed PIC-GAN framework has shown superior reconstruction performance in terms of reducing aliasing artifacts and restoring tissue structures as compared to other conventional and state-of-the-art reconstruction methods.
Collapse
Affiliation(s)
- Jun Lv
- School of Computer and Control Engineering, Yantai University, Yantai 264005, China;
| | - Chengyan Wang
- Human Phenome Institute, Fudan University, Shanghai 201203, China
| | - Guang Yang
- Cardiovascular Research Centre, Royal Brompton Hospital, London SW3 6NP, UK
- National Heart and Lung Institute, Imperial College London, London SW7 2AZ, UK
| |
Collapse
|
26
|
Li G, Lv J, Tong X, Wang C, Yang G. High-Resolution Pelvic MRI Reconstruction Using a Generative Adversarial Network With Attention and Cyclic Loss. IEEE ACCESS 2021; 9:105951-105964. [DOI: 10.1109/access.2021.3099695] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/29/2023]
|