1
|
Shin H, Kim T, Lee J, Chun SY, Cho S, Shin D. Sparse-view CBCT reconstruction using meta-learned neural attenuation field and hash-encoding regularization. Comput Biol Med 2025; 189:109900. [PMID: 40024186 DOI: 10.1016/j.compbiomed.2025.109900] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2024] [Revised: 02/09/2025] [Accepted: 02/18/2025] [Indexed: 03/04/2025]
Abstract
Cone beam computed tomography (CBCT) is an emerging medical imaging technique to visualize the internal anatomical structures of patients. During a CBCT scan, several projection images of different angles or views are collectively utilized to reconstruct a tomographic image. However, reducing the number of projections in a CBCT scan while preserving the quality of a reconstructed image is challenging due to the nature of an ill-posed inverse problem. Recently, a neural attenuation field (NAF) method was proposed by adopting a neural radiance field algorithm as a new way for CBCT reconstruction, demonstrating fast and promising results using only 50 views. However, decreasing the number of projections is still preferable to reduce potential radiation exposure, and a faster reconstruction time is required considering a typical scan time. In this work, we propose a fast and accurate sparse-view CBCT reconstruction (FACT) method to provide better reconstruction quality and faster optimization speed in the minimal number of view acquisitions (< 50 views). In the FACT method, we meta-trained a neural network and a hash-encoder using a few scans (= 15), and a new regularization technique is utilized to reconstruct the details of an anatomical structure. In conclusion, we have shown that the FACT method produced better, and faster reconstruction results over the other conventional algorithms based on CBCT scans of different body parts (chest, head, and abdomen) and CT vendors (Siemens, Phillips, and GE).
Collapse
Affiliation(s)
- Heejun Shin
- Artificial Intelligence Engineering Division, Radisen Co. Ltd., Seoul, Republic of Korea
| | - Taehee Kim
- Artificial Intelligence Engineering Division, Radisen Co. Ltd., Seoul, Republic of Korea
| | - Jongho Lee
- Laboratory for Imaging Science and Technology, Department of Electrical and Computer Engineering, Seoul National University, Seoul, Republic of Korea
| | - Se Young Chun
- Intelligent Computational Imaging Laboratory, Department of Electrical and Computer Engineering, Seoul National University, Seoul, Republic of Korea
| | - Seungryong Cho
- Medical Imaging and Radiotherapy Laboratory, Department of Nuclear and Quantum Engineering, Korea Advanced Institute of Science and Technology, Daejean, Republic of Korea
| | - Dongmyung Shin
- Artificial Intelligence Engineering Division, Radisen Co. Ltd., Seoul, Republic of Korea.
| |
Collapse
|
2
|
Xu S, Fu J, Sun Y, Cong P, Xiang X. Deep Radon Prior: A fully unsupervised framework for sparse-view CT reconstruction. Comput Biol Med 2025; 189:109853. [PMID: 40056836 DOI: 10.1016/j.compbiomed.2025.109853] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2024] [Revised: 02/08/2025] [Accepted: 02/11/2025] [Indexed: 03/10/2025]
Abstract
BACKGROUND Sparse-view computed tomography (CT) substantially reduces radiation exposure but often introduces severe artifacts that compromise image fidelity. Recent advances in deep learning for solving inverse problems have shown considerable promise in enhancing CT reconstruction; however, most approaches heavily rely on high-quality training datasets and lack interpretability. METHOD To address these challenges, this paper introduces a novel, fully unsupervised deep learning framework that mitigates the dependency on extensive labeled data and improves the interpretability of the reconstruction process. Specifically, we propose the Deep Radon Prior (DRP) framework, inspired by the Deep Image Prior (DIP), which integrates a neural network as an implicit prior into the iterative reconstruction process. This integration facilitates the image domain and the Radon domain gradient feedback and progressively optimizes the neural network through multiple stages, effectively narrowing the solution space in the Radon domain for under-constrained imaging protocols. RESULTS We discuss the convergence properties of DRP and validate our approach experimentally, demonstrating its ability to produce high-fidelity images while significantly reducing artifacts. Results indicate that DRP achieves comparable or superior performance to supervised methods, thereby addressing the inherent challenges of sparse-view CT and substantially enhancing image quality. CONCLUSIONS The introduction of DRP represents a significant advancement in sparse-view CT imaging by leveraging the inherent deep self-correlation of the Radon domain, enabling effective cooperation with neural network manifolds for image reconstruction. This paradigm shift toward fully unsupervised learning offers a scalable and insightful approach to medical imaging, potentially redefining the landscape of CT reconstruction.
Collapse
Affiliation(s)
- Shuo Xu
- Institute of Nuclear and New Energy Technology, Tsinghua University, Beijing, China; Beijing Key Laboratory of Nuclear Detection Technology, Beijing, China; Hefei Meyer Optoelectronic Technology INC, Hefei, Anhui Province, China.
| | - Jintao Fu
- Institute of Nuclear and New Energy Technology, Tsinghua University, Beijing, China; Beijing Key Laboratory of Nuclear Detection Technology, Beijing, China.
| | - Yuewen Sun
- Institute of Nuclear and New Energy Technology, Tsinghua University, Beijing, China; Beijing Key Laboratory of Nuclear Detection Technology, Beijing, China.
| | - Peng Cong
- Institute of Nuclear and New Energy Technology, Tsinghua University, Beijing, China; Beijing Key Laboratory of Nuclear Detection Technology, Beijing, China.
| | - Xincheng Xiang
- Institute of Nuclear and New Energy Technology, Tsinghua University, Beijing, China; Beijing Key Laboratory of Nuclear Detection Technology, Beijing, China.
| |
Collapse
|
3
|
Li Z, Sun Z, Lv L, Liu Y, Wang X, Xu J, Xing J, Babyn P, Sun FR. Ultra-sparse view lung CT image reconstruction using generative adversarial networks and compressed sensing. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2025:8953996251329214. [PMID: 40296779 DOI: 10.1177/08953996251329214] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/30/2025]
Abstract
X-ray ionizing radiation from Computed Tomography (CT) scanning increases cancer risk for patients, thus making sparse view CT, which diminishes X-ray exposure by lowering the number of projections, highly significant in diagnostic imaging. However, reducing the number of projections inherently degrades image quality, negatively impacting clinical diagnosis. Consequently, attaining reconstructed images that meet diagnostic imaging criteria for sparse view CT is challenging. This paper presents a novel network (CSUF), specifically designed for ultra-sparse view lung CT image reconstruction. The CSUF network consists of three cohesive components including (1) a compressed sensing-based CT image reconstruction module (VdCS module), (2) a U-shaped end-to-end network, CT-RDNet, enhanced with a self-attention mechanism, acting as the generator in a Generative Adversarial Network (GAN) for CT image restoration and denoising, and (3) a feedback loop. The VdCS module enriches CT-RDNet with enhanced features, while CT-RDNet supplies the VdCS module with prior images infused with rich details and minimized artifacts, facilitated by the feedback loop. Engineering simulation experimental results demonstrate the robustness of the CSUF network and its potential to deliver lung CT images with diagnostic imaging quality even under ultra-sparse view conditions.
Collapse
Affiliation(s)
- Zhaoguang Li
- School of Integrated Circuits, Shandong University, Jinan, China
| | - Zhengxiang Sun
- Faculty of Science, The University of Sydney, NSW, Australia
| | - Lin Lv
- School of Integrated Circuits, Shandong University, Jinan, China
| | - Yuhan Liu
- School of Integrated Circuits, Shandong University, Jinan, China
| | - Xiuying Wang
- Faculty of Engineering, The University of Sydney, NSW, Australia
| | - Jingjing Xu
- School of Integrated Circuits, Shandong University, Jinan, China
| | - Jianping Xing
- School of Integrated Circuits, Shandong University, Jinan, China
| | - Paul Babyn
- Department of Medical Imaging, University of Saskatchewan and Saskatoon Health Region, Saskatoon, Canada
| | - Feng-Rong Sun
- School of Integrated Circuits, Shandong University, Jinan, China
| |
Collapse
|
4
|
Li Y, Fu X, Li H, Zhao S, Jin R, Zhou SK. 3DGR-CT: Sparse-view CT reconstruction with a 3D Gaussian representation. Med Image Anal 2025; 103:103585. [PMID: 40279825 DOI: 10.1016/j.media.2025.103585] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/31/2024] [Revised: 03/16/2025] [Accepted: 04/04/2025] [Indexed: 04/29/2025]
Abstract
Sparse-view computed tomography (CT) reduces radiation exposure by acquiring fewer projections, making it a valuable tool in clinical scenarios where low-dose radiation is essential. However, this often results in increased noise and artifacts due to limited data. In this paper we propose a novel 3D Gaussian representation (3DGR) based method for sparse-view CT reconstruction. Inspired by recent success in novel view synthesis driven by 3D Gaussian splatting, we leverage the efficiency and expressiveness of 3D Gaussian representation as an alternative to implicit neural representation. To unleash the potential of 3DGR for CT imaging scenario, we propose two key innovations: (i) FBP-image-guided Guassian initialization and (ii) efficient integration with a differentiable CT projector. Extensive experiments and ablations on diverse datasets demonstrate the proposed 3DGR-CT consistently outperforms state-of-the-art counterpart methods, achieving higher reconstruction accuracy with faster convergence. Furthermore, we showcase the potential of 3DGR-CT for real-time physical simulation, which holds important clinical applications while challenging for implicit neural representations. Code available at: https://github.com/SigmaLDC/3DGR-CT.
Collapse
Affiliation(s)
- Yingtai Li
- School of Biomedical Engineering, Division of Life Sciences and Medicine, University of Science and Technology of China (USTC), Hefei, 230026, Anhui, China; China and Center for Medical Imaging, Robotics, Analytic Computing & Learning (MIRACLE), Suzhou Institute for Advance Research, USTC, Suzhou, 215123, Jiangsu, China
| | - Xueming Fu
- School of Biomedical Engineering, Division of Life Sciences and Medicine, University of Science and Technology of China (USTC), Hefei, 230026, Anhui, China; China and Center for Medical Imaging, Robotics, Analytic Computing & Learning (MIRACLE), Suzhou Institute for Advance Research, USTC, Suzhou, 215123, Jiangsu, China
| | - Han Li
- School of Biomedical Engineering, Division of Life Sciences and Medicine, University of Science and Technology of China (USTC), Hefei, 230026, Anhui, China; China and Center for Medical Imaging, Robotics, Analytic Computing & Learning (MIRACLE), Suzhou Institute for Advance Research, USTC, Suzhou, 215123, Jiangsu, China
| | - Shang Zhao
- School of Biomedical Engineering, Division of Life Sciences and Medicine, University of Science and Technology of China (USTC), Hefei, 230026, Anhui, China; China and Center for Medical Imaging, Robotics, Analytic Computing & Learning (MIRACLE), Suzhou Institute for Advance Research, USTC, Suzhou, 215123, Jiangsu, China
| | - Ruiyang Jin
- School of Biomedical Engineering, Division of Life Sciences and Medicine, University of Science and Technology of China (USTC), Hefei, 230026, Anhui, China; China and Center for Medical Imaging, Robotics, Analytic Computing & Learning (MIRACLE), Suzhou Institute for Advance Research, USTC, Suzhou, 215123, Jiangsu, China
| | - S Kevin Zhou
- School of Biomedical Engineering, Division of Life Sciences and Medicine, University of Science and Technology of China (USTC), Hefei, 230026, Anhui, China; China and Center for Medical Imaging, Robotics, Analytic Computing & Learning (MIRACLE), Suzhou Institute for Advance Research, USTC, Suzhou, 215123, Jiangsu, China; Key Laboratory of Precision and Intelligent Chemistry, USTC, Hefei, 230026, Anhui, China; Key Laboratory of Intelligent Information Processing of Chinese Academy of Sciences (CAS), Institute of Computing Technology, CAS, Beijing, 100190, China.
| |
Collapse
|
5
|
Shao W. Machine Learning in Microwave Medical Imaging and Lesion Detection. Diagnostics (Basel) 2025; 15:986. [PMID: 40310355 PMCID: PMC12025532 DOI: 10.3390/diagnostics15080986] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/10/2025] [Revised: 04/04/2025] [Accepted: 04/08/2025] [Indexed: 05/02/2025] Open
Abstract
Machine learning (ML) techniques have attracted many microwave researchers and engineers for their potential to improve performance in microwave- and millimeter-wave-based medical applications. This paper reviews ML algorithms, data acquisition, training techniques, and applications that have emerged in recent years. It also reviews state-of-the-art ML techniques applied for the detection of various organ diseases with microwave signals, achieving more successful results than using traditional methods alone, such as a higher diagnosis accuracy or spatial resolution and significantly improved efficiency. Challenges and the outlook of using ML in future microwave medical applications are also discussed.
Collapse
|
6
|
Zhao X, Du Y, Peng Y. DLPVI: Deep learning framework integrating projection, view-by-view backprojection, and image domains for high- and ultra-sparse-view CBCT reconstruction. Comput Med Imaging Graph 2025; 121:102508. [PMID: 39921927 DOI: 10.1016/j.compmedimag.2025.102508] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/28/2024] [Revised: 01/07/2025] [Accepted: 01/30/2025] [Indexed: 02/10/2025]
Abstract
This study proposes a deep learning framework, DLPVI, which integrates projection, view-by-view backprojection (VVBP), and image domains to improve the quality of high-sparse-view and ultra-sparse-view cone-beam computed tomography (CBCT) images. The DLPVI comprises a projection domain sub-framework, a VVBP domain sub-framework, and a Transformer-based image domain model. First, full-view projections were restored from sparse-view projections via the projection domain sub-framework, then filtered and view-by-view backprojected to generate VVBP raw data. Next, the VVBP raw data was processed by the VVBP domain sub-framework to suppress residual noise and artifacts, and produce CBCT axial images. Finally, the axial images were further refined using the image domain model. The DLPVI was trained, validated, and tested on CBCT data from 163, 30, and 30 real patients respectively. Quantitative metrics including root-mean-square error (RMSE), peak signal-to-noise ratio (PSNR), structural similarity (SSIM), and feature similarity (FSIM) were calculated to evaluate the method performance. The DLPVI was compared with 15 state-of-the-art (SOTA) methods, including 2 projection domain models, 10 image domain models, and 3 projection-image dual-domain frameworks, on 1/8 high-sparse-view and 1/16 ultra-sparse-view reconstruction tasks. Statistical analysis was conducted using the Kruskal-Wallis test, followed by the post-hoc Dunn's test. Experimental results demonstrated that the DLPVI outperformed all 15 SOTA methods for both tasks, with statistically significant improvements (p < 0.05 in Kruskal-Wallis test and p < 0.05/15 in Dunn's test). The proposed DLPVI effectively improves the quality of high- and ultra-sparse-view CBCT images.
Collapse
Affiliation(s)
- Xuzhi Zhao
- School of Electronic and Information Engineering, Beijing Jiaotong University, Beijing, China
| | - Yi Du
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education/Beijing), Department of Radiation Oncology, Peking University Cancer Hospital & Institute, Beijing, China; Institute of Medical Technology, Peking University Health Science Center, Beijing, China.
| | - Yahui Peng
- School of Electronic and Information Engineering, Beijing Jiaotong University, Beijing, China
| |
Collapse
|
7
|
Long Y, Zhong Q, Lu J, Xiong C. A novel detail-enhanced wavelet domain feature compensation network for sparse-view X-ray computed laminography. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2025; 33:488-498. [PMID: 39973784 DOI: 10.1177/08953996251319183] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/21/2025]
Abstract
BackgroundX-ray Computed Laminography (CL) is a popular industrial tool for non-destructive visualization of flat objects. However, high-quality CL imaging requires a large number of projections, resulting in a long imaging time. Reducing the number of projections allows acceleration of the imaging process, but decreases the quality of reconstructed images.ObjectiveOur objective is to build a deep learning network for sparse-view CL reconstruction.MethodsConsidering complementarities of feature extraction in different domains, we design an encoder-decoder network that enables to compensate the missing information during spatial domain feature extraction in wavelet domain. Also, a detail-enhanced module is developed to highlight details. Additionally, Swin Transformer and convolution operators are combined to better capture features.ResultsA total of 3200 pairs of 16-view and 1024-view CL images (2880 pairs for training, 160 pairs for validation, and 160 pairs for testing) of solder joints have been employed to investigate the performance of the proposed network. It is observed that the proposed network obtains the highest image quality with PSNR and SSIM of 37.875 ± 0.908 dB, 0.992 ± 0.002, respectively. Also, it achieves competitive results on the AAPM dataset.ConclusionsThis study demonstrates the effectiveness and generalization of the proposed network for sparse-view CL reconstruction.
Collapse
Affiliation(s)
- Yawu Long
- Huawei Technologies Co. Ltd, Shenzhen, China
| | | | - Jin Lu
- Huawei Technologies Co. Ltd, Shenzhen, China
| | | |
Collapse
|
8
|
Ma X, Zou M, Fang X, Luo G, Wang W, Dong S, Li X, Wang K, Dong Q, Tian Y, Li S. Convergent-Diffusion Denoising Model for multi-scenario CT Image Reconstruction. Comput Med Imaging Graph 2025; 120:102491. [PMID: 39787736 DOI: 10.1016/j.compmedimag.2024.102491] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/03/2024] [Revised: 10/27/2024] [Accepted: 12/31/2024] [Indexed: 01/12/2025]
Abstract
A generic and versatile CT Image Reconstruction (CTIR) scheme can efficiently mitigate imaging noise resulting from inherent physical limitations, substantially bolstering the dependability of CT imaging diagnostics across a wider spectrum of patient cases. Current CTIR techniques often concentrate on distinct areas such as Low-Dose CT denoising (LDCTD), Sparse-View CT reconstruction (SVCTR), and Metal Artifact Reduction (MAR). Nevertheless, due to the intricate nature of multi-scenario CTIR, these techniques frequently narrow their focus to specific tasks, resulting in limited generalization capabilities for diverse scenarios. We propose a novel Convergent-Diffusion Denoising Model (CDDM) for multi-scenario CTIR, which utilizes a stepwise denoising process to converge toward an imaging-noise-free image with high generalization. CDDM uses a diffusion-based process based on a priori decay distribution to steadily correct imaging noise, thus avoiding the overfitting of individual samples. Within CDDM, a domain-correlated sampling network (DS-Net) provides an innovative sinogram-guided noise prediction scheme to leverage both image and sinogram (i.e., dual-domain) information. DS-Net analyzes the correlation of the dual-domain representations for sampling the noise distribution, introducing sinogram semantics to avoid secondary artifacts. Experimental results validate the practical applicability of our scheme across various CTIR scenarios, including LDCTD, MAR, and SVCTR, with the support of sinogram knowledge.
Collapse
Affiliation(s)
- Xinghua Ma
- The Faculty of Computing, Harbin Institute of Technology, Harbin, Heilongjiang, China; The Computational Bioscience Research Center, King Abdullah University of Science and Technology, Thuwal, Makkah, Saudi Arabia
| | - Mingye Zou
- The Faculty of Computing, Harbin Institute of Technology, Harbin, Heilongjiang, China
| | - Xinyan Fang
- The Faculty of Computing, Harbin Institute of Technology, Harbin, Heilongjiang, China
| | - Gongning Luo
- The Faculty of Computing, Harbin Institute of Technology, Harbin, Heilongjiang, China; The Computational Bioscience Research Center, King Abdullah University of Science and Technology, Thuwal, Makkah, Saudi Arabia
| | - Wei Wang
- The Faculty of Computing, Harbin Institute of Technology, Shenzhen, Guangdong, China.
| | - Suyu Dong
- The College of Computer and Control Engineering, Northeast Forestry University, Harbin, Heilongjiang, China.
| | - Xiangyu Li
- The Faculty of Computing, Harbin Institute of Technology, Harbin, Heilongjiang, China.
| | - Kuanquan Wang
- The Faculty of Computing, Harbin Institute of Technology, Harbin, Heilongjiang, China
| | - Qing Dong
- The Department of Thoracic Surgery at No. 4 Affiliated Hospital, Harbin Medical University, Harbin, Heilongjiang, China
| | - Ye Tian
- The Department of Cardiology at No. 1 Affiliated Hospital, Harbin Medical University, Harbin, Heilongjiang, China
| | - Shuo Li
- The Department of Computer and Data Science, Case Western Reserve University, Cleveland, OH, USA; The Department of Biomedical Engineering, Case Western Reserve University, Cleveland, OH, USA
| |
Collapse
|
9
|
Hu D, Zhang C, Fei X, Yao Y, Xi Y, Liu J, Zhang Y, Coatrieux G, Coatrieux JL, Chen Y. DPI-MoCo: Deep Prior Image Constrained Motion Compensation Reconstruction for 4D CBCT. IEEE TRANSACTIONS ON MEDICAL IMAGING 2025; 44:1243-1256. [PMID: 39423082 DOI: 10.1109/tmi.2024.3483451] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/21/2024]
Abstract
4D cone-beam computed tomography (CBCT) plays a critical role in adaptive radiation therapy for lung cancer. However, extremely sparse sampling projection data will cause severe streak artifacts in 4D CBCT images. Existing deep learning (DL) methods heavily rely on large labeled training datasets which are difficult to obtain in practical scenarios. Restricted by this dilemma, DL models often struggle with simultaneously retaining dynamic motions, removing streak degradations, and recovering fine details. To address the above challenging problem, we introduce a Deep Prior Image Constrained Motion Compensation framework (DPI-MoCo) that decouples the 4D CBCT reconstruction into two sub-tasks including coarse image restoration and structural detail fine-tuning. In the first stage, the proposed DPI-MoCo combines the prior image guidance, generative adversarial network, and contrastive learning to globally suppress the artifacts while maintaining the respiratory movements. After that, to further enhance the local anatomical structures, the motion estimation and compensation technique is adopted. Notably, our framework is performed without the need for paired datasets, ensuring practicality in clinical cases. In the Monte Carlo simulation dataset, the DPI-MoCo achieves competitive quantitative performance compared to the state-of-the-art (SOTA) methods. Furthermore, we test DPI-MoCo in clinical lung cancer datasets, and experiments validate that DPI-MoCo not only restores small anatomical structures and lesions but also preserves motion information.
Collapse
|
10
|
Zhang S, Tuo M, Jin S, Gu Y. An efficient and high-quality scheme for cone-beam CT reconstruction from sparse-view data. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2025; 33:420-435. [PMID: 39973789 DOI: 10.1177/08953996241313121] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/21/2025]
Abstract
Computed tomography (CT) is capable of generating detailed cross-sectional images of the scanned objects non-destructively. So far, CT has become an increasingly vital tool for 3D modelling of cultural relics. Compressed sensing (CS)-based CT reconstruction algorithms, such as the algebraic reconstruction technique (ART) regularized by total variation (TV), enable accurate reconstructions from sparse-view data, which consequently reduces both scanning time and costs. However, the implementation of the ART-TV is considerably slow, particularly in cone-beam reconstruction. In this paper, we propose an efficient and high-quality scheme for cone-beam CT reconstruction based on the traditional ART-TV algorithm. Our scheme employs Joseph's projection method for the computation of the system matrix. By exploiting the geometric symmetry of the cone-beam rays, we are able to compute the weight coefficients of the system matrix for two symmetric rays simultaneously. We then employ multi-threading technology to speed up the reconstruction of ART, and utilize graphics processing units (GPUs) to accelerate the TV minimization. Experimental results demonstrate that, for a typical reconstruction of a 512 × 512 × 512 volume from 60 views of 512 × 512 projection images, our scheme achieves a speedup of 14 × compared to a single-threaded CPU implementation. Furthermore, high-quality reconstructions of ART-TV are obtained by using Joseph's projection compared with that using traditional Siddon's projection.
Collapse
Affiliation(s)
- Shunli Zhang
- School of Information Science and Technology, Northwest University, Xi'an, China
| | - Mingxiu Tuo
- School of Information Science and Technology, Northwest University, Xi'an, China
| | - Siyu Jin
- School of Information Science and Technology, Northwest University, Xi'an, China
| | - Yikuan Gu
- School of Information Science and Technology, Northwest University, Xi'an, China
| |
Collapse
|
11
|
Li Y, Sun X, Wang S, Guo L, Qin Y, Pan J, Chen P. TD-STrans: Tri-domain sparse-view CT reconstruction based on sparse transformer. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2025; 260:108575. [PMID: 39733746 DOI: 10.1016/j.cmpb.2024.108575] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/04/2024] [Revised: 12/15/2024] [Accepted: 12/24/2024] [Indexed: 12/31/2024]
Abstract
BACKGROUND AND OBJECTIVE Sparse-view computed tomography (CT) speeds up scanning and reduces radiation exposure in medical diagnosis. However, when the projection views are severely under-sampled, deep learning-based reconstruction methods often suffer from over-smoothing of the reconstructed images due to the lack of high-frequency information. To address this issue, we introduce frequency domain information into the popular projection-image domain reconstruction, proposing a Tri-Domain sparse-view CT reconstruction model based on Sparse Transformer (TD-STrans). METHODS TD-STrans integrates three essential modules: the projection recovery module completes the sparse-view projection, the Fourier domain filling module mitigates artifacts and over-smoothing by filling in missing high-frequency details; the image refinement module further enhances and preserves image details. Additionally, a multi-domain joint loss function is designed to simultaneously enhance the reconstruction quality in the projection domain, image domain, and frequency domain, thereby further improving the preservation of image details. RESULTS The results of simulation experiments on the lymph node dataset and real experiments on the walnut dataset consistently demonstrate the effectiveness of TD-STrans in artifact removal, suppression of over-smoothing, and preservation of structural fidelity. CONCLUSION The reconstruction results of TD-STrans indicate that sparse transformer across multiple domains can alleviate over-smoothing and detail loss caused by reduced views, offering a novel solution for ultra-sparse-view CT imaging.
Collapse
Affiliation(s)
- Yu Li
- Department of Information and Communication Engineering, North University of China, Taiyuan 030051, China; The State Key Lab for Electronic Testing Technology, North University of China, Taiyuan 030051, China
| | - Xueqin Sun
- Department of Information and Communication Engineering, North University of China, Taiyuan 030051, China; The State Key Lab for Electronic Testing Technology, North University of China, Taiyuan 030051, China
| | - Sukai Wang
- The State Key Lab for Electronic Testing Technology, North University of China, Taiyuan 030051, China; Department of computer science and technology, North University of China, Taiyuan 030051, China
| | - Lina Guo
- Department of Information and Communication Engineering, North University of China, Taiyuan 030051, China; The State Key Lab for Electronic Testing Technology, North University of China, Taiyuan 030051, China
| | - Yingwei Qin
- Department of Information and Communication Engineering, North University of China, Taiyuan 030051, China; The State Key Lab for Electronic Testing Technology, North University of China, Taiyuan 030051, China
| | - Jinxiao Pan
- Department of Information and Communication Engineering, North University of China, Taiyuan 030051, China; The State Key Lab for Electronic Testing Technology, North University of China, Taiyuan 030051, China
| | - Ping Chen
- Department of Information and Communication Engineering, North University of China, Taiyuan 030051, China; The State Key Lab for Electronic Testing Technology, North University of China, Taiyuan 030051, China.
| |
Collapse
|
12
|
Zhang R, Szczykutowicz TP, Toia GV. Artificial Intelligence in Computed Tomography Image Reconstruction: A Review of Recent Advances. J Comput Assist Tomogr 2025:00004728-990000000-00429. [PMID: 40008975 DOI: 10.1097/rct.0000000000001734] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/12/2024] [Accepted: 01/07/2025] [Indexed: 02/27/2025]
Abstract
The development of novel image reconstruction algorithms has been pivotal in enhancing image quality and reducing radiation dose in computed tomography (CT) imaging. Traditional techniques like filtered back projection perform well under ideal conditions but fail to generate high-quality images under low-dose, sparse-view, and limited-angle conditions. Iterative reconstruction methods improve upon filtered back projection by incorporating system models and assumptions about the patient, yet they can suffer from patchy image textures. The emergence of artificial intelligence (AI), particularly deep learning, has further advanced CT reconstruction. AI techniques have demonstrated great potential in reducing radiation dose while preserving image quality and noise texture. Moreover, AI has exhibited unprecedented performance in addressing challenging CT reconstruction problems, including low-dose CT, sparse-view CT, limited-angle CT, and interior tomography. This review focuses on the latest advances in AI-based CT reconstruction under these challenging conditions.
Collapse
Affiliation(s)
- Ran Zhang
- Departments of Radiology and Medical Physics, University of Wisconsin, Madison, WI
| | | | | |
Collapse
|
13
|
Zhang J, Li Z, Pan J, Wang S, Wu W. Trustworthy Limited Data CT Reconstruction Using Progressive Artifact Image Learning. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2025; 34:1163-1178. [PMID: 40031253 DOI: 10.1109/tip.2025.3534559] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/05/2025]
Abstract
The reconstruction of limited data computed tomography (CT) aims to obtain high-quality images from a reduced set of projection views acquired from sparse views or limited angles. This approach is utilized to reduce radiation exposure or expedite the scanning process. Deep Learning (DL) techniques have been incorporated into limited data CT reconstruction tasks and achieve remarkable performance. However, these DL methods suffer from various limitations. Firstly, the distribution inconsistency between the simulation data and the real data hinders the generalization of these DL-based methods. Secondly, these DL-based methods could be unstable due to lack of kernel awareness. This paper addresses these issues by proposing an unrolling framework called Progressive Artifact Image Learning (PAIL) for limited data CT reconstruction. The proposed PAIL primarily consists of three key modules, i.e., a residual domain module (RDM), an image domain module (IDM), and a wavelet domain module (WDM). The RDM is designed to refine features from residual images and suppress the observable artifacts from the reconstructed images. This module could effectively alleviate the effects of distribution inconsistency among different data sets by transferring the optimization space from the original data domain to the residual data domain. The IDM is designed to suppress the unobservable artifacts in the image space. The RDM and IDM collaborate with each other during the iterative optimization process, progressively removing artifacts and reconstructing the underlying CT image. Furthermore, in order to void the potential hallucinations generated by the RDM and IDM, an additional WDM is incorporated into the network to enhance its stability. This is achieved by making the network become kernel-aware via integrating wavelet-based compressed sensing. The effectiveness of the proposed PAIL method has been consistently verified on two simulated CT data sets, a clinical cardiac data set and a sheep lung data set. Compared to other state-of-the-art methods, the proposed PAIL method achieves superior performance in various limited data CT reconstruction tasks, demonstrating its promising generalization and stability.
Collapse
|
14
|
Wu W, Long Y, Gao Z, Yang G, Cheng F, Zhang J. Multi-Level Noise Sampling From Single Image for Low-Dose Tomography Reconstruction. IEEE J Biomed Health Inform 2025; 29:1256-1268. [PMID: 39527416 DOI: 10.1109/jbhi.2024.3486726] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2024]
Abstract
Low-dose digital radiography (DR) and computed tomography (CT) become increasingly popular due to reduced radiation dose. However, they often result in degraded images with lower signal-to-noise ratios, creating an urgent need for effective denoising techniques. The recent advancement of the single-image-based denoising approach provides a promising solution without requirement of pairwise training data, which are scarce in medical imaging. These methods typically rely on sampling image pairs from a noisy image for inter-supervised denoising. Although enjoying simplicity, the generated image pairs are at the same noise level and only include partial information about the input images. This study argues that generating image pairs at different noise levels while fully using the information of the input image is preferable since it could provide richer multi-perspective clues to guide the denoising process. To this end, we present a novel Multi-Level Noise Sampling (MNS) method for low-dose tomography denoising. Specifically, MNS method generates multi-level noisy sub-images by partitioning the high-dimensional input space into multiple low-dimensional sub-spaces with a simple yet effective strategy. The superiority of the MNS method in single-image-based denoising over the competing methods has been investigated and verified theoretically. Moreover, to bridge the gap between self-supervised and supervised denoising networks, we introduce an optimization function that leverages prior knowledge of multi-level noisy sub-images to guide the training process. Through extensive quantitative and qualitative experiments conducted on large-scale clinical low-dose CT and DR datasets, we validate the effectiveness and superiority of our MNS approach over other state-of-the-art supervised and self-supervised methods.
Collapse
|
15
|
Zhao X, Du Y, Peng Y. Deep Learning-Based Multi-View Projection Synthesis Approach for Improving the Quality of Sparse-View CBCT in Image-Guided Radiotherapy. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025:10.1007/s10278-025-01390-0. [PMID: 39849201 DOI: 10.1007/s10278-025-01390-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/09/2024] [Revised: 12/16/2024] [Accepted: 12/19/2024] [Indexed: 01/25/2025]
Abstract
While radiation hazards induced by cone-beam computed tomography (CBCT) in image-guided radiotherapy (IGRT) can be reduced by sparse-view sampling, the image quality is inevitably degraded. We propose a deep learning-based multi-view projection synthesis (DLMPS) approach to improve the quality of sparse-view low-dose CBCT images. In the proposed DLMPS approach, linear interpolation was first applied to sparse-view projections and the projections were rearranged into sinograms; these sinograms were processed with a sinogram restoration model and then rearranged back into projections. The sinogram restoration model was modified from the 2D U-Net by incorporating dynamic convolutional layers and residual learning techniques. The DLMPS approach was trained, validated, and tested on CBCT data from 163, 30, and 30 real patients respectively. Sparse-view projection datasets with 1/4 and 1/8 of the original sampling rate were simulated, and the corresponding full-view projection datasets were restored via the DLMPS approach. Tomographic images were reconstructed using the Feldkamp-Davis-Kress algorithm. Quantitative metrics including root-mean-square error (RMSE), peak signal-to-noise ratio (PSNR), structural similarity (SSIM), and feature similarity (FSIM) were calculated in both the projection and image domains to evaluate the performance of the DLMPS approach. The DLMPS approach was compared with 11 state-of-the-art (SOTA) models, including CNN and Transformer architectures. For 1/4 sparse-view reconstruction task, the proposed DLMPS approach achieved averaged RMSE, PSNR, SSIM, and FSIM values of 0.0271, 45.93 dB, 0.9817, and 0.9587 in the projection domain, and 0.000885, 37.63 dB, 0.9074, and 0.9885 in the image domain, respectively. For 1/8 sparse-view reconstruction task, the DLMPS approach achieved averaged RMSE, PSNR, SSIM, and FSIM values of 0.0304, 44.85 dB, 0.9785, and 0.9524 in the projection domain, and 0.001057, 36.05 dB, 0.8786, and 0.9774 in the image domain, respectively. The DLMPS approach outperformed all the 11 SOTA models in both the projection and image domains for 1/4 and 1/8 sparse-view reconstruction tasks. The proposed DLMPS approach effectively improves the quality of sparse-view CBCT images in IGRT by accurately synthesizing missing projections, exhibiting potential in substantially reducing imaging dose to patients with minimal loss of image quality.
Collapse
Affiliation(s)
- Xuzhi Zhao
- School of Electronic and Information Engineering, Beijing Jiaotong University, Beijing, China
| | - Yi Du
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education/Beijing), Department of Radiation Oncology, Peking University Cancer Hospital & Institute, Beijing, China.
- Institute of Medical Technology, Peking University Health Science Center, Beijing, China.
| | - Yahui Peng
- School of Electronic and Information Engineering, Beijing Jiaotong University, Beijing, China.
| |
Collapse
|
16
|
Sun C, Liu Y, Yang H. An efficient deep unrolling network for sparse-view CT reconstruction via alternating optimization of dense-view sinograms and images. Phys Med Biol 2025; 70:025006. [PMID: 39662047 DOI: 10.1088/1361-6560/ad9dac] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/07/2024] [Accepted: 12/11/2024] [Indexed: 12/13/2024]
Abstract
Objective. Recently, there have been many advancements in deep unrolling methods for sparse-view computed tomography (SVCT) reconstruction. These methods combine model-based and deep learning-based reconstruction techniques, improving the interpretability and achieving significant results. However, they are often computationally expensive, particularly for clinical raw projection data with large sizes. This study aims to address this issue while maintaining the quality of the reconstructed image.Approach. The SVCT reconstruction task is decomposed into two subproblems using the proximal gradient method: optimizing dense-view sinograms and optimizing images. Then dense-view sinogram inpainting, image-residual learning, and image-refinement modules are performed at each iteration stage using deep neural networks. Unlike previous unrolling methods, the proposed method focuses on optimizing dense-view sinograms instead of full-view sinograms. This approach not only reduces computational resources and runtime but also minimizes the challenge for the network to perform sinogram inpainting when the sparse ratio is extremely small, thereby decreasing the propagation of estimation error from the sinogram domain to the image domain.Main results. The proposed method successfully reconstructs an image (512 × 512 pixels) from real-size (2304 × 736) projection data, with 3.39 M training parameters and an inference time of 0.09 s per slice on a GPU. The proposed method also achieves superior quantitative and qualitative results compared with state-of-the-art deep unrolling methods on datasets with sparse ratios of 1/12 and 1/18, especially in suppressing artifacts and preserving structural details. Additionally, results show that using dense-view sinogram inpainting not only accelerates the computational speed but also leads to faster network convergence and further improvements in reconstruction results.Significance. This research presents an efficient dual-domain deep unrolling technique that produces excellent results in SVCT reconstruction while requiring small computational resources. These findings have important implications for speeding up deep unrolling CT reconstruction methods and making them more practical for processing clinical CT projection data.
Collapse
Affiliation(s)
- Chang Sun
- School of Information and Communication Engineering, Beijing University of Posts and Telecommunications, Beijing 100876, People's Republic of China
| | - Yitong Liu
- School of Information and Communication Engineering, Beijing University of Posts and Telecommunications, Beijing 100876, People's Republic of China
| | - Hongwen Yang
- School of Information and Communication Engineering, Beijing University of Posts and Telecommunications, Beijing 100876, People's Republic of China
| |
Collapse
|
17
|
Huang J, Yang L, Wang F, Wu Y, Nan Y, Wu W, Wang C, Shi K, Aviles-Rivero AI, Schönlieb CB, Zhang D, Yang G. Enhancing global sensitivity and uncertainty quantification in medical image reconstruction with Monte Carlo arbitrary-masked mamba. Med Image Anal 2025; 99:103334. [PMID: 39255733 DOI: 10.1016/j.media.2024.103334] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2024] [Revised: 08/05/2024] [Accepted: 09/01/2024] [Indexed: 09/12/2024]
Abstract
Deep learning has been extensively applied in medical image reconstruction, where Convolutional Neural Networks (CNNs) and Vision Transformers (ViTs) represent the predominant paradigms, each possessing distinct advantages and inherent limitations: CNNs exhibit linear complexity with local sensitivity, whereas ViTs demonstrate quadratic complexity with global sensitivity. The emerging Mamba has shown superiority in learning visual representation, which combines the advantages of linear scalability and global sensitivity. In this study, we introduce MambaMIR, an Arbitrary-Masked Mamba-based model with wavelet decomposition for joint medical image reconstruction and uncertainty estimation. A novel Arbitrary Scan Masking (ASM) mechanism "masks out" redundant information to introduce randomness for further uncertainty estimation. Compared to the commonly used Monte Carlo (MC) dropout, our proposed MC-ASM provides an uncertainty map without the need for hyperparameter tuning and mitigates the performance drop typically observed when applying dropout to low-level tasks. For further texture preservation and better perceptual quality, we employ the wavelet transformation into MambaMIR and explore its variant based on the Generative Adversarial Network, namely MambaMIR-GAN. Comprehensive experiments have been conducted for multiple representative medical image reconstruction tasks, demonstrating that the proposed MambaMIR and MambaMIR-GAN outperform other baseline and state-of-the-art methods in different reconstruction tasks, where MambaMIR achieves the best reconstruction fidelity and MambaMIR-GAN has the best perceptual quality. In addition, our MC-ASM provides uncertainty maps as an additional tool for clinicians, while mitigating the typical performance drop caused by the commonly used dropout.
Collapse
Affiliation(s)
- Jiahao Huang
- Bioengineering Department and Imperial-X, Imperial College London, London W12 7SL, United Kingdom; National Heart and Lung Institute, Imperial College London, London SW7 2AZ, United Kingdom; Cardiovascular Research Centre, Royal Brompton Hospital, London SW3 6NP, United Kingdom
| | - Liutao Yang
- College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, Nanjing, China
| | - Fanwen Wang
- Bioengineering Department and Imperial-X, Imperial College London, London W12 7SL, United Kingdom; National Heart and Lung Institute, Imperial College London, London SW7 2AZ, United Kingdom; Cardiovascular Research Centre, Royal Brompton Hospital, London SW3 6NP, United Kingdom
| | - Yinzhe Wu
- Bioengineering Department and Imperial-X, Imperial College London, London W12 7SL, United Kingdom; National Heart and Lung Institute, Imperial College London, London SW7 2AZ, United Kingdom; Cardiovascular Research Centre, Royal Brompton Hospital, London SW3 6NP, United Kingdom
| | - Yang Nan
- Bioengineering Department and Imperial-X, Imperial College London, London W12 7SL, United Kingdom; National Heart and Lung Institute, Imperial College London, London SW7 2AZ, United Kingdom
| | - Weiwen Wu
- School of Biomedical Engineering, Shenzhen Campus of Sun Yat-sen University, Guangdong, China
| | - Chengyan Wang
- Human Phenome Institute, Fudan University, Shanghai, China
| | - Kuangyu Shi
- Department of Nuclear Medicine, Inselspital, University of Bern, Bern, Switzerland; Department of Informatics, Technical University of Munich, Munich, Germany
| | - Angelica I Aviles-Rivero
- Department of Applied Mathematics and Theoretical Physics, University of Cambridge, Cambridge, United Kingdom
| | - Carola-Bibiane Schönlieb
- Department of Applied Mathematics and Theoretical Physics, University of Cambridge, Cambridge, United Kingdom
| | - Daoqiang Zhang
- College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, Nanjing, China
| | - Guang Yang
- Bioengineering Department and Imperial-X, Imperial College London, London W12 7SL, United Kingdom; National Heart and Lung Institute, Imperial College London, London SW7 2AZ, United Kingdom; Cardiovascular Research Centre, Royal Brompton Hospital, London SW3 6NP, United Kingdom; School of Biomedical Engineering & Imaging Sciences, King's College London, London WC2R 2LS, United Kingdom.
| |
Collapse
|
18
|
Amadita K, Gray F, Gee E, Ekpo E, Jimenez Y. CT metal artefact reduction for hip and shoulder implants using novel algorithms and machine learning: A systematic review with pairwise and network meta-analyses. Radiography (Lond) 2025; 31:36-52. [PMID: 39509906 DOI: 10.1016/j.radi.2024.10.009] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/02/2024] [Revised: 09/25/2024] [Accepted: 10/14/2024] [Indexed: 11/15/2024]
Abstract
INTRODUCTION Many tools have been developed to reduce metal artefacts in computed tomography (CT) images resulting from metallic prosthesis; however, their relative effectiveness in preserving image quality is poorly understood. This paper reviews the literature on novel metal artefact reduction (MAR) methods targeting large metal artefacts in fan-beam CT to examine their effectiveness in reducing metal artefacts and effect on image quality. METHODS The PRISMA checklist was used to search for articles in five electronic databases (MEDLINE, Scopus, Web of Science, IEEE, EMBASE). Studies that assessed the effectiveness of recently developed MAR method on fan-beam CT images of hip and shoulder implants were reviewed. Study quality was assessed using the National Institute of Health (NIH) tool. Meta-analyses were conducted in R, and results that could not be meta-analysed were synthesised narratively. RESULTS Thirty-six studies were reviewed. Of these, 20 studies proposed statistical algorithms and 16 used machine learning (ML), and there were 19 novel comparators. Network meta-analysis of 19 studies showed that Recurrent Neural Network MAR (RNN-MAR) is more effective in reducing noise (LogOR 20.7; 95 % CI 12.6-28.9) without compromising image quality (LogOR 4.4; 95 % CI -13.8-22.5). The network meta-analysis and narrative synthesis showed novel MAR methods reduce noise more effectively than baseline algorithms, with five out of 23 ML methods significantly more effective than Filtered Back Projection (FBP) (p < 0.05). Computation time varied, but ML methods were faster than statistical algorithms. CONCLUSION ML tools are more effective in reducing metal artefacts without compromising image quality and are computationally faster than statistical algorithms. Overall, novel MAR methods were also more effective in reducing noise than the baseline reconstructions. IMPLICATIONS FOR PRACTICE Implementation research is needed to establish the clinical suitability of ML MAR in practice.
Collapse
Affiliation(s)
- K Amadita
- Discipline of Medical Imaging Science, Faculty of Medicine and Health, University of Sydney, NSW 2006, Australia.
| | - F Gray
- Discipline of Medical Imaging Science, Faculty of Medicine and Health, University of Sydney, NSW 2006, Australia.
| | - E Gee
- Discipline of Medical Imaging Science, Faculty of Medicine and Health, University of Sydney, NSW 2006, Australia.
| | - E Ekpo
- Discipline of Medical Imaging Science, Faculty of Medicine and Health, University of Sydney, NSW 2006, Australia.
| | - Y Jimenez
- Discipline of Medical Imaging Science, Faculty of Medicine and Health, University of Sydney, NSW 2006, Australia.
| |
Collapse
|
19
|
Zhang Y, Hu D, Li W, Zhang W, Chen G, Chen RC, Chen Y, Gao H. 2V-CBCT: Two-Orthogonal-Projection Based CBCT Reconstruction and Dose Calculation for Radiation Therapy Using Real Projection Data. IEEE TRANSACTIONS ON MEDICAL IMAGING 2025; 44:284-296. [PMID: 39106129 PMCID: PMC11846251 DOI: 10.1109/tmi.2024.3439573] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/09/2024]
Abstract
This work demonstrates the feasibility of two-orthogonal-projection-based CBCT (2V-CBCT) reconstruction and dose calculation for radiation therapy (RT) using real projection data, which is the first 2V-CBCT feasibility study with real projection data, to the best of our knowledge. RT treatments are often delivered in multiple fractions, for which on-board CBCT is desirable to calculate the delivered dose per fraction for the purpose of RT delivery quality assurance and adaptive RT. However, not all RT treatments/fractions have CBCT acquired, but two orthogonal projections are always available. The question to be addressed in this work is the feasibility of 2V-CBCT for the purpose of RT dose calculation. 2V-CBCT is a severely ill-posed inverse problem for which we propose a coarse-to-fine learning strategy. First, a 3D deep neural network that can extract and exploit the inter-slice and intra-slice information is adopted to predict the initial 3D volumes. Then, a 2D deep neural network is utilized to fine-tune the initial 3D volumes slice-by-slice. During the fine-tuning stage, a perceptual loss based on multi-frequency features is employed to enhance the image reconstruction. Dose calculation results from both photon and proton RT demonstrate that 2V-CBCT provides comparable accuracy with full-view CBCT based on real projection data.
Collapse
|
20
|
Zhang X, Fang C, Qiao Z. MAFA-Uformer: Multi-attention and dual-branch feature aggregation U-shaped transformer for sparse-view CT reconstruction. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2025; 33:157-166. [PMID: 39973771 DOI: 10.1177/08953996241300016] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/21/2025]
Abstract
BACKGROUND Although computed tomography (CT) is widely employed in disease detection, X-ray radiation may pose a risk to the health of patients. Reducing the projection views is a common method, however, the reconstructed images often suffer from streak artifacts. PURPOSE In previous related works, it can be found that the convolutional neural network (CNN) is proficient in extracting local features, while the Transformer is adept at capturing global information. To suppress streak artifacts for sparse-view CT, this study aims to develop a method that combines the advantages of CNN and Transformer. METHODS In this paper, we propose a Multi-Attention and Dual-Branch Feature Aggregation U-shaped Transformer network (MAFA-Uformer), which consists of two branches: CNN and Transformer. Firstly, with a coordinate attention mechanism, the Transformer branch can capture the overall structure and orientation information to provide a global context understanding of the image under reconstruction. Secondly, the CNN branch focuses on extracting crucial local features of images through channel spatial attention, thus enhancing detail recognition capabilities. Finally, through a feature fusion module, the global information from the Transformer and the local features from the CNN are integrated effectively. RESULTS Experimental results demonstrate that our method achieves outstanding performance in terms of peak signal-to-noise ratio (PSNR), structural similarity (SSIM), and root mean square error (RMSE). Compared with Restormer, our model achieves significant improvements: PSNR increases by 0.76 dB, SSIM improves by 0.44%, and RMSE decreases by 8.55%. CONCLUSION Our method not only effectively suppresses artifacts but also better preserves details and features, thereby providing robust support for accurate diagnosis of CT images.
Collapse
Affiliation(s)
- Xuan Zhang
- School of Computer and Information Technology, Shanxi University, Taiyuan, Shanxi, China
| | - Chenyun Fang
- School of Computer and Information Technology, Shanxi University, Taiyuan, Shanxi, China
| | - Zhiwei Qiao
- School of Computer and Information Technology, Shanxi University, Taiyuan, Shanxi, China
| |
Collapse
|
21
|
Zhou Y, Chen T, Hou J, Xie H, Dvornek NC, Zhou SK, Wilson DL, Duncan JS, Liu C, Zhou B. Cascaded Multi-path Shortcut Diffusion Model for Medical Image Translation. Med Image Anal 2024; 98:103300. [PMID: 39226710 PMCID: PMC11979896 DOI: 10.1016/j.media.2024.103300] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/28/2024] [Revised: 05/29/2024] [Accepted: 08/06/2024] [Indexed: 09/05/2024]
Abstract
Image-to-image translation is a vital component in medical imaging processing, with many uses in a wide range of imaging modalities and clinical scenarios. Previous methods include Generative Adversarial Networks (GANs) and Diffusion Models (DMs), which offer realism but suffer from instability and lack uncertainty estimation. Even though both GAN and DM methods have individually exhibited their capability in medical image translation tasks, the potential of combining a GAN and DM to further improve translation performance and to enable uncertainty estimation remains largely unexplored. In this work, we address these challenges by proposing a Cascade Multi-path Shortcut Diffusion Model (CMDM) for high-quality medical image translation and uncertainty estimation. To reduce the required number of iterations and ensure robust performance, our method first obtains a conditional GAN-generated prior image that will be used for the efficient reverse translation with a DM in the subsequent step. Additionally, a multi-path shortcut diffusion strategy is employed to refine translation results and estimate uncertainty. A cascaded pipeline further enhances translation quality, incorporating residual averaging between cascades. We collected three different medical image datasets with two sub-tasks for each dataset to test the generalizability of our approach. Our experimental results found that CMDM can produce high-quality translations comparable to state-of-the-art methods while providing reasonable uncertainty estimations that correlate well with the translation error.
Collapse
Affiliation(s)
- Yinchi Zhou
- Department of Biomedical Engineering, Yale University, New Haven, CT, USA.
| | - Tianqi Chen
- Department of Computer Science, University of California Irvine, Irvine, CA, USA
| | - Jun Hou
- Department of Computer Science, University of California Irvine, Irvine, CA, USA
| | - Huidong Xie
- Department of Biomedical Engineering, Yale University, New Haven, CT, USA
| | - Nicha C Dvornek
- Department of Biomedical Engineering, Yale University, New Haven, CT, USA; Department of Radiology and Biomedical Imaging, Yale School of Medicine, New Haven, CT, USA
| | - S Kevin Zhou
- School of Biomedical Engineering & Suzhou Institute for Advanced Research, University of Science and Technology of China, Suzhou, China
| | - David L Wilson
- Department of Biomedical Engineering, Case Western Reserve University, Cleveland, OH, USA
| | - James S Duncan
- Department of Biomedical Engineering, Yale University, New Haven, CT, USA; Department of Radiology and Biomedical Imaging, Yale School of Medicine, New Haven, CT, USA; Department of Electrical Engineering, Yale University, New Haven, CT, USA
| | - Chi Liu
- Department of Biomedical Engineering, Yale University, New Haven, CT, USA; Department of Radiology and Biomedical Imaging, Yale School of Medicine, New Haven, CT, USA
| | - Bo Zhou
- Department of Radiology, Northwestern University, Chicago, IL, USA.
| |
Collapse
|
22
|
Yun S, Lee S, Choi DI, Lee T, Cho S. TMAA-net: tensor-domain multi-planal anti-aliasing network for sparse-view CT image reconstruction. Phys Med Biol 2024; 69:225012. [PMID: 39481239 DOI: 10.1088/1361-6560/ad8da2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/22/2024] [Accepted: 10/31/2024] [Indexed: 11/02/2024]
Abstract
Objective.Among various deep-network-based sparse-view CT image reconstruction studies, the sinogram upscaling network has been predominantly employed to synthesize additional view information. However, the performance of the sinogram-based network is limited in terms of removing aliasing streak artifacts and recovering low-contrast small structures. In this study, we used a view-by-view back-projection (VVBP) tensor-domain network to overcome such limitations of the sinogram-based approaches.Approach.The proposed method offers advantages of addressing the aliasing artifacts directly in the 3D tensor domain over the 2D sinogram. In the tensor-domain network, the multi-planal anti-aliasing modules were used to remove artifacts within the coronal and sagittal tensor planes. In addition, the data-fidelity-based refinement module was also implemented to successively process output images of the tensor network to recover image sharpness and textures.Main result.The proposed method showed outperformance in terms of removing aliasing artifacts and recovering low-contrast details compared to other state-of-the-art sinogram-based networks. The performance was validated for both numerical and clinical projection data in a circular fan-beam CT configuration.Significance.We observed that view-by-view aliasing artifacts in sparse-view CT exhibit distinct patterns within the tensor planes, making them effectively removable in high-dimensional representations. Additionally, we demonstrated that the co-domain characteristics of tensor space processing offer higher generalization performance for aliasing artifact removal compared to conventional sinogram-domain processing.
Collapse
Affiliation(s)
- Sungho Yun
- Department of Nuclear and Quantum Engineering, Korea Advanced Institute of Science and Technology, Daejeon 34141, Republic of Korea
| | - Seoyoung Lee
- Department of Nuclear and Quantum Engineering, Korea Advanced Institute of Science and Technology, Daejeon 34141, Republic of Korea
| | - Da-In Choi
- Department of Nuclear and Quantum Engineering, Korea Advanced Institute of Science and Technology, Daejeon 34141, Republic of Korea
| | - Taewon Lee
- Department of Semiconductor Engineering, Hoseo University, Asan 31499, Republic of Korea
| | - Seungryong Cho
- Department of Nuclear and Quantum Engineering, Korea Advanced Institute of Science and Technology, Daejeon 34141, Republic of Korea
- KAIST Institute for IT Convergence, Korea Advanced Institute of Science and Technology, Daejeon 34141, Republic of Korea
| |
Collapse
|
23
|
Sun C, Salimi Y, Angeliki N, Boudabbous S, Zaidi H. An efficient dual-domain deep learning network for sparse-view CT reconstruction. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 256:108376. [PMID: 39173481 DOI: 10.1016/j.cmpb.2024.108376] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/23/2024] [Revised: 08/02/2024] [Accepted: 08/15/2024] [Indexed: 08/24/2024]
Abstract
BACKGROUND AND OBJECTIVE We develop an efficient deep-learning based dual-domain reconstruction method for sparse-view CT reconstruction with small training parameters and comparable running time. We aim to investigate the model's capability and its clinical value by performing objective and subjective quality assessments using clinical CT projection data acquired on commercial scanners. METHODS We designed two lightweight networks, namely Sino-Net and Img-Net, to restore the projection and image signal from the DD-Net reconstructed images in the projection and image domains, respectively. The proposed network has small training parameters and comparable running time among dual-domain based reconstruction networks and is easy to train (end-to-end). We prospectively collected clinical thoraco-abdominal CT projection data acquired on a Siemens Biograph 128 Edge CT scanner to train and validate the proposed network. Further, we quantitatively evaluated the CT Hounsfield unit (HU) values on 21 organs and anatomic structures, such as the liver, aorta, and ribcage. We also analyzed the noise properties and compared the signal-to-noise ratio (SNR) and the contrast-to-noise ratio (CNR) of the reconstructed images. Besides, two radiologists conducted the subjective qualitative evaluation including the confidence and conspicuity of anatomic structures, and the overall image quality using a 1-5 likert scoring system. RESULTS Objective and subjective evaluation showed that the proposed algorithm achieves competitive results in eliminating noise and artifacts, restoring fine structure details, and recovering edges and contours of anatomic structures using 384 views (1/6 sparse rate). The proposed method exhibited good computational cost performance on clinical projection data. CONCLUSION This work presents an efficient dual-domain learning network for sparse-view CT reconstruction on raw projection data from a commercial scanner. The study also provides insights for designing an organ-based image quality assessment pipeline for sparse-view reconstruction tasks, potentially benefiting organ-specific dose reduction by sparse-view imaging.
Collapse
Affiliation(s)
- Chang Sun
- Beijing University of Posts and Telecommunications, School of Information and Communication Engineering, 100876 Beijing, China; Geneva University Hospital, Division of Nuclear Medicine and Molecular Imaging, CH-1211 Geneva, Switzerland
| | - Yazdan Salimi
- Geneva University Hospital, Division of Nuclear Medicine and Molecular Imaging, CH-1211 Geneva, Switzerland
| | - Neroladaki Angeliki
- Geneva University Hospital, Division of Radiology, CH-1211, Geneva, Switzerland
| | - Sana Boudabbous
- Geneva University Hospital, Division of Radiology, CH-1211, Geneva, Switzerland
| | - Habib Zaidi
- Geneva University Hospital, Division of Nuclear Medicine and Molecular Imaging, CH-1211 Geneva, Switzerland; Department of Nuclear Medicine and Molecular Imaging, University of Groningen, University Medical Center Groningen, Groningen, Netherlands; Department of Nuclear Medicine, University of Southern Denmark, Odense, Denmark; University Research and Innovation Center, Óbuda University, Budapest, Hungary.
| |
Collapse
|
24
|
Xu R, Liu Y, Li Z, Gui Z. Sparse-view CT reconstruction based on group-based sparse representation using weighted guided image filtering. BIOMED ENG-BIOMED TE 2024; 69:431-439. [PMID: 38598849 DOI: 10.1515/bmt-2023-0581] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/12/2023] [Accepted: 03/21/2024] [Indexed: 04/12/2024]
Abstract
OBJECTIVES In the past, guided image filtering (GIF)-based methods often utilized total variation (TV)-based methods to reconstruct guidance images. And they failed to reconstruct the intricate details of complex clinical images accurately. To address these problems, we propose a new sparse-view CT reconstruction method based on group-based sparse representation using weighted guided image filtering. METHODS In each iteration of the proposed algorithm, the result constrained by the group-based sparse representation (GSR) is used as the guidance image. Then, the weighted guided image filtering (WGIF) was used to transfer the important features from the guidance image to the reconstruction of the SART method. RESULTS Three representative slices were tested under 64 projection views, and the proposed method yielded the best visual effect. For the shoulder case, the PSNR can achieve 48.82, which is far superior to other methods. CONCLUSIONS The experimental results demonstrate that our method is more effective in preserving structures, suppressing noise, and reducing artifacts compared to other methods.
Collapse
Affiliation(s)
- Rong Xu
- School of Information and Communication Engineering, 66291 North University of China , Taiyuan, China
| | - Yi Liu
- School of Information and Communication Engineering, 66291 North University of China , Taiyuan, China
| | - Zhiyuan Li
- School of Information and Communication Engineering, 66291 North University of China , Taiyuan, China
| | - Zhiguo Gui
- School of Information and Communication Engineering, 66291 North University of China , Taiyuan, China
| |
Collapse
|
25
|
Wu W, Pan J, Wang Y, Wang S, Zhang J. Multi-Channel Optimization Generative Model for Stable Ultra-Sparse-View CT Reconstruction. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:3461-3475. [PMID: 38466593 DOI: 10.1109/tmi.2024.3376414] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/13/2024]
Abstract
Score-based generative model (SGM) has risen to prominence in sparse-view CT reconstruction due to its impressive generation capability. The consistency of data is crucial in guiding the reconstruction process in SGM-based reconstruction methods. However, the existing data consistency policy exhibits certain limitations. Firstly, it employs partial data from the reconstructed image of the iteration process for image updates, which leads to secondary artifacts with compromising image quality. Moreover, the updates to the SGM and data consistency are considered as distinct stages, disregarding their interdependent relationship. Additionally, the reference image used to compute gradients in the reconstruction process is derived from the intermediate result rather than ground truth. Motivated by the fact that a typical SGM yields distinct outcomes with different random noise inputs, we propose a Multi-channel Optimization Generative Model (MOGM) for stable ultra-sparse-view CT reconstruction by integrating a novel data consistency term into the stochastic differential equation model. Notably, the unique aspect of this data consistency component is its exclusive reliance on original data for effectively confining generation outcomes. Furthermore, we pioneer an inference strategy that traces back from the current iteration result to ground truth, enhancing reconstruction stability through foundational theoretical support. We also establish a multi-channel optimization reconstruction framework, where conventional iterative techniques are employed to seek the reconstruction solution. Quantitative and qualitative assessments on 23 views datasets from numerical simulation, clinical cardiac and sheep's lung underscore the superiority of MOGM over alternative methods. Reconstructing from just 10 and 7 views, our method consistently demonstrates exceptional performance.
Collapse
|
26
|
Xu K, Lu S, Huang B, Wu W, Liu Q. Stage-by-Stage Wavelet Optimization Refinement Diffusion Model for Sparse-View CT Reconstruction. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:3412-3424. [PMID: 38236666 DOI: 10.1109/tmi.2024.3355455] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/29/2024]
Abstract
Diffusion model has emerged as a potential tool to tackle the challenge of sparse-view CT reconstruction, displaying superior performance compared to conventional methods. Nevertheless, these prevailing diffusion models predominantly focus on the sinogram or image domains, which can lead to instability during model training, potentially culminating in convergence towards local minimal solutions. The wavelet transform serves to disentangle image contents and features into distinct frequency-component bands at varying scales, adeptly capturing diverse directional structures. Employing the wavelet transform as a guiding sparsity prior significantly enhances the robustness of diffusion models. In this study, we present an innovative approach named the Stage-by-stage Wavelet Optimization Refinement Diffusion (SWORD) model for sparse-view CT reconstruction. Specifically, we establish a unified mathematical model integrating low-frequency and high-frequency generative models, achieving the solution with an optimization procedure. Furthermore, we perform the low-frequency and high-frequency generative models on wavelet's decomposed components rather than the original sinogram, ensuring the stability of model training. Our method is rooted in established optimization theory, comprising three distinct stages, including low-frequency generation, high-frequency refinement and domain transform. The experimental results demonstrated that the proposed method outperformed existing state-of-the-art methods both quantitatively and qualitatively.
Collapse
|
27
|
Li Z, Chang D, Zhang Z, Luo F, Liu Q, Zhang J, Yang G, Wu W. Dual-Domain Collaborative Diffusion Sampling for Multi-Source Stationary Computed Tomography Reconstruction. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:3398-3411. [PMID: 38941197 DOI: 10.1109/tmi.2024.3420411] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/30/2024]
Abstract
The multi-source stationary CT, where both the detector and X-ray source are fixed, represents a novel imaging system with high temporal resolution that has garnered significant interest. Limited space within the system restricts the number of X-ray sources, leading to sparse-view CT imaging challenges. Recent diffusion models for reconstructing sparse-view CT have generally focused separately on sinogram or image domains. Sinogram-centric models effectively estimate missing projections but may introduce artifacts, lacking mechanisms to ensure image correctness. Conversely, image-domain models, while capturing detailed image features, often struggle with complex data distribution, leading to inaccuracies in projections. Addressing these issues, the Dual-domain Collaborative Diffusion Sampling (DCDS) model integrates sinogram and image domain diffusion processes for enhanced sparse-view reconstruction. This model combines the strengths of both domains in an optimized mathematical framework. A collaborative diffusion mechanism underpins this model, improving sinogram recovery and image generative capabilities. This mechanism facilitates feedback-driven image generation from the sinogram domain and uses image domain results to complete missing projections. Optimization of the DCDS model is further achieved through the alternative direction iteration method, focusing on data consistency updates. Extensive testing, including numerical simulations, real phantoms, and clinical cardiac datasets, demonstrates the DCDS model's effectiveness. It consistently outperforms various state-of-the-art benchmarks, delivering exceptional reconstruction quality and precise sinogram.
Collapse
|
28
|
Rai S, Bhatt JS, Patra SK. An AI-Based Low-Risk Lung Health Image Visualization Framework Using LR-ULDCT. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024; 37:2047-2062. [PMID: 38491236 PMCID: PMC11522248 DOI: 10.1007/s10278-024-01062-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/23/2023] [Revised: 01/18/2024] [Accepted: 02/12/2024] [Indexed: 03/18/2024]
Abstract
In this article, we propose an AI-based low-risk visualization framework for lung health monitoring using low-resolution ultra-low-dose CT (LR-ULDCT). We present a novel deep cascade processing workflow to achieve diagnostic visualization on LR-ULDCT (<0.3 mSv) at par high-resolution CT (HRCT) of 100 mSV radiation technology. To this end, we build a low-risk and affordable deep cascade network comprising three sequential deep processes: restoration, super-resolution (SR), and segmentation. Given degraded LR-ULDCT, the first novel network unsupervisedly learns restoration function from augmenting patch-based dictionaries and residuals. The restored version is then super-resolved (SR) for target (sensor) resolution. Here, we combine perceptual and adversarial losses in novel GAN to establish the closeness between probability distributions of generated SR-ULDCT and restored LR-ULDCT. Thus SR-ULDCT is presented to the segmentation network that first separates the chest portion from SR-ULDCT followed by lobe-wise colorization. Finally, we extract five lobes to account for the presence of ground glass opacity (GGO) in the lung. Hence, our AI-based system provides low-risk visualization of input degraded LR-ULDCT to various stages, i.e., restored LR-ULDCT, restored SR-ULDCT, and segmented SR-ULDCT, and achieves diagnostic power of HRCT. We perform case studies by experimenting on real datasets of COVID-19, pneumonia, and pulmonary edema/congestion while comparing our results with state-of-the-art. Ablation experiments are conducted for better visualizing different operating pipelines. Finally, we present a verification report by fourteen (14) experienced radiologists and pulmonologists.
Collapse
Affiliation(s)
- Swati Rai
- Indian Institute of Information Technology Vadodara, Vadodara, India.
| | - Jignesh S Bhatt
- Indian Institute of Information Technology Vadodara, Vadodara, India
| | | |
Collapse
|
29
|
Liu Y, Zhou X, Wei C, Xu Q. Sparse-View Spectral CT Reconstruction and Material Decomposition Based on Multi-Channel SGM. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:3425-3435. [PMID: 38865221 DOI: 10.1109/tmi.2024.3413085] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2024]
Abstract
In medical applications, the diffusion of contrast agents in tissue can reflect the physiological function of organisms, so it is valuable to quantify the distribution and content of contrast agents in the body over a period. Spectral CT has the advantages of multi-energy projection acquisition and material decomposition, which can quantify K-edge contrast agents. However, multiple repetitive spectral CT scans can cause excessive radiation doses. Sparse-view scanning is commonly used to reduce dose and scan time, but its reconstructed images are usually accompanied by streaking artifacts, which leads to inaccurate quantification of the contrast agents. To solve this problem, an unsupervised sparse-view spectral CT reconstruction and material decomposition algorithm based on the multi-channel score-based generative model (SGM) is proposed in this paper. First, multi-energy images and tissue images are used as multi-channel input data for SGM training. Secondly, the organism is multiply scanned in sparse views, and the trained SGM is utilized to generate multi-energy images and tissue images driven by sparse-view projections. After that, a material decomposition algorithm using tissue images generated by SGM as prior images for solving contrast agent images is established. Finally, the distribution and content of the contrast agents are obtained. The comparison and evaluation of this method are given in this paper, and a series of mouse scanning experiments are carried out to verify the effectiveness of the method.
Collapse
|
30
|
Li G, Deng Z, Ge Y, Luo S. HEAL: High-Frequency Enhanced and Attention-Guided Learning Network for Sparse-View CT Reconstruction. Bioengineering (Basel) 2024; 11:646. [PMID: 39061728 PMCID: PMC11273693 DOI: 10.3390/bioengineering11070646] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/14/2024] [Revised: 06/08/2024] [Accepted: 06/18/2024] [Indexed: 07/28/2024] Open
Abstract
X-ray computed tomography (CT) imaging technology has become an indispensable diagnostic tool in clinical examination. However, it poses a risk of ionizing radiation, making the reduction of radiation dose one of the current research hotspots in CT imaging. Sparse-view imaging, as one of the main methods for reducing radiation dose, has made significant progress in recent years. In particular, sparse-view reconstruction methods based on deep learning have shown promising results. Nevertheless, efficiently recovering image details under ultra-sparse conditions remains a challenge. To address this challenge, this paper proposes a high-frequency enhanced and attention-guided learning Network (HEAL). HEAL includes three optimization strategies to achieve detail enhancement: Firstly, we introduce a dual-domain progressive enhancement module, which leverages fidelity constraints within each domain and consistency constraints across domains to effectively narrow the solution space. Secondly, we incorporate both channel and spatial attention mechanisms to improve the network's feature-scaling process. Finally, we propose a high-frequency component enhancement regularization term that integrates residual learning with direction-weighted total variation, utilizing directional cues to effectively distinguish between noise and textures. The HEAL network is trained, validated and tested under different ultra-sparse configurations of 60 views and 30 views, demonstrating its advantages in reconstruction accuracy and detail enhancement.
Collapse
Affiliation(s)
- Guang Li
- Jiangsu Key Laboratory for Biomaterials and Devices, School of Biological Science and Medical Engineering, Southeast University, Nanjing 210096, China; (G.L.); (Z.D.)
| | - Zhenhao Deng
- Jiangsu Key Laboratory for Biomaterials and Devices, School of Biological Science and Medical Engineering, Southeast University, Nanjing 210096, China; (G.L.); (Z.D.)
| | - Yongshuai Ge
- Research Center for Medical Artificial Intelligence, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
- Paul C Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
- Key Laboratory of Biomedical Imaging Science and System, Chinese Academy of Sciences, Shenzhen 518055, China
| | - Shouhua Luo
- Jiangsu Key Laboratory for Biomaterials and Devices, School of Biological Science and Medical Engineering, Southeast University, Nanjing 210096, China; (G.L.); (Z.D.)
| |
Collapse
|
31
|
Yao L, Wang J, Wu Z, Du Q, Yang X, Li M, Zheng J. Parallel processing model for low-dose computed tomography image denoising. Vis Comput Ind Biomed Art 2024; 7:14. [PMID: 38865022 PMCID: PMC11169366 DOI: 10.1186/s42492-024-00165-8] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/19/2024] [Accepted: 05/20/2024] [Indexed: 06/13/2024] Open
Abstract
Low-dose computed tomography (LDCT) has gained increasing attention owing to its crucial role in reducing radiation exposure in patients. However, LDCT-reconstructed images often suffer from significant noise and artifacts, negatively impacting the radiologists' ability to accurately diagnose. To address this issue, many studies have focused on denoising LDCT images using deep learning (DL) methods. However, these DL-based denoising methods have been hindered by the highly variable feature distribution of LDCT data from different imaging sources, which adversely affects the performance of current denoising models. In this study, we propose a parallel processing model, the multi-encoder deep feature transformation network (MDFTN), which is designed to enhance the performance of LDCT imaging for multisource data. Unlike traditional network structures, which rely on continual learning to process multitask data, the approach can simultaneously handle LDCT images within a unified framework from various imaging sources. The proposed MDFTN consists of multiple encoders and decoders along with a deep feature transformation module (DFTM). During forward propagation in network training, each encoder extracts diverse features from its respective data source in parallel and the DFTM compresses these features into a shared feature space. Subsequently, each decoder performs an inverse operation for multisource loss estimation. Through collaborative training, the proposed MDFTN leverages the complementary advantages of multisource data distribution to enhance its adaptability and generalization. Numerous experiments were conducted on two public datasets and one local dataset, which demonstrated that the proposed network model can simultaneously process multisource data while effectively suppressing noise and preserving fine structures. The source code is available at https://github.com/123456789ey/MDFTN .
Collapse
Affiliation(s)
- Libing Yao
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230026, China
- Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China
| | - Jiping Wang
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230026, China
- Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China
| | - Zhongyi Wu
- Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China.
| | - Qiang Du
- Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China
| | - Xiaodong Yang
- Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China
| | - Ming Li
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230026, China.
- Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China.
| | - Jian Zheng
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230026, China
- Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China
| |
Collapse
|
32
|
Zhang Y, Zhang R, Cao R, Xu F, Jiang F, Meng J, Ma F, Guo Y, Liu J. Unsupervised low-dose CT denoising using bidirectional contrastive network. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 251:108206. [PMID: 38723435 DOI: 10.1016/j.cmpb.2024.108206] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/07/2024] [Revised: 04/16/2024] [Accepted: 04/29/2024] [Indexed: 05/31/2024]
Abstract
BACKGROUND AND OBJECTIVE Low-dose computed tomography (LDCT) scans significantly reduce radiation exposure, but introduce higher levels of noise and artifacts that compromise image quality and diagnostic accuracy. Supervised learning methods have proven effective in denoising LDCT images, but are hampered by the need for large, paired datasets, which pose significant challenges in data acquisition. This study aims to develop a robust unsupervised LDCT denoising method that overcomes the reliance on paired LDCT and normal-dose CT (NDCT) samples, paving the way for more accessible and practical denoising techniques. METHODS We propose a novel unsupervised network model, Bidirectional Contrastive Unsupervised Denoising (BCUD), for LDCT denoising. This model innovatively combines a bidirectional network structure with contrastive learning theory to map the precise mutual correspondence between the noisy LDCT image domain and the clean NDCT image domain. Specifically, we employ dual encoders and discriminators for domain-specific data generation, and use unique projection heads for each domain to adaptively learn customized embedded representations. We then align corresponding features across domains within the learned embedding spaces to achieve effective noise reduction. This approach fundamentally improves the model's ability to match features in latent space, thereby improving noise reduction while preserving fine image detail. RESULTS Through extensive experimental validation on the AAPM-Mayo public dataset and real-world clinical datasets, the proposed BCUD method demonstrated superior performance. It achieved a peak signal-to-noise ratio (PSNR) of 31.387 dB, a structural similarity index measure (SSIM) of 0.886, an information fidelity criterion (IFC) of 2.305, and a visual information fidelity (VIF) of 0.373. Notably, subjective evaluation by radiologists resulted in a mean score of 4.23, highlighting its advantages over existing methods in terms of clinical applicability. CONCLUSIONS This paper presents an innovative unsupervised LDCT denoising method using a bidirectional contrastive network, which greatly improves clinical applicability by eliminating the need for perfectly matched image pairs. The method sets a new benchmark in unsupervised LDCT image denoising, excelling in noise reduction and preservation of fine structural details.
Collapse
Affiliation(s)
- Yuanke Zhang
- School of Computer Science, Qufu Normal University, Rizhao 276826, China; Shandong Provincial Key Laboratory of Data Security and Intelligent Computing, Qufu Normal University, Rizhao 276826, China.
| | - Rui Zhang
- School of Computer Science, Qufu Normal University, Rizhao 276826, China
| | - Rujuan Cao
- School of Computer Science, Qufu Normal University, Rizhao 276826, China
| | - Fan Xu
- School of Computer Science, Qufu Normal University, Rizhao 276826, China
| | - Fengjuan Jiang
- School of Computer Science, Qufu Normal University, Rizhao 276826, China
| | - Jing Meng
- School of Computer Science, Qufu Normal University, Rizhao 276826, China; Shandong Provincial Key Laboratory of Data Security and Intelligent Computing, Qufu Normal University, Rizhao 276826, China
| | - Fei Ma
- School of Computer Science, Qufu Normal University, Rizhao 276826, China; Shandong Provincial Key Laboratory of Data Security and Intelligent Computing, Qufu Normal University, Rizhao 276826, China
| | - Yanfei Guo
- School of Computer Science, Qufu Normal University, Rizhao 276826, China; Shandong Provincial Key Laboratory of Data Security and Intelligent Computing, Qufu Normal University, Rizhao 276826, China
| | - Jianlei Liu
- School of Computer Science, Qufu Normal University, Rizhao 276826, China; Shandong Provincial Key Laboratory of Data Security and Intelligent Computing, Qufu Normal University, Rizhao 276826, China
| |
Collapse
|
33
|
Ries A, Dorosti T, Thalhammer J, Sasse D, Sauter A, Meurer F, Benne A, Lasser T, Pfeiffer F, Schaff F, Pfeiffer D. Improving image quality of sparse-view lung tumor CT images with U-Net. Eur Radiol Exp 2024; 8:54. [PMID: 38698099 PMCID: PMC11065797 DOI: 10.1186/s41747-024-00450-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/20/2023] [Accepted: 02/09/2024] [Indexed: 05/05/2024] Open
Abstract
BACKGROUND We aimed to improve the image quality (IQ) of sparse-view computed tomography (CT) images using a U-Net for lung metastasis detection and determine the best tradeoff between number of views, IQ, and diagnostic confidence. METHODS CT images from 41 subjects aged 62.8 ± 10.6 years (mean ± standard deviation, 23 men), 34 with lung metastasis, 7 healthy, were retrospectively selected (2016-2018) and forward projected onto 2,048-view sinograms. Six corresponding sparse-view CT data subsets at varying levels of undersampling were reconstructed from sinograms using filtered backprojection with 16, 32, 64, 128, 256, and 512 views. A dual-frame U-Net was trained and evaluated for each subsampling level on 8,658 images from 22 diseased subjects. A representative image per scan was selected from 19 subjects (12 diseased, 7 healthy) for a single-blinded multireader study. These slices, for all levels of subsampling, with and without U-Net postprocessing, were presented to three readers. IQ and diagnostic confidence were ranked using predefined scales. Subjective nodule segmentation was evaluated using sensitivity and Dice similarity coefficient (DSC); clustered Wilcoxon signed-rank test was used. RESULTS The 64-projection sparse-view images resulted in 0.89 sensitivity and 0.81 DSC, while their counterparts, postprocessed with the U-Net, had improved metrics (0.94 sensitivity and 0.85 DSC) (p = 0.400). Fewer views led to insufficient IQ for diagnosis. For increased views, no substantial discrepancies were noted between sparse-view and postprocessed images. CONCLUSIONS Projection views can be reduced from 2,048 to 64 while maintaining IQ and the confidence of the radiologists on a satisfactory level. RELEVANCE STATEMENT Our reader study demonstrates the benefit of U-Net postprocessing for regular CT screenings of patients with lung metastasis to increase the IQ and diagnostic confidence while reducing the dose. KEY POINTS • Sparse-projection-view streak artifacts reduce the quality and usability of sparse-view CT images. • U-Net-based postprocessing removes sparse-view artifacts while maintaining diagnostically accurate IQ. • Postprocessed sparse-view CTs drastically increase radiologists' confidence in diagnosing lung metastasis.
Collapse
Affiliation(s)
- Annika Ries
- Chair of Biomedical Physics, Department of Physics, School of Natural Sciences, Technical University of Munich, Garching, 85748, Germany
- Munich Institute of Biomedical Engineering, Technical University of Munich, 85748, Garching, Germany
| | - Tina Dorosti
- Chair of Biomedical Physics, Department of Physics, School of Natural Sciences, Technical University of Munich, Garching, 85748, Germany.
- Munich Institute of Biomedical Engineering, Technical University of Munich, 85748, Garching, Germany.
- Department of Diagnostic and Interventional Radiology, School of Medicine, Klinikum Rechts Der Isar, Technical University of Munich, 81675, Munich, Germany.
| | - Johannes Thalhammer
- Chair of Biomedical Physics, Department of Physics, School of Natural Sciences, Technical University of Munich, Garching, 85748, Germany
- Munich Institute of Biomedical Engineering, Technical University of Munich, 85748, Garching, Germany
- Department of Diagnostic and Interventional Radiology, School of Medicine, Klinikum Rechts Der Isar, Technical University of Munich, 81675, Munich, Germany
- Institute for Advanced Study, Technical University of Munich, 85748, Garching, Germany
| | - Daniel Sasse
- Department of Diagnostic and Interventional Radiology, School of Medicine, Klinikum Rechts Der Isar, Technical University of Munich, 81675, Munich, Germany
| | - Andreas Sauter
- Department of Diagnostic and Interventional Radiology, School of Medicine, Klinikum Rechts Der Isar, Technical University of Munich, 81675, Munich, Germany
| | - Felix Meurer
- Department of Diagnostic and Interventional Radiology, School of Medicine, Klinikum Rechts Der Isar, Technical University of Munich, 81675, Munich, Germany
| | - Ashley Benne
- Department of Diagnostic and Interventional Radiology, School of Medicine, Klinikum Rechts Der Isar, Technical University of Munich, 81675, Munich, Germany
- Institute for Advanced Study, Technical University of Munich, 85748, Garching, Germany
| | - Tobias Lasser
- Munich Institute of Biomedical Engineering, Technical University of Munich, 85748, Garching, Germany
- Computational Imaging and Inverse Problems, Department of Computer Science, School of Computation, Information, and Technology, Technical University of Munich, 85748, Garching, Germany
| | - Franz Pfeiffer
- Chair of Biomedical Physics, Department of Physics, School of Natural Sciences, Technical University of Munich, Garching, 85748, Germany
- Munich Institute of Biomedical Engineering, Technical University of Munich, 85748, Garching, Germany
- Department of Diagnostic and Interventional Radiology, School of Medicine, Klinikum Rechts Der Isar, Technical University of Munich, 81675, Munich, Germany
- Institute for Advanced Study, Technical University of Munich, 85748, Garching, Germany
| | - Florian Schaff
- Chair of Biomedical Physics, Department of Physics, School of Natural Sciences, Technical University of Munich, Garching, 85748, Germany
- Munich Institute of Biomedical Engineering, Technical University of Munich, 85748, Garching, Germany
| | - Daniela Pfeiffer
- Department of Diagnostic and Interventional Radiology, School of Medicine, Klinikum Rechts Der Isar, Technical University of Munich, 81675, Munich, Germany
- Institute for Advanced Study, Technical University of Munich, 85748, Garching, Germany
| |
Collapse
|
34
|
Zhang J, Wang Z, Cao T, Cao G, Ren W, Jiang J. Robust residual-guided iterative reconstruction for sparse-view CT in small animal imaging. Phys Med Biol 2024; 69:105010. [PMID: 38507796 DOI: 10.1088/1361-6560/ad360a] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/15/2023] [Accepted: 03/20/2024] [Indexed: 03/22/2024]
Abstract
Objective. We introduce a robust image reconstruction algorithm named residual-guided Golub-Kahan iterative reconstruction technique (RGIRT) designed for sparse-view computed tomography (CT), which aims at high-fidelity image reconstruction from a limited number of projection views.Approach. RGIRT utilizes an inner-outer dual iteration framework, with a flexible least square QR (FLSQR) algorithm implemented in the inner iteration and a restarted iterative scheme applied in the outer iteration. The inner FLSQR employs a flexible Golub-Kahan bidiagonalization method to reduce the size of the inverse problem, and a weighted generalized cross-validation method to adaptively estimate the regularization hyper-parameter. The inner iteration efficiently yields the intermediate reconstruction result, while the outer iteration minimizes the residual and refines the solution by using the result obtained from the inner iteration.Main results. The reconstruction performance of RGIRT is evaluated and compared to other reference methods (FBPConvNet, SART-TV, and FLSQR) using projection data from both numerical phantoms and real experimental Micro-CT data. The experimental findings, from testing various numbers of projection views and different noise levels, underscore the robustness of RGIRT. Meanwhile, theoretical analysis confirms the convergence of residual for our approach.Significance. We propose a robust iterative reconstruction algorithm for x-ray CT scans with sparse views, thereby shortening scanning time and mitigating excessive ionizing radiation exposure to small animals.
Collapse
Affiliation(s)
- Jianru Zhang
- School of Information Science and Technology, ShanghaiTech University, Shanghai, 201210, People's Republic of China
- School of Mathematics, University of Birmingham, Edgbaston, B15 2TT, United Kingdom
| | - Zhe Wang
- School of Biomedical Engineering & State Key Laboratory of Advanced Medical Materials and Devices, ShanghaiTech University, Shanghai, 201210, People's Republic of China
| | - Tuoyu Cao
- United Imaging Healthcare Co., Ltd, Shanghai, 201807, People's Republic of China
| | - Guohua Cao
- School of Biomedical Engineering & State Key Laboratory of Advanced Medical Materials and Devices, ShanghaiTech University, Shanghai, 201210, People's Republic of China
| | - Wuwei Ren
- School of Information Science and Technology, ShanghaiTech University, Shanghai, 201210, People's Republic of China
| | - Jiahua Jiang
- Institute of Mathematical Science, ShanghaiTech University, Shanghai, 201210, People's Republic of China
- School of Mathematics, University of Birmingham, Edgbaston, B15 2TT, United Kingdom
| |
Collapse
|
35
|
Zhang C, Liu L, Dai J, Liu X, He W, Chan Y, Xie Y, Chi F, Liang X. XTransCT: ultra-fast volumetric CT reconstruction using two orthogonal x-ray projections for image-guided radiation therapy via a transformer network. Phys Med Biol 2024; 69:085010. [PMID: 38471171 DOI: 10.1088/1361-6560/ad3320] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/16/2023] [Accepted: 03/12/2024] [Indexed: 03/14/2024]
Abstract
Objective.The aim of this study was to reconstruct volumetric computed tomography (CT) images in real-time from ultra-sparse two-dimensional x-ray projections, facilitating easier navigation and positioning during image-guided radiation therapy.Approach.Our approach leverages a voxel-sapce-searching Transformer model to overcome the limitations of conventional CT reconstruction techniques, which require extensive x-ray projections and lead to high radiation doses and equipment constraints.Main results.The proposed XTransCT algorithm demonstrated superior performance in terms of image quality, structural accuracy, and generalizability across different datasets, including a hospital set of 50 patients, the large-scale public LIDC-IDRI dataset, and the LNDb dataset for cross-validation. Notably, the algorithm achieved an approximately 300% improvement in reconstruction speed, with a rate of 44 ms per 3D image reconstruction compared to former 3D convolution-based methods.Significance.The XTransCT architecture has the potential to impact clinical practice by providing high-quality CT images faster and with substantially reduced radiation exposure for patients. The model's generalizability suggests it has the potential applicable in various healthcare settings.
Collapse
Affiliation(s)
- Chulong Zhang
- Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, 518055, Guangdong, People's Republic of China
| | - Lin Liu
- Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, 518055, Guangdong, People's Republic of China
| | - Jingjing Dai
- Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, 518055, Guangdong, People's Republic of China
| | - Xuan Liu
- Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, 518055, Guangdong, People's Republic of China
| | - Wenfeng He
- Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, 518055, Guangdong, People's Republic of China
| | - Yinping Chan
- Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, 518055, Guangdong, People's Republic of China
| | - Yaoqin Xie
- Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, 518055, Guangdong, People's Republic of China
| | - Feng Chi
- State Key Laboratory of Oncology in South China, Guangdong Provincial Clinical Research Center for Cancer, Sun Yat-sen University Cancer Center, 510060, People's Republic of China
| | - Xiaokun Liang
- Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, 518055, Guangdong, People's Republic of China
| |
Collapse
|
36
|
Li Y, Feng J, Xiang J, Li Z, Liang D. AIRPORT: A Data Consistency Constrained Deep Temporal Extrapolation Method To Improve Temporal Resolution In Contrast Enhanced CT Imaging. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:1605-1618. [PMID: 38133967 DOI: 10.1109/tmi.2023.3344712] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/24/2023]
Abstract
Typical tomographic image reconstruction methods require that the imaged object is static and stationary during the time window to acquire a minimally complete data set. The violation of this requirement leads to temporal-averaging errors in the reconstructed images. For a fixed gantry rotation speed, to reduce the errors, it is desired to reconstruct images using data acquired over a narrower angular range, i.e., with a higher temporal resolution. However, image reconstruction with a narrower angular range violates the data sufficiency condition, resulting in severe data-insufficiency-induced errors. The purpose of this work is to decouple the trade-off between these two types of errors in contrast-enhanced computed tomography (CT) imaging. We demonstrated that using the developed data consistency constrained deep temporal extrapolation method (AIRPORT), the entire time-varying imaged object can be accurately reconstructed with 40 frames-per-second temporal resolution, the time window needed to acquire a single projection view data using a typical C-arm cone-beam CT system. AIRPORT is applicable to general non-sparse imaging tasks using a single short-scan data acquisition.
Collapse
|
37
|
Jung J, Han J, Han JM, Ko J, Yoon J, Hwang JS, Park JI, Hwang G, Jung JH, Hwang DDJ. Prediction of neovascular age-related macular degeneration recurrence using optical coherence tomography images with a deep neural network. Sci Rep 2024; 14:5854. [PMID: 38462646 PMCID: PMC10925587 DOI: 10.1038/s41598-024-56309-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/27/2023] [Accepted: 03/05/2024] [Indexed: 03/12/2024] Open
Abstract
Neovascular age-related macular degeneration (nAMD) can result in blindness if left untreated, and patients often require repeated anti-vascular endothelial growth factor injections. Although, the treat-and-extend method is becoming popular to reduce vision loss attributed to recurrence, it may pose a risk of overtreatment. This study aimed to develop a deep learning model based on DenseNet201 to predict nAMD recurrence within 3 months after confirming dry-up 1 month following three loading injections in treatment-naïve patients. A dataset of 1076 spectral domain optical coherence tomography (OCT) images from 269 patients diagnosed with nAMD was used. The performance of the model was compared with that of 6 ophthalmologists, using 100 randomly selected samples. The DenseNet201-based model achieved 53.0% accuracy in predicting nAMD recurrence using a single pre-injection image and 60.2% accuracy after viewing all the images immediately after the 1st, 2nd, and 3rd injections. The model outperformed experienced ophthalmologists, with an average accuracy of 52.17% using a single pre-injection image and 53.3% after examining four images before and after three loading injections. In conclusion, the artificial intelligence model demonstrated a promising ability to predict nAMD recurrence using OCT images and outperformed experienced ophthalmologists. These findings suggest that deep learning models can assist in nAMD recurrence prediction, thus improving patient outcomes and optimizing treatment strategies.
Collapse
Affiliation(s)
- Juho Jung
- Department of Applied Artificial Intelligence, Sungkyunkwan University, Seoul, Korea
| | - Jinyoung Han
- Department of Applied Artificial Intelligence, Sungkyunkwan University, Seoul, Korea
- Department of Human-Artificial Intelligence Interaction, Sungkyunkwan University, Seoul, Korea
| | - Jeong Mo Han
- Department of Ophthalmology, Seoul National University College of Medicine, Seoul, Korea
- Kong Eye Center, Seoul, Korea
| | | | | | | | - Ji In Park
- Department of Medicine, Kangwon National University Hospital, Kangwon National University School of Medicine, Chuncheon, Gangwon-do, Korea
| | - Gyudeok Hwang
- Department of Ophthalmology, Hangil Eye Hospital, 35 Bupyeong-Daero, Bupyeong-gu, Incheon, 21388, Korea
| | - Jae Ho Jung
- Department of Ophthalmology, Seoul National University College of Medicine, Seoul, Korea
| | - Daniel Duck-Jin Hwang
- Department of Ophthalmology, Hangil Eye Hospital, 35 Bupyeong-Daero, Bupyeong-gu, Incheon, 21388, Korea.
- Department of Ophthalmology, Catholic Kwandong University College of Medicine, Incheon, Korea.
- Lux Mind, Incheon, Korea.
| |
Collapse
|
38
|
Chen H, Li Q, Zhou L, Li F. Deep learning-based algorithms for low-dose CT imaging: A review. Eur J Radiol 2024; 172:111355. [PMID: 38325188 DOI: 10.1016/j.ejrad.2024.111355] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/19/2023] [Revised: 01/05/2024] [Accepted: 01/31/2024] [Indexed: 02/09/2024]
Abstract
The computed tomography (CT) technique is extensively employed as an imaging modality in clinical settings. The radiation dose of CT, however, is significantly high, thereby raising concerns regarding the potential radiation damage it may cause. The reduction of X-ray exposure dose in CT scanning may result in a significant decline in imaging quality, thereby elevating the risk of missed diagnosis and misdiagnosis. The reduction of CT radiation dose and acquisition of high-quality images to meet clinical diagnostic requirements have always been a critical research focus and challenge in the field of CT. Over the years, scholars have conducted extensive research on enhancing low-dose CT (LDCT) imaging algorithms, among which deep learning-based algorithms have demonstrated superior performance. In this review, we initially introduced the conventional algorithms for CT image reconstruction along with their respective advantages and disadvantages. Subsequently, we provided a detailed description of four aspects concerning the application of deep neural networks in LDCT imaging process: preprocessing in the projection domain, post-processing in the image domain, dual-domain processing imaging, and direct deep learning-based reconstruction (DLR). Furthermore, an analysis was conducted to evaluate the merits and demerits of each method. The commercial and clinical applications of the LDCT-DLR algorithm were also presented in an overview. Finally, we summarized the existing issues pertaining to LDCT-DLR and concluded the paper while outlining prospective trends for algorithmic advancement.
Collapse
Affiliation(s)
- Hongchi Chen
- School of Medical Information Engineering, Gannan Medical University, Ganzhou 341000, China
| | - Qiuxia Li
- School of Medical Information Engineering, Gannan Medical University, Ganzhou 341000, China
| | - Lazhen Zhou
- School of Medical Information Engineering, Gannan Medical University, Ganzhou 341000, China
| | - Fangzuo Li
- School of Medical Information Engineering, Gannan Medical University, Ganzhou 341000, China; Key Laboratory of Prevention and Treatment of Cardiovascular and Cerebrovascular Diseases, Ministry of Education, Gannan Medical University, Ganzhou 341000, China.
| |
Collapse
|
39
|
Kang Y, Liu J, Wu F, Wang K, Qiang J, Hu D, Zhang Y. Deep convolutional dictionary learning network for sparse view CT reconstruction with a group sparse prior. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 244:108010. [PMID: 38199137 DOI: 10.1016/j.cmpb.2024.108010] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/19/2023] [Revised: 12/25/2023] [Accepted: 01/05/2024] [Indexed: 01/12/2024]
Abstract
Purpose Numerous techniques based on deep learning have been utilized in sparse view computed tomography (CT) imaging. Nevertheless, the majority of techniques are instinctively constructed utilizing state-of-the-art opaque convolutional neural networks (CNNs) and lack interpretability. Moreover, CNNs tend to focus on local receptive fields and neglect nonlocal self-similarity prior information. Obtaining diagnostically valuable images from sparsely sampled projections is a challenging and ill-posed task. Method To address this issue, we propose a unique and understandable model named DCDL-GS for sparse view CT imaging. This model relies on a network comprised of convolutional dictionary learning and a nonlocal group sparse prior. To enhance the quality of image reconstruction, we utilize a neural network in conjunction with a statistical iterative reconstruction framework and perform a set number of iterations. Inspired by group sparsity priors, we adopt a novel group thresholding operation to improve the feature representation and constraint ability and obtain a theoretical interpretation. Furthermore, our DCDL-GS model incorporates filtered backprojection (FBP) reconstruction, fast sliding window nonlocal self-similarity operations, and a lightweight and interpretable convolutional dictionary learning network to enhance the applicability of the model. Results The efficiency of our proposed DCDL-GS model in preserving edges and recovering features is demonstrated by the visual results obtained on the LDCT-P and UIH datasets. Compared to the results of the most advanced techniques, the quantitative results are enhanced, with increases of 0.6-0.8 dB for the peak signal-to-noise ratio (PSNR), 0.005-0.01 for the structural similarity index measure (SSIM), and 1-1.3 for the regulated Fréchet inception distance (rFID) on the test dataset. The quantitative results also show the effectiveness of our proposed deep convolution iterative reconstruction module and nonlocal group sparse prior. Conclusion In this paper, we create a consolidated and enhanced mathematical model by integrating projection data and prior knowledge of images into a deep iterative model. The model is more practical and interpretable than existing approaches. The results from the experiment show that the proposed model performs well in comparison to the others.
Collapse
Affiliation(s)
- Yanqin Kang
- College of Computer and Information, Anhui Polytechnic University, Wuhu, China; Key Laboratory of Computer Network and Information Integration (Southeast University) Ministry of Education Nanjing, China
| | - Jin Liu
- College of Computer and Information, Anhui Polytechnic University, Wuhu, China; Key Laboratory of Computer Network and Information Integration (Southeast University) Ministry of Education Nanjing, China.
| | - Fan Wu
- College of Computer and Information, Anhui Polytechnic University, Wuhu, China
| | - Kun Wang
- College of Computer and Information, Anhui Polytechnic University, Wuhu, China
| | - Jun Qiang
- College of Computer and Information, Anhui Polytechnic University, Wuhu, China
| | - Dianlin Hu
- Key Laboratory of Computer Network and Information Integration (Southeast University) Ministry of Education Nanjing, China; School of Computer Science and Engineering, Southeast University, Nanjing, China
| | - Yikun Zhang
- Key Laboratory of Computer Network and Information Integration (Southeast University) Ministry of Education Nanjing, China; School of Computer Science and Engineering, Southeast University, Nanjing, China
| |
Collapse
|
40
|
Liu A, Gang GJ, Stayman JW. Fourier Diffusion for Sparse CT Reconstruction. PROCEEDINGS OF SPIE--THE INTERNATIONAL SOCIETY FOR OPTICAL ENGINEERING 2024; 12925:1292516. [PMID: 39247536 PMCID: PMC11378968 DOI: 10.1117/12.3008622] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/10/2024]
Abstract
Sparse CT reconstruction continues to be an area of interest in a number of novel imaging systems. Many different approaches have been tried including model-based methods, compressed sensing approaches, and most recently deep-learning-based processing. Diffusion models, in particular, have become extremely popular due to their ability to effectively encode rich information about images and to allow for posterior sampling to generate many possible outputs. One drawback of diffusion models is that their recurrent structure tends to be computationally expensive. In this work we apply a new Fourier diffusion approach that permits processing with many fewer time steps than the standard scalar diffusion model. We present an extension of the Fourier diffusion technique and evaluate it in a simulated breast cone-beam CT system with a sparse view acquisition.
Collapse
Affiliation(s)
- Anqi Liu
- Biomedical Engineering, Johns Hopkins University, Baltimore, MD, USA
| | - Grace J Gang
- Radiology, University of Pennsylvania, Philadelphia, PA, USA
| | - J Webster Stayman
- Biomedical Engineering, Johns Hopkins University, Baltimore, MD, USA
| |
Collapse
|
41
|
Leo M, Carcagnì P, Signore L, Corcione F, Benincasa G, Laukkanen MO, Distante C. Convolutional Neural Networks in the Diagnosis of Colon Adenocarcinoma. AI 2024; 5:324-341. [DOI: 10.3390/ai5010016] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/04/2025] Open
Abstract
Colorectal cancer is one of the most lethal cancers because of late diagnosis and challenges in the selection of therapy options. The histopathological diagnosis of colon adenocarcinoma is hindered by poor reproducibility and a lack of standard examination protocols required for appropriate treatment decisions. In the current study, using state-of-the-art approaches on benchmark datasets, we analyzed different architectures and ensembling strategies to develop the most efficient network combinations to improve binary and ternary classification. We propose an innovative two-stage pipeline approach to diagnose colon adenocarcinoma grading from histological images in a similar manner to a pathologist. The glandular regions were first segmented by a transformer architecture with subsequent classification using a convolutional neural network (CNN) ensemble, which markedly improved the learning efficiency and shortened the learning time. Moreover, we prepared and published a dataset for clinical validation of the developed artificial neural network, which suggested the discovery of novel histological phenotypic alterations in adenocarcinoma sections that could have prognostic value. Therefore, AI could markedly improve the reproducibility, efficiency, and accuracy of colon cancer diagnosis, which are required for precision medicine to personalize the treatment of cancer patients.
Collapse
Affiliation(s)
- Marco Leo
- Institute of Applied Sciences and Intelligent Systems (ISASI), National Research Council (CNR) of Italy, 73100 Lecce, Italy
| | - Pierluigi Carcagnì
- Institute of Applied Sciences and Intelligent Systems (ISASI), National Research Council (CNR) of Italy, 73100 Lecce, Italy
| | - Luca Signore
- Dipartimento di Ingegneria per L’Innovazione, Università del Salento, 73100 Lecce, Italy
| | | | | | - Mikko O. Laukkanen
- Department of Translational Medical Sciences, University of Naples Federico II, 80131 Naples, Italy
| | - Cosimo Distante
- Institute of Applied Sciences and Intelligent Systems (ISASI), National Research Council (CNR) of Italy, 73100 Lecce, Italy
- Dipartimento di Ingegneria per L’Innovazione, Università del Salento, 73100 Lecce, Italy
| |
Collapse
|
42
|
Lin J, Li J, Dou J, Zhong L, Di J, Qin Y. Dual-Domain Reconstruction Network Incorporating Multi-Level Wavelet Transform and Recurrent Convolution for Sparse View Computed Tomography Imaging. Tomography 2024; 10:133-158. [PMID: 38250957 PMCID: PMC11154272 DOI: 10.3390/tomography10010011] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/11/2023] [Revised: 01/08/2024] [Accepted: 01/10/2024] [Indexed: 01/23/2024] Open
Abstract
Sparse view computed tomography (SVCT) aims to reduce the number of X-ray projection views required for reconstructing the cross-sectional image of an object. While SVCT significantly reduces X-ray radiation dose and speeds up scanning, insufficient projection data give rise to issues such as severe streak artifacts and blurring in reconstructed images, thereby impacting the diagnostic accuracy of CT detection. To address this challenge, a dual-domain reconstruction network incorporating multi-level wavelet transform and recurrent convolution is proposed in this paper. The dual-domain network is composed of a sinogram domain network (SDN) and an image domain network (IDN). Multi-level wavelet transform is employed in both IDN and SDN to decompose sinograms and CT images into distinct frequency components, which are then processed through separate network branches to recover detailed information within their respective frequency bands. To capture global textures, artifacts, and shallow features in sinograms and CT images, a recurrent convolution unit (RCU) based on convolutional long and short-term memory (Conv-LSTM) is designed, which can model their long-range dependencies through recurrent calculation. Additionally, a self-attention-based multi-level frequency feature normalization fusion (MFNF) block is proposed to assist in recovering high-frequency components by aggregating low-frequency components. Finally, an edge loss function based on the Laplacian of Gaussian (LoG) is designed as the regularization term for enhancing the recovery of high-frequency edge structures. The experimental results demonstrate the effectiveness of our approach in reducing artifacts and enhancing the reconstruction of intricate structural details across various sparse views and noise levels. Our method excels in both performance and robustness, as evidenced by its superior outcomes in numerous qualitative and quantitative assessments, surpassing contemporary state-of-the-art CNNs or Transformer-based reconstruction methods.
Collapse
Affiliation(s)
- Juncheng Lin
- Institute of Advanced Photonics Technology, School of Information Engineering, Guangdong University of Technology, Guangzhou 510006, China; (J.L.); (J.L.); (J.D.); (L.Z.); (Y.Q.)
- Guangdong Provincial Key Laboratory of Information Photonics Technology, Guangdong University of Technology, Guangzhou 510006, China
- Key Laboratory of Photonic Technology for Integrated Sensing and Communication, Ministry of Education, Guangdong University of Technology, Guangzhou 510006, China
| | - Jialin Li
- Institute of Advanced Photonics Technology, School of Information Engineering, Guangdong University of Technology, Guangzhou 510006, China; (J.L.); (J.L.); (J.D.); (L.Z.); (Y.Q.)
- Guangdong Provincial Key Laboratory of Information Photonics Technology, Guangdong University of Technology, Guangzhou 510006, China
- Key Laboratory of Photonic Technology for Integrated Sensing and Communication, Ministry of Education, Guangdong University of Technology, Guangzhou 510006, China
| | - Jiazhen Dou
- Institute of Advanced Photonics Technology, School of Information Engineering, Guangdong University of Technology, Guangzhou 510006, China; (J.L.); (J.L.); (J.D.); (L.Z.); (Y.Q.)
- Guangdong Provincial Key Laboratory of Information Photonics Technology, Guangdong University of Technology, Guangzhou 510006, China
- Key Laboratory of Photonic Technology for Integrated Sensing and Communication, Ministry of Education, Guangdong University of Technology, Guangzhou 510006, China
| | - Liyun Zhong
- Institute of Advanced Photonics Technology, School of Information Engineering, Guangdong University of Technology, Guangzhou 510006, China; (J.L.); (J.L.); (J.D.); (L.Z.); (Y.Q.)
- Guangdong Provincial Key Laboratory of Information Photonics Technology, Guangdong University of Technology, Guangzhou 510006, China
- Key Laboratory of Photonic Technology for Integrated Sensing and Communication, Ministry of Education, Guangdong University of Technology, Guangzhou 510006, China
| | - Jianglei Di
- Institute of Advanced Photonics Technology, School of Information Engineering, Guangdong University of Technology, Guangzhou 510006, China; (J.L.); (J.L.); (J.D.); (L.Z.); (Y.Q.)
- Guangdong Provincial Key Laboratory of Information Photonics Technology, Guangdong University of Technology, Guangzhou 510006, China
- Key Laboratory of Photonic Technology for Integrated Sensing and Communication, Ministry of Education, Guangdong University of Technology, Guangzhou 510006, China
| | - Yuwen Qin
- Institute of Advanced Photonics Technology, School of Information Engineering, Guangdong University of Technology, Guangzhou 510006, China; (J.L.); (J.L.); (J.D.); (L.Z.); (Y.Q.)
- Guangdong Provincial Key Laboratory of Information Photonics Technology, Guangdong University of Technology, Guangzhou 510006, China
- Key Laboratory of Photonic Technology for Integrated Sensing and Communication, Ministry of Education, Guangdong University of Technology, Guangzhou 510006, China
| |
Collapse
|
43
|
Liu P, Fang C, Qiao Z. A dense and U-shaped transformer with dual-domain multi-loss function for sparse-view CT reconstruction. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2024; 32:207-228. [PMID: 38306086 DOI: 10.3233/xst-230184] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/03/2024]
Abstract
OBJECTIVE CT image reconstruction from sparse-view projections is an important imaging configuration for low-dose CT, as it can reduce radiation dose. However, the CT images reconstructed from sparse-view projections by traditional analytic algorithms suffer from severe sparse artifacts. Therefore, it is of great value to develop advanced methods to suppress these artifacts. In this work, we aim to use a deep learning (DL)-based method to suppress sparse artifacts. METHODS Inspired by the good performance of DenseNet and Transformer architecture in computer vision tasks, we propose a Dense U-shaped Transformer (D-U-Transformer) to suppress sparse artifacts. This architecture exploits the advantages of densely connected convolutions in capturing local context and Transformer in modelling long-range dependencies, and applies channel attention to fusion features. Moreover, we design a dual-domain multi-loss function with learned weights for the optimization of the model to further improve image quality. RESULTS Experimental results of our proposed D-U-Transformer yield performance improvements on the well-known Mayo Clinic LDCT dataset over several representative DL-based models in terms of artifact suppression and image feature preservation. Extensive internal ablation experiments demonstrate the effectiveness of the components in the proposed model for sparse-view computed tomography (SVCT) reconstruction. SIGNIFICANCE The proposed method can effectively suppress sparse artifacts and achieve high-precision SVCT reconstruction, thus promoting clinical CT scanning towards low-dose radiation and high-quality imaging. The findings of this work can be applied to denoising and artifact removal tasks in CT and other medical images.
Collapse
Affiliation(s)
- Peng Liu
- School of Computer and Information Technology, Shanxi University, Taiyuan, China
- Department of Big Data and Intelligent Engineering, Shanxi Institute of Technology, Yangquan, China
| | - Chenyun Fang
- School of Computer and Information Technology, Shanxi University, Taiyuan, China
| | - Zhiwei Qiao
- School of Computer and Information Technology, Shanxi University, Taiyuan, China
| |
Collapse
|
44
|
Kim S, Kim B, Lee J, Baek J. Sparsier2Sparse: Self-supervised convolutional neural network-based streak artifacts reduction in sparse-view CT images. Med Phys 2023; 50:7731-7747. [PMID: 37303108 DOI: 10.1002/mp.16552] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/26/2022] [Revised: 05/26/2023] [Accepted: 05/29/2023] [Indexed: 06/13/2023] Open
Abstract
BACKGROUND Sparse-view computed tomography (CT) has attracted a lot of attention for reducing both scanning time and radiation dose. However, sparsely-sampled projection data generate severe streak artifacts in the reconstructed images. In recent decades, many sparse-view CT reconstruction techniques based on fully-supervised learning have been proposed and have shown promising results. However, it is not feasible to acquire pairs of full-view and sparse-view CT images in real clinical practice. PURPOSE In this study, we propose a novel self-supervised convolutional neural network (CNN) method to reduce streak artifacts in sparse-view CT images. METHODS We generate the training dataset using only sparse-view CT data and train CNN based on self-supervised learning. Since the streak artifacts can be estimated using prior images under the same CT geometry system, we acquire prior images by iteratively applying the trained network to given sparse-view CT images. We then subtract the estimated steak artifacts from given sparse-view CT images to produce the final results. RESULTS We validated the imaging performance of the proposed method using extended cardiac-torso (XCAT) and the 2016 AAPM Low-Dose CT Grand Challenge dataset from Mayo Clinic. From the results of visual inspection and modulation transfer function (MTF), the proposed method preserved the anatomical structures effectively and showed higher image resolution compared to the various streak artifacts reduction methods for all projection views. CONCLUSIONS We propose a new framework for streak artifacts reduction when only the sparse-view CT data are given. Although we do not use any information of full-view CT data for CNN training, the proposed method achieved the highest performance in preserving fine details. By overcoming the limitation of dataset requirements on fully-supervised-based methods, we expect that our framework can be utilized in the medical imaging field.
Collapse
Affiliation(s)
- Seongjun Kim
- School of Integrated Technology, Yonsei University, Incheon, South Korea
| | - Byeongjoon Kim
- Department of Artificial Intelligence, College of Computing, Yonsei University, Seoul, South Korea
| | - Jooho Lee
- Department of Artificial Intelligence, College of Computing, Yonsei University, Seoul, South Korea
| | - Jongduk Baek
- Department of Artificial Intelligence, College of Computing, Yonsei University, Seoul, South Korea
- Bareunex Imaging, Inc., Seoul, South Korea
| |
Collapse
|
45
|
Ikuta M, Zhang J. A Deep Convolutional Gated Recurrent Unit for CT Image Reconstruction. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2023; 34:10612-10625. [PMID: 35522637 DOI: 10.1109/tnnls.2022.3169569] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Computed tomography (CT) is one of the most important medical imaging technologies in use today. Most commercial CT products use a technique known as the filtered backprojection (FBP) that is fast and can produce decent image quality when an X-ray dose is high. However, the FBP is not good enough on low-dose X-ray CT imaging because the CT image reconstruction problem becomes more stochastic. A more effective reconstruction technique proposed recently and implemented in a limited number of CT commercial products is an iterative reconstruction (IR). The IR technique is based on a Bayesian formulation of the CT image reconstruction problem with an explicit model of the CT scanning, including its stochastic nature, and a prior model that incorporates our knowledge about what a good CT image should look like. However, constructing such prior knowledge is more complicated than it seems. In this article, we propose a novel neural network for CT image reconstruction. The network is based on the IR formulation and constructed with a recurrent neural network (RNN). Specifically, we transform the gated recurrent unit (GRU) into a neural network performing CT image reconstruction. We call it "GRU reconstruction." This neural network conducts concurrent dual-domain learning. Many deep learning (DL)-based methods in medical imaging are single-domain learning, but dual-domain learning performs better because it learns from both the sinogram and the image domain. In addition, we propose backpropagation through stage (BPTS) as a new RNN backpropagation algorithm. It is similar to the backpropagation through time (BPTT) of an RNN; however, it is tailored for iterative optimization. Results from extensive experiments indicate that our proposed method outperforms conventional model-based methods, single-domain DL methods, and state-of-the-art DL techniques in terms of the root mean squared error (RMSE), the peak signal-to-noise ratio (PSNR), and the structure similarity (SSIM) and in terms of visual appearance.
Collapse
|
46
|
Subbakrishna Adishesha A, Vanselow DJ, La Riviere P, Cheng KC, Huang SX. Sinogram domain angular upsampling of sparse-view micro-CT with dense residual hierarchical transformer and attention-weighted loss. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2023; 242:107802. [PMID: 37738839 PMCID: PMC11158828 DOI: 10.1016/j.cmpb.2023.107802] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/08/2023] [Revised: 07/23/2023] [Accepted: 09/04/2023] [Indexed: 09/24/2023]
Abstract
Reduced angular sampling is a key strategy for increasing scanning efficiency of micron-scale computed tomography (micro-CT). Despite boosting throughput, this strategy introduces noise and extrapolation artifacts due to undersampling. In this work, we present a solution to this issue, by proposing a novel Dense Residual Hierarchical Transformer (DRHT) network to recover high-quality sinograms from 2×, 4× and 8× undersampled scans. DRHT is trained to utilize limited information available from sparsely angular sampled scans and once trained, it can be applied to recover higher-resolution sinograms from shorter scan sessions. Our proposed DRHT model aggregates the benefits of a hierarchical- multi-scale structure along with the combination of local and global feature extraction through dense residual convolutional blocks and non-overlapping window transformer blocks respectively. We also propose a novel noise-aware loss function named KL-L1 to improve sinogram restoration to full resolution. KL-L1, a weighted combination of pixel-level and distribution-level cost functions, leverages inconsistencies in noise distribution and uses learnable spatial weight maps to improve the training of the DRHT model. We present ablation studies and evaluations of our method against other state-of-the-art (SOTA) models over multiple datasets. Our proposed DRHT network achieves an average increase in peak signal to noise ratio (PSNR) of 17.73 dB and a structural similarity index (SSIM) of 0.161, for 8× upsampling, across the three diverse datasets, compared to their respective Bicubic interpolated versions. This novel approach can be utilized to decrease radiation exposure to patients and reduce imaging time for large-scale CT imaging projects.
Collapse
Affiliation(s)
| | - Daniel J Vanselow
- Penn State College of Medicine, The Jake Gittlen Laboratories for Cancer Research, Hershey, 17033, PA, USA; Penn State College of Medicine, Division of Experimental Pathology, Department of Pathology, Hershey, 17033, PA, USA.
| | - Patrick La Riviere
- University of Chicago, Department of Radiology, Chicago, 60637, IL, USA.
| | - Keith C Cheng
- Penn State College of Medicine, The Jake Gittlen Laboratories for Cancer Research, Hershey, 17033, PA, USA; Penn State College of Medicine, Division of Experimental Pathology, Department of Pathology, Hershey, 17033, PA, USA.
| | - Sharon X Huang
- Pennsylvania State University, College of Information Sciences and Technology, University Park, 16802, PA, USA.
| |
Collapse
|
47
|
Hu D, Zhang Y, Li W, Zhang W, Reddy K, Ding Q, Zhang X, Chen Y, Gao H. SEA-Net: Structure-Enhanced Attention Network for Limited-Angle CBCT Reconstruction of Clinical Projection Data. IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT 2023; 72:4507613. [PMID: 38957474 PMCID: PMC11218899 DOI: 10.1109/tim.2023.3318712] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/04/2024]
Abstract
This work aims to improve limited-angle (LA) cone beam computed tomography (CBCT) by developing deep learning (DL) methods for real clinical CBCT projection data, which is the first feasibility study of clinical-projection-data-based LA-CBCT, to the best of our knowledge. In radiation therapy (RT), CBCT is routinely used as the on-board imaging modality for patient setup. Compared to diagnostic CT, CBCT has a long acquisition time, e.g., 60 seconds for a full 360° rotation, which is subject to the motion artifact. Therefore, the LA-CBCT, if achievable, is of the great interest for the purpose of RT, for its proportionally reduced scanning time in addition to the radiation dose. However, LA-CBCT suffers from severe wedge artifacts and image distortions. Targeting at real clinical projection data, we have explored various DL methods such as image/data/hybrid-domain methods and finally developed a so-called Structure-Enhanced Attention Network (SEA-Net) method that has the best image quality from clinical projection data among the DL methods we have implemented. Specifically, the proposed SEA-Net employs a specialized structure enhancement sub-network to promote texture preservation. Based on the observation that the distribution of wedge artifacts in reconstruction images is non-uniform, the spatial attention module is utilized to emphasize the relevant regions while ignores the irrelevant ones, which leads to more accurate texture restoration.
Collapse
Affiliation(s)
- Dianlin Hu
- Department of Radiation Oncology, University of Kansas Medical Center, KS 66160, USA
- Jiangsu Provincial Joint International Research Laboratory of Medical Information Processing, the Laboratory of Image Science and Technology, the School of Computer Science and Engineering, and the Key Laboratory of New Generation Artificial Intelligence Technology and Its Interdisciplinary Applications (Southeast University), Ministry of Education, Nanjing 210096, China
| | - Yikun Zhang
- Department of Radiation Oncology, University of Kansas Medical Center, KS 66160, USA
- Jiangsu Provincial Joint International Research Laboratory of Medical Information Processing, the Laboratory of Image Science and Technology, the School of Computer Science and Engineering, and the Key Laboratory of New Generation Artificial Intelligence Technology and Its Interdisciplinary Applications (Southeast University), Ministry of Education, Nanjing 210096, China
| | - Wangyao Li
- Department of Radiation Oncology, University of Kansas Medical Center, KS 66160, USA
| | - Weijie Zhang
- Department of Radiation Oncology, University of Kansas Medical Center, KS 66160, USA
| | - Krishna Reddy
- Department of Radiation Oncology, University of Kansas Medical Center, KS 66160, USA
| | - Qiaoqiao Ding
- Institute of Natural Sciences & School of Mathematical Sciences & MOE-LSC & SJTU-GenSci Joint Lab, Shanghai Jiao Tong University, Shanghai, 200240, China
| | - Xiaoqun Zhang
- Institute of Natural Sciences & School of Mathematical Sciences & MOE-LSC & SJTU-GenSci Joint Lab, Shanghai Jiao Tong University, Shanghai, 200240, China
| | - Yang Chen
- Jiangsu Provincial Joint International Research Laboratory of Medical Information Processing, the Laboratory of Image Science and Technology, the School of Computer Science and Engineering, and the Key Laboratory of New Generation Artificial Intelligence Technology and Its Interdisciplinary Applications (Southeast University), Ministry of Education, Nanjing 210096, China
| | - Hao Gao
- Department of Radiation Oncology, University of Kansas Medical Center, KS 66160, USA
| |
Collapse
|
48
|
Li Q, Li R, Wang T, Cheng Y, Qiang Y, Wu W, Zhao J, Zhang D. A cascade-based dual-domain data correction network for sparse view CT image reconstruction. Comput Biol Med 2023; 165:107345. [PMID: 37603960 DOI: 10.1016/j.compbiomed.2023.107345] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/03/2023] [Revised: 07/18/2023] [Accepted: 08/07/2023] [Indexed: 08/23/2023]
Abstract
Computed tomography (CT) provides non-invasive anatomical structures of the human body and is also widely used for clinical diagnosis, but excessive ionizing radiation in X-rays can cause harm to the human body. Therefore, the researchers obtained sparse sinograms reconstructed sparse view CT images (SVCT) by reducing the amount of X-ray projection, thereby reducing the radiological effects caused by radiation. This paper proposes a cascade-based dual-domain data correction network (CDDCN), which can effectively combine the complementary information contained in the sinogram domain and the image domain to reconstruct high-quality CT images from sparse view sinograms. Specifically, several encoder-decoder subnets are cascaded in the sinogram domain to reconstruct artifact-free and noise-free CT images. In the encoder-decoder subnets, spatial-channel domain learning is designed to achieve efficient feature fusion through a group merging structure, providing continuous and elaborate pixel-level features and improving feature extraction efficiency. At the same time, to ensure that the original sinogram data collected can be retained, a sinogram data consistency layer is proposed to ensure the fidelity of the sinogram data. To further maintain the consistency between the reconstructed image and the reference image, a multi-level composite loss function is designed for regularization to compensate for excessive smoothing and distortion of the image caused by pixel loss and preserve image details and texture. Quantitative and qualitative analysis shows that CDDCN achieves competitive results in artifact removal, edge preservation, detail restoration, and visual improvement for sparsely sampled data under different views.
Collapse
Affiliation(s)
- Qing Li
- College of Computer Science and Technology (College of Data Science), Taiyuan University of Technology, Taiyuan, 030024, China
| | - Runrui Li
- College of Computer Science and Technology (College of Data Science), Taiyuan University of Technology, Taiyuan, 030024, China
| | - Tao Wang
- College of Computer Science and Technology (College of Data Science), Taiyuan University of Technology, Taiyuan, 030024, China
| | - Yubin Cheng
- College of Computer Science and Technology (College of Data Science), Taiyuan University of Technology, Taiyuan, 030024, China
| | - Yan Qiang
- College of Computer Science and Technology (College of Data Science), Taiyuan University of Technology, Taiyuan, 030024, China.
| | - Wei Wu
- Department of Clinical Laboratory, Affiliated People's Hospital of Shanxi Medical University, Shanxi Provincial People's Hospital, Taiyuan, 030012, China
| | - Juanjuan Zhao
- College of Computer Science and Technology (College of Data Science), Taiyuan University of Technology, Taiyuan, 030024, China; School of Information Engineering, Jinzhong College of Information, Jinzhong, 030800, China
| | - Dongxu Zhang
- College of Computer Science and Technology (College of Data Science), Taiyuan University of Technology, Taiyuan, 030024, China
| |
Collapse
|
49
|
Duan X, Ding XF, Li N, Wu FX, Chen X, Zhu N. Sparse2Noise: Low-dose synchrotron X-ray tomography without high-quality reference data. Comput Biol Med 2023; 165:107473. [PMID: 37690288 DOI: 10.1016/j.compbiomed.2023.107473] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2023] [Revised: 08/30/2023] [Accepted: 09/04/2023] [Indexed: 09/12/2023]
Abstract
BACKGROUND Synchrotron radiation computed tomography (SR-CT) holds promise for high-resolution in vivo imaging. Notably, the reconstruction of SR-CT images necessitates a large set of data to be captured with sufficient photons from multiple angles, resulting in high radiation dose received by the object. Reducing the number of projections and/or photon flux is a straightforward means to lessen the radiation dose, however, compromises data completeness, thus introducing noises and artifacts. Deep learning (DL)-based supervised methods effectively denoise and remove artifacts, but they heavily depend on high-quality paired data acquired at high doses. Although algorithms exist for training without high-quality references, they struggle to effectively eliminate persistent artifacts present in real-world data. METHODS This work presents a novel low-dose imaging strategy namely Sparse2Noise, which combines the reconstruction data from paired sparse-view CT scan (normal-flux) and full-view CT scan (low-flux) using a convolutional neural network (CNN). Sparse2Noise does not require high-quality reconstructed data as references and allows for fresh training on data with very small size. Sparse2Noise was evaluated by both simulated and experimental data. RESULTS Sparse2Noise effectively reduces noise and ring artifacts while maintaining high image quality, outperforming state-of-the-art image denoising methods at same dose levels. Furthermore, Sparse2Noise produces impressive high image quality for ex vivo rat hindlimb imaging with the acceptable low radiation dose (i.e., 0.5 Gy with the isotropic voxel size of 26 μm). CONCLUSIONS This work represents a significant advance towards in vivo SR-CT imaging. It is noteworthy that Sparse2Noise can also be used for denoising in conventional CT and/or phase-contrast CT.
Collapse
Affiliation(s)
- Xiaoman Duan
- Division of Biomedical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada
| | - Xiao Fan Ding
- Division of Biomedical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada
| | - Naitao Li
- Division of Biomedical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada
| | - Fang-Xiang Wu
- Division of Biomedical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada; Department of Computer Science, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada; Department of Mechanical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada
| | - Xiongbiao Chen
- Division of Biomedical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada; Department of Mechanical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada.
| | - Ning Zhu
- Division of Biomedical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada; Canadian Light Source, Saskatoon, S7N 2V3, SK, Canada; Department of Chemical and Biological Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada.
| |
Collapse
|
50
|
Amirian M, Montoya-Zegarra JA, Herzig I, Eggenberger Hotz P, Lichtensteiger L, Morf M, Züst A, Paysan P, Peterlik I, Scheib S, Füchslin RM, Stadelmann T, Schilling FP. Mitigation of motion-induced artifacts in cone beam computed tomography using deep convolutional neural networks. Med Phys 2023; 50:6228-6242. [PMID: 36995003 DOI: 10.1002/mp.16405] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/04/2022] [Revised: 02/25/2023] [Accepted: 03/19/2023] [Indexed: 03/31/2023] Open
Abstract
BACKGROUND Cone beam computed tomography (CBCT) is often employed on radiation therapy treatment devices (linear accelerators) used in image-guided radiation therapy (IGRT). For each treatment session, it is necessary to obtain the image of the day in order to accurately position the patient and to enable adaptive treatment capabilities including auto-segmentation and dose calculation. Reconstructed CBCT images often suffer from artifacts, in particular those induced by patient motion. Deep-learning based approaches promise ways to mitigate such artifacts. PURPOSE We propose a novel deep-learning based approach with the goal to reduce motion induced artifacts in CBCT images and improve image quality. It is based on supervised learning and includes neural network architectures employed as pre- and/or post-processing steps during CBCT reconstruction. METHODS Our approach is based on deep convolutional neural networks which complement the standard CBCT reconstruction, which is performed either with the analytical Feldkamp-Davis-Kress (FDK) method, or with an iterative algebraic reconstruction technique (SART-TV). The neural networks, which are based on refined U-net architectures, are trained end-to-end in a supervised learning setup. Labeled training data are obtained by means of a motion simulation, which uses the two extreme phases of 4D CT scans, their deformation vector fields, as well as time-dependent amplitude signals as input. The trained networks are validated against ground truth using quantitative metrics, as well as by using real patient CBCT scans for a qualitative evaluation by clinical experts. RESULTS The presented novel approach is able to generalize to unseen data and yields significant reductions in motion induced artifacts as well as improvements in image quality compared with existing state-of-the-art CBCT reconstruction algorithms (up to +6.3 dB and +0.19 improvements in peak signal-to-noise ratio, PSNR, and structural similarity index measure, SSIM, respectively), as evidenced by validation with an unseen test dataset, and confirmed by a clinical evaluation on real patient scans (up to 74% preference for motion artifact reduction over standard reconstruction). CONCLUSIONS For the first time, it is demonstrated, also by means of clinical evaluation, that inserting deep neural networks as pre- and post-processing plugins in the existing 3D CBCT reconstruction and trained end-to-end yield significant improvements in image quality and reduction of motion artifacts.
Collapse
Affiliation(s)
- Mohammadreza Amirian
- Centre for Artificial Intelligence CAI, Zurich University of Applied Sciences ZHAW, Winterthur, Switzerland
- Institute of Neural Information Processing, Ulm University, Ulm, Germany
| | - Javier A Montoya-Zegarra
- Centre for Artificial Intelligence CAI, Zurich University of Applied Sciences ZHAW, Winterthur, Switzerland
| | - Ivo Herzig
- Institute for Applied Mathematics and Physics IAMP, Zurich University of Applied Sciences ZHAW, Winterthur, Switzerland
| | - Peter Eggenberger Hotz
- Institute for Applied Mathematics and Physics IAMP, Zurich University of Applied Sciences ZHAW, Winterthur, Switzerland
| | - Lukas Lichtensteiger
- Institute for Applied Mathematics and Physics IAMP, Zurich University of Applied Sciences ZHAW, Winterthur, Switzerland
| | - Marco Morf
- Institute for Applied Mathematics and Physics IAMP, Zurich University of Applied Sciences ZHAW, Winterthur, Switzerland
| | - Alexander Züst
- Institute for Applied Mathematics and Physics IAMP, Zurich University of Applied Sciences ZHAW, Winterthur, Switzerland
| | - Pascal Paysan
- Varian Medical Systems Imaging Laboratory GmbH, Baden, Switzerland
| | - Igor Peterlik
- Varian Medical Systems Imaging Laboratory GmbH, Baden, Switzerland
| | - Stefan Scheib
- Varian Medical Systems Imaging Laboratory GmbH, Baden, Switzerland
| | - Rudolf Marcel Füchslin
- Institute for Applied Mathematics and Physics IAMP, Zurich University of Applied Sciences ZHAW, Winterthur, Switzerland
- European Centre for Living Technology, Venice, Italy
| | - Thilo Stadelmann
- Centre for Artificial Intelligence CAI, Zurich University of Applied Sciences ZHAW, Winterthur, Switzerland
- European Centre for Living Technology, Venice, Italy
| | - Frank-Peter Schilling
- Centre for Artificial Intelligence CAI, Zurich University of Applied Sciences ZHAW, Winterthur, Switzerland
| |
Collapse
|