1
|
Zhao H, Zhu W, Jin L, Xiong Y, Deng X, Li Y, Zou W. Calcium deblooming in coronary computed tomography angiography via semantic-oriented generative adversarial network. Comput Med Imaging Graph 2025; 122:102515. [PMID: 40020506 DOI: 10.1016/j.compmedimag.2025.102515] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/16/2024] [Revised: 01/09/2025] [Accepted: 02/17/2025] [Indexed: 03/03/2025]
Abstract
Calcium blooming artifact produced by calcified plaque in coronary computed tomography angiography (CCTA) is a significant contributor to false-positive results for radiologists. Most previous research focused on general noise reduction of CT images, while performance was limited when facing the blooming artifact. To address this problem, we designed an automated and robust semantics-oriented adversarial network that fully exploits the calcified plaques as semantic regions in the CCTA. The semantic features were extracted using a feature extraction module and implemented through a global-local fusion module, a generator with a semantic similarity module, and a matrix discriminator. The effectiveness of our network was validated both on a virtual and a clinical dataset. The clinical dataset consists of 372 CCTA and corresponding coronary angiogram (CAG) results, with the assistance of two cardiac radiologists (with 10 and 21 years of experience) for clinical evaluation. The proposed method effectively reduces artifacts for three major coronary arteries and significantly improves the specificity and positive predictive value for the diagnosis of coronary stenosis.
Collapse
Affiliation(s)
- Huiyu Zhao
- State Key Laboratory of Advanced Optical Communication Systems and Networks, Intelligent Microwave Lightwave Integration Innovation Center (imLic), Department of Electronic Engineering, Shanghai Jiao Tong University, Shanghai, 200240, China.
| | - Wangshu Zhu
- Institute of Diagnostic and Interventional Radiology, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, #600 Yishan Rd, Shanghai, China.
| | - Luyuan Jin
- State Key Laboratory of Advanced Optical Communication Systems and Networks, Intelligent Microwave Lightwave Integration Innovation Center (imLic), Department of Electronic Engineering, Shanghai Jiao Tong University, Shanghai, 200240, China
| | - Yijia Xiong
- Institute of Diagnostic and Interventional Radiology, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, #600 Yishan Rd, Shanghai, China
| | - Xiao Deng
- State Key Laboratory of Advanced Optical Communication Systems and Networks, Intelligent Microwave Lightwave Integration Innovation Center (imLic), Department of Electronic Engineering, Shanghai Jiao Tong University, Shanghai, 200240, China
| | - Yuehua Li
- Institute of Diagnostic and Interventional Radiology, Shanghai Sixth People's Hospital Affiliated to Shanghai Jiao Tong University School of Medicine, #600 Yishan Rd, Shanghai, China.
| | - Weiwen Zou
- State Key Laboratory of Advanced Optical Communication Systems and Networks, Intelligent Microwave Lightwave Integration Innovation Center (imLic), Department of Electronic Engineering, Shanghai Jiao Tong University, Shanghai, 200240, China.
| |
Collapse
|
2
|
Zhang MX, Liu PF, Zhang MD, Su PG, Shang HS, Zhu JT, Wang DY, Ji XY, Liao QM. Deep learning in nuclear medicine: from imaging to therapy. Ann Nucl Med 2025; 39:424-440. [PMID: 40080372 DOI: 10.1007/s12149-025-02031-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/25/2024] [Accepted: 02/24/2025] [Indexed: 03/15/2025]
Abstract
BACKGROUND Deep learning, a leading technology in artificial intelligence (AI), has shown remarkable potential in revolutionizing nuclear medicine. OBJECTIVE This review presents recent advancements in deep learning applications, particularly in nuclear medicine imaging, lesion detection, and radiopharmaceutical therapy. RESULTS Leveraging various neural network architectures, deep learning has significantly enhanced the accuracy of image reconstruction, lesion segmentation, and diagnosis, improving the efficiency of disease detection and treatment planning. The integration of deep learning with functional imaging techniques such as positron emission tomography (PET) and single-photon emission computed tomography (SPECT) enable more precise diagnostics, while facilitating the development of personalized treatment strategies. Despite its promising outlook, there are still some limitations and challenges, particularly in model interpretability, generalization across diverse datasets, multimodal data fusion, and the ethical and legal issues faced in its application. CONCLUSION As technological advancements continue, deep learning is poised to drive substantial changes in nuclear medicine, particularly in the areas of precision healthcare, real-time treatment monitoring, and clinical decision-making. Future research will likely focus on overcoming these challenges and further enhancing model transparency, thus improving clinical applicability.
Collapse
Affiliation(s)
- Meng-Xin Zhang
- Department of Microbiology and Immunology, Henan Provincial Research Center of Engineering Technology for Nuclear Protein Medical Detection, Zhengzhou Health College, Zhengzhou, 45000, Henan, China
- Department of Nuclear Medicine, Henan International Joint Laboratory for Nuclear Protein Regulation, The First Affiliated Hospital, Henan University College of Medicine, Ximen St, Kaifeng, 475004, Henan, China
| | - Peng-Fei Liu
- Department of Microbiology and Immunology, Henan Provincial Research Center of Engineering Technology for Nuclear Protein Medical Detection, Zhengzhou Health College, Zhengzhou, 45000, Henan, China
- Department of Nuclear Medicine, Henan International Joint Laboratory for Nuclear Protein Regulation, The First Affiliated Hospital, Henan University College of Medicine, Ximen St, Kaifeng, 475004, Henan, China
| | - Meng-Di Zhang
- Department of Microbiology and Immunology, Henan Provincial Research Center of Engineering Technology for Nuclear Protein Medical Detection, Zhengzhou Health College, Zhengzhou, 45000, Henan, China
- Department of Nuclear Medicine, Henan International Joint Laboratory for Nuclear Protein Regulation, The First Affiliated Hospital, Henan University College of Medicine, Ximen St, Kaifeng, 475004, Henan, China
| | - Pei-Gen Su
- Department of Microbiology and Immunology, Henan Provincial Research Center of Engineering Technology for Nuclear Protein Medical Detection, Zhengzhou Health College, Zhengzhou, 45000, Henan, China
- School of Medical Technology, Qiqihar Medical University, Qiqihar, 161006, Heilongjiang, China
| | - He-Shan Shang
- Department of Microbiology and Immunology, Henan Provincial Research Center of Engineering Technology for Nuclear Protein Medical Detection, Zhengzhou Health College, Zhengzhou, 45000, Henan, China
- School of Computer and Information Engineering, Henan University, Kaifeng, 475004, Henan, China
| | - Jiang-Tao Zhu
- Faculty of Basic Medical Subjects, Shu-Qing Medical College of Zhengzhou, Zhengzhou, 450064, Henan, China.
- Department of Surgery, Faculty of Clinical Medicine, Zhengzhou Shu-Qing Medical College, Gongming Rd, Mazhai Town, Zhengzhou, 450064, Henan, China.
| | - Da-Yong Wang
- Department of Microbiology and Immunology, Henan Provincial Research Center of Engineering Technology for Nuclear Protein Medical Detection, Zhengzhou Health College, Zhengzhou, 45000, Henan, China.
- Department of Nuclear Medicine, Henan International Joint Laboratory for Nuclear Protein Regulation, The First Affiliated Hospital, Henan University College of Medicine, Ximen St, Kaifeng, 475004, Henan, China.
| | - Xin-Ying Ji
- Department of Microbiology and Immunology, Henan Provincial Research Center of Engineering Technology for Nuclear Protein Medical Detection, Zhengzhou Health College, Zhengzhou, 45000, Henan, China.
- Department of Nuclear Medicine, Henan International Joint Laboratory for Nuclear Protein Regulation, The First Affiliated Hospital, Henan University College of Medicine, Ximen St, Kaifeng, 475004, Henan, China.
- Faculty of Basic Medical Subjects, Shu-Qing Medical College of Zhengzhou, Zhengzhou, 450064, Henan, China.
| | - Qi-Ming Liao
- Department of Medical Informatics and Computer, Shu-Qing Medical College of Zhengzhou, Gong-Ming Rd, Mazhai Town, Erqi District, Zhengzhou, 450064, Henan, China.
| |
Collapse
|
3
|
Sun Q, He N, Yang P, Zhao X. Low dose computed tomography reconstruction with momentum-based frequency adjustment network. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2025; 263:108673. [PMID: 40023964 DOI: 10.1016/j.cmpb.2025.108673] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/11/2024] [Revised: 11/29/2024] [Accepted: 02/13/2025] [Indexed: 03/04/2025]
Abstract
BACKGROUND AND OBJECTIVE Recent investigations into Low-Dose Computed Tomography (LDCT) reconstruction methods have brought Model-Based Data-Driven (MBDD) approaches to the forefront. One prominent architecture within MBDD entails the integration of Model-Based Iterative Reconstruction (MBIR) with Deep Learning (DL). While this approach offers the advantage of harnessing information from sinogram and image domains, it also reveals several deficiencies. First and foremost, the efficacy of DL methods within the realm of MBDD necessitates meticulous enhancement, as it directly impacts the computational cost and the quality of reconstructed images. Next, high computational costs and a high number of iterations limit the development of MBDD methods. Last but not least, CT reconstruction is sensitive to pixel accuracy, and the role of loss functions within DL methods is crucial for meeting this requirement. METHODS This paper advances MBDD methods through three principal contributions. Firstly, we introduce an innovative Frequency Adjustment Network (FAN) that effectively adjusts both high and low-frequency components during the inference phase, resulting in substantial enhancements in reconstruction performance. Second, we develop the Momentum-based Frequency Adjustment Network (MFAN), which leverages momentum terms as an extrapolation strategy to facilitate the amplification of changes throughout successive iterations, culminating in a rapid convergence framework. Lastly, we delve into the visual properties of CT images and present a unique loss function named Focal Detail Loss (FDL). The FDL function preserves fine details throughout the training phase, significantly improving reconstruction quality. RESULTS Through a series of experiments validation on the AAPM-Mayo public dataset and real-world piglet datasets, the aforementioned three contributions demonstrated superior performance. MFAN achieved convergence in 10 iterations as an iteration method, faster than other methods. Ablation studies further highlight the advanced performance of each contribution. CONCLUSIONS This paper presents an MBDD-based LDCT reconstruction method using a momentum-based frequency adjustment network with a focal detail loss function. This approach significantly reduces the number of iterations required for convergence while achieving superior reconstruction results in visual and numerical analyses.
Collapse
Affiliation(s)
- Qixiang Sun
- School of Mathematical Sciences, Capital Normal University, Beijing, 100048, China
| | - Ning He
- Smart City College, Beijing Union University, Beijing, 100101, China
| | - Ping Yang
- School of Mathematical Sciences, Capital Normal University, Beijing, 100048, China
| | - Xing Zhao
- School of Mathematical Sciences, Capital Normal University, Beijing, 100048, China.
| |
Collapse
|
4
|
Saidulu N, Muduli PR. Asymmetric Convolution-based GAN Framework for Low-Dose CT Image Denoising. Comput Biol Med 2025; 190:109965. [PMID: 40107022 DOI: 10.1016/j.compbiomed.2025.109965] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/10/2024] [Revised: 02/08/2025] [Accepted: 02/28/2025] [Indexed: 03/22/2025]
Abstract
Noise reduction is essential to improve the diagnostic quality of low-dose CT (LDCT) images. In this regard, data-driven denoising methods based on generative adversarial networks (GAN) have shown promising results. However, custom designs with 2D convolution may not preserve the correlation of the local and global pixels, which results in the loss of high-frequency (edges/ boundaries of lesions) anatomical details. A recent state-of-the-art method demonstrates that using primitive GAN-based methods may introduce structural (shape) distortion. To address this issue, we develop a novel asymmetric convolution-based generator network (ACGNet), which is constructed by using one-dimensional (1D) asymmetric convolutions and a dynamic attention module (DAM). The 1D asymmetric convolutions (1 × 3 & 3 × 1) can intensify the representation power of square convolution kernels (3 × 3) in horizontal and vertical directions. Consequently, we integrated the highlighted low-level CT voxel details via purposed attention DAM with high-level CT-scan features. As a result, ACGNet efficiently preserves the local and global pixel relations in denoised LDCT images. Furthermore, we propose a novel neural structure preserving loss (NSPL) through which ACGNet learns the neighborhood structure of CT images, preventing structural (shape) distortion. In addition, the ACGNet can reconstruct the CT images with human-perceived quality via back-propagated gradients due to the feature-based NSPL loss. Finally, we include differential content loss in network optimization to restore high-frequency lesion boundaries. The proposed method outperforms many state-of-the-art methods on two publicly accessible datasets: the Mayo 2016 dataset (PSNR: 35.2015 dB, SSIM: 0.9560), and Low-dose CT image and projection dataset (PSNR: 35.2825 dB, SSIM: 0.9566).
Collapse
Affiliation(s)
- Naragoni Saidulu
- Department of Electronics Engineering, Indian Institute of Technology (BHU) Varanasi, India.
| | - Priya Ranjan Muduli
- Department of Electronics Engineering, Indian Institute of Technology (BHU) Varanasi, India.
| |
Collapse
|
5
|
Li Z, Sun Z, Lv L, Liu Y, Wang X, Xu J, Xing J, Babyn P, Sun FR. Ultra-sparse view lung CT image reconstruction using generative adversarial networks and compressed sensing. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2025:8953996251329214. [PMID: 40296779 DOI: 10.1177/08953996251329214] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/30/2025]
Abstract
X-ray ionizing radiation from Computed Tomography (CT) scanning increases cancer risk for patients, thus making sparse view CT, which diminishes X-ray exposure by lowering the number of projections, highly significant in diagnostic imaging. However, reducing the number of projections inherently degrades image quality, negatively impacting clinical diagnosis. Consequently, attaining reconstructed images that meet diagnostic imaging criteria for sparse view CT is challenging. This paper presents a novel network (CSUF), specifically designed for ultra-sparse view lung CT image reconstruction. The CSUF network consists of three cohesive components including (1) a compressed sensing-based CT image reconstruction module (VdCS module), (2) a U-shaped end-to-end network, CT-RDNet, enhanced with a self-attention mechanism, acting as the generator in a Generative Adversarial Network (GAN) for CT image restoration and denoising, and (3) a feedback loop. The VdCS module enriches CT-RDNet with enhanced features, while CT-RDNet supplies the VdCS module with prior images infused with rich details and minimized artifacts, facilitated by the feedback loop. Engineering simulation experimental results demonstrate the robustness of the CSUF network and its potential to deliver lung CT images with diagnostic imaging quality even under ultra-sparse view conditions.
Collapse
Affiliation(s)
- Zhaoguang Li
- School of Integrated Circuits, Shandong University, Jinan, China
| | - Zhengxiang Sun
- Faculty of Science, The University of Sydney, NSW, Australia
| | - Lin Lv
- School of Integrated Circuits, Shandong University, Jinan, China
| | - Yuhan Liu
- School of Integrated Circuits, Shandong University, Jinan, China
| | - Xiuying Wang
- Faculty of Engineering, The University of Sydney, NSW, Australia
| | - Jingjing Xu
- School of Integrated Circuits, Shandong University, Jinan, China
| | - Jianping Xing
- School of Integrated Circuits, Shandong University, Jinan, China
| | - Paul Babyn
- Department of Medical Imaging, University of Saskatchewan and Saskatoon Health Region, Saskatoon, Canada
| | - Feng-Rong Sun
- School of Integrated Circuits, Shandong University, Jinan, China
| |
Collapse
|
6
|
Du Y, Liu Y, Wu H, Kang J, Gui Z, Zhang P, Ren Y. Combination of edge enhancement and cold diffusion model for low dose CT image denoising. BIOMED ENG-BIOMED TE 2025; 70:157-169. [PMID: 39501464 DOI: 10.1515/bmt-2024-0362] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/17/2024] [Accepted: 10/16/2024] [Indexed: 04/05/2025]
Abstract
OBJECTIVES Since the quality of low dose CT (LDCT) images is often severely affected by noise and artifacts, it is very important to maintain high quality CT images while effectively reducing the radiation dose. METHODS In recent years, the representation of diffusion models to produce high quality images and stable trainability has attracted wide attention. With the extension of the cold diffusion model to the classical diffusion model, its application has greater flexibility. Inspired by the cold diffusion model, we proposes a low dose CT image denoising method, called CECDM, based on the combination of edge enhancement and cold diffusion model. The LDCT image is taken as the end point (forward) of the diffusion process and the starting point (reverse) of the sampling process. Improved sobel operator and Convolution Block Attention Module are added to the network, and compound loss function is adopted. RESULTS The experimental results show that CECDM can effectively remove noise and artifacts from LDCT images while the inference time of a single image is reduced to 0.41 s. CONCLUSIONS Compared with the existing LDCT image post-processing methods, CECDM has a significant improvement in all indexes.
Collapse
Affiliation(s)
- Yinglin Du
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, China
| | - Yi Liu
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, China
| | - Han Wu
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, China
| | - Jiaqi Kang
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, China
| | - Zhiguo Gui
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, China
| | - Pengcheng Zhang
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, China
| | - Yali Ren
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, China
| |
Collapse
|
7
|
Minhaz AT, Murali A, Örge FH, Wilson DL, Bayat M. Improved biometric quantification in 3D ultrasound biomicroscopy via generative adversarial networks-based image enhancement. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025:10.1007/s10278-025-01488-5. [PMID: 40210809 DOI: 10.1007/s10278-025-01488-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/20/2024] [Revised: 03/06/2025] [Accepted: 03/18/2025] [Indexed: 04/12/2025]
Abstract
This study addresses the limitations of inexpensive, high-frequency ultrasound biomicroscopy (UBM) systems in visualizing small ocular structures and anatomical landmarks, especially outside the focal area, by improving image quality and visibility of important ocular structures for clinical ophthalmology applications. We developed a generative adversarial network (GAN) method for the 3D ultrasound biomicroscopy (3D-UBM) imaging system, called Spatially variant Deconvolution GAN (SDV-GAN). We employed spatially varying deconvolution and patch blending to enhance the original UBM images. This computationally expensive iterative deconvolution process yielded paired original and enhanced images for training the SDV-GAN. SDV-GAN achieved high performance metrics, with a structural similarity index measure (SSIM) of 0.96 and a peak signal-to-noise ratio (PSNR) of 36.92 dB. Structures were more clearly seen with no noticeable artifacts in the test images. SDV-GAN deconvolution improved biometric measurements made from UBM images, giving significant differences in angle opening distance (AOD, p < 0.0001) and angle recess area (ARA, p < 0.0001) measurements before and after SDV-GAN deconvolution. With clearer identification of apex, SDV-GAN improved inter-reader agreement in ARA measurements in images before and after deconvolution (intraclass correlation coefficient, [ICC] of 0.62 and 0.73, respectively). Real-time enhancement was achieved with an inference time of ~ 40 ms/frame (25 frames/s) on a standard GPU, compared to ~ 93 ms/frame (11 frames/s) using iterative deconvolution. SDV-GAN effectively enhanced UBM images, improving visibility and assessment of important ocular structures. Its real-time processing capabilities highlight the clinical potential of GAN enhancement in facilitating accurate diagnosis and treatment planning in ophthalmology using existing scanners.
Collapse
Affiliation(s)
- Ahmed Tahseen Minhaz
- Department of Biomedical Engineering, Lerner Research Institute, Cleveland Clinic, Cleveland, OH, USA
| | - Archana Murali
- School of Medicine, Case Western Reserve University, Cleveland, OH, USA
| | - Faruk H Örge
- Department of Biomedical Engineering, Case Western Reserve University, Cleveland, OH, USA
- Department of Ophthalmology and Visual Sciences, Case Western Reserve University, Cleveland, OH, USA
- Center for Pediatric Ophthalmology and Adult Strabismus, Rainbow Babies and Children's Hospital and University Hospitals Cleveland Medical Center Eye Institute, Cleveland, OH, USA
| | - David L Wilson
- Department of Biomedical Engineering, Case Western Reserve University, Cleveland, OH, USA
- Department of Radiology, Case Western Reserve University, Cleveland, OH, USA
| | - Mahdi Bayat
- Institute for Health Informatics, University of Minnesota, Minneapolis, MN, USA.
| |
Collapse
|
8
|
Klug M, Sobeh T, Green M, Mayer A, Kirshenboim Z, Konen E, Marom EM. Denoised Ultra-Low-Dose Chest CT to Assess Pneumonia in Individuals Who Are Immunocompromised. Radiol Cardiothorac Imaging 2025; 7:e240189. [PMID: 40079757 DOI: 10.1148/ryct.240189] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/15/2025]
Abstract
Purpose To evaluate the accuracy of chest ultra-low-dose CT (ULDCT) as compared with normal-dose CT in the evaluation of pneumonia in individuals who are immunocompromised. Materials and Methods This prospective study included 54 adults who were immunocompromised (median age, 62 years [IQR, 47.75-69.25 years]; 34 [63%] male participants) referred for a chest CT scan between September 2020 and December 2022 to evaluate for pneumonia. Each participant underwent two scans: normal-dose CT (120 kVp and automatic current modulation) and ULDCT (100 kVp and constant current of 10 mA). ULDCT images underwent a postprocessing procedure using an artificial intelligence algorithm to reduce image noise. Two radiologists, blinded to all clinical information, examined the images obtained from the three methods (normal-dose CT, ULDCT, and denoised ULDCT) for the presence of pneumonia and associated findings. The normal-dose CT was used as the reference standard, and sensitivity, specificity, positive and negative predictive values, and accuracy were calculated. Results The median effective radiation dose of ULDCT scans (0.12 mSV) was 1.95% of that of the normal-dose CT (6.15 mSV). Ten of the 54 participants were correctly identified as having no pneumonia, with similar accuracy between denoised ULDCT and ULDCT (100% vs 96%-98%, respectively). Both methods allowed for detection of pneumonia and features associated with invasive fungal pneumonia, but accuracy was slightly better with denoised ULDCT (accuracy, 100% vs 91%-98%). Fine details were better visualized in denoised ULDCT images: tree-in-bud pattern (accuracy, 93% vs 78%-80%), interlobular septal thickening (accuracy, 78%-83% vs 61%-67%), and intralobular septal thickening (accuracy, 85%-87% vs 0%). Conclusion Denoised ULDCT imaging showed better accuracy than ULDCT in identifying lungs with or without pneumonia in individuals who were immunocompromised. Keywords: CT, Pulmonary, Lung, Infection, Technology Assessment Supplemental material is available for this article. © RSNA, 2025.
Collapse
Affiliation(s)
- Maximiliano Klug
- Division of Diagnostic Imaging, The Chaim Sheba Medical Center, Tel Hashomer, 2 Derech Sheba St, Ramat Gan 5265601, Israel
- Faculty of Medicine, Tel Aviv University, Tel Aviv, Israel
| | - Tamer Sobeh
- Division of Diagnostic Imaging, The Chaim Sheba Medical Center, Tel Hashomer, 2 Derech Sheba St, Ramat Gan 5265601, Israel
- Faculty of Medicine, Tel Aviv University, Tel Aviv, Israel
| | - Michael Green
- Division of Diagnostic Imaging, The Chaim Sheba Medical Center, Tel Hashomer, 2 Derech Sheba St, Ramat Gan 5265601, Israel
| | - Arnaldo Mayer
- Division of Diagnostic Imaging, The Chaim Sheba Medical Center, Tel Hashomer, 2 Derech Sheba St, Ramat Gan 5265601, Israel
| | - Zehavit Kirshenboim
- Division of Diagnostic Imaging, The Chaim Sheba Medical Center, Tel Hashomer, 2 Derech Sheba St, Ramat Gan 5265601, Israel
- Faculty of Medicine, Tel Aviv University, Tel Aviv, Israel
| | - Eli Konen
- Division of Diagnostic Imaging, The Chaim Sheba Medical Center, Tel Hashomer, 2 Derech Sheba St, Ramat Gan 5265601, Israel
- Faculty of Medicine, Tel Aviv University, Tel Aviv, Israel
| | - Edith Michelle Marom
- Division of Diagnostic Imaging, The Chaim Sheba Medical Center, Tel Hashomer, 2 Derech Sheba St, Ramat Gan 5265601, Israel
- Faculty of Medicine, Tel Aviv University, Tel Aviv, Israel
| |
Collapse
|
9
|
Zhu J, Sun H, Chen W, Zhi S, Liu C, Zhao M, Zhang Y, Zhou T, Lam YL, Peng T, Qin J, Zhao L, Cai J, Ren G. Feature-targeted deep learning framework for pulmonary tumorous Cone-beam CT (CBCT) enhancement with multi-task customized perceptual loss and feature-guided CycleGAN. Comput Med Imaging Graph 2025; 121:102487. [PMID: 39891955 DOI: 10.1016/j.compmedimag.2024.102487] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/01/2023] [Revised: 12/21/2024] [Accepted: 12/30/2024] [Indexed: 02/03/2025]
Abstract
Thoracic Cone-beam computed tomography (CBCT) is routinely collected during image-guided radiation therapy (IGRT) to provide updated patient anatomy information for lung cancer treatments. However, CBCT images often suffer from streaking artifacts and noise caused by under-rate sampling projections and low-dose exposure, resulting in loss of lung anatomy which contains crucial pulmonary tumorous and functional information. While recent deep learning-based CBCT enhancement methods have shown promising results in suppressing artifacts, they have limited performance on preserving anatomical details containing crucial tumorous information due to lack of targeted guidance. To address this issue, we propose a novel feature-targeted deep learning framework which generates ultra-quality pulmonary imaging from CBCT of lung cancer patients via a multi-task customized feature-to-feature perceptual loss function and a feature-guided CycleGAN. The framework comprises two main components: a multi-task learning feature-selection network (MTFS-Net) for building up a customized feature-to-feature perceptual loss function (CFP-loss); and a feature-guided CycleGan network. Our experiments showed that the proposed framework can generate synthesized CT (sCT) images for the lung that achieved a high similarity to CT images, with an average SSIM index of 0.9747 and an average PSNR index of 38.5995 globally, and an average Pearman's coefficient of 0.8929 within the tumor region on multi-institutional datasets. The sCT images also achieved visually pleasing performance with effective artifacts suppression, noise reduction, and distinctive anatomical details preservation. Functional imaging tests further demonstrated the pulmonary texture correction performance of the sCT images, and the similarity of the functional imaging generated from sCT and CT images has reached an average DSC value of 0.9147, SCC value of 0.9615 and R value of 0.9661. Comparison experiments with pixel-to-pixel loss also showed that the proposed perceptual loss significantly enhances the performance of involved generative models. Our experiment results indicate that the proposed framework outperforms the state-of-the-art models for pulmonary CBCT enhancement. This framework holds great promise for generating high-quality pulmonary imaging from CBCT that is suitable for supporting further analysis of lung cancer treatment.
Collapse
Affiliation(s)
- Jiarui Zhu
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, 999077, Hong Kong SAR
| | - Hongfei Sun
- Department of Radiation Oncology, Xijing Hospital, Fourth Military Medical University, Xian 710032, China
| | - Weixing Chen
- School of Computer Science and Engineering, Sun Yat-Sen University, Guangzhou 510006, China
| | - Shaohua Zhi
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, 999077, Hong Kong SAR
| | - Chenyang Liu
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, 999077, Hong Kong SAR
| | - Mayang Zhao
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, 999077, Hong Kong SAR
| | - Yuanpeng Zhang
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, 999077, Hong Kong SAR
| | - Ta Zhou
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, 999077, Hong Kong SAR
| | - Yu Lap Lam
- Department of Clinical Oncology, Queen Mary Hospital, 999077, Hong Kong SAR
| | - Tao Peng
- School of Future Science and Engineering, Soochow University, Suzhou 215299, China
| | - Jing Qin
- School of Nursing, The Hong Kong Polytechnic University, 999077, Hong Kong SAR
| | - Lina Zhao
- Department of Radiation Oncology, Xijing Hospital, Fourth Military Medical University, Xian 710032, China.
| | - Jing Cai
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, 999077, Hong Kong SAR.
| | - Ge Ren
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, 999077, Hong Kong SAR; Research Institute for Intelligent Wearable Systems, The Hong Kong Polytechnic University, 999077, Hong Kong SAR; The Hong Kong Polytechnic University Shenzhen Research Institute, Shenzhen, China.
| |
Collapse
|
10
|
Zhang J, Ye L, Gong W, Chen M, Liu G, Cheng Y. A Novel Network for Low-Dose CT Denoising Based on Dual-Branch Structure and Multi-Scale Residual Attention. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025; 38:1245-1264. [PMID: 39261373 PMCID: PMC11950452 DOI: 10.1007/s10278-024-01254-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/09/2024] [Revised: 08/15/2024] [Accepted: 08/22/2024] [Indexed: 09/13/2024]
Abstract
Deep learning-based denoising of low-dose medical CT images has received great attention both from academic researchers and physicians in recent years, and has shown important application value in clinical practice. In this work, a novel two-branch and multi-scale residual attention-based network for low-dose CT image denoising is proposed. It adopts a two-branch framework structure, to extract and fuse image features at shallow and deep levels respectively, to recover image texture and structure information as much as possible. We propose the adaptive dynamic convolution block (ADCB) in the local information extraction layer. It can effectively extract the detailed information of low-dose CT denoising and enables the network to better capture the local details and texture features of the image, thereby improving the denoising effect and image quality. Multi-scale edge enhancement attention block (MEAB) is proposed in the global information extraction layer, to perform feature fusion through dilated convolution and a multi-dimensional attention mechanism. A multi-scale residual convolution block (MRCB) is proposed to integrate feature information and improve the robustness and generalization of the network. To demonstrate the effectiveness of our method, extensive comparison experiments are conducted and the performances evaluated on two publicly available datasets. Our model achieves 29.3004 PSNR, 0.8659 SSIM, and 14.0284 RMSE on the AAPM-Mayo dataset. It is evaluated by adding four different noise levels σ = 15, 30, 45, and 60 on the Qin_LUNG_CT dataset and achieves the best results. Ablation studies show that the proposed ADCB, MEAB, and MRCB modules improve the denoising performances significantly. The source code is available at https://github.com/Ye111-cmd/LDMANet .
Collapse
Affiliation(s)
- Ju Zhang
- College of Information Science and Technology, Hangzhou Normal University, Hangzhou, 310030, China
| | - Lieli Ye
- College of Information Science and Technology, Hangzhou Normal University, Hangzhou, 310030, China
| | - Weiwei Gong
- College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou, 310023, China
| | - Mingyang Chen
- College of Information Science and Technology, Hangzhou Normal University, Hangzhou, 310030, China
| | - Guangyu Liu
- College of Information Science and Technology, Hangzhou Normal University, Hangzhou, 310030, China
| | - Yun Cheng
- Department of Medical Imaging, Zhejiang Hospital, Hangzhou, 310058, China.
| |
Collapse
|
11
|
Phipps B, Hadoux X, Sheng B, Campbell JP, Liu TYA, Keane PA, Cheung CY, Chung TY, Wong TY, van Wijngaarden P. AI image generation technology in ophthalmology: Use, misuse and future applications. Prog Retin Eye Res 2025; 106:101353. [PMID: 40107410 DOI: 10.1016/j.preteyeres.2025.101353] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2024] [Revised: 03/12/2025] [Accepted: 03/13/2025] [Indexed: 03/22/2025]
Abstract
BACKGROUND AI-powered image generation technology holds the potential to reshape medical practice, yet it remains an unfamiliar technology for both medical researchers and clinicians alike. Given the adoption of this technology relies on clinician understanding and acceptance, we sought to demystify its use in ophthalmology. To this end, we present a literature review on image generation technology in ophthalmology, examining both its theoretical applications and future role in clinical practice. METHODS First, we consider the key model designs used for image synthesis, including generative adversarial networks, autoencoders, and diffusion models. We then perform a survey of the literature for image generation technology in ophthalmology prior to September 2024, presenting both the type of model used and its clinical application. Finally, we discuss the limitations of this technology, the risks of its misuse and the future directions of research in this field. RESULTS Applications of this technology include improving AI diagnostic models, inter-modality image transformation, more accurate treatment and disease prognostication, image denoising, and individualised education. Key barriers to its adoption include bias in generative models, risks to patient data security, computational and logistical barriers to development, challenges with model explainability, inconsistent use of validation metrics between studies and misuse of synthetic images. Looking forward, researchers are placing a further emphasis on clinically grounded metrics, the development of image generation foundation models and the implementation of methods to ensure data provenance. CONCLUSION Compared to other medical applications of AI, image generation is still in its infancy. Yet, it holds the potential to revolutionise ophthalmology across research, education and clinical practice. This review aims to guide ophthalmic researchers wanting to leverage this technology, while also providing an insight for clinicians on how it may change ophthalmic practice in the future.
Collapse
Affiliation(s)
- Benjamin Phipps
- Centre for Eye Research Australia, Royal Victorian Eye and Ear Hospital, East Melbourne, 3002, VIC, Australia; Ophthalmology, Department of Surgery, University of Melbourne, Parkville, 3010, VIC, Australia.
| | - Xavier Hadoux
- Centre for Eye Research Australia, Royal Victorian Eye and Ear Hospital, East Melbourne, 3002, VIC, Australia; Ophthalmology, Department of Surgery, University of Melbourne, Parkville, 3010, VIC, Australia
| | - Bin Sheng
- Department of Computer Science and Engineering, Shanghai Jiao Tong University, Shanghai, China
| | - J Peter Campbell
- Department of Ophthalmology, Casey Eye Institute, Oregon Health and Science University, Portland, USA
| | - T Y Alvin Liu
- Retina Division, Wilmer Eye Institute, Johns Hopkins University, Baltimore, MD, 21287, USA
| | - Pearse A Keane
- NIHR Biomedical Research Centre at Moorfields Eye Hospital NHS Foundation Trust, London, UK; Institute of Ophthalmology, University College London, London, UK
| | - Carol Y Cheung
- Department of Ophthalmology and Visual Sciences, The Chinese University of Hong Kong, 999077, China
| | - Tham Yih Chung
- Yong Loo Lin School of Medicine, National University of Singapore, Singapore; Centre for Innovation and Precision Eye Health, Department of Ophthalmology, Yong Loo Lin School of Medicine, National University of Singapore and National University Health System, Singapore; Singapore Eye Research Institute, Singapore National Eye Centre, Singapore; Eye Academic Clinical Program (Eye ACP), Duke NUS Medical School, Singapore
| | - Tien Y Wong
- Singapore Eye Research Institute, Singapore National Eye Centre, Singapore; Tsinghua Medicine, Tsinghua University, Beijing, China; Beijing Visual Science and Translational Eye Research Institute, Beijing Tsinghua Changgung Hospital, Beijing, China
| | - Peter van Wijngaarden
- Centre for Eye Research Australia, Royal Victorian Eye and Ear Hospital, East Melbourne, 3002, VIC, Australia; Ophthalmology, Department of Surgery, University of Melbourne, Parkville, 3010, VIC, Australia; Florey Institute of Neuroscience & Mental Health, Parkville, VIC, Australia
| |
Collapse
|
12
|
Deebani W, Aziz L, Aziz A, Basri WS, Alawad WM, Althubiti SA. Synergistic transfer learning and adversarial networks for breast cancer diagnosis: benign vs. invasive classification. Sci Rep 2025; 15:7461. [PMID: 40032913 DOI: 10.1038/s41598-025-90288-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/08/2024] [Accepted: 02/11/2025] [Indexed: 03/05/2025] Open
Abstract
Current breast cancer diagnosis methods often face limitations such as high cost, time consumption, and inter-observer variability. To address these challenges, this research proposes a novel deep learning framework that leverages generative adversarial networks (GANs) for data augmentation and transfer learning to enhance breast cancer classification using convolutional neural networks (CNNs). The framework uses a two-stage augmentation approach. First, a conditional Wasserstein GAN (cWGAN) generates synthetic breast cancer images based on clinical data, enhancing training stability and enabling targeted feature incorporation. Second, traditional augmentation techniques (e.g., rotation, flipping, cropping) are applied to both original and synthetic images. A multi-scale transfer learning technique is also employed, integrating three pre-trained CNNs (DenseNet-201, NasNetMobile, ResNet-101) with a multi-scale feature enrichment scheme, allowing the model to capture features at various scales. The framework was evaluated on the BreakHis dataset, achieving an accuracy of 99.2% for binary classification and 98.5% for multi-class classification, significantly outperforming existing methods. This framework offers a more efficient, cost-effective, and accurate approach for breast cancer diagnosis. Future work will focus on generalizing the framework to clinical datasets and integrating it into diagnostic workflows.
Collapse
Affiliation(s)
- Wejdan Deebani
- Department of Mathematics, College of Science and Arts, King Abdul Aziz University, 21911, Rabigh, Saudi Arabia
| | - Lubna Aziz
- Department of Artificial Intelligence, FEST Iqra University Karachi, Karachi, Pakistan.
- Faculty of Computing, Universiti Teknologi Malaysia, Johor Bahru, Johor, Malaysia.
| | - Arshad Aziz
- Department of Artificial Intelligence, FEST Iqra University Karachi, Karachi, Pakistan
| | - Wael Sh Basri
- College of Business Administration, Management Information System, Northern Border University, Arar, Saudi Arabia
| | - Wedad M Alawad
- Department of Information Technology, College of Computer, Qassim University, Buraydah, 51452, Saudi Arabia
| | - Sara A Althubiti
- Department of Computer Science, College of Computer and Information Sciences, Majmaah University, 11952, Al-Majmaah, Saudi Arabia
| |
Collapse
|
13
|
Ma X, Zou M, Fang X, Luo G, Wang W, Dong S, Li X, Wang K, Dong Q, Tian Y, Li S. Convergent-Diffusion Denoising Model for multi-scenario CT Image Reconstruction. Comput Med Imaging Graph 2025; 120:102491. [PMID: 39787736 DOI: 10.1016/j.compmedimag.2024.102491] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/03/2024] [Revised: 10/27/2024] [Accepted: 12/31/2024] [Indexed: 01/12/2025]
Abstract
A generic and versatile CT Image Reconstruction (CTIR) scheme can efficiently mitigate imaging noise resulting from inherent physical limitations, substantially bolstering the dependability of CT imaging diagnostics across a wider spectrum of patient cases. Current CTIR techniques often concentrate on distinct areas such as Low-Dose CT denoising (LDCTD), Sparse-View CT reconstruction (SVCTR), and Metal Artifact Reduction (MAR). Nevertheless, due to the intricate nature of multi-scenario CTIR, these techniques frequently narrow their focus to specific tasks, resulting in limited generalization capabilities for diverse scenarios. We propose a novel Convergent-Diffusion Denoising Model (CDDM) for multi-scenario CTIR, which utilizes a stepwise denoising process to converge toward an imaging-noise-free image with high generalization. CDDM uses a diffusion-based process based on a priori decay distribution to steadily correct imaging noise, thus avoiding the overfitting of individual samples. Within CDDM, a domain-correlated sampling network (DS-Net) provides an innovative sinogram-guided noise prediction scheme to leverage both image and sinogram (i.e., dual-domain) information. DS-Net analyzes the correlation of the dual-domain representations for sampling the noise distribution, introducing sinogram semantics to avoid secondary artifacts. Experimental results validate the practical applicability of our scheme across various CTIR scenarios, including LDCTD, MAR, and SVCTR, with the support of sinogram knowledge.
Collapse
Affiliation(s)
- Xinghua Ma
- The Faculty of Computing, Harbin Institute of Technology, Harbin, Heilongjiang, China; The Computational Bioscience Research Center, King Abdullah University of Science and Technology, Thuwal, Makkah, Saudi Arabia
| | - Mingye Zou
- The Faculty of Computing, Harbin Institute of Technology, Harbin, Heilongjiang, China
| | - Xinyan Fang
- The Faculty of Computing, Harbin Institute of Technology, Harbin, Heilongjiang, China
| | - Gongning Luo
- The Faculty of Computing, Harbin Institute of Technology, Harbin, Heilongjiang, China; The Computational Bioscience Research Center, King Abdullah University of Science and Technology, Thuwal, Makkah, Saudi Arabia
| | - Wei Wang
- The Faculty of Computing, Harbin Institute of Technology, Shenzhen, Guangdong, China.
| | - Suyu Dong
- The College of Computer and Control Engineering, Northeast Forestry University, Harbin, Heilongjiang, China.
| | - Xiangyu Li
- The Faculty of Computing, Harbin Institute of Technology, Harbin, Heilongjiang, China.
| | - Kuanquan Wang
- The Faculty of Computing, Harbin Institute of Technology, Harbin, Heilongjiang, China
| | - Qing Dong
- The Department of Thoracic Surgery at No. 4 Affiliated Hospital, Harbin Medical University, Harbin, Heilongjiang, China
| | - Ye Tian
- The Department of Cardiology at No. 1 Affiliated Hospital, Harbin Medical University, Harbin, Heilongjiang, China
| | - Shuo Li
- The Department of Computer and Data Science, Case Western Reserve University, Cleveland, OH, USA; The Department of Biomedical Engineering, Case Western Reserve University, Cleveland, OH, USA
| |
Collapse
|
14
|
Hu D, Zhang C, Fei X, Yao Y, Xi Y, Liu J, Zhang Y, Coatrieux G, Coatrieux JL, Chen Y. DPI-MoCo: Deep Prior Image Constrained Motion Compensation Reconstruction for 4D CBCT. IEEE TRANSACTIONS ON MEDICAL IMAGING 2025; 44:1243-1256. [PMID: 39423082 DOI: 10.1109/tmi.2024.3483451] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/21/2024]
Abstract
4D cone-beam computed tomography (CBCT) plays a critical role in adaptive radiation therapy for lung cancer. However, extremely sparse sampling projection data will cause severe streak artifacts in 4D CBCT images. Existing deep learning (DL) methods heavily rely on large labeled training datasets which are difficult to obtain in practical scenarios. Restricted by this dilemma, DL models often struggle with simultaneously retaining dynamic motions, removing streak degradations, and recovering fine details. To address the above challenging problem, we introduce a Deep Prior Image Constrained Motion Compensation framework (DPI-MoCo) that decouples the 4D CBCT reconstruction into two sub-tasks including coarse image restoration and structural detail fine-tuning. In the first stage, the proposed DPI-MoCo combines the prior image guidance, generative adversarial network, and contrastive learning to globally suppress the artifacts while maintaining the respiratory movements. After that, to further enhance the local anatomical structures, the motion estimation and compensation technique is adopted. Notably, our framework is performed without the need for paired datasets, ensuring practicality in clinical cases. In the Monte Carlo simulation dataset, the DPI-MoCo achieves competitive quantitative performance compared to the state-of-the-art (SOTA) methods. Furthermore, we test DPI-MoCo in clinical lung cancer datasets, and experiments validate that DPI-MoCo not only restores small anatomical structures and lesions but also preserves motion information.
Collapse
|
15
|
Jia L, Jia B, Li Z, Zhang Y, Gui Z. A cross-type multi-dimensional network based on feature enhancement and triple interactive attention for LDCT denoising. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2025; 33:393-404. [PMID: 39973791 DOI: 10.1177/08953996241306696] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/21/2025]
Abstract
BackgroundNumerous deep leaning methods for low-dose computed technology (CT) image denoising have been proposed, achieving impressive results. However, issues such as loss of structure and edge information and low denoising efficiency still exist.ObjectiveTo improve image denoising quality, an enhanced multi-dimensional hybrid attention LDCT image denoising network based on edge detection is proposed in this paper.MethodsIn our network, we employ a trainable Sobel convolution to design an edge enhancement module and fuse an enhanced triplet attention network (ETAN) after each 3 × 3 convolutional layer to extract richer features more comprehensively and suppress useless information. During the training process, we adopt a strategy that combines total variation loss (TVLoss) with mean squared error (MSE) loss to reduce high-frequency artifacts in image reconstruction and balance image denoising and detail preservation.ResultsCompared with other advanced algorithms (CT-former, REDCNN and EDCNN), our proposed model achieves the best PSNR and SSIM values in CT image of the abdomen, which are 34.8211and 0.9131, respectively.ConclusionThrough comparative experiments with other related algorithms, it can be seen that the algorithm proposed in this article has achieved significant improvements in both subjective vision and objective indicators.
Collapse
Affiliation(s)
- Lina Jia
- School of Physics and Electronic Engineering, Shanxi University, Taiyuan, China
- Shanxi Key Laboratory of Wireless Communication and Detection, Taiyuan, China
| | - Beibei Jia
- School of Physics and Electronic Engineering, Shanxi University, Taiyuan, China
| | - Zongyang Li
- School of Physics and Electronic Engineering, Shanxi University, Taiyuan, China
| | - Yizhuo Zhang
- School of Physics and Electronic Engineering, Shanxi University, Taiyuan, China
| | - Zhiguo Gui
- State Key Laboratory of Dynamic Measurement Technology, North University of China, Taiyuan, China
| |
Collapse
|
16
|
Shi J, Pelt DM, Batenburg KJ. Multi-stage deep learning artifact reduction for parallel-beam computed tomography. JOURNAL OF SYNCHROTRON RADIATION 2025; 32:442-456. [PMID: 39960472 DOI: 10.1107/s1600577525000359] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/19/2024] [Accepted: 01/14/2025] [Indexed: 03/11/2025]
Abstract
Computed tomography (CT) using synchrotron radiation is a powerful technique that, compared with laboratory CT techniques, boosts high spatial and temporal resolution while also providing access to a range of contrast-formation mechanisms. The acquired projection data are typically processed by a computational pipeline composed of multiple stages. Artifacts introduced during data acquisition can propagate through the pipeline and degrade image quality in the reconstructed images. Recently, deep learning has shown significant promise in enhancing image quality for images representing scientific data. This success has driven increasing adoption of deep learning techniques in CT imaging. Various approaches have been proposed to incorporate deep learning into computational pipelines, but each has limitations in addressing artifacts effectively and efficiently in synchrotron CT, either in properly addressing the specific artifacts or in computational efficiency. Recognizing these challenges, we introduce a novel method that incorporates separate deep learning models at each stage of the tomography pipeline - projection, sinogram and reconstruction - to address specific artifacts locally in a data-driven way. Our approach includes bypass connections that feed both the outputs from previous stages and raw data to subsequent stages, minimizing the risk of error propagation. Extensive evaluations on both simulated and real-world datasets illustrate that our approach effectively reduces artifacts and outperforms comparison methods.
Collapse
Affiliation(s)
- Jiayang Shi
- Leiden University, Einsteinweg 55, 2333 CC Leiden, The Netherlands
| | - Daniël M Pelt
- Leiden University, Einsteinweg 55, 2333 CC Leiden, The Netherlands
| | | |
Collapse
|
17
|
Zhang R, Szczykutowicz TP, Toia GV. Artificial Intelligence in Computed Tomography Image Reconstruction: A Review of Recent Advances. J Comput Assist Tomogr 2025:00004728-990000000-00429. [PMID: 40008975 DOI: 10.1097/rct.0000000000001734] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/12/2024] [Accepted: 01/07/2025] [Indexed: 02/27/2025]
Abstract
The development of novel image reconstruction algorithms has been pivotal in enhancing image quality and reducing radiation dose in computed tomography (CT) imaging. Traditional techniques like filtered back projection perform well under ideal conditions but fail to generate high-quality images under low-dose, sparse-view, and limited-angle conditions. Iterative reconstruction methods improve upon filtered back projection by incorporating system models and assumptions about the patient, yet they can suffer from patchy image textures. The emergence of artificial intelligence (AI), particularly deep learning, has further advanced CT reconstruction. AI techniques have demonstrated great potential in reducing radiation dose while preserving image quality and noise texture. Moreover, AI has exhibited unprecedented performance in addressing challenging CT reconstruction problems, including low-dose CT, sparse-view CT, limited-angle CT, and interior tomography. This review focuses on the latest advances in AI-based CT reconstruction under these challenging conditions.
Collapse
Affiliation(s)
- Ran Zhang
- Departments of Radiology and Medical Physics, University of Wisconsin, Madison, WI
| | | | | |
Collapse
|
18
|
Yang F, Zhao F, Liu Y, Liu M, Liu M. Dual-Domain Self-Supervised Deep Learning with Graph Convolution for Low-Dose Computed Tomography Reconstruction. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025:10.1007/s10278-024-01314-4. [PMID: 39966223 DOI: 10.1007/s10278-024-01314-4] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/01/2024] [Revised: 10/09/2024] [Accepted: 10/23/2024] [Indexed: 02/20/2025]
Abstract
X-ray computed tomography (CT) is a commonly used imaging modality in clinical practice. Recent years have seen increasing public concern regarding the ionizing radiation from CT. Low-dose CT (LDCT) has been proven to be effective in reducing patients' radiation exposure, but it results in CT images with low signal-to-noise ratio (SNR), failing to meet the image quality required for diagnosis. To enhance the SNR of LDCT images, numerous denoising strategies based on deep learning have been introduced, leading to notable advancements. Despite these advancements, most methods have relied on a supervised training paradigm. The challenge in acquiring aligned pairs of low-dose and normal-dose images in a clinical setting has limited their applicability. Recently, some self-supervised deep learning methods have enabled denoising using only noisy samples. However, these techniques are based on overly simplistic assumptions about noise and focus solely on CT sinogram denoising or image denoising, compromising their effectiveness. To address this, we introduce the Dual-Domain Self-supervised framework, termed DDoS, to accomplish effective LDCT denoising and reconstruction. The framework includes denoising in the sinogram domain, filtered back-projection reconstruction, and denoising in the image domain. By identifying the statistical characteristics of sinogram noise and CT image noise, we develop sinogram-denoising and CT image-denoising networks that are fully adapted to these characteristics. Both networks utilize a unified hybrid architecture that combines graph convolution and incorporates multiple channel attention modules, facilitating the extraction of local and non-local multi-scale features. Comprehensive experiments on two large-scale LDCT datasets demonstrate the superiority of DDoS framework over existing state-of-the-art methods.
Collapse
Affiliation(s)
- Feng Yang
- College of Nuclear Technology and Automation Engineering, Chengdu University of Technology, No. 1 East 3rd Road, Erxianqiao, Chengdu, 610059, Sichuan, China
- Department of Radiation Oncology, Sichuan Cancer Hospital & Institute, Sichuan Cancer Center, Affiliated Cancer Hospital of University of Electronic Science and Technology of China, No. 55, Section 4, Renmin South Road, Chengdu, 610054, Sichuan, China
| | - Feixiang Zhao
- College of Nuclear Technology and Automation Engineering, Chengdu University of Technology, No. 1 East 3rd Road, Erxianqiao, Chengdu, 610059, Sichuan, China
| | - Yanhua Liu
- School of Mechanical and Electrical Engineering, Chengdu University of Technology, No. 1 East 3rd Road, Erxianqiao, Chengdu, 610059, Sichuan, China
| | - Min Liu
- College of Nuclear Technology and Automation Engineering, Chengdu University of Technology, No. 1 East 3rd Road, Erxianqiao, Chengdu, 610059, Sichuan, China
- Department of Radiation Oncology, Sichuan Cancer Hospital & Institute, Sichuan Cancer Center, Affiliated Cancer Hospital of University of Electronic Science and Technology of China, No. 55, Section 4, Renmin South Road, Chengdu, 610054, Sichuan, China
| | - Mingzhe Liu
- School of Data Science and Artificial Intelligence, Wenzhou University of Technology, Ouhai District, Wenzhou, 325000, Zhejiang, China.
| |
Collapse
|
19
|
Soltanpour S, Chang A, Madularu D, Kulkarni P, Ferris C, Joslin C. 3D Wasserstein Generative Adversarial Network with Dense U-Net-Based Discriminator for Preclinical fMRI Denoising. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025:10.1007/s10278-025-01434-5. [PMID: 39939477 DOI: 10.1007/s10278-025-01434-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/28/2024] [Revised: 01/21/2025] [Accepted: 01/28/2025] [Indexed: 02/14/2025]
Abstract
Functional magnetic resonance imaging (fMRI) is extensively used in clinical and preclinical settings to study brain function; however, fMRI data is inherently noisy due to physiological processes, hardware, and external noise. Denoising is one of the main preprocessing steps in any fMRI analysis pipeline. This process is challenging in preclinical data in comparison to clinical data due to variations in brain geometry, image resolution, and low signal-to-noise ratios. In this paper, we propose a structure-preserved algorithm based on a 3D Wasserstein generative adversarial network with a 3D dense U-net-based discriminator called 3D U-WGAN. We apply a 4D data configuration to effectively denoise temporal and spatial information in analyzing preclinical fMRI data. GAN-based denoising methods often utilize a discriminator to identify significant differences between denoised and noise-free images, focusing on global or local features. To refine the fMRI denoising model, our method employs a 3D dense U-Net discriminator to learn both global and local distinctions. To tackle potential oversmoothing, we introduce an adversarial loss and enhance perceptual similarity by measuring feature space distances. Experiments illustrate that 3D U-WGAN significantly improves image quality in resting-state and task preclinical fMRI data, enhancing signal-to-noise ratio without introducing excessive structural changes in existing methods. The proposed method outperforms state-of-the-art methods when applied to simulated and real data in a fMRI analysis pipeline.
Collapse
Affiliation(s)
- Sima Soltanpour
- School of Information Technology, Carleton University, 1125 Colonel By Dr, Ottawa, Ontario, K1S 5B6, Canada.
| | - Arnold Chang
- Center for Translational NeuroImaging (CTNI), Northeastern University, 360 Huntington Ave, Boston, MA, 02115, USA
| | - Dan Madularu
- Department of Psychology, Carleton University, 1125 Colonel By Dr, Ottawa, Ontario, K1S 5B6, Canada
- Tessellis Ltd., 350 Legget Drive, Ottawa, Ontario, K2K 0G7, Canada
| | - Praveen Kulkarni
- Center for Translational NeuroImaging (CTNI), Northeastern University, 360 Huntington Ave, Boston, MA, 02115, USA
| | - Craig Ferris
- Center for Translational NeuroImaging (CTNI), Northeastern University, 360 Huntington Ave, Boston, MA, 02115, USA
| | - Chris Joslin
- School of Information Technology, Carleton University, 1125 Colonel By Dr, Ottawa, Ontario, K1S 5B6, Canada
| |
Collapse
|
20
|
Yu X, Hu D, Yao Q, Fu Y, Zhong Y, Wang J, Tian M, Zhang H. Diffused Multi-scale Generative Adversarial Network for low-dose PET images reconstruction. Biomed Eng Online 2025; 24:16. [PMID: 39924498 PMCID: PMC11807330 DOI: 10.1186/s12938-025-01348-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2024] [Accepted: 01/29/2025] [Indexed: 02/11/2025] Open
Abstract
PURPOSE The aim of this study is to convert low-dose PET (L-PET) images to full-dose PET (F-PET) images based on our Diffused Multi-scale Generative Adversarial Network (DMGAN) to offer a potential balance between reducing radiation exposure and maintaining diagnostic performance. METHODS The proposed method includes two modules: the diffusion generator and the u-net discriminator. The goal of the first module is to get different information from different levels, enhancing the generalization ability of the generator to the image and improving the stability of the training. Generated images are inputted into the u-net discriminator, extracting details from both overall and specific perspectives to enhance the quality of the generated F-PET images. We conducted evaluations encompassing both qualitative assessments and quantitative measures. In terms of quantitative comparisons, we employed two metrics, structure similarity index measure (SSIM) and peak signal-to-noise ratio (PSNR) to evaluate the performance of diverse methods. RESULTS Our proposed method achieved the highest PSNR and SSIM scores among the compared methods, which improved PSNR by at least 6.2% compared to the other methods. Compared to other methods, the synthesized full-dose PET image generated by our method exhibits a more accurate voxel-wise metabolic intensity distribution, resulting in a clearer depiction of the epilepsy focus. CONCLUSIONS The proposed method demonstrates improved restoration of original details from low-dose PET images compared to other models trained on the same datasets. This method offers a potential balance between minimizing radiation exposure and preserving diagnostic performance.
Collapse
Affiliation(s)
- Xiang Yu
- Polytechnic Institute, Zhejiang University, Hangzhou, China
| | - Daoyan Hu
- The College of Biomedical Engineering and Instrument Science of Zhejiang University, Hangzhou, China
| | - Qiong Yao
- Department of Nuclear Medicine and Medical PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, China
| | - Yu Fu
- College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China
| | - Yan Zhong
- Department of Nuclear Medicine and Medical PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, China
| | - Jing Wang
- Department of Nuclear Medicine and Medical PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, China
| | - Mei Tian
- Human Phenome Institute, Fudan University, 825 Zhangheng Road, Shanghai, 201203, China.
| | - Hong Zhang
- The College of Biomedical Engineering and Instrument Science of Zhejiang University, Hangzhou, China.
- Department of Nuclear Medicine and Medical PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, China.
- Key Laboratory for Biomedical Engineering of Ministry of Education, Zhejiang University, Hangzhou, China.
| |
Collapse
|
21
|
Morovati B, Li M, Han S, Zhou L, Wang D, Wang G, Yu H. Patch-based dual-domain photon-counting CT data correction with residual-based WGAN-ViT. Phys Med Biol 2025; 70:045008. [PMID: 39874670 PMCID: PMC11800073 DOI: 10.1088/1361-6560/adaf71] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/24/2024] [Revised: 01/08/2025] [Accepted: 01/28/2025] [Indexed: 01/30/2025]
Abstract
Objective.x-ray photon-counting detectors have recently gained popularity due to their capabilities in energy discrimination power, noise suppression, and resolution refinement. The latest extremity photon-counting computed tomography (PCCT) scanner leverages these advantages for tissue characterization, material decomposition, beam hardening correction, and metal artifact reduction. However, technical challenges such as charge splitting and pulse pileup can distort the energy spectrum and compromise image quality. Also, there is a clinical need to balance radiation dose and imaging speed for contrast-enhancement and other studies. This paper aims to address these challenges by developing a dual-domain correction approach to enhance PCCT reconstruction quality quantitatively and qualitatively.Approach.We propose a novel correction method that operates in both projection and image domains. In the projection domain, we employ a residual-based Wasserstein generative adversarial network to capture local and global features, suppressing pulse pileup, charge splitting, and data noise. This is facilitated with traditional filtering methods in the image domain to enhance signal-to-noise ratio while preserving texture across each energy channel. To address GPU memory constraints, our approach utilizes a patch-based volumetric refinement network.Main results.Our dual-domain correction approach demonstrates significant fidelity improvements across both projection and image domains. Experiments on simulated and real datasets reveal that the proposed model effectively suppresses noise and preserves intricate details, outperforming the state-of-the-art methods.Significance.This approach highlights the potential of dual-domain PCCT data correction to enhance image quality for clinical applications, showing promise for advancing PCCT image fidelity and applicability in preclinical/clinical environments.
Collapse
Affiliation(s)
- Bahareh Morovati
- Department of Electrical and Computer Engineering, University of Massachusetts Lowell, Lowell, MA 01854, United States of America
| | - Mengzhou Li
- Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY 12180, United States of America
| | - Shuo Han
- Department of Electrical and Computer Engineering, University of Massachusetts Lowell, Lowell, MA 01854, United States of America
| | - Li Zhou
- Department of Electrical and Computer Engineering, University of Massachusetts Lowell, Lowell, MA 01854, United States of America
| | - Dayang Wang
- Department of Electrical and Computer Engineering, University of Massachusetts Lowell, Lowell, MA 01854, United States of America
| | - Ge Wang
- Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY 12180, United States of America
| | - Hengyong Yu
- Department of Electrical and Computer Engineering, University of Massachusetts Lowell, Lowell, MA 01854, United States of America
| |
Collapse
|
22
|
Hein D, Holmin S, Prochazka V, Yin Z, Danielsson M, Persson M, Wang G. Syn2Real: synthesis of CT image ring artifacts for deep learning-based correction. Phys Med Biol 2025; 70:04NT01. [PMID: 39842097 DOI: 10.1088/1361-6560/adad2c] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/23/2024] [Accepted: 01/22/2025] [Indexed: 01/24/2025]
Abstract
Objective. We strive to overcome the challenges posed by ring artifacts in x-ray computed tomography (CT) by developing a novel approach for generating training data for deep learning-based methods. Training such networks require large, high quality, datasets that are often generated in the data domain, time-consuming and expensive. Our objective is to develop a technique for synthesizing realistic ring artifacts directly in the image domain, enabling scalable production of training data without relying on specific imaging system physics.Approach. We develop 'Syn2Real,' a computationally efficient pipeline that generates realistic ring artifacts directly in the image domain. To demonstrate the effectiveness of our approach, we train two versions of UNet, vanilla and a high capacity version with self-attention layers that we call UNetpp, withℓ2and perceptual losses, as well as a diffusion model, on energy-integrating CT images with and without these synthetic ring artifacts.Main Results.Despite being trained on conventional single-energy CT images, our models effectively correct ring artifacts across various monoenergetic images, at different energy levels and slice thicknesses, from a prototype photon-counting CT system. This generalizability validates the realism and versatility of our ring artifact generation process.Significance.Ring artifacts in x-ray CT pose a unique challenge to image quality and clinical utility. By focusing on data generation, our work provides a foundation for developing more robust and adaptable ring artifact correction methods for pre-clinical, clinical and other CT applications.
Collapse
Affiliation(s)
- Dennis Hein
- Department of Physics, KTH Royal Institute of Technology, Stockholm, Sweden
- MedTechLabs, BioClinicum, Karolinska University Hospital, Stockholm, Sweden
| | - Staffan Holmin
- Department of Clinical Neuroscience, Karolinska Institute, Stockholm, Sweden
- Department of Neuroradiology, Karolinska University Hospital, Stockholm Sweden
| | | | - Zhye Yin
- GE HealthCare, Waukesha, WI, United States of America
| | - Mats Danielsson
- Department of Physics, KTH Royal Institute of Technology, Stockholm, Sweden
- MedTechLabs, BioClinicum, Karolinska University Hospital, Stockholm, Sweden
| | - Mats Persson
- Department of Physics, KTH Royal Institute of Technology, Stockholm, Sweden
- MedTechLabs, BioClinicum, Karolinska University Hospital, Stockholm, Sweden
| | - Ge Wang
- Department of Biomedical Engineering, School of Engineering, Biomedical Imaging Center, Center for Biotechnology and Interdisciplinary Studies, Rensselaer Polytechnic Institute, Troy, NY, United States of America
| |
Collapse
|
23
|
Yang Z, Xia W, Lu Z, Chen Y, Li X, Zhang Y. Hypernetwork-Based Physics-Driven Personalized Federated Learning for CT Imaging. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2025; 36:3136-3150. [PMID: 38100342 DOI: 10.1109/tnnls.2023.3338867] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/17/2023]
Abstract
In clinical practice, computed tomography (CT) is an important noninvasive inspection technology to provide patients' anatomical information. However, its potential radiation risk is an unavoidable problem that raises people's concerns. Recently, deep learning (DL)-based methods have achieved promising results in CT reconstruction, but these methods usually require the centralized collection of large amounts of data for training from specific scanning protocols, which leads to serious domain shift and privacy concerns. To relieve these problems, in this article, we propose a hypernetwork-based physics-driven personalized federated learning method (HyperFed) for CT imaging. The basic assumption of the proposed HyperFed is that the optimization problem for each domain can be divided into two subproblems: local data adaption and global CT imaging problems, which are implemented by an institution-specific physics-driven hypernetwork and a global-sharing imaging network, respectively. Learning stable and effective invariant features from different data distributions is the main purpose of global-sharing imaging network. Inspired by the physical process of CT imaging, we carefully design physics-driven hypernetwork for each domain to obtain hyperparameters from specific physical scanning protocol to condition the global-sharing imaging network, so that we can achieve personalized local CT reconstruction. Experiments show that HyperFed achieves competitive performance in comparison with several other state-of-the-art methods. It is believed as a promising direction to improve CT imaging quality and personalize the needs of different institutions or scanners without data sharing. Related codes have been released at https://github.com/Zi-YuanYang/HyperFed.
Collapse
|
24
|
Zhao X, Du Y, Peng Y. Deep Learning-Based Multi-View Projection Synthesis Approach for Improving the Quality of Sparse-View CBCT in Image-Guided Radiotherapy. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2025:10.1007/s10278-025-01390-0. [PMID: 39849201 DOI: 10.1007/s10278-025-01390-0] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/09/2024] [Revised: 12/16/2024] [Accepted: 12/19/2024] [Indexed: 01/25/2025]
Abstract
While radiation hazards induced by cone-beam computed tomography (CBCT) in image-guided radiotherapy (IGRT) can be reduced by sparse-view sampling, the image quality is inevitably degraded. We propose a deep learning-based multi-view projection synthesis (DLMPS) approach to improve the quality of sparse-view low-dose CBCT images. In the proposed DLMPS approach, linear interpolation was first applied to sparse-view projections and the projections were rearranged into sinograms; these sinograms were processed with a sinogram restoration model and then rearranged back into projections. The sinogram restoration model was modified from the 2D U-Net by incorporating dynamic convolutional layers and residual learning techniques. The DLMPS approach was trained, validated, and tested on CBCT data from 163, 30, and 30 real patients respectively. Sparse-view projection datasets with 1/4 and 1/8 of the original sampling rate were simulated, and the corresponding full-view projection datasets were restored via the DLMPS approach. Tomographic images were reconstructed using the Feldkamp-Davis-Kress algorithm. Quantitative metrics including root-mean-square error (RMSE), peak signal-to-noise ratio (PSNR), structural similarity (SSIM), and feature similarity (FSIM) were calculated in both the projection and image domains to evaluate the performance of the DLMPS approach. The DLMPS approach was compared with 11 state-of-the-art (SOTA) models, including CNN and Transformer architectures. For 1/4 sparse-view reconstruction task, the proposed DLMPS approach achieved averaged RMSE, PSNR, SSIM, and FSIM values of 0.0271, 45.93 dB, 0.9817, and 0.9587 in the projection domain, and 0.000885, 37.63 dB, 0.9074, and 0.9885 in the image domain, respectively. For 1/8 sparse-view reconstruction task, the DLMPS approach achieved averaged RMSE, PSNR, SSIM, and FSIM values of 0.0304, 44.85 dB, 0.9785, and 0.9524 in the projection domain, and 0.001057, 36.05 dB, 0.8786, and 0.9774 in the image domain, respectively. The DLMPS approach outperformed all the 11 SOTA models in both the projection and image domains for 1/4 and 1/8 sparse-view reconstruction tasks. The proposed DLMPS approach effectively improves the quality of sparse-view CBCT images in IGRT by accurately synthesizing missing projections, exhibiting potential in substantially reducing imaging dose to patients with minimal loss of image quality.
Collapse
Affiliation(s)
- Xuzhi Zhao
- School of Electronic and Information Engineering, Beijing Jiaotong University, Beijing, China
| | - Yi Du
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education/Beijing), Department of Radiation Oncology, Peking University Cancer Hospital & Institute, Beijing, China.
- Institute of Medical Technology, Peking University Health Science Center, Beijing, China.
| | - Yahui Peng
- School of Electronic and Information Engineering, Beijing Jiaotong University, Beijing, China.
| |
Collapse
|
25
|
Zhou H, Liu W, Zhou Y, Song W, Zhang F, Zhu Y. Dual-domain Wasserstein Generative Adversarial Network with Hybrid Loss for Low-dose CT Imaging. Phys Med Biol 2025; 70:025018. [PMID: 39761646 DOI: 10.1088/1361-6560/ada687] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/01/2024] [Accepted: 01/06/2025] [Indexed: 01/21/2025]
Abstract
Objective.Low-dose computed tomography (LDCT) has gained significant attention in hospitals and clinics as a popular imaging modality for reducing the risk of x-ray radiation. However, reconstructed LDCT images often suffer from undesired noise and artifacts, which can negatively impact diagnostic accuracy. This study aims to develop a novel approach to improve LDCT imaging performance.Approach.A dual-domain Wasserstein generative adversarial network (DWGAN) with hybrid loss is proposed as an effective and integrated deep neural network (DNN) for LDCT imaging. The DWGAN comprises two key components: a generator (G) network and a discriminator (D) network. TheGnetwork is a dual-domain DNN designed to predict high-quality images by integrating three essential components: the projection-domain denoising module, filtered back-projection-based reconstruction layer, and image-domain enhancement module. TheDnetwork is a shallow convolutional neural network used to differentiate between real (label) and generated images. To prevent the reconstructed images from becoming excessively smooth and to preserve both structural and textural details, a hybrid loss function with weighting coefficients is incorporated into the DWGAN.Main results.Numerical experiments demonstrate that the proposed DWGAN can effectively suppress noise and better preserve image details compared with existing methods. Moreover, its application to head CT data confirms the superior performance of the DWGAN in restoring structural and textural details.Significance.The proposed DWGAN framework exhibits excellent performance in recovering structural and textural details in LDCT images. Furthermore, the framework can be applied to other tomographic imaging techniques that suffer from image distortion problems.
Collapse
Affiliation(s)
- Haichuan Zhou
- School of Medical Technology and Engineering, Henan University of Science and Technology, Luoyang 471000, People's Republic of China
| | - Wei Liu
- School of Medical Technology and Engineering, Henan University of Science and Technology, Luoyang 471000, People's Republic of China
| | - Yu Zhou
- School of Medical Technology and Engineering, Henan University of Science and Technology, Luoyang 471000, People's Republic of China
| | - Weidong Song
- School of Medical Technology and Engineering, Henan University of Science and Technology, Luoyang 471000, People's Republic of China
| | - Fengshou Zhang
- School of Medical Technology and Engineering, Henan University of Science and Technology, Luoyang 471000, People's Republic of China
| | - Yining Zhu
- School of Mathematical Sciences, Capital Normal University, Beijing 100048, People's Republic of China
| |
Collapse
|
26
|
Bhuyan SS, Sateesh V, Mukul N, Galvankar A, Mahmood A, Nauman M, Rai A, Bordoloi K, Basu U, Samuel J. Generative Artificial Intelligence Use in Healthcare: Opportunities for Clinical Excellence and Administrative Efficiency. J Med Syst 2025; 49:10. [PMID: 39820845 PMCID: PMC11739231 DOI: 10.1007/s10916-024-02136-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/27/2024] [Accepted: 12/19/2024] [Indexed: 01/19/2025]
Abstract
Generative Artificial Intelligence (Gen AI) has transformative potential in healthcare to enhance patient care, personalize treatment options, train healthcare professionals, and advance medical research. This paper examines various clinical and non-clinical applications of Gen AI. In clinical settings, Gen AI supports the creation of customized treatment plans, generation of synthetic data, analysis of medical images, nursing workflow management, risk prediction, pandemic preparedness, and population health management. By automating administrative tasks such as medical documentations, Gen AI has the potential to reduce clinician burnout, freeing more time for direct patient care. Furthermore, application of Gen AI may enhance surgical outcomes by providing real-time feedback and automation of certain tasks in operating rooms. The generation of synthetic data opens new avenues for model training for diseases and simulation, enhancing research capabilities and improving predictive accuracy. In non-clinical contexts, Gen AI improves medical education, public relations, revenue cycle management, healthcare marketing etc. Its capacity for continuous learning and adaptation enables it to drive ongoing improvements in clinical and operational efficiencies, making healthcare delivery more proactive, predictive, and precise.
Collapse
Affiliation(s)
- Soumitra S Bhuyan
- Edward J. Bloustein School of Planning and Public Policy, Rutgers, The State University of New Jersey, 255, Civic Square Building 33 Livingston Ave #400, New Brunswick, NJ, 08901, USA.
| | - Vidyoth Sateesh
- Edward J. Bloustein School of Planning and Public Policy, Rutgers, The State University of New Jersey, 255, Civic Square Building 33 Livingston Ave #400, New Brunswick, NJ, 08901, USA
| | - Naya Mukul
- School of Social Policy, Rice University, Houston, TX, USA
| | | | - Asos Mahmood
- Center for Health System Improvement, College of Medicine, The University of Tennessee Health Science Center, Memphis, TN, USA
| | | | - Akash Rai
- Edward J. Bloustein School of Planning and Public Policy, Rutgers, The State University of New Jersey, 255, Civic Square Building 33 Livingston Ave #400, New Brunswick, NJ, 08901, USA
| | - Kahuwa Bordoloi
- Department of Psychology and Counselling, St. Joseph's University, Bangalore, India
| | - Urmi Basu
- Insight Biopharma, Princeton, NJ, USA
| | - Jim Samuel
- Edward J. Bloustein School of Planning and Public Policy, Rutgers, The State University of New Jersey, 255, Civic Square Building 33 Livingston Ave #400, New Brunswick, NJ, 08901, USA
| |
Collapse
|
27
|
Kyung S, Won J, Pak S, Kim S, Lee S, Park K, Hong GS, Kim N. Generative Adversarial Network With Robust Discriminator Through Multi-Task Learning for Low-Dose CT Denoising. IEEE TRANSACTIONS ON MEDICAL IMAGING 2025; 44:499-518. [PMID: 39186436 DOI: 10.1109/tmi.2024.3449647] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/28/2024]
Abstract
Reducing the dose of radiation in computed tomography (CT) is vital to decreasing secondary cancer risk. However, the use of low-dose CT (LDCT) images is accompanied by increased noise that can negatively impact diagnoses. Although numerous deep learning algorithms have been developed for LDCT denoising, several challenges persist, including the visual incongruence experienced by radiologists, unsatisfactory performances across various metrics, and insufficient exploration of the networks' robustness in other CT domains. To address such issues, this study proposes three novel accretions. First, we propose a generative adversarial network (GAN) with a robust discriminator through multi-task learning that simultaneously performs three vision tasks: restoration, image-level, and pixel-level decisions. The more multi-tasks that are performed, the better the denoising performance of the generator, which means multi-task learning enables the discriminator to provide more meaningful feedback to the generator. Second, two regulatory mechanisms, restoration consistency (RC) and non-difference suppression (NDS), are introduced to improve the discriminator's representation capabilities. These mechanisms eliminate irrelevant regions and compare the discriminator's results from the input and restoration, thus facilitating effective GAN training. Lastly, we incorporate residual fast Fourier transforms with convolution (Res-FFT-Conv) blocks into the generator to utilize both frequency and spatial representations. This approach provides mixed receptive fields by using spatial (or local), spectral (or global), and residual connections. Our model was evaluated using various pixel- and feature-space metrics in two denoising tasks. Additionally, we conducted visual scoring with radiologists. The results indicate superior performance in both quantitative and qualitative measures compared to state-of-the-art denoising techniques.
Collapse
|
28
|
Wang S, Yang Y, Stevens GM, Yin Z, Wang AS. Emulating Low-Dose PCCT Image Pairs With Independent Noise for Self-Supervised Spectral Image Denoising. IEEE TRANSACTIONS ON MEDICAL IMAGING 2025; 44:530-542. [PMID: 39196747 DOI: 10.1109/tmi.2024.3449817] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/30/2024]
Abstract
Photon counting CT (PCCT) acquires spectral measurements and enables generation of material decomposition (MD) images that provide distinct advantages in various clinical situations. However, noise amplification is observed in MD images, and denoising is typically applied. Clean or high-quality references are rare in clinical scans, often making supervised learning (Noise2Clean) impractical. Noise2Noise is a self-supervised counterpart, using noisy images and corresponding noisy references with zero-mean, independent noise. PCCT counts transmitted photons separately, and raw measurements are assumed to follow a Poisson distribution in each energy bin, providing the possibility to create noise-independent pairs. The approach is to use binomial selection to split the counts into two low-dose scans with independent noise. We prove that the reconstructed spectral images inherit the noise independence from counts domain through noise propagation analysis and also validated it in numerical simulation and experimental phantom scans. The method offers the flexibility to split measurements into desired dose levels while ensuring the reconstructed images share identical underlying features, thereby strengthening the model's robustness for input dose levels and capability of preserving fine details. In both numerical simulation and experimental phantom scans, we demonstrated that Noise2Noise with binomial selection outperforms other common self-supervised learning methods based on different presumptive conditions.
Collapse
|
29
|
Xue H, Yao Y, Teng Y. Noise-assisted hybrid attention networks for low-dose PET and CT denoising. Med Phys 2025; 52:444-453. [PMID: 39431968 DOI: 10.1002/mp.17430] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/14/2024] [Revised: 07/25/2024] [Accepted: 09/04/2024] [Indexed: 10/22/2024] Open
Abstract
BACKGROUND Positron emission tomography (PET) and computed tomography (CT) play a vital role in tumor-related medical diagnosis, assessment, and treatment planning. However, full-dose PET and CT pose the risk of excessive radiation exposure to patients, whereas low-dose images compromise image quality, impacting subsequent tumor recognition and disease diagnosis. PURPOSE To solve such problems, we propose a Noise-Assisted Hybrid Attention Network (NAHANet) to reconstruct full-dose PET and CT images from low-dose PET (LDPET) and CT (LDCT) images to reduce patient radiation risks while ensuring the performance of subsequent tumor recognition. METHODS NAHANet contains two branches: the noise feature prediction branch (NFPB) and the cascaded reconstruction branch. Among them, NFPB providing noise features for the cascade reconstruction branch. The cascaded reconstruction branch comprises a shallow feature extraction module and a reconstruction module which contains a series of cascaded noise feature fusion blocks (NFFBs). Among these, the NFFB fuses the features extracted from low-dose images with the noise features obtained by NFPB to improve the feature extraction capability. To validate the effectiveness of the NAHANet method, we performed experiments using two public available datasets: the Ultra-low Dose PET Imaging Challenge dataset and Low Dose CT Grand Challenge dataset. RESULTS As a result, the proposed NAHANet achieved higher performance on common indicators. For example, on the CT dataset, the PSNR and SSIM indicators were improved by 4.1 dB and 0.06 respectively, and the rMSE indicator was reduced by 5.46 compared with the LDCT; on the PET dataset, the PSNR and SSIM was improved by 3.37 dB and 0.02, and the rMSE was reduced by 9.04 compared with the LDPET. CONCLUSIONS This paper proposes a transformer-based denoising algorithm, which utilizes hybrid attention to extract high-level features of low dose images and fuses noise features to optimize the denoising performance of the network, achieving good performance improvements on low-dose CT and PET datasets.
Collapse
Affiliation(s)
- Hengzhi Xue
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
| | - Yudong Yao
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
- Department of Electrical and Computer Engineering, Steven Institute of Technology, Hoboken, New Jersey, USA
| | - Yueyang Teng
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
- Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Shenyang, China
| |
Collapse
|
30
|
Dou Q, Wang Z, Feng X, Campbell‐Washburn AE, Mugler JP, Meyer CH. MRI denoising with a non-blind deep complex-valued convolutional neural network. NMR IN BIOMEDICINE 2025; 38:e5291. [PMID: 39523816 PMCID: PMC11605166 DOI: 10.1002/nbm.5291] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/03/2023] [Revised: 10/10/2024] [Accepted: 10/26/2024] [Indexed: 11/16/2024]
Abstract
MR images with high signal-to-noise ratio (SNR) provide more diagnostic information. Various methods for MRI denoising have been developed, but the majority of them operate on the magnitude image and neglect the phase information. Therefore, the goal of this work is to design and implement a complex-valued convolutional neural network (CNN) for MRI denoising. A complex-valued CNN incorporating the noise level map (non-blindℂ $$ \mathbb{C} $$ DnCNN) was trained with ground truth and simulated noise-corrupted image pairs. The proposed method was validated using both simulated and in vivo data collected from low-field scanners. Its denoising performance was quantitively and qualitatively evaluated, and it was compared with the real-valued CNN and several other algorithms. For the simulated noise-corrupted testing dataset, the complex-valued models had superior normalized root-mean-square error, peak SNR, structural similarity index, and phase ABSD. By incorporating the noise level map, the non-blindℂ $$ \mathbb{C} $$ DnCNN showed better performance in dealing with spatially varying parallel imaging noise. For in vivo low-field data, the non-blindℂ $$ \mathbb{C} $$ DnCNN significantly improved the SNR and visual quality of the image. The proposed non-blindℂ $$ \mathbb{C} $$ DnCNN provides an efficient and effective approach for MRI denoising. This is the first application of non-blindℂ $$ \mathbb{C} $$ DnCNN to medical imaging. The method holds the potential to enable improved low-field MRI, facilitating enhanced diagnostic imaging in under-resourced areas.
Collapse
Affiliation(s)
- Quan Dou
- Department of Biomedical EngineeringUniversity of VirginiaCharlottesvilleVirginiaUSA
| | - Zhixing Wang
- Department of Biomedical EngineeringUniversity of VirginiaCharlottesvilleVirginiaUSA
| | - Xue Feng
- Department of Biomedical EngineeringUniversity of VirginiaCharlottesvilleVirginiaUSA
| | - Adrienne E. Campbell‐Washburn
- Cardiovascular Branch, Division of Intramural ResearchNational Heart, Lung, and Blood Institute, National Institutes of HealthBethesdaMarylandUSA
| | - John P. Mugler
- Department of Radiology and Medical ImagingUniversity of VirginiaCharlottesvilleVirginiaUSA
| | - Craig H. Meyer
- Department of Biomedical EngineeringUniversity of VirginiaCharlottesvilleVirginiaUSA
- Department of Radiology and Medical ImagingUniversity of VirginiaCharlottesvilleVirginiaUSA
| |
Collapse
|
31
|
Shen Y, Liang N, Zhong X, Ren J, Zheng Z, Li L, Yan B. CT image super-resolution under the guidance of deep gradient information. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2025; 33:58-71. [PMID: 39973779 DOI: 10.1177/08953996241289225] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/21/2025]
Abstract
Due to the hardware constraints of Computed Tomography (CT) imaging, acquiring high-resolution (HR) CT images in clinical settings poses a significant challenge. In recent years, convolutional neural networks have shown great potential in CT super-resolution (SR) problems. However, the reconstruction results of many deep learning-based SR methods have structural distortion and detail ambiguity. In this paper, a new SR network based on generative adversarial learning is proposed. The network consists of gradient branch and SR branch. Gradient branch is used to recover HR gradient maps. The network merges gradient image features of the gradient branch into the SR branch, offering gradient information guidance for super-resolution (SR) reconstruction. Further, the loss function of the network combines the image space loss function with the gradient loss and the gradient variance loss to further generate a more realistic detail texture. Compared to other comparison algorithms, the structural similarity index of the SR results obtained by the proposed method on simulation and experimental data has increased by 1.8% and 1.4%, respectively. The experimental results demonstrate that the proposed CT SR network exhibits superior performance in terms of structure preservation and detail restoration.
Collapse
Affiliation(s)
- Ye Shen
- These authors contributed equally to this work and should be considered co-first authors
| | - Ningning Liang
- These authors contributed equally to this work and should be considered co-first authors
| | | | | | | | | | | |
Collapse
|
32
|
Liu X, Xie Y, Liu C, Cheng J, Diao S, Tan S, Liang X. Diffusion probabilistic priors for zero-shot low-dose CT image denoising. Med Phys 2025; 52:329-345. [PMID: 39413369 DOI: 10.1002/mp.17431] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/05/2024] [Revised: 08/17/2024] [Accepted: 09/03/2024] [Indexed: 10/18/2024] Open
Abstract
BACKGROUND Denoising low-dose computed tomography (CT) images is a critical task in medical image computing. Supervised deep learning-based approaches have made significant advancements in this area in recent years. However, these methods typically require pairs of low-dose and normal-dose CT images for training, which are challenging to obtain in clinical settings. Existing unsupervised deep learning-based methods often require training with a large number of low-dose CT images or rely on specially designed data acquisition processes to obtain training data. PURPOSE To address these limitations, we propose a novel unsupervised method that only utilizes normal-dose CT images during training, enabling zero-shot denoising of low-dose CT images. METHODS Our method leverages the diffusion model, a powerful generative model. We begin by training a cascaded unconditional diffusion model capable of generating high-quality normal-dose CT images from low-resolution to high-resolution. The cascaded architecture makes the training of high-resolution diffusion models more feasible. Subsequently, we introduce low-dose CT images into the reverse process of the diffusion model as likelihood, combined with the priors provided by the diffusion model and iteratively solve multiple maximum a posteriori (MAP) problems to achieve denoising. Additionally, we propose methods to adaptively adjust the coefficients that balance the likelihood and prior in MAP estimations, allowing for adaptation to different noise levels in low-dose CT images. RESULTS We test our method on low-dose CT datasets of different regions with varying dose levels. The results demonstrate that our method outperforms the state-of-the-art unsupervised method and surpasses several supervised deep learning-based methods. Our method achieves PSNR of 45.02 and 35.35 dB on the abdomen CT dataset and the chest CT dataset, respectively, surpassing the best unsupervised algorithm Noise2Sim in the comparative methods by 0.39 and 0.85 dB, respectively. CONCLUSIONS We propose a novel low-dose CT image denoising method based on diffusion model. Our proposed method only requires normal-dose CT images as training data, greatly alleviating the data scarcity issue faced by most deep learning-based methods. At the same time, as an unsupervised algorithm, our method achieves very good qualitative and quantitative results. The Codes are available in https://github.com/DeepXuan/Dn-Dp.
Collapse
Affiliation(s)
- Xuan Liu
- School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan, China
| | - Yaoqin Xie
- Institute of Biomedical and Health Engineering, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Chenbin Liu
- Radiation Oncology Department, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital & Shenzhen Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Shenzhen, China
| | - Jun Cheng
- School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan, China
| | - Songhui Diao
- Institute of Biomedical and Health Engineering, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Shan Tan
- School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan, China
| | - Xiaokun Liang
- Institute of Biomedical and Health Engineering, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| |
Collapse
|
33
|
Zhang X, Fang C, Qiao Z. MAFA-Uformer: Multi-attention and dual-branch feature aggregation U-shaped transformer for sparse-view CT reconstruction. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2025; 33:157-166. [PMID: 39973771 DOI: 10.1177/08953996241300016] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/21/2025]
Abstract
BACKGROUND Although computed tomography (CT) is widely employed in disease detection, X-ray radiation may pose a risk to the health of patients. Reducing the projection views is a common method, however, the reconstructed images often suffer from streak artifacts. PURPOSE In previous related works, it can be found that the convolutional neural network (CNN) is proficient in extracting local features, while the Transformer is adept at capturing global information. To suppress streak artifacts for sparse-view CT, this study aims to develop a method that combines the advantages of CNN and Transformer. METHODS In this paper, we propose a Multi-Attention and Dual-Branch Feature Aggregation U-shaped Transformer network (MAFA-Uformer), which consists of two branches: CNN and Transformer. Firstly, with a coordinate attention mechanism, the Transformer branch can capture the overall structure and orientation information to provide a global context understanding of the image under reconstruction. Secondly, the CNN branch focuses on extracting crucial local features of images through channel spatial attention, thus enhancing detail recognition capabilities. Finally, through a feature fusion module, the global information from the Transformer and the local features from the CNN are integrated effectively. RESULTS Experimental results demonstrate that our method achieves outstanding performance in terms of peak signal-to-noise ratio (PSNR), structural similarity (SSIM), and root mean square error (RMSE). Compared with Restormer, our model achieves significant improvements: PSNR increases by 0.76 dB, SSIM improves by 0.44%, and RMSE decreases by 8.55%. CONCLUSION Our method not only effectively suppresses artifacts but also better preserves details and features, thereby providing robust support for accurate diagnosis of CT images.
Collapse
Affiliation(s)
- Xuan Zhang
- School of Computer and Information Technology, Shanxi University, Taiyuan, Shanxi, China
| | - Chenyun Fang
- School of Computer and Information Technology, Shanxi University, Taiyuan, Shanxi, China
| | - Zhiwei Qiao
- School of Computer and Information Technology, Shanxi University, Taiyuan, Shanxi, China
| |
Collapse
|
34
|
Eulig E, Jäger F, Maier J, Ommer B, Kachelrieß M. Reconstructing and analyzing the invariances of low-dose CT image denoising networks. Med Phys 2025; 52:188-200. [PMID: 39348044 PMCID: PMC11700010 DOI: 10.1002/mp.17413] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/25/2024] [Revised: 08/08/2024] [Accepted: 08/26/2024] [Indexed: 10/01/2024] Open
Abstract
BACKGROUND Deep learning-based methods led to significant advancements in many areas of medical imaging, most of which are concerned with the reduction of artifacts caused by motion, scatter, or noise. However, with most neural networks being black boxes, they remain notoriously difficult to interpret, hindering their clinical implementation. In particular, it has been shown that networks exhibit invariances w.r.t. input features, that is, they learn to ignore certain information in the input data. PURPOSE To improve the interpretability of deep learning-based low-dose CT image denoising networks. METHODS We learn a complete data representation of low-dose input images using a conditional variational autoencoder (cVAE). In this representation, invariances of any given denoising network are then disentangled from the information it is not invariant to using a conditional invertible neural network (cINN). At test time, image-space invariances are generated by applying the inverse of the cINN and subsequent decoding using the cVAE. We propose two methods to analyze sampled invariances and to find those that correspond to alterations of anatomical structures. RESULTS The proposed method is applied to four popular deep learning-based low-dose CT image denoising networks. We find that the networks are not only invariant to noise amplitude and realizations, but also to anatomical structures. CONCLUSIONS The proposed method is capable of reconstructing and analyzing invariances of deep learning-based low-dose CT image denoising networks. This is an important step toward interpreting deep learning-based methods for medical imaging, which is essential for their clinical implementation.
Collapse
Affiliation(s)
- Elias Eulig
- Division of X‐Ray Imaging and Computed TomographyGerman Cancer Research Center (DKFZ)HeidelbergGermany
- Faculty of Physics and AstronomyHeidelberg UniversityHeidelbergGermany
| | - Fabian Jäger
- Division of X‐Ray Imaging and Computed TomographyGerman Cancer Research Center (DKFZ)HeidelbergGermany
- Faculty of Physics and AstronomyHeidelberg UniversityHeidelbergGermany
| | - Joscha Maier
- Division of X‐Ray Imaging and Computed TomographyGerman Cancer Research Center (DKFZ)HeidelbergGermany
| | | | - Marc Kachelrieß
- Division of X‐Ray Imaging and Computed TomographyGerman Cancer Research Center (DKFZ)HeidelbergGermany
- Medical Faculty HeidelbergHeidelberg UniversityHeidelbergGermany
| |
Collapse
|
35
|
Song Z, Xue L, Xu J, Zhang B, Jin C, Yang J, Zou C. Real-World Low-Dose CT Image Denoising by Patch Similarity Purification. IEEE TRANSACTIONS ON IMAGE PROCESSING : A PUBLICATION OF THE IEEE SIGNAL PROCESSING SOCIETY 2024; PP:196-208. [PMID: 40030715 DOI: 10.1109/tip.2024.3515878] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/05/2025]
Abstract
Reducing the radiation dose in CT scanning is important to alleviate the damage to the human health in clinical scenes. A promising way is to replace the normal-dose CT (NDCT) imaging by low-dose CT (LDCT) imaging with lower tube voltage and tube current. This often brings severe noise to the LDCT images, which adversely affects the diagnosis accuracy. Most of existing LDCT image denoising networks are trained either with synthetic LDCT images or real-world LDCT and NDCT image pairs with huge spatial misalignment. However, the synthetic noise is very different from the complex noise in real-world LDCT images, while the huge spatial misalignment brings inaccurate predictions of tissue structures in the denoised LDCT images. To well utilize real-world LDCT and NDCT image pairs for LDCT image denoising, in this paper, we introduce a new Patch Similarity Purification (PSP) strategy to construct high-quality training dataset for network training. Specifically, our PSP strategy first perform binarization for each pair of image patches cropped from the corresponding LDCT and NDCT image pairs. For each pair of binary masks, it then computes their similarity ratio by common mask calculation, and the patch pair can be selected as a training sample if their mask similarity ratio is higher than a threshold. By using our PSP strategy, each training set of our Rabbit and Patient datasets contain hundreds of thousands of real-world LDCT and NDCT image patch pairs with negligible misalignment. Extensive experiments demonstrate the usefulness of our PSP strategy on purifying the training data and the effectiveness of training LDCT image denoising networks on our datasets. The code and dataset are provided at https://github.com/TuTusong/PSP.
Collapse
|
36
|
Tang Y, Lyu T, Jin H, Du Q, Wang J, Li Y, Li M, Chen Y, Zheng J. Domain adaptive noise reduction with iterative knowledge transfer and style generalization learning. Med Image Anal 2024; 98:103327. [PMID: 39191093 DOI: 10.1016/j.media.2024.103327] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/30/2023] [Revised: 08/20/2024] [Accepted: 08/21/2024] [Indexed: 08/29/2024]
Abstract
Low-dose computed tomography (LDCT) denoising tasks face significant challenges in practical imaging scenarios. Supervised methods encounter difficulties in real-world scenarios as there are no paired data for training. Moreover, when applied to datasets with varying noise patterns, these methods may experience decreased performance owing to the domain gap. Conversely, unsupervised methods do not require paired data and can be directly trained on real-world data. However, they often exhibit inferior performance compared to supervised methods. To address this issue, it is necessary to leverage the strengths of these supervised and unsupervised methods. In this paper, we propose a novel domain adaptive noise reduction framework (DANRF), which integrates both knowledge transfer and style generalization learning to effectively tackle the domain gap problem. Specifically, an iterative knowledge transfer method with knowledge distillation is selected to train the target model using unlabeled target data and a pre-trained source model trained with paired simulation data. Meanwhile, we introduce the mean teacher mechanism to update the source model, enabling it to adapt to the target domain. Furthermore, an iterative style generalization learning process is also designed to enrich the style diversity of the training dataset. We evaluate the performance of our approach through experiments conducted on multi-source datasets. The results demonstrate the feasibility and effectiveness of our proposed DANRF model in multi-source LDCT image processing tasks. Given its hybrid nature, which combines the advantages of supervised and unsupervised learning, and its ability to bridge domain gaps, our approach is well-suited for improving practical low-dose CT imaging in clinical settings. Code for our proposed approach is publicly available at https://github.com/tyfeiii/DANRF.
Collapse
Affiliation(s)
- Yufei Tang
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230026, China; Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China
| | - Tianling Lyu
- Research Center of Augmented Intelligence, Zhejiang Lab, Hangzhou, 310000, China
| | - Haoyang Jin
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230026, China; Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China
| | - Qiang Du
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230026, China; Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China
| | - Jiping Wang
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230026, China; Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China
| | - Yunxiang Li
- Nanovision Technology Co., Ltd., Beiqing Road, Haidian District, Beijing, 100094, China
| | - Ming Li
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230026, China; Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China.
| | - Yang Chen
- Laboratory of Image Science and Technology, the School of Computer Science and Engineering, Southeast University, Nanjing, 210096, China
| | - Jian Zheng
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230026, China; Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China; Shandong Laboratory of Advanced Biomaterials and Medical Devices in Weihai, Weihai, 264200, China.
| |
Collapse
|
37
|
Eulig E, Ommer B, Kachelrieß M. Benchmarking deep learning-based low-dose CT image denoising algorithms. Med Phys 2024; 51:8776-8788. [PMID: 39287517 DOI: 10.1002/mp.17379] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/09/2024] [Revised: 08/05/2024] [Accepted: 08/17/2024] [Indexed: 09/19/2024] Open
Abstract
BACKGROUND Long-lasting efforts have been made to reduce radiation dose and thus the potential radiation risk to the patient for computed tomography (CT) acquisitions without severe deterioration of image quality. To this end, various techniques have been employed over the years including iterative reconstruction methods and noise reduction algorithms. PURPOSE Recently, deep learning-based methods for noise reduction became increasingly popular and a multitude of papers claim ever improving performance both quantitatively and qualitatively. However, the lack of a standardized benchmark setup and inconsistencies in experimental design across studies hinder the verifiability and reproducibility of reported results. METHODS In this study, we propose a benchmark setup to overcome those flaws and improve reproducibility and verifiability of experimental results in the field. We perform a comprehensive and fair evaluation of several state-of-the-art methods using this standardized setup. RESULTS Our evaluation reveals that most deep learning-based methods show statistically similar performance, and improvements over the past years have been marginal at best. CONCLUSIONS This study highlights the need for a more rigorous and fair evaluation of novel deep learning-based methods for low-dose CT image denoising. Our benchmark setup is a first and important step towards this direction and can be used by future researchers to evaluate their algorithms.
Collapse
Affiliation(s)
- Elias Eulig
- Division of X-Ray Imaging and Computed Tomography, German Cancer Research Center (DKFZ), Heidelberg, Germany
- Faculty of Physics and Astronomy, Heidelberg University, Heidelberg, Germany
| | | | - Marc Kachelrieß
- Division of X-Ray Imaging and Computed Tomography, German Cancer Research Center (DKFZ), Heidelberg, Germany
- Medical Faculty Heidelberg, Heidelberg University, Heidelberg, Germany
| |
Collapse
|
38
|
Yun S, Lee S, Choi DI, Lee T, Cho S. TMAA-net: tensor-domain multi-planal anti-aliasing network for sparse-view CT image reconstruction. Phys Med Biol 2024; 69:225012. [PMID: 39481239 DOI: 10.1088/1361-6560/ad8da2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/22/2024] [Accepted: 10/31/2024] [Indexed: 11/02/2024]
Abstract
Objective.Among various deep-network-based sparse-view CT image reconstruction studies, the sinogram upscaling network has been predominantly employed to synthesize additional view information. However, the performance of the sinogram-based network is limited in terms of removing aliasing streak artifacts and recovering low-contrast small structures. In this study, we used a view-by-view back-projection (VVBP) tensor-domain network to overcome such limitations of the sinogram-based approaches.Approach.The proposed method offers advantages of addressing the aliasing artifacts directly in the 3D tensor domain over the 2D sinogram. In the tensor-domain network, the multi-planal anti-aliasing modules were used to remove artifacts within the coronal and sagittal tensor planes. In addition, the data-fidelity-based refinement module was also implemented to successively process output images of the tensor network to recover image sharpness and textures.Main result.The proposed method showed outperformance in terms of removing aliasing artifacts and recovering low-contrast details compared to other state-of-the-art sinogram-based networks. The performance was validated for both numerical and clinical projection data in a circular fan-beam CT configuration.Significance.We observed that view-by-view aliasing artifacts in sparse-view CT exhibit distinct patterns within the tensor planes, making them effectively removable in high-dimensional representations. Additionally, we demonstrated that the co-domain characteristics of tensor space processing offer higher generalization performance for aliasing artifact removal compared to conventional sinogram-domain processing.
Collapse
Affiliation(s)
- Sungho Yun
- Department of Nuclear and Quantum Engineering, Korea Advanced Institute of Science and Technology, Daejeon 34141, Republic of Korea
| | - Seoyoung Lee
- Department of Nuclear and Quantum Engineering, Korea Advanced Institute of Science and Technology, Daejeon 34141, Republic of Korea
| | - Da-In Choi
- Department of Nuclear and Quantum Engineering, Korea Advanced Institute of Science and Technology, Daejeon 34141, Republic of Korea
| | - Taewon Lee
- Department of Semiconductor Engineering, Hoseo University, Asan 31499, Republic of Korea
| | - Seungryong Cho
- Department of Nuclear and Quantum Engineering, Korea Advanced Institute of Science and Technology, Daejeon 34141, Republic of Korea
- KAIST Institute for IT Convergence, Korea Advanced Institute of Science and Technology, Daejeon 34141, Republic of Korea
| |
Collapse
|
39
|
Takeda T, Koyama Y, Ikeno H, Matsuishi S, Hirosaki N. Exploring new useful phosphors by combining experiments with machine learning. SCIENCE AND TECHNOLOGY OF ADVANCED MATERIALS 2024; 25:2421761. [PMID: 39525501 PMCID: PMC11544735 DOI: 10.1080/14686996.2024.2421761] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 08/09/2024] [Revised: 10/20/2024] [Accepted: 10/22/2024] [Indexed: 11/16/2024]
Abstract
New phosphors are consistently in demand for advances in solid-state lighting and displays. Conventional trial-and-error exploration experiments for new phosphors require considerable time. If a phosphor host suitable for the target luminescent property can be proposed using computational science, the speed of development of new phosphors will significantly increase, and unexpected/overlooked compositions could be proposed as candidates. As a more practical approach for developing new phosphors with target luminescent properties, we looked at combining experiments with machine learning on the topics of emission wavelength, full width at half maximum (FWHM) of the emission peak, temperature dependence of the emission spectrum (thermal quenching), new phosphors with new chemical composition or crystal structure, and high-throughput experiments.
Collapse
Affiliation(s)
- Takashi Takeda
- Research Center for Electronic and Optical Materials, National Institute for Materials Science (NIMS), Tsukuba, Japan
| | - Yukinori Koyama
- Center for Basic Research on Materials, National Institute for Materials Science (NIMS), Tsukuba, Japan
| | - Hidekazu Ikeno
- Department of Materials Science, Graduate School of Engineering, Osaka Metropolitan University, Sakai, Japan
| | - Satoru Matsuishi
- Research Center for Materials Nanoarchitectonics, National Institute for Materials Science (NIMS), Tsukuba, Japan
| | - Naoto Hirosaki
- Research Center for Electronic and Optical Materials, National Institute for Materials Science (NIMS), Tsukuba, Japan
| |
Collapse
|
40
|
Williams MC, Weir-McCall JR, Baldassarre LA, De Cecco CN, Choi AD, Dey D, Dweck MR, Isgum I, Kolossvary M, Leipsic J, Lin A, Lu MT, Motwani M, Nieman K, Shaw L, van Assen M, Nicol E. Artificial Intelligence and Machine Learning for Cardiovascular Computed Tomography (CCT): A White Paper of the Society of Cardiovascular Computed Tomography (SCCT). J Cardiovasc Comput Tomogr 2024; 18:519-532. [PMID: 39214777 DOI: 10.1016/j.jcct.2024.08.003] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 12/29/2023] [Revised: 08/03/2024] [Accepted: 08/05/2024] [Indexed: 09/04/2024]
Affiliation(s)
| | | | - Lauren A Baldassarre
- Section of Cardiovascular Medicine and Department of Radiology and Biomedical Imaging, Yale School of Medicine, New Haven, CT, USA
| | | | - Andrew D Choi
- The George Washington University School of Medicine, Washington, USA
| | - Damini Dey
- Biomedical Imaging Research Institute, Cedars-Sinai Medical Center, Los Angeles, USA
| | - Marc R Dweck
- Centre for Cardiovascular Science, University of Edinburgh, Edinburgh, UK
| | - Ivana Isgum
- Amsterdam University Medical Center, University of Amsterdam, Netherlands
| | - Márton Kolossvary
- Gottsegen National Cardiovascular Center, Budapest, Hungary, and Physiological Controls Research Center, University Research and Innovation Center, Óbuda University, Budapest, Hungary
| | | | - Andrew Lin
- Victorian Heart Institute and Monash Health Heart, Victorian Heart Hospital, Monash University, Australia
| | - Michael T Lu
- Massachusetts General Hospital Cardiovascular Imaging Research Center/Harvard Medical School, USA
| | | | | | - Leslee Shaw
- Icahn School of Medicine at Mount Sinai, New York, USA
| | | | - Edward Nicol
- Royal Brompton Hospital, Guys and St Thomas' NHS Foundation Trust, London, UK; School of Biomedical Engineering and Imaging Sciences, King's College London, UK
| |
Collapse
|
41
|
Kim W, Jeon SY, Byun G, Yoo H, Choi JH. A systematic review of deep learning-based denoising for low-dose computed tomography from a perceptual quality perspective. Biomed Eng Lett 2024; 14:1153-1173. [PMID: 39465112 PMCID: PMC11502640 DOI: 10.1007/s13534-024-00419-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/12/2024] [Revised: 08/03/2024] [Accepted: 08/18/2024] [Indexed: 10/29/2024] Open
Abstract
Low-dose computed tomography (LDCT) scans are essential in reducing radiation exposure but often suffer from significant image noise that can impair diagnostic accuracy. While deep learning approaches have enhanced LDCT denoising capabilities, the predominant reliance on objective metrics like PSNR and SSIM has resulted in over-smoothed images that lack critical detail. This paper explores advanced deep learning methods tailored specifically to improve perceptual quality in LDCT images, focusing on generating diagnostic-quality images preferred in clinical practice. We review and compare current methodologies, including perceptual loss functions and generative adversarial networks, addressing the significant limitations of current benchmarks and the subjective nature of perceptual quality evaluation. Through a systematic analysis, this study underscores the urgent need for developing methods that balance both perceptual and diagnostic quality, proposing new directions for future research in the field.
Collapse
Affiliation(s)
- Wonjin Kim
- Department of Mechanical Engineering, Korean Advanced Institute of Science and Technology, 291, Daehak-ro, Yuseong-gu, Daejeon, 34141 Korea
- AI Analysis Team, Dotter Inc., 225 Gasan Digital 1-ro, Geumchoen-gu, Seoul, 08501 Korea
| | - Sun-Young Jeon
- Department of Artificial Intelligence and Software, Ewha Womans University, 52, Ewhayeodae-gil, Seodaemun-gu, Seoul, 03760 Korea
| | - Gyuri Byun
- Department of Artificial Intelligence and Software, Ewha Womans University, 52, Ewhayeodae-gil, Seodaemun-gu, Seoul, 03760 Korea
| | - Hongki Yoo
- Department of Mechanical Engineering, Korean Advanced Institute of Science and Technology, 291, Daehak-ro, Yuseong-gu, Daejeon, 34141 Korea
| | - Jang-Hwan Choi
- Department of Artificial Intelligence and Software, Ewha Womans University, 52, Ewhayeodae-gil, Seodaemun-gu, Seoul, 03760 Korea
- Computational Medicine, Graduate Program in System Health Science and Engineering, Ewha Womans University, 52, Ewhayeodae-gil, Seodaemun-gu, Seoul, 03760 Korea
| |
Collapse
|
42
|
Shi Y, Gao Y, Xu Q, Li Y, Mou X, Liang Z. Learned Tensor Neural Network Texture Prior for Photon-Counting CT Reconstruction. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:3830-3842. [PMID: 38753483 DOI: 10.1109/tmi.2024.3402079] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/18/2024]
Abstract
Photon-counting computed tomography (PCCT) reconstructs multiple energy-channel images to describe the same object, where there exists a strong correlation among different channel images. In addition, reconstruction of each channel image suffers photon count starving problem. To make full use of the correlation among different channel images to suppress the data noise and enhance the texture details in reconstructing each channel image, this paper proposes a tensor neural network (TNN) architecture to learn a multi-channel texture prior for PCCT reconstruction. Specifically, we first learn a spatial texture prior in each individual channel image by modeling the relationship between the center pixels and its corresponding neighbor pixels using a neural network. Then, we merge the single channel spatial texture prior into multi-channel neural network to learn the spectral local correlation information among different channel images. Since our proposed TNN is trained on a series of unpaired small spatial-spectral cubes which are extracted from one single reference multi-channel image, the local correlation in the spatial-spectral cubes is considered by TNN. To boost the TNN performance, a low-rank representation is also employed to consider the global correlation among different channel images. Finally, we integrate the learned TNN and the low-rank representation as priors into Bayesian reconstruction framework. To evaluate the performance of the proposed method, four references are considered. One is simulated images from ultra-high-resolution CT. One is spectral images from dual-energy CT. The other two are animal tissue and preclinical mouse images from a custom-made PCCT systems. Our TNN prior Bayesian reconstruction demonstrated better performance than other state-of-the-art competing algorithms, in terms of not only preserving texture feature but also suppressing image noise in each channel image.
Collapse
|
43
|
Choi K. Self-supervised learning for CT image denoising and reconstruction: a review. Biomed Eng Lett 2024; 14:1207-1220. [PMID: 39465103 PMCID: PMC11502646 DOI: 10.1007/s13534-024-00424-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/15/2024] [Revised: 08/28/2024] [Accepted: 09/03/2024] [Indexed: 10/29/2024] Open
Abstract
This article reviews the self-supervised learning methods for CT image denoising and reconstruction. Currently, deep learning has become a dominant tool in medical imaging as well as computer vision. In particular, self-supervised learning approaches have attracted great attention as a technique for learning CT images without clean/noisy references. After briefly reviewing the fundamentals of CT image denoising and reconstruction, we examine the progress of deep learning in CT image denoising and reconstruction. Finally, we focus on the theoretical and methodological evolution of self-supervised learning for image denoising and reconstruction.
Collapse
Affiliation(s)
- Kihwan Choi
- Department of Applied Artificial Intelligence, Seoul National University of Science and Technology, 232 Gongneung-ro, Nowon-gu, Seoul, 01811 Republic of Korea
| |
Collapse
|
44
|
Lu Y, Xu Z, Hyung Choi M, Kim J, Jung SW. Cross-Domain Denoising for Low-Dose Multi-Frame Spiral Computed Tomography. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:3949-3963. [PMID: 38787677 DOI: 10.1109/tmi.2024.3405024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/26/2024]
Abstract
Computed tomography (CT) has been used worldwide as a non-invasive test to assist in diagnosis. However, the ionizing nature of X-ray exposure raises concerns about potential health risks such as cancer. The desire for lower radiation doses has driven researchers to improve reconstruction quality. Although previous studies on low-dose computed tomography (LDCT) denoising have demonstrated the effectiveness of learning-based methods, most were developed on the simulated data. However, the real-world scenario differs significantly from the simulation domain, especially when using the multi-slice spiral scanner geometry. This paper proposes a two-stage method for the commercially available multi-slice spiral CT scanners that better exploits the complete reconstruction pipeline for LDCT denoising across different domains. Our approach makes good use of the high redundancy of multi-slice projections and the volumetric reconstructions while leveraging the over-smoothing issue in conventional cascaded frameworks caused by aggressive denoising. The dedicated design also provides a more explicit interpretation of the data flow. Extensive experiments on various datasets showed that the proposed method could remove up to 70% of noise without compromised spatial resolution, while subjective evaluations by two experienced radiologists further supported its superior performance against state-of-the-art methods in clinical practice. Code is available at https://github.com/YCL92/TMD-LDCT.
Collapse
|
45
|
Fan Y, Qin T, Sun Q, Wang M, Liang B. A Review of Factors Affecting Radiation Dose and Image Quality in Coronary CTA Performed with Wide-Detector CT. Tomography 2024; 10:1730-1743. [PMID: 39590936 PMCID: PMC11598146 DOI: 10.3390/tomography10110127] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2024] [Revised: 10/19/2024] [Accepted: 10/22/2024] [Indexed: 11/28/2024] Open
Abstract
Compared with traditional invasive coronary angiography (ICA), coronary CT angiography (CCTA) has the advantages of being rapid, economical, and minimally invasive. The wide-detector CT, with its superior temporal resolution and robust three-dimensional reconstruction technology, thus enables CCTA in patients with high heart rates and arrhythmias, leading to a high potential for clinical application. This paper systematically summarizes wide-detector CT hardware configurations of various vendors routinely used for CCTA examinations and reviews the effects of patient heart rate and heart rate variability, scanning modality, reconstruction algorithms, tube voltage, and scanning field of view on image quality and radiation dose. In addition, novel technologies in the field of CT applied to CCTA examinations are also presented. Since this examination has a diagnostic accuracy that is highly consistent with ICA, it can be further used as a routine examination tool for coronary artery disease in clinical practice.
Collapse
Affiliation(s)
- Yihan Fan
- School of Medical Imaging, Bengbu Medical University, Bengbu 233000, China; (Y.F.); (T.Q.); (Q.S.)
| | - Tian Qin
- School of Medical Imaging, Bengbu Medical University, Bengbu 233000, China; (Y.F.); (T.Q.); (Q.S.)
| | - Qingting Sun
- School of Medical Imaging, Bengbu Medical University, Bengbu 233000, China; (Y.F.); (T.Q.); (Q.S.)
| | - Mengting Wang
- The Second Hospital of Anhui Medical University, Hefei 230000, China;
| | - Baohui Liang
- School of Medical Imaging, Bengbu Medical University, Bengbu 233000, China; (Y.F.); (T.Q.); (Q.S.)
| |
Collapse
|
46
|
Wang B, Deng F, Jiang P, Wang S, Han X, Zhang Z. WiTUnet: A U-shaped architecture integrating CNN and Transformer for improved feature alignment and local information fusion. Sci Rep 2024; 14:25525. [PMID: 39462127 PMCID: PMC11512998 DOI: 10.1038/s41598-024-76886-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/29/2024] [Accepted: 10/17/2024] [Indexed: 10/28/2024] Open
Abstract
Low-dose computed tomography (LDCT) has emerged as the preferred technology for diagnostic medical imaging due to the potential health risks associated with X-ray radiation and conventional computed tomography (CT) techniques. While LDCT utilizes a lower radiation dose compared to standard CT, it results in increased image noise, which can impair the accuracy of diagnoses. To mitigate this issue, advanced deep learning-based LDCT denoising algorithms have been developed. These primarily utilize Convolutional Neural Networks (CNNs) or Transformer Networks and often employ the Unet architecture, which enhances image detail by integrating feature maps from the encoder and decoder via skip connections. However, existing methods focus excessively on the optimization of the encoder and decoder structures while overlooking potential enhancements to the Unet architecture itself. This oversight can be problematic due to significant differences in feature map characteristics between the encoder and decoder, where simple fusion strategies may hinder effective image reconstruction. In this paper, we introduce WiTUnet, a novel LDCT image denoising method that utilizes nested, dense skip pathway in place of traditional skip connections to improve feature integration. Additionally, to address the high computational demands of conventional Transformers on large images, WiTUnet incorporates a windowed Transformer structure that processes images in smaller, non-overlapping segments, significantly reducing computational load. Moreover, our approach includes a Local Image Perception Enhancement (LiPe) module within both the encoder and decoder to replace the standard multi-layer perceptron (MLP) in Transformers, thereby improving the capture and representation of local image features. Through extensive experimental comparisons, WiTUnet has demonstrated superior performance over existing methods in critical metrics such as Peak Signal-to-Noise Ratio (PSNR), Structural Similarity (SSIM), and Root Mean Square Error (RMSE), significantly enhancing noise removal and image quality. The code is available on github https://github.com/woldier/WiTUNet .
Collapse
Affiliation(s)
- Bin Wang
- College of Computer Science and Cyber Security, Chengdu University of Technology, Chengdu, 610059, Sichuan, China
| | - Fei Deng
- College of Computer Science and Cyber Security, Chengdu University of Technology, Chengdu, 610059, Sichuan, China.
| | - Peifan Jiang
- College of Geophysics, Chengdu University of Technology, Chengdu, 610059, Sichuan, China
| | - Shuang Wang
- College of Geophysics, Chengdu University of Technology, Chengdu, 610059, Sichuan, China
| | - Xiao Han
- College of Computer Science and Cyber Security, Chengdu University of Technology, Chengdu, 610059, Sichuan, China
| | - Zhixuan Zhang
- College of Mechanical and Vehicle Engineering, Changchun University, Changchun, 629100, Jilin, China
| |
Collapse
|
47
|
廖 静, 彭 声, 王 永, 边 兆. [A dual-domain cone beam computed tomography sparse-view reconstruction method based on generative projection interpolation]. NAN FANG YI KE DA XUE XUE BAO = JOURNAL OF SOUTHERN MEDICAL UNIVERSITY 2024; 44:2044-2054. [PMID: 39523105 PMCID: PMC11526453 DOI: 10.12122/j.issn.1673-4254.2024.10.23] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Subscribe] [Scholar Register] [Received: 05/31/2024] [Indexed: 11/16/2024]
Abstract
OBJECTIVE To propose a dual-domain CBCT reconstruction framework (DualSFR-Net) based on generative projection interpolation to reduce artifacts in sparse-view cone beam computed tomography (CBCT) reconstruction. METHODS The proposed method DualSFR-Net consists of a generative projection interpolation module, a domain transformation module, and an image restoration module. The generative projection interpolation module includes a sparse projection interpolation network (SPINet) based on a generative adversarial network and a full-view projection restoration network (FPRNet). SPINet performs projection interpolation to synthesize full-view projection data from the sparse-view projection data, while FPRNet further restores the synthesized full-view projection data. The domain transformation module introduces the FDK reconstruction and forward projection operators to complete the forward and gradient backpropagation processes. The image restoration module includes an image restoration network FIRNet that fine-tunes the domain-transformed images to eliminate residual artifacts and noise. RESULTS Validation experiments conducted on a dental CT dataset demonstrated that DualSFR-Net was capable to reconstruct high-quality CBCT images under sparse-view sampling protocols. Quantitatively, compared to the current best methods, the DualSFR-Net method improved the PSNR by 0.6615 and 0.7658 and increased the SSIM by 0.0053 and 0.0134 under 2-fold and 4-fold sparse protocols, respectively. CONCLUSION The proposed generative projection interpolation-based dual-domain CBCT sparse-view reconstruction method can effectively reduce stripe artifacts to improve image quality and enables efficient joint training for dual-domain imaging networks in sparse-view CBCT reconstruction.
Collapse
|
48
|
Shi Y, Xia W, Wang G, Mou X. Blind CT Image Quality Assessment Using DDPM-Derived Content and Transformer-Based Evaluator. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:3559-3569. [PMID: 38913529 PMCID: PMC11560125 DOI: 10.1109/tmi.2024.3418652] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/26/2024]
Abstract
Lowering radiation dose per view and utilizing sparse views per scan are two common CT scan modes, albeit often leading to distorted images characterized by noise and streak artifacts. Blind image quality assessment (BIQA) strives to evaluate perceptual quality in alignment with what radiologists perceive, which plays an important role in advancing low-dose CT reconstruction techniques. An intriguing direction involves developing BIQA methods that mimic the operational characteristic of the human visual system (HVS). The internal generative mechanism (IGM) theory reveals that the HVS actively deduces primary content to enhance comprehension. In this study, we introduce an innovative BIQA metric that emulates the active inference process of IGM. Initially, an active inference module, implemented as a denoising diffusion probabilistic model (DDPM), is constructed to anticipate the primary content. Then, the dissimilarity map is derived by assessing the interrelation between the distorted image and its primary content. Subsequently, the distorted image and dissimilarity map are combined into a multi-channel image, which is inputted into a transformer-based image quality evaluator. By leveraging the DDPM-derived primary content, our approach achieves competitive performance on a low-dose CT dataset.
Collapse
|
49
|
Sharma V, Awate SP. Adversarial EM for variational deep learning: Application to semi-supervised image quality enhancement in low-dose PET and low-dose CT. Med Image Anal 2024; 97:103291. [PMID: 39121545 DOI: 10.1016/j.media.2024.103291] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/23/2022] [Revised: 07/23/2024] [Accepted: 07/25/2024] [Indexed: 08/12/2024]
Abstract
In positron emission tomography (PET) and X-ray computed tomography (CT), reducing radiation dose can cause significant degradation in image quality. For image quality enhancement in low-dose PET and CT, we propose a novel theoretical adversarial and variational deep neural network (DNN) framework relying on expectation maximization (EM) based learning, termed adversarial EM (AdvEM). AdvEM proposes an encoder-decoder architecture with a multiscale latent space, and generalized-Gaussian models enabling datum-specific robust statistical modeling in latent space and image space. The model robustness is further enhanced by including adversarial learning in the training protocol. Unlike typical variational-DNN learning, AdvEM proposes latent-space sampling from the posterior distribution, and uses a Metropolis-Hastings scheme. Unlike existing schemes for PET or CT image enhancement which train using pairs of low-dose images with their corresponding normal-dose versions, we propose a semi-supervised AdvEM (ssAdvEM) framework that enables learning using a small number of normal-dose images. AdvEM and ssAdvEM enable per-pixel uncertainty estimates for their outputs. Empirical analyses on real-world PET and CT data involving many baselines, out-of-distribution data, and ablation studies show the benefits of the proposed framework.
Collapse
Affiliation(s)
- Vatsala Sharma
- Computer Science and Engineering (CSE) Department, Indian Institute of Technology (IIT) Bombay, Mumbai, India.
| | - Suyash P Awate
- Computer Science and Engineering (CSE) Department, Indian Institute of Technology (IIT) Bombay, Mumbai, India
| |
Collapse
|
50
|
Chen S, Qiu J, Zhang H, Yu Y, Chen H, Sun Y. Speech Fatigue Recognition Under Small Samples Based on Generative Adversarial Networks and BLSTM. INT J PATTERN RECOGN 2024; 38. [DOI: 10.1142/s0218001424580059] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/05/2025]
Abstract
To address the issue of low accuracy in speech fatigue recognition (SFR) under small samples, a method for small-sample SFR based on generative adversarial networks (GANs) is proposed. First, we enable the generator and discriminator to adversarially train and learn the features of the samples, and use the generator to generate high-quality simulated samples to expand our dataset. Then, we transfer discriminator parameters to fatigue identification network to accelerate network training speed. Furthermore, we use a bidirectional long short-term memory network (BLSTM) to further learn temporal fatigue features and improve the recognition rate of fatigue. 720 speech samples from a self-made Chinese speech database (SUSP-SFD) were chosen for training and testing. The results indicate that compared with traditional SFR methods, like convolutional neural networks (CNNs) and long short-term memory network (LSTM), our method improved the SFR rate by about 2.3–6.7%, verifying the effectiveness of the method.
Collapse
Affiliation(s)
- Shuxi Chen
- School of Computer and Information Engineering, Nantong Institute of Technology, Yongxing Road 211, Nantong 226002, P. R. China
| | - Jianlin Qiu
- School of Information Science and Technology, Nantong University, Seyuan Road 9, Nantong 226019, P. R. China
| | - Haifei Zhang
- School of Computer and Information Engineering, Nantong Institute of Technology, Yongxing Road 211, Nantong 226002, P. R. China
| | - Yifan Yu
- School of Computer and Information Engineering, Nantong Institute of Technology, Yongxing Road 211, Nantong 226002, P. R. China
| | - Hao Chen
- School of Electrical Engineering and Computer Science, Royal Institute of Technology (KTH), Brinellvägen 8, 114 28 Stockholm, Sweden
| | - Yiyang Sun
- School of Computer and Information Engineering, Nantong Institute of Technology, Yongxing Road 211, Nantong 226002, P. R. China
| |
Collapse
|