1
|
Chen K, Liu J, Wan R, Ho-Fun Lee V, Vardhanabhuti V, Yan H, Li H. Unsupervised Domain Adaptation for Low-Dose CT Reconstruction via Bayesian Uncertainty Alignment. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2025; 36:8525-8539. [PMID: 38985555 DOI: 10.1109/tnnls.2024.3409573] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/12/2024]
Abstract
Low-dose computed tomography (LDCT) image reconstruction techniques can reduce patient radiation exposure while maintaining acceptable imaging quality. Deep learning (DL) is widely used in this problem, but the performance of testing data (also known as target domain) is often degraded in clinical scenarios due to the variations that were not encountered in training data (also known as source domain). Unsupervised domain adaptation (UDA) of LDCT reconstruction has been proposed to solve this problem through distribution alignment. However, existing UDA methods fail to explore the usage of uncertainty quantification, which is crucial for reliable intelligent medical systems in clinical scenarios with unexpected variations. Moreover, existing direct alignment for different patients would lead to content mismatch issues. To address these issues, we propose to leverage a probabilistic reconstruction framework to conduct a joint discrepancy minimization between source and target domains in both the latent and image spaces. In the latent space, we devise a Bayesian uncertainty alignment to reduce the epistemic gap between the two domains. This approach reduces the uncertainty level of target domain data, making it more likely to render well-reconstructed results on target domains. In the image space, we propose a sharpness-aware distribution alignment (SDA) to achieve a match of second-order information, which can ensure that the reconstructed images from the target domain have similar sharpness to normal-dose CT (NDCT) images from the source domain. Experimental results on two simulated datasets and one clinical low-dose imaging dataset show that our proposed method outperforms other methods in quantitative and visualized performance.
Collapse
|
2
|
Yu X, Hu D, Yao Q, Fu Y, Zhong Y, Wang J, Tian M, Zhang H. Diffused Multi-scale Generative Adversarial Network for low-dose PET images reconstruction. Biomed Eng Online 2025; 24:16. [PMID: 39924498 PMCID: PMC11807330 DOI: 10.1186/s12938-025-01348-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/25/2024] [Accepted: 01/29/2025] [Indexed: 02/11/2025] Open
Abstract
PURPOSE The aim of this study is to convert low-dose PET (L-PET) images to full-dose PET (F-PET) images based on our Diffused Multi-scale Generative Adversarial Network (DMGAN) to offer a potential balance between reducing radiation exposure and maintaining diagnostic performance. METHODS The proposed method includes two modules: the diffusion generator and the u-net discriminator. The goal of the first module is to get different information from different levels, enhancing the generalization ability of the generator to the image and improving the stability of the training. Generated images are inputted into the u-net discriminator, extracting details from both overall and specific perspectives to enhance the quality of the generated F-PET images. We conducted evaluations encompassing both qualitative assessments and quantitative measures. In terms of quantitative comparisons, we employed two metrics, structure similarity index measure (SSIM) and peak signal-to-noise ratio (PSNR) to evaluate the performance of diverse methods. RESULTS Our proposed method achieved the highest PSNR and SSIM scores among the compared methods, which improved PSNR by at least 6.2% compared to the other methods. Compared to other methods, the synthesized full-dose PET image generated by our method exhibits a more accurate voxel-wise metabolic intensity distribution, resulting in a clearer depiction of the epilepsy focus. CONCLUSIONS The proposed method demonstrates improved restoration of original details from low-dose PET images compared to other models trained on the same datasets. This method offers a potential balance between minimizing radiation exposure and preserving diagnostic performance.
Collapse
Affiliation(s)
- Xiang Yu
- Polytechnic Institute, Zhejiang University, Hangzhou, China
| | - Daoyan Hu
- The College of Biomedical Engineering and Instrument Science of Zhejiang University, Hangzhou, China
| | - Qiong Yao
- Department of Nuclear Medicine and Medical PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, China
| | - Yu Fu
- College of Information Science and Electronic Engineering, Zhejiang University, Hangzhou, China
| | - Yan Zhong
- Department of Nuclear Medicine and Medical PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, China
| | - Jing Wang
- Department of Nuclear Medicine and Medical PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, China
| | - Mei Tian
- Human Phenome Institute, Fudan University, 825 Zhangheng Road, Shanghai, 201203, China.
| | - Hong Zhang
- The College of Biomedical Engineering and Instrument Science of Zhejiang University, Hangzhou, China.
- Department of Nuclear Medicine and Medical PET Center, The Second Affiliated Hospital of Zhejiang University School of Medicine, Hangzhou, 310009, China.
- Key Laboratory for Biomedical Engineering of Ministry of Education, Zhejiang University, Hangzhou, China.
| |
Collapse
|
3
|
Xue H, Yao Y, Teng Y. Noise-assisted hybrid attention networks for low-dose PET and CT denoising. Med Phys 2025; 52:444-453. [PMID: 39431968 DOI: 10.1002/mp.17430] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/14/2024] [Revised: 07/25/2024] [Accepted: 09/04/2024] [Indexed: 10/22/2024] Open
Abstract
BACKGROUND Positron emission tomography (PET) and computed tomography (CT) play a vital role in tumor-related medical diagnosis, assessment, and treatment planning. However, full-dose PET and CT pose the risk of excessive radiation exposure to patients, whereas low-dose images compromise image quality, impacting subsequent tumor recognition and disease diagnosis. PURPOSE To solve such problems, we propose a Noise-Assisted Hybrid Attention Network (NAHANet) to reconstruct full-dose PET and CT images from low-dose PET (LDPET) and CT (LDCT) images to reduce patient radiation risks while ensuring the performance of subsequent tumor recognition. METHODS NAHANet contains two branches: the noise feature prediction branch (NFPB) and the cascaded reconstruction branch. Among them, NFPB providing noise features for the cascade reconstruction branch. The cascaded reconstruction branch comprises a shallow feature extraction module and a reconstruction module which contains a series of cascaded noise feature fusion blocks (NFFBs). Among these, the NFFB fuses the features extracted from low-dose images with the noise features obtained by NFPB to improve the feature extraction capability. To validate the effectiveness of the NAHANet method, we performed experiments using two public available datasets: the Ultra-low Dose PET Imaging Challenge dataset and Low Dose CT Grand Challenge dataset. RESULTS As a result, the proposed NAHANet achieved higher performance on common indicators. For example, on the CT dataset, the PSNR and SSIM indicators were improved by 4.1 dB and 0.06 respectively, and the rMSE indicator was reduced by 5.46 compared with the LDCT; on the PET dataset, the PSNR and SSIM was improved by 3.37 dB and 0.02, and the rMSE was reduced by 9.04 compared with the LDPET. CONCLUSIONS This paper proposes a transformer-based denoising algorithm, which utilizes hybrid attention to extract high-level features of low dose images and fuses noise features to optimize the denoising performance of the network, achieving good performance improvements on low-dose CT and PET datasets.
Collapse
Affiliation(s)
- Hengzhi Xue
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
| | - Yudong Yao
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
- Department of Electrical and Computer Engineering, Steven Institute of Technology, Hoboken, New Jersey, USA
| | - Yueyang Teng
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
- Key Laboratory of Intelligent Computing in Medical Image, Ministry of Education, Shenyang, China
| |
Collapse
|
4
|
Wirtensohn S, Schmid C, Berthe D, John D, Heck L, Taphorn K, Flenner S, Herzen J. Self-supervised denoising of grating-based phase-contrast computed tomography. Sci Rep 2024; 14:32169. [PMID: 39741166 DOI: 10.1038/s41598-024-83517-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2024] [Accepted: 12/16/2024] [Indexed: 01/02/2025] Open
Abstract
In the last decade, grating-based phase-contrast computed tomography (gbPC-CT) has received growing interest. It provides additional information about the refractive index decrement in the sample. This signal shows an increased soft-tissue contrast. However, the resolution dependence of the signal poses a challenge: its contrast enhancement is overcompensated by the low resolution in low-dose applications such as clinical computed tomography. As a result, the implementation of gbPC-CT is currently tied to a higher dose. To reduce the dose, we introduce the self-supervised deep learning network Noise2Inverse into the field of gbPC-CT. We evaluate the behavior of the Noise2Inverse parameters on the phase-contrast results. Afterward, we compare its results with other denoising methods, namely the Statistical Iterative Reconstruction, Block Matching 3D, and Patchwise Phase Retrieval. In the example of Noise2Inverse, we show that deep learning networks can deliver superior denoising results with respect to the investigated image quality metrics. Their application allows to increase the resolution while maintaining the dose. At higher resolutions, gbPC-CT can naturally deliver higher contrast than conventional absorption-based CT. Therefore, the application of machine learning-based denoisers shifts the dose-normalized image quality in favor of gbPC-CT, bringing it one step closer to medical application.
Collapse
Affiliation(s)
- Sami Wirtensohn
- Research Group Biomedical Imaging Physics, Department of Physics, TUM School of Natural Sciences, Technical University of Munich, 85748, Garching, Germany.
- Chair of Biomedical Physics, Department of Physics, TUM School of Natural Sciences, Technical University of Munich, 85748, Garching, Germany.
- Munich Institute of Biomedical Engineering, Technical University of Munich, 85748, Garching, Germany.
- Institute of Materials Physics, Helmholtz-Zentrum Hereon, 21502, Geesthacht, Germany.
| | - Clemens Schmid
- Chair of Biomedical Physics, Department of Physics, TUM School of Natural Sciences, Technical University of Munich, 85748, Garching, Germany
- Munich Institute of Biomedical Engineering, Technical University of Munich, 85748, Garching, Germany
- Paul Scherer Institute, Forschungsstrasse 111, 5232, Villigen, Switzerland
| | - Daniel Berthe
- Chair of Biomedical Physics, Department of Physics, TUM School of Natural Sciences, Technical University of Munich, 85748, Garching, Germany
- Munich Institute of Biomedical Engineering, Technical University of Munich, 85748, Garching, Germany
| | - Dominik John
- Research Group Biomedical Imaging Physics, Department of Physics, TUM School of Natural Sciences, Technical University of Munich, 85748, Garching, Germany
- Chair of Biomedical Physics, Department of Physics, TUM School of Natural Sciences, Technical University of Munich, 85748, Garching, Germany
- Munich Institute of Biomedical Engineering, Technical University of Munich, 85748, Garching, Germany
- Institute of Materials Physics, Helmholtz-Zentrum Hereon, 21502, Geesthacht, Germany
| | - Lisa Heck
- Chair of Biomedical Physics, Department of Physics, TUM School of Natural Sciences, Technical University of Munich, 85748, Garching, Germany
- Munich Institute of Biomedical Engineering, Technical University of Munich, 85748, Garching, Germany
| | - Kirsten Taphorn
- Chair of Biomedical Physics, Department of Physics, TUM School of Natural Sciences, Technical University of Munich, 85748, Garching, Germany
- Munich Institute of Biomedical Engineering, Technical University of Munich, 85748, Garching, Germany
| | - Silja Flenner
- Institute of Materials Physics, Helmholtz-Zentrum Hereon, 21502, Geesthacht, Germany
| | - Julia Herzen
- Research Group Biomedical Imaging Physics, Department of Physics, TUM School of Natural Sciences, Technical University of Munich, 85748, Garching, Germany
- Chair of Biomedical Physics, Department of Physics, TUM School of Natural Sciences, Technical University of Munich, 85748, Garching, Germany
- Munich Institute of Biomedical Engineering, Technical University of Munich, 85748, Garching, Germany
| |
Collapse
|
5
|
Duan X, Ding XF, Khoz S, Chen X, Zhu N. Development of a low-dose strategy for propagation-based imaging helical computed tomography (PBI-HCT): high image quality and reduced radiation dose. Biomed Phys Eng Express 2024; 11:015049. [PMID: 39681007 DOI: 10.1088/2057-1976/ad9f66] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/15/2024] [Accepted: 12/16/2024] [Indexed: 12/18/2024]
Abstract
Background. Propagation-based imaging computed tomography (PBI-CT) has been recently emerging for visualizing low-density materials due to its excellent image contrast and high resolution. Based on this, PBI-CT with a helical acquisition mode (PBI-HCT) offers superior imaging quality (e.g., fewer ring artifacts) and dose uniformity, making it ideal for biomedical imaging applications. However, the excessive radiation dose associated with high-resolution PBI-HCT may potentially harm objects or hosts being imaged, especially in live animal imaging, raising a great need to reduce radiation dose.Methods. In this study, we strategically integrated Sparse2Noise (a deep learning approach) with PBI-HCT imaging to reduce radiation dose without compromising image quality. Sparse2Noise uses paired low-dose noisy images with different photon fluxes and projection numbers for high-quality reconstruction via a convolutional neural network (CNN). Then, we examined the imaging quality and radiation dose of PBI-HCT imaging using Sparse2Noise, as compared to when Sparse2Noise was used in low-dose PBI-CT imaging (circular scanning mode). Furthermore, we conducted a comparison study on the use of Sparse2Noise versus two other state-of-the-art low-dose imaging algorithms (i.e., Noise2Noise and Noise2Inverse) for imaging low-density materials using PBI-HCT at equivalent dose levels.Results. Sparse2Noise allowed for a 90% dose reduction in PBI-HCT imaging while maintaining high image quality. As compared to PBI-CT imaging, the use of Sparse2Noise in PBI-HCT imaging shows more effective by reducing additional radiation dose (30%-36%). Furthermore, helical scanning mode also enhances the performance of existing low-dose algorithms (Noise2Noise and Noise2Inverse); nevertheless, Sparse2Noise shows significantly higher signal-to-noise ratio (SNR) value compared to Noise2Noise and Noise2Inverse at the same radiation dose level.Conclusions and significance. Our proposed low-dose imaging strategy Sparse2Noise can be effectively applied to PBI-HCT imaging technique and requires lower dose for acceptable quality imaging. This would represent a significant advance imaging for low-density materials imaging and for future live animals imaging applications.
Collapse
Affiliation(s)
- Xiaoman Duan
- Division of Biomedical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada
| | - Xiao Fan Ding
- Division of Biomedical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada
| | - Samira Khoz
- Department of Mechanical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada
| | - Xiongbiao Chen
- Division of Biomedical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada
- Department of Mechanical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada
| | - Ning Zhu
- Division of Biomedical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada
- Department of Chemical and Biological Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada
- Canadian Light Source, Saskatoon, S7N 2V3, SK, Canada
| |
Collapse
|
6
|
Tang Y, Lyu T, Jin H, Du Q, Wang J, Li Y, Li M, Chen Y, Zheng J. Domain adaptive noise reduction with iterative knowledge transfer and style generalization learning. Med Image Anal 2024; 98:103327. [PMID: 39191093 DOI: 10.1016/j.media.2024.103327] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/30/2023] [Revised: 08/20/2024] [Accepted: 08/21/2024] [Indexed: 08/29/2024]
Abstract
Low-dose computed tomography (LDCT) denoising tasks face significant challenges in practical imaging scenarios. Supervised methods encounter difficulties in real-world scenarios as there are no paired data for training. Moreover, when applied to datasets with varying noise patterns, these methods may experience decreased performance owing to the domain gap. Conversely, unsupervised methods do not require paired data and can be directly trained on real-world data. However, they often exhibit inferior performance compared to supervised methods. To address this issue, it is necessary to leverage the strengths of these supervised and unsupervised methods. In this paper, we propose a novel domain adaptive noise reduction framework (DANRF), which integrates both knowledge transfer and style generalization learning to effectively tackle the domain gap problem. Specifically, an iterative knowledge transfer method with knowledge distillation is selected to train the target model using unlabeled target data and a pre-trained source model trained with paired simulation data. Meanwhile, we introduce the mean teacher mechanism to update the source model, enabling it to adapt to the target domain. Furthermore, an iterative style generalization learning process is also designed to enrich the style diversity of the training dataset. We evaluate the performance of our approach through experiments conducted on multi-source datasets. The results demonstrate the feasibility and effectiveness of our proposed DANRF model in multi-source LDCT image processing tasks. Given its hybrid nature, which combines the advantages of supervised and unsupervised learning, and its ability to bridge domain gaps, our approach is well-suited for improving practical low-dose CT imaging in clinical settings. Code for our proposed approach is publicly available at https://github.com/tyfeiii/DANRF.
Collapse
Affiliation(s)
- Yufei Tang
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230026, China; Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China
| | - Tianling Lyu
- Research Center of Augmented Intelligence, Zhejiang Lab, Hangzhou, 310000, China
| | - Haoyang Jin
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230026, China; Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China
| | - Qiang Du
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230026, China; Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China
| | - Jiping Wang
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230026, China; Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China
| | - Yunxiang Li
- Nanovision Technology Co., Ltd., Beiqing Road, Haidian District, Beijing, 100094, China
| | - Ming Li
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230026, China; Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China.
| | - Yang Chen
- Laboratory of Image Science and Technology, the School of Computer Science and Engineering, Southeast University, Nanjing, 210096, China
| | - Jian Zheng
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei, 230026, China; Medical Imaging Department, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Sciences, Suzhou, 215163, China; Shandong Laboratory of Advanced Biomaterials and Medical Devices in Weihai, Weihai, 264200, China.
| |
Collapse
|
7
|
Mileto A, Yu L, Revels JW, Kamel S, Shehata MA, Ibarra-Rovira JJ, Wong VK, Roman-Colon AM, Lee JM, Elsayes KM, Jensen CT. State-of-the-Art Deep Learning CT Reconstruction Algorithms in Abdominal Imaging. Radiographics 2024; 44:e240095. [PMID: 39612283 PMCID: PMC11618294 DOI: 10.1148/rg.240095] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/09/2023] [Revised: 05/18/2023] [Accepted: 05/21/2023] [Indexed: 12/01/2024]
Abstract
The implementation of deep neural networks has spurred the creation of deep learning reconstruction (DLR) CT algorithms. DLR CT techniques encompass a spectrum of deep learning-based methodologies that operate during the different steps of the image creation, prior to or after the traditional image formation process (eg, filtered backprojection [FBP] or iterative reconstruction [IR]), or alternatively by fully replacing FBP or IR techniques. DLR algorithms effectively facilitate the reduction of image noise associated with low photon counts from reduced radiation dose protocols. DLR methods have emerged as an effective solution to ameliorate limitations observed with prior CT image reconstruction algorithms, including FBP and IR algorithms, which are not able to preserve image texture and diagnostic performance at low radiation dose levels. An additional advantage of DLR algorithms is their high reconstruction speed, hence targeting the ideal triad of features for a CT image reconstruction (ie, the ability to consistently provide diagnostic-quality images and achieve radiation dose imaging levels as low as reasonably possible, with high reconstruction speed). An accumulated body of evidence supports the clinical use of DLR algorithms in abdominal imaging across multiple CT imaging tasks. The authors explore the technical aspects of DLR CT algorithms and examine various approaches to image synthesis in DLR creation. The clinical applications of DLR algorithms are highlighted across various abdominal CT imaging domains, with emphasis on the supporting evidence for diverse clinical tasks. An overview of the current limitations of and outlook for DLR algorithms for CT is provided. ©RSNA, 2024.
Collapse
Affiliation(s)
- Achille Mileto
- From the Department of Radiology, University of Washington School of
Medicine, Seattle, Wash (A.M.); Department of Radiology, Mayo Clinic, Rochester,
Minn (L.Y.); Department of Radiology, New York University Grossman School of
Medicine, NYU Langone Health, New York, NY (J.W.R.); Departments of Radiation
Oncology (S.K.) and Abdominal Imaging (M.A.S., J.J.I.R., V.K.W., K.M.E.,
C.T.J.), The University of Texas MD Anderson Cancer Center, 1400 Pressler St,
Unit 1473, Houston, TX 77030-4009; Department of Radiology, Texas
Children's Hospital, Houston, Tex (A.M.R.C.); and Department of
Radiology, Seoul National University College of Medicine, Seoul, South Korea
(J.M.L.)
| | - Lifeng Yu
- From the Department of Radiology, University of Washington School of
Medicine, Seattle, Wash (A.M.); Department of Radiology, Mayo Clinic, Rochester,
Minn (L.Y.); Department of Radiology, New York University Grossman School of
Medicine, NYU Langone Health, New York, NY (J.W.R.); Departments of Radiation
Oncology (S.K.) and Abdominal Imaging (M.A.S., J.J.I.R., V.K.W., K.M.E.,
C.T.J.), The University of Texas MD Anderson Cancer Center, 1400 Pressler St,
Unit 1473, Houston, TX 77030-4009; Department of Radiology, Texas
Children's Hospital, Houston, Tex (A.M.R.C.); and Department of
Radiology, Seoul National University College of Medicine, Seoul, South Korea
(J.M.L.)
| | - Jonathan W. Revels
- From the Department of Radiology, University of Washington School of
Medicine, Seattle, Wash (A.M.); Department of Radiology, Mayo Clinic, Rochester,
Minn (L.Y.); Department of Radiology, New York University Grossman School of
Medicine, NYU Langone Health, New York, NY (J.W.R.); Departments of Radiation
Oncology (S.K.) and Abdominal Imaging (M.A.S., J.J.I.R., V.K.W., K.M.E.,
C.T.J.), The University of Texas MD Anderson Cancer Center, 1400 Pressler St,
Unit 1473, Houston, TX 77030-4009; Department of Radiology, Texas
Children's Hospital, Houston, Tex (A.M.R.C.); and Department of
Radiology, Seoul National University College of Medicine, Seoul, South Korea
(J.M.L.)
| | - Serageldin Kamel
- From the Department of Radiology, University of Washington School of
Medicine, Seattle, Wash (A.M.); Department of Radiology, Mayo Clinic, Rochester,
Minn (L.Y.); Department of Radiology, New York University Grossman School of
Medicine, NYU Langone Health, New York, NY (J.W.R.); Departments of Radiation
Oncology (S.K.) and Abdominal Imaging (M.A.S., J.J.I.R., V.K.W., K.M.E.,
C.T.J.), The University of Texas MD Anderson Cancer Center, 1400 Pressler St,
Unit 1473, Houston, TX 77030-4009; Department of Radiology, Texas
Children's Hospital, Houston, Tex (A.M.R.C.); and Department of
Radiology, Seoul National University College of Medicine, Seoul, South Korea
(J.M.L.)
| | - Mostafa A. Shehata
- From the Department of Radiology, University of Washington School of
Medicine, Seattle, Wash (A.M.); Department of Radiology, Mayo Clinic, Rochester,
Minn (L.Y.); Department of Radiology, New York University Grossman School of
Medicine, NYU Langone Health, New York, NY (J.W.R.); Departments of Radiation
Oncology (S.K.) and Abdominal Imaging (M.A.S., J.J.I.R., V.K.W., K.M.E.,
C.T.J.), The University of Texas MD Anderson Cancer Center, 1400 Pressler St,
Unit 1473, Houston, TX 77030-4009; Department of Radiology, Texas
Children's Hospital, Houston, Tex (A.M.R.C.); and Department of
Radiology, Seoul National University College of Medicine, Seoul, South Korea
(J.M.L.)
| | - Juan J. Ibarra-Rovira
- From the Department of Radiology, University of Washington School of
Medicine, Seattle, Wash (A.M.); Department of Radiology, Mayo Clinic, Rochester,
Minn (L.Y.); Department of Radiology, New York University Grossman School of
Medicine, NYU Langone Health, New York, NY (J.W.R.); Departments of Radiation
Oncology (S.K.) and Abdominal Imaging (M.A.S., J.J.I.R., V.K.W., K.M.E.,
C.T.J.), The University of Texas MD Anderson Cancer Center, 1400 Pressler St,
Unit 1473, Houston, TX 77030-4009; Department of Radiology, Texas
Children's Hospital, Houston, Tex (A.M.R.C.); and Department of
Radiology, Seoul National University College of Medicine, Seoul, South Korea
(J.M.L.)
| | - Vincenzo K. Wong
- From the Department of Radiology, University of Washington School of
Medicine, Seattle, Wash (A.M.); Department of Radiology, Mayo Clinic, Rochester,
Minn (L.Y.); Department of Radiology, New York University Grossman School of
Medicine, NYU Langone Health, New York, NY (J.W.R.); Departments of Radiation
Oncology (S.K.) and Abdominal Imaging (M.A.S., J.J.I.R., V.K.W., K.M.E.,
C.T.J.), The University of Texas MD Anderson Cancer Center, 1400 Pressler St,
Unit 1473, Houston, TX 77030-4009; Department of Radiology, Texas
Children's Hospital, Houston, Tex (A.M.R.C.); and Department of
Radiology, Seoul National University College of Medicine, Seoul, South Korea
(J.M.L.)
| | - Alicia M. Roman-Colon
- From the Department of Radiology, University of Washington School of
Medicine, Seattle, Wash (A.M.); Department of Radiology, Mayo Clinic, Rochester,
Minn (L.Y.); Department of Radiology, New York University Grossman School of
Medicine, NYU Langone Health, New York, NY (J.W.R.); Departments of Radiation
Oncology (S.K.) and Abdominal Imaging (M.A.S., J.J.I.R., V.K.W., K.M.E.,
C.T.J.), The University of Texas MD Anderson Cancer Center, 1400 Pressler St,
Unit 1473, Houston, TX 77030-4009; Department of Radiology, Texas
Children's Hospital, Houston, Tex (A.M.R.C.); and Department of
Radiology, Seoul National University College of Medicine, Seoul, South Korea
(J.M.L.)
| | - Jeong Min Lee
- From the Department of Radiology, University of Washington School of
Medicine, Seattle, Wash (A.M.); Department of Radiology, Mayo Clinic, Rochester,
Minn (L.Y.); Department of Radiology, New York University Grossman School of
Medicine, NYU Langone Health, New York, NY (J.W.R.); Departments of Radiation
Oncology (S.K.) and Abdominal Imaging (M.A.S., J.J.I.R., V.K.W., K.M.E.,
C.T.J.), The University of Texas MD Anderson Cancer Center, 1400 Pressler St,
Unit 1473, Houston, TX 77030-4009; Department of Radiology, Texas
Children's Hospital, Houston, Tex (A.M.R.C.); and Department of
Radiology, Seoul National University College of Medicine, Seoul, South Korea
(J.M.L.)
| | - Khaled M. Elsayes
- From the Department of Radiology, University of Washington School of
Medicine, Seattle, Wash (A.M.); Department of Radiology, Mayo Clinic, Rochester,
Minn (L.Y.); Department of Radiology, New York University Grossman School of
Medicine, NYU Langone Health, New York, NY (J.W.R.); Departments of Radiation
Oncology (S.K.) and Abdominal Imaging (M.A.S., J.J.I.R., V.K.W., K.M.E.,
C.T.J.), The University of Texas MD Anderson Cancer Center, 1400 Pressler St,
Unit 1473, Houston, TX 77030-4009; Department of Radiology, Texas
Children's Hospital, Houston, Tex (A.M.R.C.); and Department of
Radiology, Seoul National University College of Medicine, Seoul, South Korea
(J.M.L.)
| | - Corey T. Jensen
- From the Department of Radiology, University of Washington School of
Medicine, Seattle, Wash (A.M.); Department of Radiology, Mayo Clinic, Rochester,
Minn (L.Y.); Department of Radiology, New York University Grossman School of
Medicine, NYU Langone Health, New York, NY (J.W.R.); Departments of Radiation
Oncology (S.K.) and Abdominal Imaging (M.A.S., J.J.I.R., V.K.W., K.M.E.,
C.T.J.), The University of Texas MD Anderson Cancer Center, 1400 Pressler St,
Unit 1473, Houston, TX 77030-4009; Department of Radiology, Texas
Children's Hospital, Houston, Tex (A.M.R.C.); and Department of
Radiology, Seoul National University College of Medicine, Seoul, South Korea
(J.M.L.)
| |
Collapse
|
8
|
Kim W, Jeon SY, Byun G, Yoo H, Choi JH. A systematic review of deep learning-based denoising for low-dose computed tomography from a perceptual quality perspective. Biomed Eng Lett 2024; 14:1153-1173. [PMID: 39465112 PMCID: PMC11502640 DOI: 10.1007/s13534-024-00419-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/12/2024] [Revised: 08/03/2024] [Accepted: 08/18/2024] [Indexed: 10/29/2024] Open
Abstract
Low-dose computed tomography (LDCT) scans are essential in reducing radiation exposure but often suffer from significant image noise that can impair diagnostic accuracy. While deep learning approaches have enhanced LDCT denoising capabilities, the predominant reliance on objective metrics like PSNR and SSIM has resulted in over-smoothed images that lack critical detail. This paper explores advanced deep learning methods tailored specifically to improve perceptual quality in LDCT images, focusing on generating diagnostic-quality images preferred in clinical practice. We review and compare current methodologies, including perceptual loss functions and generative adversarial networks, addressing the significant limitations of current benchmarks and the subjective nature of perceptual quality evaluation. Through a systematic analysis, this study underscores the urgent need for developing methods that balance both perceptual and diagnostic quality, proposing new directions for future research in the field.
Collapse
Affiliation(s)
- Wonjin Kim
- Department of Mechanical Engineering, Korean Advanced Institute of Science and Technology, 291, Daehak-ro, Yuseong-gu, Daejeon, 34141 Korea
- AI Analysis Team, Dotter Inc., 225 Gasan Digital 1-ro, Geumchoen-gu, Seoul, 08501 Korea
| | - Sun-Young Jeon
- Department of Artificial Intelligence and Software, Ewha Womans University, 52, Ewhayeodae-gil, Seodaemun-gu, Seoul, 03760 Korea
| | - Gyuri Byun
- Department of Artificial Intelligence and Software, Ewha Womans University, 52, Ewhayeodae-gil, Seodaemun-gu, Seoul, 03760 Korea
| | - Hongki Yoo
- Department of Mechanical Engineering, Korean Advanced Institute of Science and Technology, 291, Daehak-ro, Yuseong-gu, Daejeon, 34141 Korea
| | - Jang-Hwan Choi
- Department of Artificial Intelligence and Software, Ewha Womans University, 52, Ewhayeodae-gil, Seodaemun-gu, Seoul, 03760 Korea
- Computational Medicine, Graduate Program in System Health Science and Engineering, Ewha Womans University, 52, Ewhayeodae-gil, Seodaemun-gu, Seoul, 03760 Korea
| |
Collapse
|
9
|
Choi K. Self-supervised learning for CT image denoising and reconstruction: a review. Biomed Eng Lett 2024; 14:1207-1220. [PMID: 39465103 PMCID: PMC11502646 DOI: 10.1007/s13534-024-00424-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/15/2024] [Revised: 08/28/2024] [Accepted: 09/03/2024] [Indexed: 10/29/2024] Open
Abstract
This article reviews the self-supervised learning methods for CT image denoising and reconstruction. Currently, deep learning has become a dominant tool in medical imaging as well as computer vision. In particular, self-supervised learning approaches have attracted great attention as a technique for learning CT images without clean/noisy references. After briefly reviewing the fundamentals of CT image denoising and reconstruction, we examine the progress of deep learning in CT image denoising and reconstruction. Finally, we focus on the theoretical and methodological evolution of self-supervised learning for image denoising and reconstruction.
Collapse
Affiliation(s)
- Kihwan Choi
- Department of Applied Artificial Intelligence, Seoul National University of Science and Technology, 232 Gongneung-ro, Nowon-gu, Seoul, 01811 Republic of Korea
| |
Collapse
|
10
|
Lu Y, Xu Z, Hyung Choi M, Kim J, Jung SW. Cross-Domain Denoising for Low-Dose Multi-Frame Spiral Computed Tomography. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:3949-3963. [PMID: 38787677 DOI: 10.1109/tmi.2024.3405024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/26/2024]
Abstract
Computed tomography (CT) has been used worldwide as a non-invasive test to assist in diagnosis. However, the ionizing nature of X-ray exposure raises concerns about potential health risks such as cancer. The desire for lower radiation doses has driven researchers to improve reconstruction quality. Although previous studies on low-dose computed tomography (LDCT) denoising have demonstrated the effectiveness of learning-based methods, most were developed on the simulated data. However, the real-world scenario differs significantly from the simulation domain, especially when using the multi-slice spiral scanner geometry. This paper proposes a two-stage method for the commercially available multi-slice spiral CT scanners that better exploits the complete reconstruction pipeline for LDCT denoising across different domains. Our approach makes good use of the high redundancy of multi-slice projections and the volumetric reconstructions while leveraging the over-smoothing issue in conventional cascaded frameworks caused by aggressive denoising. The dedicated design also provides a more explicit interpretation of the data flow. Extensive experiments on various datasets showed that the proposed method could remove up to 70% of noise without compromised spatial resolution, while subjective evaluations by two experienced radiologists further supported its superior performance against state-of-the-art methods in clinical practice. Code is available at https://github.com/YCL92/TMD-LDCT.
Collapse
|
11
|
Shi Y, Xia W, Wang G, Mou X. Blind CT Image Quality Assessment Using DDPM-Derived Content and Transformer-Based Evaluator. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:3559-3569. [PMID: 38913529 PMCID: PMC11560125 DOI: 10.1109/tmi.2024.3418652] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/26/2024]
Abstract
Lowering radiation dose per view and utilizing sparse views per scan are two common CT scan modes, albeit often leading to distorted images characterized by noise and streak artifacts. Blind image quality assessment (BIQA) strives to evaluate perceptual quality in alignment with what radiologists perceive, which plays an important role in advancing low-dose CT reconstruction techniques. An intriguing direction involves developing BIQA methods that mimic the operational characteristic of the human visual system (HVS). The internal generative mechanism (IGM) theory reveals that the HVS actively deduces primary content to enhance comprehension. In this study, we introduce an innovative BIQA metric that emulates the active inference process of IGM. Initially, an active inference module, implemented as a denoising diffusion probabilistic model (DDPM), is constructed to anticipate the primary content. Then, the dissimilarity map is derived by assessing the interrelation between the distorted image and its primary content. Subsequently, the distorted image and dissimilarity map are combined into a multi-channel image, which is inputted into a transformer-based image quality evaluator. By leveraging the DDPM-derived primary content, our approach achieves competitive performance on a low-dose CT dataset.
Collapse
|
12
|
Kumschier T, Thalhammer J, Schmid C, Haeusele J, Koehler T, Pfeiffer F, Lasser T, Schaff F. Streak artefact removal in x-ray dark-field computed tomography using a convolutional neural network. Med Phys 2024; 51:7404-7414. [PMID: 39012833 DOI: 10.1002/mp.17305] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/04/2023] [Revised: 06/24/2024] [Accepted: 07/01/2024] [Indexed: 07/18/2024] Open
Abstract
BACKGROUND Computed tomography (CT) relies on the attenuation of x-rays, and is, hence, of limited use for weakly attenuating organs of the body, such as the lung. X-ray dark-field (DF) imaging is a recently developed technology that utilizes x-ray optical gratings to enable small-angle scattering as an alternative contrast mechanism. The DF signal provides structural information about the micromorphology of an object, complementary to the conventional attenuation signal. A first human-scale x-ray DF CT has been developed by our group. Despite specialized processing algorithms, reconstructed images remain affected by streaking artifacts, which often hinder image interpretation. In recent years, convolutional neural networks have gained popularity in the field of CT reconstruction, amongst others for streak artefact removal. PURPOSE Reducing streak artifacts is essential for the optimization of image quality in DF CT, and artefact free images are a prerequisite for potential future clinical application. The purpose of this paper is to demonstrate the feasibility of CNN post-processing for artefact reduction in x-ray DF CT and how multi-rotation scans can serve as a pathway for training data. METHODS We employed a supervised deep-learning approach using a three-dimensional dual-frame UNet in order to remove streak artifacts. Required training data were obtained from the experimental x-ray DF CT prototype at our institute. Two different operating modes were used to generate input and corresponding ground truth data sets. Clinically relevant scans at dose-compatible radiation levels were used as input data, and extended scans with substantially fewer artifacts were used as ground truth data. The latter is neither dose-, nor time-compatible and, therefore, unfeasible for clinical imaging of patients. RESULTS The trained CNN was able to greatly reduce streak artifacts in DF CT images. The network was tested against images with entirely different, previously unseen image characteristics. In all cases, CNN processing substantially increased the image quality, which was quantitatively confirmed by increased image quality metrics. Fine details are preserved during processing, despite the output images appearing smoother than the ground truth images. CONCLUSIONS Our results showcase the potential of a neural network to reduce streak artifacts in x-ray DF CT. The image quality is successfully enhanced in dose-compatible x-ray DF CT, which plays an essential role for the adoption of x-ray DF CT into modern clinical radiology.
Collapse
Affiliation(s)
- Tom Kumschier
- Chair of Biomedical Physics, Department of Physics, School of Natural Sciences, Technical University of Munich, Garching, Germany
- Munich School of BioEngineering, Technical University of Munich, Garching, Germany
| | - Johannes Thalhammer
- Chair of Biomedical Physics, Department of Physics, School of Natural Sciences, Technical University of Munich, Garching, Germany
- Munich School of BioEngineering, Technical University of Munich, Garching, Germany
- Institute for Advanced Study, Technical University of Munich, Garching, Germany
| | - Clemens Schmid
- Chair of Biomedical Physics, Department of Physics, School of Natural Sciences, Technical University of Munich, Garching, Germany
- Munich School of BioEngineering, Technical University of Munich, Garching, Germany
| | - Jakob Haeusele
- Chair of Biomedical Physics, Department of Physics, School of Natural Sciences, Technical University of Munich, Garching, Germany
- Munich School of BioEngineering, Technical University of Munich, Garching, Germany
| | - Thomas Koehler
- Institute for Advanced Study, Technical University of Munich, Garching, Germany
- Philips Research, Hamburg, Germany
| | - Franz Pfeiffer
- Chair of Biomedical Physics, Department of Physics, School of Natural Sciences, Technical University of Munich, Garching, Germany
- Munich School of BioEngineering, Technical University of Munich, Garching, Germany
- Department of Diagnostic and Interventional Radiology, School of Medicine & Klinikum rechts der Isar, Technical University of Munich, Munich, Germany
| | - Tobias Lasser
- Munich School of BioEngineering, Technical University of Munich, Garching, Germany
- Computational Imaging and Inverse Problems, Department of Computer Science, School of Computation, Information and Technology, Technical University of Munich, Garching, Germany
| | - Florian Schaff
- Chair of Biomedical Physics, Department of Physics, School of Natural Sciences, Technical University of Munich, Garching, Germany
- Munich School of BioEngineering, Technical University of Munich, Garching, Germany
| |
Collapse
|
13
|
Xu Y, Sun L, Peng W, Jia S, Morrison K, Perer A, Zandifar A, Visweswaran S, Eslami M, Batmanghelich K. MedSyn: Text-Guided Anatomy-Aware Synthesis of High-Fidelity 3-D CT Images. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:3648-3660. [PMID: 38900619 PMCID: PMC11656526 DOI: 10.1109/tmi.2024.3415032] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/22/2024]
Abstract
This paper introduces an innovative methodology for producing high-quality 3D lung CT images guided by textual information. While diffusion-based generative models are increasingly used in medical imaging, current state-of-the-art approaches are limited to low-resolution outputs and underutilize radiology reports' abundant information. The radiology reports can enhance the generation process by providing additional guidance and offering fine-grained control over the synthesis of images. Nevertheless, expanding text-guided generation to high-resolution 3D images poses significant memory and anatomical detail-preserving challenges. Addressing the memory issue, we introduce a hierarchical scheme that uses a modified UNet architecture. We start by synthesizing low-resolution images conditioned on the text, serving as a foundation for subsequent generators for complete volumetric data. To ensure the anatomical plausibility of the generated samples, we provide further guidance by generating vascular, airway, and lobular segmentation masks in conjunction with the CT images. The model demonstrates the capability to use textual input and segmentation tasks to generate synthesized images. Algorithmic comparative assessments and blind evaluations conducted by 10 board-certified radiologists indicate that our approach exhibits superior performance compared to the most advanced models based on GAN and diffusion techniques, especially in accurately retaining crucial anatomical features such as fissure lines and airways. This innovation introduces novel possibilities. This study focuses on two main objectives: (1) the development of a method for creating images based on textual prompts and anatomical components, and (2) the capability to generate new images conditioning on anatomical elements. The advancements in image generation can be applied to enhance numerous downstream tasks.
Collapse
|
14
|
Bousse A, Kandarpa VSS, Rit S, Perelli A, Li M, Wang G, Zhou J, Wang G. Systematic Review on Learning-based Spectral CT. ARXIV 2024:arXiv:2304.07588v9. [PMID: 37461421 PMCID: PMC10350100] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Subscribe] [Scholar Register] [Indexed: 07/24/2023]
Abstract
Spectral computed tomography (CT) has recently emerged as an advanced version of medical CT and significantly improves conventional (single-energy) CT. Spectral CT has two main forms: dual-energy computed tomography (DECT) and photon-counting computed tomography (PCCT), which offer image improvement, material decomposition, and feature quantification relative to conventional CT. However, the inherent challenges of spectral CT, evidenced by data and image artifacts, remain a bottleneck for clinical applications. To address these problems, machine learning techniques have been widely applied to spectral CT. In this review, we present the state-of-the-art data-driven techniques for spectral CT.
Collapse
Affiliation(s)
| | | | - Simon Rit
- Univ. Lyon, INSA-Lyon, Université Claude Bernard Lyon 1, UJM-Saint Étienne, CNRS, Inserm, CREATIS UMR 5220, U1294, F-69373, Lyon, France
| | - Alessandro Perelli
- School of Science and Engineering, University of Dundee, DD1 4HN Dundee, U.K
| | - Mengzhou Li
- Biomedical Imaging Center, Rensselaer Polytechnic Institute, Troy, NY 12180 USA
| | - Guobao Wang
- Department of Radiology, University of California Davis Health, Sacramento, CA 95817 USA
| | - Jian Zhou
- CTIQ, Canon Medical Research USA, Inc., Vernon Hills, IL 60061 USA
| | - Ge Wang
- Biomedical Imaging Center, Rensselaer Polytechnic Institute, Troy, NY 12180 USA
| |
Collapse
|
15
|
Liu T, Huang S, Li R, Gao P, Li W, Lu H, Song Y, Rong J. Dual and Multi-Target Cone-Beam X-ray Luminescence Computed Tomography Based on the DeepCB-XLCT Network. Bioengineering (Basel) 2024; 11:874. [PMID: 39329616 PMCID: PMC11428951 DOI: 10.3390/bioengineering11090874] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/31/2024] [Revised: 08/24/2024] [Accepted: 08/27/2024] [Indexed: 09/28/2024] Open
Abstract
BACKGROUND AND OBJECTIVE Emerging as a hybrid imaging modality, cone-beam X-ray luminescence computed tomography (CB-XLCT) has been developed using X-ray-excitable nanoparticles. In contrast to conventional bio-optical imaging techniques like bioluminescence tomography (BLT) and fluorescence molecular tomography (FMT), CB-XLCT offers the advantage of greater imaging depth while significantly reducing interference from autofluorescence and background fluorescence, owing to its utilization of X-ray-excited nanoparticles. However, due to the intricate excitation process and extensive light scattering within biological tissues, the inverse problem of CB-XLCT is fundamentally ill-conditioned. METHODS An end-to-end three-dimensional deep encoder-decoder network, termed DeepCB-XLCT, is introduced to improve the quality of CB-XLCT reconstructions. This network directly establishes a nonlinear mapping between the distribution of internal X-ray-excitable nanoparticles and the corresponding boundary fluorescent signals. To improve the fidelity of target shape restoration, the structural similarity loss (SSIM) was incorporated into the objective function of the DeepCB-XLCT network. Additionally, a loss term specifically for target regions was introduced to improve the network's emphasis on the areas of interest. As a result, the inaccuracies in reconstruction caused by the simplified linear model used in conventional methods can be effectively minimized by the proposed DeepCB-XLCT method. RESULTS AND CONCLUSIONS Numerical simulations, phantom experiments, and in vivo experiments with two targets were performed, revealing that the DeepCB-XLCT network enhances reconstruction accuracy regarding contrast-to-noise ratio and shape similarity when compared to traditional methods. In addition, the findings from the XLCT tomographic images involving three targets demonstrate its potential for multi-target CB-XLCT imaging.
Collapse
Affiliation(s)
- Tianshuai Liu
- Biomedical Engineering Department, Fourth Military Medical University, Xi’an 710032, China; (T.L.); (S.H.); (R.L.); (P.G.); (W.L.)
- Shaanxi Provincial Key Laboratory of Bioelectromagnetic Detection and Intelligent Perception, Xi’an 710032, China
| | - Shien Huang
- Biomedical Engineering Department, Fourth Military Medical University, Xi’an 710032, China; (T.L.); (S.H.); (R.L.); (P.G.); (W.L.)
- School of Software Engineering, Xi’an Jiaotong University, Xi’an 710049, China
| | - Ruijing Li
- Biomedical Engineering Department, Fourth Military Medical University, Xi’an 710032, China; (T.L.); (S.H.); (R.L.); (P.G.); (W.L.)
- Shaanxi Provincial Key Laboratory of Bioelectromagnetic Detection and Intelligent Perception, Xi’an 710032, China
| | - Peng Gao
- Biomedical Engineering Department, Fourth Military Medical University, Xi’an 710032, China; (T.L.); (S.H.); (R.L.); (P.G.); (W.L.)
- Shaanxi Provincial Key Laboratory of Bioelectromagnetic Detection and Intelligent Perception, Xi’an 710032, China
| | - Wangyang Li
- Biomedical Engineering Department, Fourth Military Medical University, Xi’an 710032, China; (T.L.); (S.H.); (R.L.); (P.G.); (W.L.)
- Shaanxi Provincial Key Laboratory of Bioelectromagnetic Detection and Intelligent Perception, Xi’an 710032, China
| | - Hongbing Lu
- Biomedical Engineering Department, Fourth Military Medical University, Xi’an 710032, China; (T.L.); (S.H.); (R.L.); (P.G.); (W.L.)
- Shaanxi Provincial Key Laboratory of Bioelectromagnetic Detection and Intelligent Perception, Xi’an 710032, China
| | - Yonghong Song
- School of Software Engineering, Xi’an Jiaotong University, Xi’an 710049, China
| | - Junyan Rong
- Biomedical Engineering Department, Fourth Military Medical University, Xi’an 710032, China; (T.L.); (S.H.); (R.L.); (P.G.); (W.L.)
- Shaanxi Provincial Key Laboratory of Bioelectromagnetic Detection and Intelligent Perception, Xi’an 710032, China
| |
Collapse
|
16
|
Chen Z, Hu B, Niu C, Chen T, Li Y, Shan H, Wang G. IQAGPT: computed tomography image quality assessment with vision-language and ChatGPT models. Vis Comput Ind Biomed Art 2024; 7:20. [PMID: 39101954 DOI: 10.1186/s42492-024-00171-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/22/2024] [Accepted: 07/24/2024] [Indexed: 08/06/2024] Open
Abstract
Large language models (LLMs), such as ChatGPT, have demonstrated impressive capabilities in various tasks and attracted increasing interest as a natural language interface across many domains. Recently, large vision-language models (VLMs) that learn rich vision-language correlation from image-text pairs, like BLIP-2 and GPT-4, have been intensively investigated. However, despite these developments, the application of LLMs and VLMs in image quality assessment (IQA), particularly in medical imaging, remains unexplored. This is valuable for objective performance evaluation and potential supplement or even replacement of radiologists' opinions. To this end, this study introduces IQAGPT, an innovative computed tomography (CT) IQA system that integrates image-quality captioning VLM with ChatGPT to generate quality scores and textual reports. First, a CT-IQA dataset comprising 1,000 CT slices with diverse quality levels is professionally annotated and compiled for training and evaluation. To better leverage the capabilities of LLMs, the annotated quality scores are converted into semantically rich text descriptions using a prompt template. Second, the image-quality captioning VLM is fine-tuned on the CT-IQA dataset to generate quality descriptions. The captioning model fuses image and text features through cross-modal attention. Third, based on the quality descriptions, users verbally request ChatGPT to rate image-quality scores or produce radiological quality reports. Results demonstrate the feasibility of assessing image quality using LLMs. The proposed IQAGPT outperformed GPT-4 and CLIP-IQA, as well as multitask classification and regression models that solely rely on images.
Collapse
Affiliation(s)
- Zhihao Chen
- Institute of Science and Technology for Brain-Inspired Intelligence, Fudan University, Shanghai, 200433, China
| | - Bin Hu
- Department of Radiology, Huashan Hospital, Fudan University, Shanghai, 200040, China
| | - Chuang Niu
- Biomedical Imaging Center, Center for Biotechnology and Interdisciplinary Studies, Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY, 12180, US
| | - Tao Chen
- Institute of Science and Technology for Brain-Inspired Intelligence, Fudan University, Shanghai, 200433, China
| | - Yuxin Li
- Department of Radiology, Huashan Hospital, Fudan University, Shanghai, 200040, China.
| | - Hongming Shan
- Institute of Science and Technology for Brain-Inspired Intelligence, Fudan University, Shanghai, 200433, China.
- MOE Frontiers Center for Brain Science, Fudan University, Shanghai, 200032, China.
- Key Laboratory of Computational Neuroscience and Brain-Inspired Intelligence (Ministry of Education), Fudan University, Shanghai, 200433, China.
| | - Ge Wang
- Biomedical Imaging Center, Center for Biotechnology and Interdisciplinary Studies, Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY, 12180, US.
| |
Collapse
|
17
|
Chi J, Sun Z, Tian S, Wang H, Wang S. A Hybrid Framework of Dual-Domain Signal Restoration and Multi-depth Feature Reinforcement for Low-Dose Lung CT Denoising. JOURNAL OF IMAGING INFORMATICS IN MEDICINE 2024; 37:1944-1959. [PMID: 38424278 PMCID: PMC11300419 DOI: 10.1007/s10278-023-00934-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/06/2023] [Revised: 09/05/2023] [Accepted: 09/06/2023] [Indexed: 03/02/2024]
Abstract
Low-dose computer tomography (LDCT) has been widely used in medical diagnosis. Various denoising methods have been presented to remove noise in LDCT scans. However, existing methods cannot achieve satisfactory results due to the difficulties in (1) distinguishing the characteristics of structures, textures, and noise confused in the image domain, and (2) representing local details and global semantics in the hierarchical features. In this paper, we propose a novel denoising method consisting of (1) a 2D dual-domain restoration framework to reconstruct noise-free structure and texture signals separately, and (2) a 3D multi-depth reinforcement U-Net model to further recover image details with enhanced hierarchical features. In the 2D dual-domain restoration framework, the convolutional neural networks are adopted in both the image domain where the image structures are well preserved through the spatial continuity, and the sinogram domain where the textures and noise are separately represented by different wavelet coefficients and processed adaptively. In the 3D multi-depth reinforcement U-Net model, the hierarchical features from the 3D U-Net are enhanced by the cross-resolution attention module (CRAM) and dual-branch graph convolution module (DBGCM). The CRAM preserves local details by integrating adjacent low-level features with different resolutions, while the DBGCM enhances global semantics by building graphs for high-level features in intra-feature and inter-feature dimensions. Experimental results on the LUNA16 dataset and 2016 NIH-AAPM-Mayo Clinic LDCT Grand Challenge dataset illustrate the proposed method outperforms the state-of-the-art methods on removing noise from LDCT images with clear structures and textures, proving its potential in clinical practice.
Collapse
Affiliation(s)
- Jianning Chi
- Faculty of Robot Science and Engineering, Northeastern University, Zhihui Street, Shenyang, 110169, Liaoning, China.
- Key Laboratory of Intelligent Computing in Medical Image of Ministry of Education, Northeastern University, Zhihui Street, Shenyang, 110169, Liaoning, China.
| | - Zhiyi Sun
- Faculty of Robot Science and Engineering, Northeastern University, Zhihui Street, Shenyang, 110169, Liaoning, China
| | - Shuyu Tian
- Graduate School, Dalian Medical University, Lyushunnan, Dalian, 116000, Liaoning, China
| | - Huan Wang
- Faculty of Robot Science and Engineering, Northeastern University, Zhihui Street, Shenyang, 110169, Liaoning, China
| | - Siqi Wang
- Faculty of Robot Science and Engineering, Northeastern University, Zhihui Street, Shenyang, 110169, Liaoning, China
| |
Collapse
|
18
|
Zhou Z, Gong H, Hsieh S, McCollough CH, Yu L. Image quality evaluation in deep-learning-based CT noise reduction using virtual imaging trial methods: Contrast-dependent spatial resolution. Med Phys 2024; 51:5399-5413. [PMID: 38555876 PMCID: PMC11321944 DOI: 10.1002/mp.17029] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2023] [Revised: 02/19/2024] [Accepted: 02/26/2024] [Indexed: 04/02/2024] Open
Abstract
BACKGROUND Deep-learning-based image reconstruction and noise reduction methods (DLIR) have been increasingly deployed in clinical CT. Accurate image quality assessment of these methods is challenging as the performance measured using physical phantoms may not represent the true performance of DLIR in patients since DLIR is trained mostly on patient images. PURPOSE In this work, we aim to develop a patient-data-based virtual imaging trial framework and, as a first application, use it to measure the spatial resolution properties of a DLIR method. METHODS The patient-data-based virtual imaging trial framework consists of five steps: (1) insertion of lesions into projection domain data using the acquisition geometry of the patient exam to simulate different lesion characteristics; (2) insertion of noise into projection domain data using a realistic photon statistical model of the CT system to simulate different dose levels; (3) creation of DLIR-processed images from projection or image data; (4) creation of ensembles of DLIR-processed patient images from a large number of noise and lesion realizations; and (5) evaluation of image quality using ensemble DLIR images. This framework was applied to measure the spatial resolution of a ResNet based deep convolutional neural network (DCNN) trained on patient images. Lesions in a cylindrical shape and different contrast levels (-500, -100, -50, -20, -10 HU) were inserted to the lower right lobe of the liver in a patient case. Multiple dose levels were simulated (50%, 25%, 12.5%). Each lesion and dose condition had 600 noise realizations. Multiple reconstruction and denoising methods were used on all the noise realizations, including the original filtered-backprojection (FBP), iterative reconstruction (IR), and the DCNN method with three different strength setting (DCNN-weak, DCNN-medium, and DCNN-strong). Mean lesion signal was calculated by performing ensemble averaging of all the noise realizations for each lesion and dose condition and then subtracting the lesion-present images from the lesion absent images. Modulation transfer functions (MTFs) both in-plane and along the z-axis were calculated based on the mean lesion signals. The standard deviations of MTFs at each condition were estimated with bootstrapping: randomly sampling (with replacement) all the DLIR/FBP/IR images from the ensemble data (600 samples) at each condition. The impact of varying lesion contrast, dose levels, and denoising strengths were evaluated. Statistical analysis with paired t-test was used to compare the z-axis and in-plane spatial resolution of five algorithms for five different contrasts and three dose levels. RESULTS The in-plane and z-axis spatial resolution degradation of DCNN becomes more severe as the contrast or radiation dose decreased, or DCNN denoising strength increased. In comparison with FBP, a 59.5% and 4.1% reduction of in-plane and z-axis MTF (in terms of spatial frequencies at 50% MTF), respectively, was observed at low contrast (-10 HU) for DCNN with the highest denoising strength at 25% routine dose level. When the dose level reduces from 50% to 12.5% of routine dose, the in-plane and z-axis MTFs reduces from 92.1% to 76.3%, and from 98.9% to 95.5%, respectively, at contrast of -100 HU, using FBP as the reference. For most conditions of contrasts and dose levels, significant differences were found among the five algorithms, with the following relationship in both in-plane and cross-plane spatial resolution: FBP > DCNN-Weak > IR > DCNN-Medium > DCNN-Strong. The spatial resolution difference among algorithms decreases at higher contrast or dose levels. CONCLUSIONS A patient-data-based virtual imaging trial framework was developed and applied to measuring the spatial resolution properties of a DCNN noise reduction method at different contrast and dose levels using real patient data. As with other non-linear image reconstruction and post-processing techniques, the evaluated DCNN method degraded the in-plane and z-axis spatial resolution at lower contrast levels, lower radiation dose, and higher denoising strength.
Collapse
Affiliation(s)
| | - Hao Gong
- Department of Radiology, Mayo Clinic, Rochester, MN, US
| | - Scott Hsieh
- Department of Radiology, Mayo Clinic, Rochester, MN, US
| | | | - Lifeng Yu
- Department of Radiology, Mayo Clinic, Rochester, MN, US
| |
Collapse
|
19
|
Wiedeman C, Lorraine P, Wang G, Do R, Simpson A, Peoples J, De Man B. Simulated deep CT characterization of liver metastases with high-resolution filtered back projection reconstruction. Vis Comput Ind Biomed Art 2024; 7:13. [PMID: 38861067 PMCID: PMC11166620 DOI: 10.1186/s42492-024-00161-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/02/2023] [Accepted: 04/14/2024] [Indexed: 06/12/2024] Open
Abstract
Early diagnosis and accurate prognosis of colorectal cancer is critical for determining optimal treatment plans and maximizing patient outcomes, especially as the disease progresses into liver metastases. Computed tomography (CT) is a frontline tool for this task; however, the preservation of predictive radiomic features is highly dependent on the scanning protocol and reconstruction algorithm. We hypothesized that image reconstruction with a high-frequency kernel could result in a better characterization of liver metastases features via deep neural networks. This kernel produces images that appear noisier but preserve more sinogram information. A simulation pipeline was developed to study the effects of imaging parameters on the ability to characterize the features of liver metastases. This pipeline utilizes a fractal approach to generate a diverse population of shapes representing virtual metastases, and then it superimposes them on a realistic CT liver region to perform a virtual CT scan using CatSim. Datasets of 10,000 liver metastases were generated, scanned, and reconstructed using either standard or high-frequency kernels. These data were used to train and validate deep neural networks to recover crafted metastases characteristics, such as internal heterogeneity, edge sharpness, and edge fractal dimension. In the absence of noise, models scored, on average, 12.2% ( α = 0.012 ) and 7.5% ( α = 0.049 ) lower squared error for characterizing edge sharpness and fractal dimension, respectively, when using high-frequency reconstructions compared to standard. However, the differences in performance were statistically insignificant when a typical level of CT noise was simulated in the clinical scan. Our results suggest that high-frequency reconstruction kernels can better preserve information for downstream artificial intelligence-based radiomic characterization, provided that noise is limited. Future work should investigate the information-preserving kernels in datasets with clinical labels.
Collapse
Affiliation(s)
- Christopher Wiedeman
- Department of Electrical and Computer Engineering, Rensselaer Polytechnic Institute, Troy, NY, 12180, USA
| | | | - Ge Wang
- Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY, 12180, USA
| | - Richard Do
- Department of Radiology, Memorial Sloan Kettering Cancer Center, New York, NY, 10065, USA
| | - Amber Simpson
- Biomedical Computing and Informatics, Queen's University, Kingston, ON, K7L 3N6, Canada
| | - Jacob Peoples
- Biomedical Computing and Informatics, Queen's University, Kingston, ON, K7L 3N6, Canada
| | - Bruno De Man
- GE Research - Healthcare, Niskayuna, NY, 12309, USA.
| |
Collapse
|
20
|
Meng M, Wang Y, Zhu M, Tao X, Mao Z, Liao J, Bian Z, Zeng D, Ma J. DDT-Net: Dose-Agnostic Dual-Task Transfer Network for Simultaneous Low-Dose CT Denoising and Simulation. IEEE J Biomed Health Inform 2024; 28:3613-3625. [PMID: 38478459 DOI: 10.1109/jbhi.2024.3376628] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/20/2024]
Abstract
Deep learning (DL) algorithms have achieved unprecedented success in low-dose CT (LDCT) imaging and are expected to be a new generation of CT reconstruction technology. However, most DL-based denoising models often lack the ability to generalize to unseen dose data. Moreover, most simulation tools for LDCT typically operate on proprietary projection data, which is generally not accessible without an established collaboration with CT manufacturers. To alleviate these issues, in this work, we propose a dose-agnostic dual-task transfer network, termed DDT-Net, for simultaneous LDCT denoising and simulation. Concretely, the dual-task learning module is constructed to integrate the LDCT denoising and simulation tasks into a unified optimization framework by learning the joint distribution of LDCT and NDCT data. We approximate the joint distribution of continuous dose level data by training DDT-Net with discrete dose data, which can be generalized to denoising and simulation of unseen dose data. In particular, the mixed-dose training strategy adopted by DDT-Net can promote the denoising performance of lower-dose data. The paired dataset simulated by DDT-Net can be used for data augmentation to further restore the tissue texture of LDCT images. Experimental results on synthetic data and clinical data show that the proposed DDT-Net outperforms competing methods in terms of denoising and generalization performance at unseen dose data, and it also provides a simulation tool that can quickly simulate realistic LDCT images at arbitrary dose levels.
Collapse
|
21
|
Sun P, Yang J, Tian X, Yuan G. Image fusion-based low-dose CBCT enhancement method for visualizing miniscrew insertion in the infrazygomatic crest. BMC Med Imaging 2024; 24:114. [PMID: 38760689 PMCID: PMC11100247 DOI: 10.1186/s12880-024-01289-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2024] [Accepted: 05/03/2024] [Indexed: 05/19/2024] Open
Abstract
Digital dental technology covers oral cone-beam computed tomography (CBCT) image processing and low-dose CBCT dental applications. A low-dose CBCT image enhancement method based on image fusion is proposed to address the need for subzygomatic small screw insertion. Specifically, firstly, a sharpening correction module is proposed, where the CBCT image is sharpened to compensate for the loss of details in the underexposed/over-exposed region. Secondly, a visibility restoration module based on type II fuzzy sets is designed, and a contrast enhancement module using curve transformation is designed. In addition to this, we propose a perceptual fusion module that fuses visibility and contrast of oral CBCT images. As a result, the problems of overexposure/underexposure, low visibility, and low contrast that occur in oral CBCT images can be effectively addressed with consistent interpretability. The proposed algorithm was analyzed in comparison experiments with a variety of algorithms, as well as ablation experiments. After analysis, compared with advanced enhancement algorithms, this algorithm achieved excellent results in low-dose CBCT enhancement and effective observation of subzygomatic small screw implantation. Compared with the best performing method, the evaluation metric is 0.07-2 higher on both datasets. The project can be found at: https://github.com/sunpeipei2024/low-dose-CBCT .
Collapse
Affiliation(s)
- Peipei Sun
- State Key Laboratory of Oral & Maxillofacial Reconstruction and Regeneration, Key Laboratory of Oral Biomedicine Ministry of Education, Hubei Key Laboratory of Stomatology, School & Hospital of Stomatology, Wuhan University, Wuhan, China
- Department of Pediatric Dentistry, School and Hospital of Stomatology, Wuhan University, Wuhan, China
| | - Jinghui Yang
- State Key Laboratory of Oral & Maxillofacial Reconstruction and Regeneration, Key Laboratory of Oral Biomedicine Ministry of Education, Hubei Key Laboratory of Stomatology, School & Hospital of Stomatology, Wuhan University, Wuhan, China
- Department of Pediatric Dentistry, School and Hospital of Stomatology, Wuhan University, Wuhan, China
| | - Xue Tian
- State Key Laboratory of Oral & Maxillofacial Reconstruction and Regeneration, Key Laboratory of Oral Biomedicine Ministry of Education, Hubei Key Laboratory of Stomatology, School & Hospital of Stomatology, Wuhan University, Wuhan, China
- Department of Pediatric Dentistry, School and Hospital of Stomatology, Wuhan University, Wuhan, China
| | - Guohua Yuan
- State Key Laboratory of Oral & Maxillofacial Reconstruction and Regeneration, Key Laboratory of Oral Biomedicine Ministry of Education, Hubei Key Laboratory of Stomatology, School & Hospital of Stomatology, Wuhan University, Wuhan, China.
- Department of Pediatric Dentistry, School and Hospital of Stomatology, Wuhan University, Wuhan, China.
- Frontier Science Center for Immunology and Metabolism, Wuhan University, Wuhan, China.
| |
Collapse
|
22
|
Kang J, Liu Y, Zhang P, Guo N, Wang L, Du Y, Gui Z. FSformer: A combined frequency separation network and transformer for LDCT denoising. Comput Biol Med 2024; 173:108378. [PMID: 38554660 DOI: 10.1016/j.compbiomed.2024.108378] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/04/2023] [Revised: 03/01/2024] [Accepted: 03/24/2024] [Indexed: 04/02/2024]
Abstract
Low-dose computed tomography (LDCT) has been widely concerned in the field of medical imaging because of its low radiation hazard to humans. However, under low-dose radiation scenarios, a large amount of noise/artifacts are present in the reconstructed image, which reduces the clarity of the image and is not conducive to diagnosis. To improve the LDCT image quality, we proposed a combined frequency separation network and Transformer (FSformer) for LDCT denoising. Firstly, FSformer decomposes the LDCT images into low-frequency images and multi-layer high-frequency images by frequency separation blocks. Then, the low-frequency components are fused with the high-frequency components of different layers to remove the noise in the high-frequency components with the help of the potential texture of low-frequency parts. Next, the estimated noise images can be obtained by using Transformer stage in the frequency aggregation denoising block. Finally, they are fed into the reconstruction prediction block to obtain improved quality images. In addition, a compound loss function with frequency loss and Charbonnier loss is used to guide the training of the network. The performance of FSformer has been validated and evaluated on AAPM Mayo dataset, real Piglet dataset and clinical dataset. Compared with previous representative models in different architectures, FSformer achieves the optimal metrics with PSNR of 33.7714 dB and SSIM of 0.9254 on Mayo dataset, the testing time is 1.825 s. The experimental results show that FSformer is a state-of-the-art (SOTA) model with noise/artifact suppression and texture/organization preservation. Moreover, the model has certain robustness and can effectively improve LDCT image quality.
Collapse
Affiliation(s)
- Jiaqi Kang
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, 030051, China; School of Information and Communication Engineering, North University of China, Taiyuan, 030051, China
| | - Yi Liu
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, 030051, China; School of Information and Communication Engineering, North University of China, Taiyuan, 030051, China
| | - Pengcheng Zhang
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, 030051, China; School of Information and Communication Engineering, North University of China, Taiyuan, 030051, China
| | - Niu Guo
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, 030051, China; School of Information and Communication Engineering, North University of China, Taiyuan, 030051, China
| | - Lei Wang
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, 030051, China; School of Information and Communication Engineering, North University of China, Taiyuan, 030051, China
| | - Yinglin Du
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, 030051, China; School of Information and Communication Engineering, North University of China, Taiyuan, 030051, China
| | - Zhiguo Gui
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, 030051, China; School of Information and Communication Engineering, North University of China, Taiyuan, 030051, China.
| |
Collapse
|
23
|
Chen Z, Niu C, Gao Q, Wang G, Shan H. LIT-Former: Linking In-Plane and Through-Plane Transformers for Simultaneous CT Image Denoising and Deblurring. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:1880-1894. [PMID: 38194396 DOI: 10.1109/tmi.2024.3351723] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/11/2024]
Abstract
This paper studies 3D low-dose computed tomography (CT) imaging. Although various deep learning methods were developed in this context, typically they focus on 2D images and perform denoising due to low-dose and deblurring for super-resolution separately. Up to date, little work was done for simultaneous in-plane denoising and through-plane deblurring, which is important to obtain high-quality 3D CT images with lower radiation and faster imaging speed. For this task, a straightforward method is to directly train an end-to-end 3D network. However, it demands much more training data and expensive computational costs. Here, we propose to link in-plane and through-plane transformers for simultaneous in-plane denoising and through-plane deblurring, termed as LIT-Former, which can efficiently synergize in-plane and through-plane sub-tasks for 3D CT imaging and enjoy the advantages of both convolution and transformer networks. LIT-Former has two novel designs: efficient multi-head self-attention modules (eMSM) and efficient convolutional feed-forward networks (eCFN). First, eMSM integrates in-plane 2D self-attention and through-plane 1D self-attention to efficiently capture global interactions of 3D self-attention, the core unit of transformer networks. Second, eCFN integrates 2D convolution and 1D convolution to extract local information of 3D convolution in the same fashion. As a result, the proposed LIT-Former synergizes these two sub-tasks, significantly reducing the computational complexity as compared to 3D counterparts and enabling rapid convergence. Extensive experimental results on simulated and clinical datasets demonstrate superior performance over state-of-the-art models. The source code is made available at https://github.com/hao1635/LIT-Former.
Collapse
|
24
|
Li X, Jing K, Yang Y, Wang Y, Ma J, Zheng H, Xu Z. Noise-Generating and Imaging Mechanism Inspired Implicit Regularization Learning Network for Low Dose CT Reconstrution. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:1677-1689. [PMID: 38145543 DOI: 10.1109/tmi.2023.3347258] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/27/2023]
Abstract
Low-dose computed tomography (LDCT) helps to reduce radiation risks in CT scanning while maintaining image quality, which involves a consistent pursuit of lower incident rays and higher reconstruction performance. Although deep learning approaches have achieved encouraging success in LDCT reconstruction, most of them treat the task as a general inverse problem in either the image domain or the dual (sinogram and image) domains. Such frameworks have not considered the original noise generation of the projection data and suffer from limited performance improvement for the LDCT task. In this paper, we propose a novel reconstruction model based on noise-generating and imaging mechanism in full-domain, which fully considers the statistical properties of intrinsic noises in LDCT and prior information in sinogram and image domains. To solve the model, we propose an optimization algorithm based on the proximal gradient technique. Specifically, we derive the approximate solutions of the integer programming problem on the projection data theoretically. Instead of hand-crafting the sinogram and image regularizers, we propose to unroll the optimization algorithm to be a deep network. The network implicitly learns the proximal operators of sinogram and image regularizers with two deep neural networks, providing a more interpretable and effective reconstruction procedure. Numerical results demonstrate our proposed method improvements of > 2.9 dB in peak signal to noise ratio, > 1.4% promotion in structural similarity metric, and > 9 HU decrements in root mean square error over current state-of-the-art LDCT methods.
Collapse
|
25
|
Xie H, Liu Q, Zhou B, Chen X, Guo X, Wang H, Li B, Rominger A, Shi K, Liu C. Unified Noise-aware Network for Low-count PET Denoising with Varying Count Levels. IEEE TRANSACTIONS ON RADIATION AND PLASMA MEDICAL SCIENCES 2024; 8:366-378. [PMID: 39391291 PMCID: PMC11463975 DOI: 10.1109/trpms.2023.3334105] [Citation(s) in RCA: 5] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/12/2024]
Abstract
As PET imaging is accompanied by substantial radiation exposure and cancer risk, reducing radiation dose in PET scans is an important topic. However, low-count PET scans often suffer from high image noise, which can negatively impact image quality and diagnostic performance. Recent advances in deep learning have shown great potential for recovering underlying signal from noisy counterparts. However, neural networks trained on a specific noise level cannot be easily generalized to other noise levels due to different noise amplitude and variances. To obtain optimal denoised results, we may need to train multiple networks using data with different noise levels. But this approach may be infeasible in reality due to limited data availability. Denoising dynamic PET images presents additional challenge due to tracer decay and continuously changing noise levels across dynamic frames. To address these issues, we propose a Unified Noise-aware Network (UNN) that combines multiple sub-networks with varying denoising power to generate optimal denoised results regardless of the input noise levels. Evaluated using large-scale data from two medical centers with different vendors, presented results showed that the UNN can consistently produce promising denoised results regardless of input noise levels, and demonstrate superior performance over networks trained on single noise level data, especially for extremely low-count data.
Collapse
Affiliation(s)
- Huidong Xie
- Department of Biomedical Engineering, Yale University
| | - Qiong Liu
- Department of Biomedical Engineering, Yale University
| | - Bo Zhou
- Department of Biomedical Engineering, Yale University
| | | | - Xueqi Guo
- Department of Biomedical Engineering, Yale University
| | - Hanzhong Wang
- Department of Nuclear Medicine, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine
| | - Biao Li
- Department of Nuclear Medicine, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine
| | - Axel Rominger
- Department of Nuclear Medicine, Inselspital, Bern University Hospital, University of Bern
| | - Kuangyu Shi
- Department of Nuclear Medicine, Inselspital, Bern University Hospital, University of Bern
- Computer Aided Medical Procedures and Augmented Reality, Institute of Informatics 116, Technical University of Munich, Munich, Germany
| | - Chi Liu
- Department of Biomedical Engineering, Yale University
- Department of Radiology and Biomedical Imaging at Yale University
| |
Collapse
|
26
|
Sherwani MK, Gopalakrishnan S. A systematic literature review: deep learning techniques for synthetic medical image generation and their applications in radiotherapy. FRONTIERS IN RADIOLOGY 2024; 4:1385742. [PMID: 38601888 PMCID: PMC11004271 DOI: 10.3389/fradi.2024.1385742] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 02/13/2024] [Accepted: 03/11/2024] [Indexed: 04/12/2024]
Abstract
The aim of this systematic review is to determine whether Deep Learning (DL) algorithms can provide a clinically feasible alternative to classic algorithms for synthetic Computer Tomography (sCT). The following categories are presented in this study: ∙ MR-based treatment planning and synthetic CT generation techniques. ∙ Generation of synthetic CT images based on Cone Beam CT images. ∙ Low-dose CT to High-dose CT generation. ∙ Attenuation correction for PET images. To perform appropriate database searches, we reviewed journal articles published between January 2018 and June 2023. Current methodology, study strategies, and results with relevant clinical applications were analyzed as we outlined the state-of-the-art of deep learning based approaches to inter-modality and intra-modality image synthesis. This was accomplished by contrasting the provided methodologies with traditional research approaches. The key contributions of each category were highlighted, specific challenges were identified, and accomplishments were summarized. As a final step, the statistics of all the cited works from various aspects were analyzed, which revealed that DL-based sCTs have achieved considerable popularity, while also showing the potential of this technology. In order to assess the clinical readiness of the presented methods, we examined the current status of DL-based sCT generation.
Collapse
Affiliation(s)
- Moiz Khan Sherwani
- Section for Evolutionary Hologenomics, Globe Institute, University of Copenhagen, Copenhagen, Denmark
| | | |
Collapse
|
27
|
Luo SH, Pan SQ, Chen GY, Xie Y, Ren B, Liu GK, Tian ZQ. Revealing the Denoising Principle of Zero-Shot N2N-Based Algorithm from 1D Spectrum to 2D Image. Anal Chem 2024; 96:4086-4092. [PMID: 38412039 DOI: 10.1021/acs.analchem.3c04608] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/29/2024]
Abstract
Denoising is a necessary step in image analysis to extract weak signals, especially those hardly identified by the naked eye. Unlike the data-driven deep-learning denoising algorithms relying on a clean image as the reference, Noise2Noise (N2N) was able to denoise the noise image, providing sufficiently noise images with the same subject but randomly distributed noise. Further, by introducing data augmentation to create a big data set and regularization to prevent model overfitting, zero-shot N2N-based denoising was proposed in which only a single noisy image was needed. Although various N2N-based denoising algorithms have been developed with high performance, their complicated black box operation prevented the lightweight. Therefore, to reveal the working function of the zero-shot N2N-based algorithm, we proposed a lightweight Peak2Peak algorithm (P2P) and qualitatively and quantitatively analyzed its denoising behavior on the 1D spectrum and 2D image. We found that the high-performance denoising originates from the trade-off balance between the loss function and regularization in the denoising module, where regularization is the switch of denoising. Meanwhile, the signal extraction is mainly from the self-supervised characteristic learning in the data augmentation module. Further, the lightweight P2P improved the denoising speed by at least ten times but with little performance loss, compared with that of the current N2N-based algorithms. In general, the visualization of P2P provides a reference for revealing the working function of zero-shot N2N-based algorithms, which would pave the way for the application of these algorithms toward real-time (in situ, in vivo, and operando) research improving both temporal and spatial resolutions. The P2P is open-source at https://github.com/3331822w/Peak2Peakand will be accessible online access at https://ramancloud.xmu.edu.cn/tutorial.
Collapse
Affiliation(s)
- Si-Heng Luo
- State Key Laboratory for Physical Chemistry of Solid Surfaces, College of Chemistry and Chemical Engineering, Xiamen University, Xiamen 361005, China
- State Key Laboratory of Marine Environmental Science, Fujian Provincial Key Laboratory for Coastal Ecology and Environmental Studies, Center for Marine Environmental Chemistry & Toxicology, College of the Environment and Ecology, Xiamen University, Xiamen 361102, China
| | - Si-Qi Pan
- State Key Laboratory of Marine Environmental Science, Fujian Provincial Key Laboratory for Coastal Ecology and Environmental Studies, Center for Marine Environmental Chemistry & Toxicology, College of the Environment and Ecology, Xiamen University, Xiamen 361102, China
| | - Gan-Yu Chen
- State Key Laboratory for Physical Chemistry of Solid Surfaces, College of Chemistry and Chemical Engineering, Xiamen University, Xiamen 361005, China
| | - Yi Xie
- Fujian Key Laboratory of Sensing and Computing for Smart City, School of Information Science and Engineering, Xiamen University, Xiamen, Fujian 361005, China
- Shenzhen Research Institute of Xiamen University, Xiamen University, Shenzhen 518000, China
| | - Bin Ren
- State Key Laboratory for Physical Chemistry of Solid Surfaces, College of Chemistry and Chemical Engineering, Xiamen University, Xiamen 361005, China
- Innovation Laboratory for Sciences and Technologies of Energy Materials of Fujian Province (IKKEM), Xiamen 361005, China
| | - Guo-Kun Liu
- State Key Laboratory of Marine Environmental Science, Fujian Provincial Key Laboratory for Coastal Ecology and Environmental Studies, Center for Marine Environmental Chemistry & Toxicology, College of the Environment and Ecology, Xiamen University, Xiamen 361102, China
| | - Zhong-Qun Tian
- State Key Laboratory for Physical Chemistry of Solid Surfaces, College of Chemistry and Chemical Engineering, Xiamen University, Xiamen 361005, China
- Innovation Laboratory for Sciences and Technologies of Energy Materials of Fujian Province (IKKEM), Xiamen 361005, China
| |
Collapse
|
28
|
Zhang K, Lin PC, Pan J, Shao R, Xu PX, Cao R, Wu CG, Crookes D, Hua L, Wang L. DeepmdQCT: A multitask network with domain invariant features and comprehensive attention mechanism for quantitative computer tomography diagnosis of osteoporosis. Comput Biol Med 2024; 170:107916. [PMID: 38237237 DOI: 10.1016/j.compbiomed.2023.107916] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/30/2023] [Revised: 12/18/2023] [Accepted: 12/29/2023] [Indexed: 02/28/2024]
Abstract
In the medical field, the application of machine learning technology in the automatic diagnosis and monitoring of osteoporosis often faces challenges related to domain adaptation in drug therapy research. The existing neural networks used for the diagnosis of osteoporosis may experience a decrease in model performance when applied to new data domains due to changes in radiation dose and equipment. To address this issue, in this study, we propose a new method for multi domain diagnostic and quantitative computed tomography (QCT) images, called DeepmdQCT. This method adopts a domain invariant feature strategy and integrates a comprehensive attention mechanism to guide the fusion of global and local features, effectively improving the diagnostic performance of multi domain CT images. We conducted experimental evaluations on a self-created OQCT dataset, and the results showed that for dose domain images, the average accuracy reached 91%, while for device domain images, the accuracy reached 90.5%. our method successfully estimated bone density values, with a fit of 0.95 to the gold standard. Our method not only achieved high accuracy in CT images in the dose and equipment fields, but also successfully estimated key bone density values, which is crucial for evaluating the effectiveness of osteoporosis drug treatment. In addition, we validated the effectiveness of our architecture in feature extraction using three publicly available datasets. We also encourage the application of the DeepmdQCT method to a wider range of medical image analysis fields to improve the performance of multi-domain images.
Collapse
Affiliation(s)
- Kun Zhang
- School of Electrical Engineering, Nantong University, Nantong, Jiangsu, 226001, China; Nantong Key Laboratory of Intelligent Control and Intelligent Computing, Nantong, Jiangsu, 226001, China; Nantong Key Laboratory of Intelligent Medicine Innovation and Transformation, Nantong, Jiangsu, 226001, China
| | - Peng-Cheng Lin
- School of Electrical Engineering, Nantong University, Nantong, Jiangsu, 226001, China
| | - Jing Pan
- Department of Radiology, Affiliated Hospital 2 of Nantong University, Nantong, Jiangsu, 226001, China
| | - Rui Shao
- School of Electrical Engineering, Nantong University, Nantong, Jiangsu, 226001, China
| | - Pei-Xia Xu
- School of Electrical Engineering, Nantong University, Nantong, Jiangsu, 226001, China
| | - Rui Cao
- Department of Radiology, Affiliated Hospital 2 of Nantong University, Nantong, Jiangsu, 226001, China
| | - Cheng-Gang Wu
- School of Electrical Engineering, Nantong University, Nantong, Jiangsu, 226001, China
| | - Danny Crookes
- School of Electronics, Electrical Engineering and Computer Science, Queen's University Belfast, Belfast, BT7 1NN, UK
| | - Liang Hua
- School of Electrical Engineering, Nantong University, Nantong, Jiangsu, 226001, China.
| | - Lin Wang
- Department of Radiology, Affiliated Hospital 2 of Nantong University, Nantong, Jiangsu, 226001, China.
| |
Collapse
|
29
|
Zhang J, Gong W, Ye L, Wang F, Shangguan Z, Cheng Y. A Review of deep learning methods for denoising of medical low-dose CT images. Comput Biol Med 2024; 171:108112. [PMID: 38387380 DOI: 10.1016/j.compbiomed.2024.108112] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/19/2023] [Revised: 01/18/2024] [Accepted: 02/04/2024] [Indexed: 02/24/2024]
Abstract
To prevent patients from being exposed to excess of radiation in CT imaging, the most common solution is to decrease the radiation dose by reducing the X-ray, and thus the quality of the resulting low-dose CT images (LDCT) is degraded, as evidenced by more noise and streaking artifacts. Therefore, it is important to maintain high quality CT image while effectively reducing radiation dose. In recent years, with the rapid development of deep learning technology, deep learning-based LDCT denoising methods have become quite popular because of their data-driven and high-performance features to achieve excellent denoising results. However, to our knowledge, no relevant article has so far comprehensively introduced and reviewed advanced deep learning denoising methods such as Transformer structures in LDCT denoising tasks. Therefore, based on the literatures related to LDCT image denoising published from year 2016-2023, and in particular from 2020 to 2023, this study presents a systematic survey of current situation, and challenges and future research directions in LDCT image denoising field. Four types of denoising networks are classified according to the network structure: CNN-based, Encoder-Decoder-based, GAN-based, and Transformer-based denoising networks, and each type of denoising network is described and summarized from the perspectives of structural features and denoising performances. Representative deep-learning denoising methods for LDCT are experimentally compared and analyzed. The study results show that CNN-based denoising methods capture image details efficiently through multi-level convolution operation, demonstrating superior denoising effects and adaptivity. Encoder-decoder networks with MSE loss, achieve outstanding results in objective metrics. GANs based methods, employing innovative generators and discriminators, obtain denoised images that exhibit perceptually a closeness to NDCT. Transformer-based methods have potential for improving denoising performances due to their powerful capability in capturing global information. Challenges and opportunities for deep learning based LDCT denoising are analyzed, and future directions are also presented.
Collapse
Affiliation(s)
- Ju Zhang
- College of Information Science and Technology, Hangzhou Normal University, Hangzhou, China.
| | - Weiwei Gong
- College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou, China.
| | - Lieli Ye
- College of Information Science and Technology, Hangzhou Normal University, Hangzhou, China.
| | - Fanghong Wang
- Zhijiang College, Zhejiang University of Technology, Shaoxing, China.
| | - Zhibo Shangguan
- College of Computer Science and Technology, Zhejiang University of Technology, Hangzhou, China.
| | - Yun Cheng
- Department of Medical Imaging, Zhejiang Hospital, Hangzhou, China.
| |
Collapse
|
30
|
Gao Q, Li Z, Zhang J, Zhang Y, Shan H. CoreDiff: Contextual Error-Modulated Generalized Diffusion Model for Low-Dose CT Denoising and Generalization. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:745-759. [PMID: 37773896 DOI: 10.1109/tmi.2023.3320812] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 10/01/2023]
Abstract
Low-dose computed tomography (CT) images suffer from noise and artifacts due to photon starvation and electronic noise. Recently, some works have attempted to use diffusion models to address the over-smoothness and training instability encountered by previous deep-learning-based denoising models. However, diffusion models suffer from long inference time due to a large number of sampling steps involved. Very recently, cold diffusion model generalizes classical diffusion models and has greater flexibility. Inspired by cold diffusion, this paper presents a novel COntextual eRror-modulated gEneralized Diffusion model for low-dose CT (LDCT) denoising, termed CoreDiff. First, CoreDiff utilizes LDCT images to displace the random Gaussian noise and employs a novel mean-preserving degradation operator to mimic the physical process of CT degradation, significantly reducing sampling steps thanks to the informative LDCT images as the starting point of the sampling process. Second, to alleviate the error accumulation problem caused by the imperfect restoration operator in the sampling process, we propose a novel ContextuaL Error-modulAted Restoration Network (CLEAR-Net), which can leverage contextual information to constrain the sampling process from structural distortion and modulate time step embedding features for better alignment with the input at the next time step. Third, to rapidly generalize the trained model to a new, unseen dose level with as few resources as possible, we devise a one-shot learning framework to make CoreDiff generalize faster and better using only one single LDCT image (un)paired with normal-dose CT (NDCT). Extensive experimental results on four datasets demonstrate that our CoreDiff outperforms competing methods in denoising and generalization performance, with clinically acceptable inference time. Source code is made available at https://github.com/qgao21/CoreDiff.
Collapse
|
31
|
Bousse A, Kandarpa VSS, Rit S, Perelli A, Li M, Wang G, Zhou J, Wang G. Systematic Review on Learning-based Spectral CT. IEEE TRANSACTIONS ON RADIATION AND PLASMA MEDICAL SCIENCES 2024; 8:113-137. [PMID: 38476981 PMCID: PMC10927029 DOI: 10.1109/trpms.2023.3314131] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/14/2024]
Abstract
Spectral computed tomography (CT) has recently emerged as an advanced version of medical CT and significantly improves conventional (single-energy) CT. Spectral CT has two main forms: dual-energy computed tomography (DECT) and photon-counting computed tomography (PCCT), which offer image improvement, material decomposition, and feature quantification relative to conventional CT. However, the inherent challenges of spectral CT, evidenced by data and image artifacts, remain a bottleneck for clinical applications. To address these problems, machine learning techniques have been widely applied to spectral CT. In this review, we present the state-of-the-art data-driven techniques for spectral CT.
Collapse
Affiliation(s)
- Alexandre Bousse
- LaTIM, Inserm UMR 1101, Université de Bretagne Occidentale, 29238 Brest, France
| | | | - Simon Rit
- Univ Lyon, INSA-Lyon, Université Claude Bernard Lyon 1, UJM-Saint Étienne, CNRS, Inserm, CREATIS UMR 5220, U1294, F-69373, Lyon, France
| | - Alessandro Perelli
- Department of Biomedical Engineering, School of Science and Engineering, University of Dundee, DD1 4HN, UK
| | - Mengzhou Li
- Biomedical Imaging Center, Rensselaer Polytechnic Institute, Troy, New York, USA
| | - Guobao Wang
- Department of Radiology, University of California Davis Health, Sacramento, USA
| | - Jian Zhou
- CTIQ, Canon Medical Research USA, Inc., Vernon Hills, 60061, USA
| | - Ge Wang
- Biomedical Imaging Center, Rensselaer Polytechnic Institute, Troy, New York, USA
| |
Collapse
|
32
|
Kim W, Lee J, Choi JH. An unsupervised two-step training framework for low-dose computed tomography denoising. Med Phys 2024; 51:1127-1144. [PMID: 37432026 DOI: 10.1002/mp.16628] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/03/2023] [Revised: 06/25/2023] [Accepted: 06/25/2023] [Indexed: 07/12/2023] Open
Abstract
BACKGROUND Although low-dose computed tomography (CT) imaging has been more widely adopted in clinical practice to reduce radiation exposure to patients, the reconstructed CT images tend to have more noise, which impedes accurate diagnosis. Recently, deep neural networks using convolutional neural networks to reduce noise in the reconstructed low-dose CT images have shown considerable improvement. However, they need a large number of paired normal- and low-dose CT images to fully train the network via supervised learning methods. PURPOSE To propose an unsupervised two-step training framework for image denoising that uses low-dose CT images of one dataset and unpaired high-dose CT images from another dataset. METHODS Our proposed framework trains the denoising network in two steps. In the first training step, we train the network using 3D volumes of CT images and predict the center CT slice from them. This pre-trained network is used in the second training step to train the denoising network and is combined with the memory-efficient denoising generative adversarial network (DenoisingGAN), which further enhances both objective and perceptual quality. RESULTS The experimental results on phantom and clinical datasets show superior performance over the existing traditional machine learning and self-supervised deep learning methods, and the results are comparable to the fully supervised learning methods. CONCLUSIONS We proposed a new unsupervised learning framework for low-dose CT denoising, convincingly improving noisy CT images from both objective and perceptual quality perspectives. Because our denoising framework does not require physics-based noise models or system-dependent assumptions, our proposed method can be easily reproduced; consequently, it can also be generally applicable to various CT scanners or dose levels.
Collapse
Affiliation(s)
- Wonjin Kim
- Division of Mechanical and Biomedical Engineering, Graduate Program in System Health Science and Engineering, Ewha Womans University, Seoul, Republic of Korea
| | - Jaayeon Lee
- Division of Mechanical and Biomedical Engineering, Graduate Program in System Health Science and Engineering, Ewha Womans University, Seoul, Republic of Korea
| | - Jang-Hwan Choi
- Division of Mechanical and Biomedical Engineering, Graduate Program in System Health Science and Engineering, Ewha Womans University, Seoul, Republic of Korea
| |
Collapse
|
33
|
Sadia RT, Chen J, Zhang J. CT image denoising methods for image quality improvement and radiation dose reduction. J Appl Clin Med Phys 2024; 25:e14270. [PMID: 38240466 PMCID: PMC10860577 DOI: 10.1002/acm2.14270] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2023] [Revised: 12/15/2023] [Accepted: 12/28/2023] [Indexed: 02/13/2024] Open
Abstract
With the ever-increasing use of computed tomography (CT), concerns about its radiation dose have become a significant public issue. To address the need for radiation dose reduction, CT denoising methods have been widely investigated and applied in low-dose CT images. Numerous noise reduction algorithms have emerged, such as iterative reconstruction and most recently, deep learning (DL)-based approaches. Given the rapid advancements in Artificial Intelligence techniques, we recognize the need for a comprehensive review that emphasizes the most recently developed methods. Hence, we have performed a thorough analysis of existing literature to provide such a review. Beyond directly comparing the performance, we focus on pivotal aspects, including model training, validation, testing, generalizability, vulnerability, and evaluation methods. This review is expected to raise awareness of the various facets involved in CT image denoising and the specific challenges in developing DL-based models.
Collapse
Affiliation(s)
- Rabeya Tus Sadia
- Department of Computer ScienceUniversity of KentuckyLexingtonKentuckyUSA
| | - Jin Chen
- Department of Medicine‐NephrologyUniversity of Alabama at BirminghamBirminghamAlabamaUSA
| | - Jie Zhang
- Department of RadiologyUniversity of KentuckyLexingtonKentuckyUSA
| |
Collapse
|
34
|
Tan XI, Liu X, Xiang K, Wang J, Tan S. Deep Filtered Back Projection for CT Reconstruction. IEEE ACCESS : PRACTICAL INNOVATIONS, OPEN SOLUTIONS 2024; 12:20962-20972. [PMID: 39211346 PMCID: PMC11361368 DOI: 10.1109/access.2024.3357355] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Indexed: 09/04/2024]
Abstract
Filtered back projection (FBP) is a classic analytical algorithm for computed tomography (CT) reconstruction, with high computational efficiency. However, images reconstructed by FBP often suffer from excessive noise and artifacts. The original FBP algorithm uses a window function to smooth signals and a linear interpolation to estimate projection values at un-sampled locations. In this study, we propose a novel framework named DeepFBP in which an optimized filter and an optimized nonlinear interpolation operator are learned with neural networks. Specifically, the learned filter can be considered as the product of an optimized window function and the ramp filter, and the learned interpolation can be considered as an optimized way to utilize projection information of nearby locations through nonlinear combination. The proposed method remains the high computational efficiency of the original FBP and achieves much better reconstruction quality at different noise levels. It also outperforms the TV-based statistical iterative algorithm, with computational time being reduced in an order of two, and state-of-the-art post-processing deep learning methods that have deeper and more complicated network structures.
Collapse
Affiliation(s)
- X I Tan
- College of Electrical and Information Engineering, Hunan University of Technology, Zhuzhou 80305, China
| | - Xuan Liu
- School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan 430074, China
| | - Kai Xiang
- School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan 430074, China
| | - Jing Wang
- Department of Radiation Oncology, University of Texas Southwestern Medical Center, Dallas, TX 75390, USA
| | - Shan Tan
- School of Artificial Intelligence and Automation, Huazhong University of Science and Technology, Wuhan 430074, China
| |
Collapse
|
35
|
Zhang Z, Li C, Wang W, Dong Z, Liu G, Dong Y, Zhang Y. Towards full-stack deep learning-empowered data processing pipeline for synchrotron tomography experiments. Innovation (N Y) 2024; 5:100539. [PMID: 38089566 PMCID: PMC10711238 DOI: 10.1016/j.xinn.2023.100539] [Citation(s) in RCA: 2] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/16/2023] [Accepted: 11/13/2023] [Indexed: 10/16/2024] Open
Abstract
Synchrotron tomography experiments are transitioning into multifunctional, cross-scale, and dynamic characterizations, enabled by new-generation synchrotron light sources and fast developments in beamline instrumentation. However, with the spatial and temporal resolving power entering a new era, this transition generates vast amounts of data, which imposes a significant burden on the data processing end. Today, as a highly accurate and efficient data processing method, deep learning shows great potential to address the big data challenge being encountered at future synchrotron beamlines. In this review, we discuss recent advances employing deep learning at different stages of the synchrotron tomography data processing pipeline. We also highlight how applications in other data-intensive fields, such as medical imaging and electron tomography, can be migrated to synchrotron tomography. Finally, we provide our thoughts on possible challenges and opportunities as well as the outlook, envisioning selected deep learning methods, curated big models, and customized learning strategies, all through an intelligent scheduling solution.
Collapse
Affiliation(s)
- Zhen Zhang
- National Synchrotron Radiation Laboratory, University of Science and Technology of China, Hefei 230029, China
| | - Chun Li
- Beijing Synchrotron Radiation Facility, Institute of High Energy Physics, Chinese Academy of Sciences, Beijing 100049, China
| | - Wenhui Wang
- Beijing Synchrotron Radiation Facility, Institute of High Energy Physics, Chinese Academy of Sciences, Beijing 100049, China
| | - Zheng Dong
- Beijing Synchrotron Radiation Facility, Institute of High Energy Physics, Chinese Academy of Sciences, Beijing 100049, China
| | - Gongfa Liu
- National Synchrotron Radiation Laboratory, University of Science and Technology of China, Hefei 230029, China
| | - Yuhui Dong
- Beijing Synchrotron Radiation Facility, Institute of High Energy Physics, Chinese Academy of Sciences, Beijing 100049, China
| | - Yi Zhang
- Beijing Synchrotron Radiation Facility, Institute of High Energy Physics, Chinese Academy of Sciences, Beijing 100049, China
| |
Collapse
|
36
|
Zhang X, Su T, Zhang Y, Cui H, Tan Y, Zhu J, Xia D, Zheng H, Liang D, Ge Y. Transferring U-Net between low-dose CT denoising tasks: a validation study with varied spatial resolutions. Quant Imaging Med Surg 2024; 14:640-652. [PMID: 38223035 PMCID: PMC10784075 DOI: 10.21037/qims-23-768] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/30/2023] [Accepted: 11/09/2023] [Indexed: 01/16/2024]
Abstract
Background Recently, deep learning techniques have been widely used in low-dose computed tomography (LDCT) imaging applications for quickly generating high quality computed tomography (CT) images at lower radiation dose levels. The purpose of this study is to validate the reproducibility of the denoising performance of a given network that has been trained in advance across varied LDCT image datasets that are acquired from different imaging systems with different spatial resolutions. Methods Specifically, LDCT images with comparable noise levels but having different spatial resolutions were prepared to train the U-Net. The number of CT images used for the network training, validation and test was 2,400, 300 and 300, respectively. Afterwards, self- and cross-validations among six selected spatial resolutions (62.5, 125, 250, 375, 500, 625 µm) were studied and compared side by side. The residual variance, peak signal to noise ratio (PSNR), normalized root mean square error (NRMSE) and structural similarity (SSIM) were measured and compared. In addition, network retraining on a small number of image set was performed to fine tune the performance of transfer learning among LDCT tasks with varied spatial resolutions. Results Results demonstrated that the U-Net trained upon LDCT images having a certain spatial resolution can effectively reduce the noise of the other LDCT images having different spatial resolutions. Regardless, results showed that image artifacts would be generated during the above cross validations. For instance, noticeable residual artifacts were presented at the margin and central areas of the object as the resolution inconsistency increased. The retraining results showed that the artifacts caused by the resolution mismatch can be greatly reduced by utilizing about only 20% of the original training data size. This quantitative improvement led to a reduction in the NRMSE from 0.1898 to 0.1263 and an increase in the SSIM from 0.7558 to 0.8036. Conclusions In conclusion, artifacts would be generated when transferring the U-Net to a LDCT denoising task with different spatial resolution. To maintain the denoising performance, it is recommended to retrain the U-Net with a small amount of datasets having the same target spatial resolution.
Collapse
Affiliation(s)
- Xin Zhang
- Research Center for Medical Artificial Intelligence, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
- University of Chinese Academy of Sciences, Beijing, China
| | - Ting Su
- Research Center for Medical Artificial Intelligence, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Yunxin Zhang
- Department of Vascular Surgery, Beijing Jishuitan Hospital, Beijing, China
| | - Han Cui
- Research Center for Medical Artificial Intelligence, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Yuhang Tan
- Research Center for Medical Artificial Intelligence, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Jiongtao Zhu
- College of Physics and Optoelectronic Engineering, Shenzhen University, Shenzhen, China
| | - Dongmei Xia
- Key Laboratory of Low-grade Energy Utilization Technologies and Systems of Ministry of Education of China, College of Power Engineering, Chongqing University, Chongqing, China
| | - Hairong Zheng
- Paul C Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Dong Liang
- Research Center for Medical Artificial Intelligence, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
- Paul C Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| | - Yongshuai Ge
- Research Center for Medical Artificial Intelligence, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
- Paul C Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
| |
Collapse
|
37
|
Zhang J, Huang X, Liu Y, Han Y, Xiang Z. GAN-based medical image small region forgery detection via a two-stage cascade framework. PLoS One 2024; 19:e0290303. [PMID: 38166011 PMCID: PMC10760893 DOI: 10.1371/journal.pone.0290303] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/01/2023] [Accepted: 08/06/2023] [Indexed: 01/04/2024] Open
Abstract
Using generative adversarial network (GAN) Goodfellow et al. (2014) for data enhancement of medical images is significantly helpful for many computer-aided diagnosis (CAD) tasks. A new GAN-based automated tampering attack, like CT-GAN Mirsky et al. (2019), has emerged. It can inject or remove lung cancer lesions to CT scans. Because the tampering region may even account for less than 1% of the original image, even state-of-the-art methods are challenging to detect the traces of such tampering. This paper proposes a two-stage cascade framework to detect GAN-based medical image small region forgery like CT-GAN. In the local detection stage, we train the detector network with small sub-images so that interference information in authentic regions will not affect the detector. We use depthwise separable convolution and residual networks to prevent the detector from over-fitting and enhance the ability to find forged regions through the attention mechanism. The detection results of all sub-images in the same image will be combined into a heatmap. In the global classification stage, using gray-level co-occurrence matrix (GLCM) can better extract features of the heatmap. Because the shape and size of the tampered region are uncertain, we use hyperplanes in an infinite-dimensional space for classification. Our method can classify whether a CT image has been tampered and locate the tampered position. Sufficient experiments show that our method can achieve excellent performance than the state-of-the-art detection methods.
Collapse
Affiliation(s)
- Jianyi Zhang
- Beijing Electronic Science and Technology Institute, Beijing, China
- University of Louisiana at Lafayette, Lafayette, Louisiana, United States of America
| | - Xuanxi Huang
- Beijing Electronic Science and Technology Institute, Beijing, China
| | - Yaqi Liu
- Beijing Electronic Science and Technology Institute, Beijing, China
| | - Yuyang Han
- Beijing Electronic Science and Technology Institute, Beijing, China
| | - Zixiao Xiang
- Beijing Electronic Science and Technology Institute, Beijing, China
| |
Collapse
|
38
|
Chao L, Wang Y, Zhang T, Shan W, Zhang H, Wang Z, Li Q. Joint denoising and interpolating network for low-dose cone-beam CT reconstruction under hybrid dose-reduction strategy. Comput Biol Med 2024; 168:107830. [PMID: 38086140 DOI: 10.1016/j.compbiomed.2023.107830] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/26/2023] [Revised: 11/12/2023] [Accepted: 12/04/2023] [Indexed: 01/10/2024]
Abstract
Cone-beam computed tomography (CBCT) is generally reconstructed with hundreds of two-dimensional X-Ray projections through the FDK algorithm, and its excessive ionizing radiation of X-Ray may impair patients' health. Two common dose-reduction strategies are to either lower the intensity of X-Ray, i.e., low-intensity CBCT, or reduce the number of projections, i.e., sparse-view CBCT. Existing efforts improve the low-dose CBCT images only under a single dose-reduction strategy. In this paper, we argue that applying the two strategies simultaneously can reduce dose in a gentle manner and avoid the extreme degradation of the projection data in a single dose-reduction strategy, especially under ultra-low-dose situations. Therefore, we develop a Joint Denoising and Interpolating Network (JDINet) in projection domain to improve the CBCT quality with the hybrid low-intensity and sparse-view projections. Specifically, JDINet mainly includes two important components, i.e., denoising module and interpolating module, to respectively suppress the noise caused by the low-intensity strategy and interpolate the missing projections caused by the sparse-view strategy. Because FDK actually utilizes the projection information after ramp-filtering, we develop a filtered structural similarity constraint to help JDINet focus on the reconstruction-required information. Afterward, we employ a Postprocessing Network (PostNet) in the reconstruction domain to refine the CBCT images that are reconstructed with denoised and interpolated projections. In general, a complete CBCT reconstruction framework is built with JDINet, FDK, and PostNet. Experiments demonstrate that our framework decreases RMSE by approximately 8 %, 15 %, and 17 %, respectively, on the 1/8, 1/16, and 1/32 dose data, compared to the latest methods. In conclusion, our learning-based framework can be deeply imbedded into the CBCT systems to promote the development of CBCT. Source code is available at https://github.com/LianyingChao/FusionLowDoseCBCT.
Collapse
Affiliation(s)
- Lianying Chao
- Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, Huazhong University of Science and Technology, Wuhan, Hubei, China; MoE Key Laboratory for Biomedical Photonics, Collaborative Innovation Center for Biomedical Engineering, School of Engineering Sciences, Huazhong University of Science and Technology, Wuhan, Hubei, China
| | - Yanli Wang
- Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, Huazhong University of Science and Technology, Wuhan, Hubei, China; MoE Key Laboratory for Biomedical Photonics, Collaborative Innovation Center for Biomedical Engineering, School of Engineering Sciences, Huazhong University of Science and Technology, Wuhan, Hubei, China
| | - TaoTao Zhang
- Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, Huazhong University of Science and Technology, Wuhan, Hubei, China; MoE Key Laboratory for Biomedical Photonics, Collaborative Innovation Center for Biomedical Engineering, School of Engineering Sciences, Huazhong University of Science and Technology, Wuhan, Hubei, China; Northern Jiangsu People's Hospital, Yangzhou, Jiangsu, China
| | - Wenqi Shan
- Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, Huazhong University of Science and Technology, Wuhan, Hubei, China; MoE Key Laboratory for Biomedical Photonics, Collaborative Innovation Center for Biomedical Engineering, School of Engineering Sciences, Huazhong University of Science and Technology, Wuhan, Hubei, China
| | - Haobo Zhang
- Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, Huazhong University of Science and Technology, Wuhan, Hubei, China; MoE Key Laboratory for Biomedical Photonics, Collaborative Innovation Center for Biomedical Engineering, School of Engineering Sciences, Huazhong University of Science and Technology, Wuhan, Hubei, China
| | - Zhiwei Wang
- Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, Huazhong University of Science and Technology, Wuhan, Hubei, China; MoE Key Laboratory for Biomedical Photonics, Collaborative Innovation Center for Biomedical Engineering, School of Engineering Sciences, Huazhong University of Science and Technology, Wuhan, Hubei, China
| | - Qiang Li
- Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, Huazhong University of Science and Technology, Wuhan, Hubei, China; MoE Key Laboratory for Biomedical Photonics, Collaborative Innovation Center for Biomedical Engineering, School of Engineering Sciences, Huazhong University of Science and Technology, Wuhan, Hubei, China.
| |
Collapse
|
39
|
Wu Q, Ji X, Gu Y, Xiang J, Quan G, Li B, Zhu J, Coatrieux G, Coatrieux JL, Chen Y. Unsharp Structure Guided Filtering for Self-Supervised Low-Dose CT Imaging. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:3283-3294. [PMID: 37235462 DOI: 10.1109/tmi.2023.3280217] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/28/2023]
Abstract
Low-dose computed tomography (LDCT) imaging faces great challenges. Although supervised learning has revealed great potential, it requires sufficient and high-quality references for network training. Therefore, existing deep learning methods have been sparingly applied in clinical practice. To this end, this paper presents a novel Unsharp Structure Guided Filtering (USGF) method, which can reconstruct high-quality CT images directly from low-dose projections without clean references. Specifically, we first employ low-pass filters to estimate the structure priors from the input LDCT images. Then, inspired by classical structure transfer techniques, deep convolutional networks are adopted to implement our imaging method which combines guided filtering and structure transfer. Finally, the structure priors serve as the guidance images to alleviate over-smoothing, as they can transfer specific structural characteristics to the generated images. Furthermore, we incorporate traditional FBP algorithms into self-supervised training to enable the transformation of projection domain data to the image domain. Extensive comparisons and analyses on three datasets demonstrate that the proposed USGF has achieved superior performance in terms of noise suppression and edge preservation, and could have a significant impact on LDCT imaging in the future.
Collapse
|
40
|
Patwari M, Gutjahr R, Marcus R, Thali Y, Calvarons AF, Raupach R, Maier A. Reducing the risk of hallucinations with interpretable deep learning models for low-dose CT denoising: comparative performance analysis. Phys Med Biol 2023; 68:19LT01. [PMID: 37733068 DOI: 10.1088/1361-6560/acfc11] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/18/2023] [Accepted: 09/21/2023] [Indexed: 09/22/2023]
Abstract
Objective.Reducing CT radiation dose is an often proposed measure to enhance patient safety, which, however results in increased image noise, translating into degradation of clinical image quality. Several deep learning methods have been proposed for low-dose CT (LDCT) denoising. The high risks posed by possible hallucinations in clinical images necessitate methods which aid the interpretation of deep learning networks. In this study, we aim to use qualitative reader studies and quantitative radiomics studies to assess the perceived quality, signal preservation and statistical feature preservation of LDCT volumes denoised by deep learning. We aim to compare interpretable deep learning methods with classical deep neural networks in clinical denoising performance.Approach.We conducted an image quality analysis study to assess the image quality of the denoised volumes based on four criteria to assess the perceived image quality. We subsequently conduct a lesion detection/segmentation study to assess the impact of denoising on signal detectability. Finally, a radiomic analysis study was performed to observe the quantitative and statistical similarity of the denoised images to standard dose CT (SDCT) images.Main results.The use of specific deep learning based algorithms generate denoised volumes which are qualitatively inferior to SDCT volumes(p< 0.05). Contrary to previous literature, denoising the volumes did not reduce the accuracy of the segmentation (p> 0.05). The denoised volumes, in most cases, generated radiomics features which were statistically similar to those generated from SDCT volumes (p> 0.05).Significance.Our results show that the denoised volumes have a lower perceived quality than SDCT volumes. Noise and denoising do not significantly affect detectability of the abdominal lesions. Denoised volumes also contain statistically identical features to SDCT volumes.
Collapse
Affiliation(s)
- Mayank Patwari
- Pattern Recognition Lab, Friedrich-Alexander Universität Erlangen-Nürnberg, D-91058 Erlangen, Germany
- CT Concepts, Siemens Healthineers AG, D-91301 Forchheim, Germany
| | - Ralf Gutjahr
- CT Concepts, Siemens Healthineers AG, D-91301 Forchheim, Germany
| | - Roy Marcus
- Balgrist University Hospital Zurich, 8008 Zurich, Switzerland
- Faculty of Medicine, University of Zurich, 8032 Zurich, Switzerland
- Cantonal Hospital of Lucerne, 6016 Lucerne, Switzerland
| | - Yannick Thali
- Spital Zofingen AG, 4800 Zofingen, Switzerland
- Cantonal Hospital of Lucerne, 6016 Lucerne, Switzerland
| | | | - Rainer Raupach
- CT Concepts, Siemens Healthineers AG, D-91301 Forchheim, Germany
| | - Andreas Maier
- Pattern Recognition Lab, Friedrich-Alexander Universität Erlangen-Nürnberg, D-91058 Erlangen, Germany
| |
Collapse
|
41
|
Choi K, Kim SH, Kim S. Self-supervised denoising of projection data for low-dose cone-beam CT. Med Phys 2023; 50:6319-6333. [PMID: 37079443 DOI: 10.1002/mp.16421] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/01/2022] [Revised: 04/03/2023] [Accepted: 04/03/2023] [Indexed: 04/21/2023] Open
Abstract
BACKGROUND Convolutional neural networks (CNNs) have shown promising results in image denoising tasks. While most existing CNN-based methods depend on supervised learning by directly mapping noisy inputs to clean targets, high-quality references are often unavailable for interventional radiology such as cone-beam computed tomography (CBCT). PURPOSE In this paper, we propose a novel self-supervised learning method that reduces noise in projections acquired by ordinary CBCT scans. METHODS With a network that partially blinds input, we are able to train the denoising model by mapping the partially blinded projections to the original projections. Additionally, we incorporate noise-to-noise learning into the self-supervised learning by mapping the adjacent projections to the original projections. With standard image reconstruction methods such as FDK-type algorithms, we can reconstruct high-quality CBCT images from the projections denoised by our projection-domain denoising method. RESULTS In the head phantom study, we measure peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM) values of the proposed method along with the other denoising methods and uncorrected low-dose CBCT data for a quantitative comparison both in projection and image domains. The PSNR and SSIM values of our self-supervised denoising approach are 27.08 and 0.839, whereas those of uncorrected CBCT images are 15.68 and 0.103, respectively. In the retrospective study, we assess the quality of interventional patient CBCT images to evaluate the projection-domain and image-domain denoising methods. Both qualitative and quantitative results indicate that our approach can effectively produce high-quality CBCT images with low-dose projections in the absence of duplicate clean or noisy references. CONCLUSIONS Our self-supervised learning strategy is capable of restoring anatomical information while efficiently removing noise in CBCT projection data.
Collapse
Affiliation(s)
- Kihwan Choi
- Bionics Research Center, Korea Institute of Science and Technology, Seoul, Republic of Korea
| | - Seung Hyoung Kim
- Department of Radiology, Yonsei University College of Medicine, Seoul, Republic of Korea
| | - Sungwon Kim
- Department of Radiology, Yonsei University College of Medicine, Seoul, Republic of Korea
| |
Collapse
|
42
|
Duan X, Ding XF, Li N, Wu FX, Chen X, Zhu N. Sparse2Noise: Low-dose synchrotron X-ray tomography without high-quality reference data. Comput Biol Med 2023; 165:107473. [PMID: 37690288 DOI: 10.1016/j.compbiomed.2023.107473] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/07/2023] [Revised: 08/30/2023] [Accepted: 09/04/2023] [Indexed: 09/12/2023]
Abstract
BACKGROUND Synchrotron radiation computed tomography (SR-CT) holds promise for high-resolution in vivo imaging. Notably, the reconstruction of SR-CT images necessitates a large set of data to be captured with sufficient photons from multiple angles, resulting in high radiation dose received by the object. Reducing the number of projections and/or photon flux is a straightforward means to lessen the radiation dose, however, compromises data completeness, thus introducing noises and artifacts. Deep learning (DL)-based supervised methods effectively denoise and remove artifacts, but they heavily depend on high-quality paired data acquired at high doses. Although algorithms exist for training without high-quality references, they struggle to effectively eliminate persistent artifacts present in real-world data. METHODS This work presents a novel low-dose imaging strategy namely Sparse2Noise, which combines the reconstruction data from paired sparse-view CT scan (normal-flux) and full-view CT scan (low-flux) using a convolutional neural network (CNN). Sparse2Noise does not require high-quality reconstructed data as references and allows for fresh training on data with very small size. Sparse2Noise was evaluated by both simulated and experimental data. RESULTS Sparse2Noise effectively reduces noise and ring artifacts while maintaining high image quality, outperforming state-of-the-art image denoising methods at same dose levels. Furthermore, Sparse2Noise produces impressive high image quality for ex vivo rat hindlimb imaging with the acceptable low radiation dose (i.e., 0.5 Gy with the isotropic voxel size of 26 μm). CONCLUSIONS This work represents a significant advance towards in vivo SR-CT imaging. It is noteworthy that Sparse2Noise can also be used for denoising in conventional CT and/or phase-contrast CT.
Collapse
Affiliation(s)
- Xiaoman Duan
- Division of Biomedical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada
| | - Xiao Fan Ding
- Division of Biomedical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada
| | - Naitao Li
- Division of Biomedical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada
| | - Fang-Xiang Wu
- Division of Biomedical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada; Department of Computer Science, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada; Department of Mechanical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada
| | - Xiongbiao Chen
- Division of Biomedical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada; Department of Mechanical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada.
| | - Ning Zhu
- Division of Biomedical Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada; Canadian Light Source, Saskatoon, S7N 2V3, SK, Canada; Department of Chemical and Biological Engineering, College of Engineering, University of Saskatchewan, Saskatoon, SK S7N 5A9, Canada.
| |
Collapse
|
43
|
Zhao F, Li D, Luo R, Liu M, Jiang X, Hu J. Self-supervised deep learning for joint 3D low-dose PET/CT image denoising. Comput Biol Med 2023; 165:107391. [PMID: 37717529 DOI: 10.1016/j.compbiomed.2023.107391] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/17/2023] [Revised: 08/08/2023] [Accepted: 08/25/2023] [Indexed: 09/19/2023]
Abstract
Deep learning (DL)-based denoising of low-dose positron emission tomography (LDPET) and low-dose computed tomography (LDCT) has been widely explored. However, previous methods have focused only on single modality denoising, neglecting the possibility of simultaneously denoising LDPET and LDCT using only one neural network, i.e., joint LDPET/LDCT denoising. Moreover, DL-based denoising methods generally require plenty of well-aligned LD-normal-dose (LD-ND) sample pairs, which can be difficult to obtain. To this end, we propose a self-supervised two-stage training framework named MAsk-then-Cycle (MAC), to achieve self-supervised joint LDPET/LDCT denoising. The first stage of MAC is masked autoencoder (MAE)-based pre-training and the second stage is self-supervised denoising training. Specifically, we propose a self-supervised denoising strategy named cycle self-recombination (CSR), which enables denoising without well-aligned sample pairs. Unlike other methods that treat noise as a homogeneous whole, CSR disentangles noise into signal-dependent and independent noises. This is more in line with the actual imaging process and allows for flexible recombination of noises and signals to generate new samples. These new samples contain implicit constraints that can improve the network's denoising ability. Based on these constraints, we design multiple loss functions to enable self-supervised training. Then we design a CSR-based denoising network to achieve joint 3D LDPET/LDCT denoising. Existing self-supervised methods generally lack pixel-level constraints on networks, which can easily lead to additional artifacts. Before denoising training, we perform MAE-based pre-training to indirectly impose pixel-level constraints on networks. Experiments on an LDPET/LDCT dataset demonstrate its superiority over existing methods. Our method is the first self-supervised joint LDPET/LDCT denoising method. It does not require any prior assumptions and is therefore more robust.
Collapse
Affiliation(s)
- Feixiang Zhao
- State Key Laboratory of Geohazard Prevention and Geoenvironment Protection, Chengdu University of Technology, Chengdu, 610000, China.
| | - Dongfen Li
- State Key Laboratory of Geohazard Prevention and Geoenvironment Protection, Chengdu University of Technology, Chengdu, 610000, China.
| | - Rui Luo
- Department of Nuclear Medicine, Mianyang Central Hospital, Mianyang, 621000, China.
| | - Mingzhe Liu
- State Key Laboratory of Geohazard Prevention and Geoenvironment Protection, Chengdu University of Technology, Chengdu, 610000, China.
| | - Xin Jiang
- School of Data Science and Artificial Intelligence, Wenzhou University of Technology, Wenzhou, 325000, China.
| | - Junjie Hu
- Machine Intelligence Laboratory, College of Computer Science, Sichuan University, Chengdu, 610065, China.
| |
Collapse
|
44
|
Zhang Y, Hao D, Lin Y, Sun W, Zhang J, Meng J, Ma F, Guo Y, Lu H, Li G, Liu J. Structure-preserving low-dose computed tomography image denoising using a deep residual adaptive global context attention network. Quant Imaging Med Surg 2023; 13:6528-6545. [PMID: 37869272 PMCID: PMC10585579 DOI: 10.21037/qims-23-194] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/17/2023] [Accepted: 08/18/2023] [Indexed: 10/24/2023]
Abstract
Background Low-dose computed tomography (LDCT) scans can effectively reduce the radiation damage to patients, but this is highly detrimental to CT image quality. Deep convolutional neural networks (CNNs) have shown their potential in improving LDCT image quality. However, the conventional CNN-based approaches rely fundamentally on the convolution operations, which are ineffective for modeling the correlations among nonlocal similar structures and the regionally distinct statistical properties in CT images. This modeling deficiency hampers the denoising performance for CT images derived in this manner. Methods In this paper, we propose an adaptive global context (AGC) modeling scheme to describe the nonlocal correlations and the regionally distinct statistics in CT images with negligible computation load. We further propose an AGC-based long-short residual encoder-decoder (AGC-LSRED) network for efficient LDCT image noise artifact-suppression tasks. Specifically, stacks of residual AGC attention blocks (RAGCBs) with long and short skip connections are constructed in the AGC-LSRED network, which allows valuable structural and positional information to be bypassed through these identity-based skip connections and thus eases the training of the deep denoising network. For training the AGC-LSRED network, we propose a compound loss that combines the L1 loss, adversarial loss, and self-supervised multi-scale perceptual loss. Results Quantitative and qualitative experimental studies were performed to verify and validate the effectiveness of the proposed method. The simulation experiments demonstrated the proposed method exhibits the best result in terms of noise suppression [root-mean-square error (RMSE) =9.02; peak signal-to-noise ratio (PSNR) =33.17] and fine structure preservation [structural similarity index (SSIM) =0.925] compared with other competitive CNN-based methods. The experiments on real data illustrated that the proposed method has advantages over other methods in terms of radiologists' subjective assessment scores (averaged scores =4.34). Conclusions With the use of the AGC modeling scheme to characterize the structural information in CT images and of residual AGC-attention blocks with long and short skip connections to ease the network training, the proposed AGC-LSRED method achieves satisfactory results in preserving fine anatomical structures and suppressing noise in LDCT images.
Collapse
Affiliation(s)
- Yuanke Zhang
- School of Computer Science, Qufu Normal University, Rizhao, China
| | - Dejing Hao
- School of Computer Science, Qufu Normal University, Rizhao, China
| | - Yingying Lin
- School of Computer Science, Qufu Normal University, Rizhao, China
| | - Wanxin Sun
- School of Computer Science, Qufu Normal University, Rizhao, China
| | - Jinke Zhang
- School of Computer Science, Qufu Normal University, Rizhao, China
| | - Jing Meng
- School of Computer Science, Qufu Normal University, Rizhao, China
| | - Fei Ma
- School of Computer Science, Qufu Normal University, Rizhao, China
| | - Yanfei Guo
- School of Computer Science, Qufu Normal University, Rizhao, China
| | - Hongbing Lu
- School of Biomedical Engineering, Fourth Military Medical University, Xi’an, China
| | - Guangshun Li
- School of Computer Science, Qufu Normal University, Rizhao, China
| | - Jianlei Liu
- School of Computer Science, Qufu Normal University, Rizhao, China
| |
Collapse
|
45
|
Yu J, Zhang H, Zhang P, Zhu Y. Unsupervised learning-based dual-domain method for low-dose CT denoising. Phys Med Biol 2023; 68:185010. [PMID: 37567225 DOI: 10.1088/1361-6560/acefa2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/12/2023] [Accepted: 08/10/2023] [Indexed: 08/13/2023]
Abstract
Objective. Low-dose CT (LDCT) is an important research topic in the field of CT imaging because of its ability to reduce radiation damage in clinical diagnosis. In recent years, deep learning techniques have been widely applied in LDCT imaging and a large number of denoising methods have been proposed. However, One major challenge of supervised deep learning-based methods is the exactly geometric pairing of datasets with different doses. Therefore, the aim of this study is to develop an unsupervised learning-based LDCT imaging method to address the aforementioned challenges.Approach. In this paper, we propose an unsupervised learning-based dual-domain method for LDCT denoising, which consists of two stages: the first stage is projection domain denoising, in which the unsupervised learning method Noise2Self is applied to denoise the projection data with statistically independent and zero-mean noise. The second stage is an iterative enhancement approach, which combines the prior information obtained from the generative model with an iterative reconstruction algorithm to enhance the details of the reconstructed image.Main results. Experimental results show that our proposed method outperforms the comparison method in terms of denoising effect. Particularly, in terms of SSIM, the denoised results obtained using our method achieve the highest SSIM.Significance. In conclusion, our unsupervised learning-based method can be a promising alternative to the traditional supervised methods for LDCT imaging, especially when the availability of the labeled datasets is limited.
Collapse
Affiliation(s)
- Jie Yu
- School of Mathematical Sciences, Capital Normal University, Beijing, 100048, People's Republic of China
| | - Huitao Zhang
- School of Mathematical Sciences, Capital Normal University, Beijing, 100048, People's Republic of China
- Shenzhen National Applied Mathematics Center, Southern University of Science and Technology, Shenzhen, 518055, People's Republic of China
| | - Peng Zhang
- School of Mathematical Sciences, Capital Normal University, Beijing, 100048, People's Republic of China
| | - Yining Zhu
- School of Mathematical Sciences, Capital Normal University, Beijing, 100048, People's Republic of China
- Shenzhen National Applied Mathematics Center, Southern University of Science and Technology, Shenzhen, 518055, People's Republic of China
| |
Collapse
|
46
|
Huang Z, Li W, Wang Y, Liu Z, Zhang Q, Jin Y, Wu R, Quan G, Liang D, Hu Z, Zhang N. MLNAN: Multi-level noise-aware network for low-dose CT imaging implemented with constrained cycle Wasserstein generative adversarial networks. Artif Intell Med 2023; 143:102609. [PMID: 37673577 DOI: 10.1016/j.artmed.2023.102609] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/05/2022] [Revised: 05/17/2023] [Accepted: 06/06/2023] [Indexed: 09/08/2023]
Abstract
Low-dose CT techniques attempt to minimize the radiation exposure of patients by estimating the high-resolution normal-dose CT images to reduce the risk of radiation-induced cancer. In recent years, many deep learning methods have been proposed to solve this problem by building a mapping function between low-dose CT images and their high-dose counterparts. However, most of these methods ignore the effect of different radiation doses on the final CT images, which results in large differences in the intensity of the noise observable in CT images. What'more, the noise intensity of low-dose CT images exists significantly differences under different medical devices manufacturers. In this paper, we propose a multi-level noise-aware network (MLNAN) implemented with constrained cycle Wasserstein generative adversarial networks to recovery the low-dose CT images under uncertain noise levels. Particularly, the noise-level classification is predicted and reused as a prior pattern in generator networks. Moreover, the discriminator network introduces noise-level determination. Under two dose-reduction strategies, experiments to evaluate the performance of proposed method are conducted on two datasets, including the simulated clinical AAPM challenge datasets and commercial CT datasets from United Imaging Healthcare (UIH). The experimental results illustrate the effectiveness of our proposed method in terms of noise suppression and structural detail preservation compared with several other deep-learning based methods. Ablation studies validate the effectiveness of the individual components regarding the afforded performance improvement. Further research for practical clinical applications and other medical modalities is required in future works.
Collapse
Affiliation(s)
- Zhenxing Huang
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
| | - Wenbo Li
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China; Shenzhen College of Advanced Technology, University of Chinese Academy of Sciences, Beijing 101408, China
| | - Yunling Wang
- Department of Radiology, First Affiliated Hospital of Xinjiang Medical University, Urumqi, 830011, China.
| | - Zhou Liu
- Department of Radiology, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital & Shenzhen Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Shenzhen, 518116, China
| | - Qiyang Zhang
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
| | - Yuxi Jin
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
| | - Ruodai Wu
- Department of Radiology, Shenzhen University General Hospital, Shenzhen University Clinical Medical Academy, Shenzhen 518055, China
| | - Guotao Quan
- Shanghai United Imaging Healthcare, Shanghai 201807, China
| | - Dong Liang
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
| | - Zhanli Hu
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
| | - Na Zhang
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China.
| |
Collapse
|
47
|
Huang J, Chen K, Ren Y, Sun J, Wang Y, Tao T, Pu X. CDDnet: Cross-domain denoising network for low-dose CT image via local and global information alignment. Comput Biol Med 2023; 163:107219. [PMID: 37422942 DOI: 10.1016/j.compbiomed.2023.107219] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/12/2023] [Revised: 05/21/2023] [Accepted: 06/25/2023] [Indexed: 07/11/2023]
Abstract
The domain shift problem has emerged as a challenge in cross-domain low-dose CT (LDCT) image denoising task, where the acquisition of a sufficient number of medical images from multiple sources may be constrained by privacy concerns. In this study, we propose a novel cross-domain denoising network (CDDnet) that incorporates both local and global information of CT images. To address the local component, a local information alignment module has been proposed to regularize the similarity between extracted target and source features from selected patches. To align the general information of the semantic structure from a global perspective, an autoencoder is adopted to learn the latent correlation between the source label and the estimated target label generated by the pre-trained denoiser. Experimental results demonstrate that our proposed CDDnet effectively alleviates the domain shift problem, outperforming other deep learning-based and domain adaptation-based methods under cross-domain scenarios.
Collapse
Affiliation(s)
- Jiaxin Huang
- School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, 611731, China
| | - Kecheng Chen
- Department of Electrical Engineering, City University of Hong Kong, 999077, Hong Kong Special Administrative Region of China
| | - Yazhou Ren
- School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, 611731, China; Shenzhen Institute for Advanced Study, University of Electronic Science and Technology of China, Shenzhen, 518110, China
| | - Jiayu Sun
- West China Hospital, Sichuan University, Chengdu, 610044, China
| | - Yanmei Wang
- Institute of Traditional Chinese Medicine, Sichuan College of Traditional Chinese Medicine (Sichuan Second Hospital of TCM), Chengdu, 610075, China
| | - Tao Tao
- Institute of Traditional Chinese Medicine, Sichuan College of Traditional Chinese Medicine (Sichuan Second Hospital of TCM), Chengdu, 610075, China
| | - Xiaorong Pu
- School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, 611731, China; Shenzhen Institute for Advanced Study, University of Electronic Science and Technology of China, Shenzhen, 518110, China; NHC Key Laboratory of Nuclear Technology Medical Transformation, Mianyang Central Hospital, Mianyang, 621000, China.
| |
Collapse
|
48
|
Gerard SE, Chaudhary MFA, Herrmann J, Christensen GE, Estépar RSJ, Reinhardt JM, Hoffman EA. Direct estimation of regional lung volume change from paired and single CT images using residual regression neural network. Med Phys 2023; 50:5698-5714. [PMID: 36929883 PMCID: PMC10743098 DOI: 10.1002/mp.16365] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/19/2022] [Revised: 02/11/2023] [Accepted: 03/01/2023] [Indexed: 03/18/2023] Open
Abstract
BACKGROUND Chest computed tomography (CT) enables characterization of pulmonary diseases by producing high-resolution and high-contrast images of the intricate lung structures. Deformable image registration is used to align chest CT scans at different lung volumes, yielding estimates of local tissue expansion and contraction. PURPOSE We investigated the utility of deep generative models for directly predicting local tissue volume change from lung CT images, bypassing computationally expensive iterative image registration and providing a method that can be utilized in scenarios where either one or two CT scans are available. METHODS A residual regression convolutional neural network, called Reg3DNet+, is proposed for directly regressing high-resolution images of local tissue volume change (i.e., Jacobian) from CT images. Image registration was performed between lung volumes at total lung capacity (TLC) and functional residual capacity (FRC) using a tissue mass- and structure-preserving registration algorithm. The Jacobian image was calculated from the registration-derived displacement field and used as the ground truth for local tissue volume change. Four separate Reg3DNet+ models were trained to predict Jacobian images using a multifactorial study design to compare the effects of network input (i.e., single image vs. paired images) and output space (i.e., FRC vs. TLC). The models were trained and evaluated on image datasets from the COPDGene study. Models were evaluated against the registration-derived Jacobian images using local, regional, and global evaluation metrics. RESULTS Statistical analysis revealed that both factors - network input and output space - were significant determinants for change in evaluation metrics. Paired-input models performed better than single-input models, and model performance was better in the output space of FRC rather than TLC. Mean structural similarity index for paired-input models was 0.959 and 0.956 for FRC and TLC output spaces, respectively, and for single-input models was 0.951 and 0.937. Global evaluation metrics demonstrated correlation between registration-derived Jacobian mean and predicted Jacobian mean: coefficient of determination (r2 ) for paired-input models was 0.974 and 0.938 for FRC and TLC output spaces, respectively, and for single-input models was 0.598 and 0.346. After correcting for effort, registration-derived lobar volume change was strongly correlated with the predicted lobar volume change: for paired-input models r2 was 0.899 for both FRC and TLC output spaces, and for single-input models r2 was 0.803 and 0.862, respectively. CONCLUSIONS Convolutional neural networks can be used to directly predict local tissue mechanics, eliminating the need for computationally expensive image registration. Networks that use paired CT images acquired at TLC and FRC allow for more accurate prediction of local tissue expansion compared to networks that use a single image. Networks that only require a single input image still show promising results, particularly after correcting for effort, and allow for local tissue expansion estimation in cases where multiple CT scans are not available. For single-input networks, the FRC image is more predictive of local tissue volume change compared to the TLC image.
Collapse
Affiliation(s)
- Sarah E. Gerard
- Roy J. Carver Department of Biomedical Engineering, University of Iowa, Iowa City, Iowa, USA
- Department of Radiology, University of Iowa, Iowa City, Iowa, USA
| | | | - Jacob Herrmann
- Roy J. Carver Department of Biomedical Engineering, University of Iowa, Iowa City, Iowa, USA
| | - Gary E. Christensen
- Department of Electrical and Computer Engineering, University of Iowa, Iowa City, Iowa, USA
- Department of Radiation Oncology, University of Iowa, Iowa City, Iowa, USA
| | | | - Joseph M. Reinhardt
- Roy J. Carver Department of Biomedical Engineering, University of Iowa, Iowa City, Iowa, USA
- Department of Radiology, University of Iowa, Iowa City, Iowa, USA
| | - Eric A. Hoffman
- Roy J. Carver Department of Biomedical Engineering, University of Iowa, Iowa City, Iowa, USA
- Department of Radiology, University of Iowa, Iowa City, Iowa, USA
| |
Collapse
|
49
|
Li M, Wang J, Chen Y, Tang Y, Wu Z, Qi Y, Jiang H, Zheng J, Tsui BMW. Low-Dose CT Image Synthesis for Domain Adaptation Imaging Using a Generative Adversarial Network With Noise Encoding Transfer Learning. IEEE TRANSACTIONS ON MEDICAL IMAGING 2023; 42:2616-2630. [PMID: 37030685 DOI: 10.1109/tmi.2023.3261822] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/19/2023]
Abstract
Deep learning (DL) based image processing methods have been successfully applied to low-dose x-ray images based on the assumption that the feature distribution of the training data is consistent with that of the test data. However, low-dose computed tomography (LDCT) images from different commercial scanners may contain different amounts and types of image noise, violating this assumption. Moreover, in the application of DL based image processing methods to LDCT, the feature distributions of LDCT images from simulation and clinical CT examination can be quite different. Therefore, the network models trained with simulated image data or LDCT images from one specific scanner may not work well for another CT scanner and image processing task. To solve such domain adaptation problem, in this study, a novel generative adversarial network (GAN) with noise encoding transfer learning (NETL), or GAN-NETL, is proposed to generate a paired dataset with a different noise style. Specifically, we proposed a method to perform noise encoding operator and incorporate it into the generator to extract a noise style. Meanwhile, with a transfer learning (TL) approach, the image noise encoding operator transformed the noise type of the source domain to that of the target domain for realistic noise generation. One public and two private datasets are used to evaluate the proposed method. Experiment results demonstrated the feasibility and effectiveness of our proposed GAN-NETL model in LDCT image synthesis. In addition, we conduct additional image denoising study using the synthesized clinical LDCT data, which verified the merit of the proposed synthesis in improving the performance of the DL based LDCT processing method.
Collapse
|
50
|
Shan H, Vimieiro RB, Borges LR, Vieira MAC, Wang G. Impact of loss functions on the performance of a deep neural network designed to restore low-dose digital mammography. Artif Intell Med 2023; 142:102555. [PMID: 37316093 PMCID: PMC10267506 DOI: 10.1016/j.artmed.2023.102555] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/26/2021] [Revised: 04/13/2023] [Accepted: 04/14/2023] [Indexed: 06/16/2023]
Abstract
Digital mammography is currently the most common imaging tool for breast cancer screening. Although the benefits of using digital mammography for cancer screening outweigh the risks associated with the x-ray exposure, the radiation dose must be kept as low as possible while maintaining the diagnostic utility of the generated images, thus minimizing patient risks. Many studies investigated the feasibility of dose reduction by restoring low-dose images using deep neural networks. In these cases, choosing the appropriate training database and loss function is crucial and impacts the quality of the results. In this work, we used a standard residual network (ResNet) to restore low-dose digital mammography images and evaluated the performance of several loss functions. For training purposes, we extracted 256,000 image patches from a dataset of 400 images of retrospective clinical mammography exams, where dose reduction factors of 75% and 50% were simulated to generate low and standard-dose pairs. We validated the network in a real scenario by using a physical anthropomorphic breast phantom to acquire real low-dose and standard full-dose images in a commercially available mammography system, which were then processed through our trained model. We benchmarked our results against an analytical restoration model for low-dose digital mammography. Objective assessment was performed through the signal-to-noise ratio (SNR) and the mean normalized squared error (MNSE), decomposed into residual noise and bias. Statistical tests revealed that the use of the perceptual loss (PL4) resulted in statistically significant differences when compared to all other loss functions. Additionally, images restored using the PL4 achieved the closest residual noise to the standard dose. On the other hand, perceptual loss PL3, structural similarity index (SSIM) and one of the adversarial losses achieved the lowest bias for both dose reduction factors. The source code of our deep neural network is available at https://github.com/WANG-AXIS/LdDMDenoising.
Collapse
Affiliation(s)
- Hongming Shan
- Institute of Science and Technology for Brain-inspired Intelligence and MOE Frontiers Center for Brain Science, Fudan University, Shanghai, China; Shanghai Center for Brain Science and Brain-inspired Technology, Shanghai, China; Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, USA.
| | - Rodrigo B Vimieiro
- Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, USA; Department of Electrical and Computer Engineering, São Carlos School of Engineering, University of São Paulo, São Carlos, Brazil.
| | - Lucas R Borges
- Department of Electrical and Computer Engineering, São Carlos School of Engineering, University of São Paulo, São Carlos, Brazil; Real Time Tomography, LLC, Villanova, USA.
| | - Marcelo A C Vieira
- Department of Electrical and Computer Engineering, São Carlos School of Engineering, University of São Paulo, São Carlos, Brazil.
| | - Ge Wang
- Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, USA.
| |
Collapse
|