51
|
Islam KT, Wijewickrema S, O’Leary S. A Deep Learning Framework for Segmenting Brain Tumors Using MRI and Synthetically Generated CT Images. SENSORS (BASEL, SWITZERLAND) 2022; 22:523. [PMID: 35062484 PMCID: PMC8780247 DOI: 10.3390/s22020523] [Citation(s) in RCA: 13] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/23/2021] [Revised: 12/26/2021] [Accepted: 12/28/2021] [Indexed: 06/14/2023]
Abstract
Multi-modal three-dimensional (3-D) image segmentation is used in many medical applications, such as disease diagnosis, treatment planning, and image-guided surgery. Although multi-modal images provide information that no single image modality alone can provide, integrating such information to be used in segmentation is a challenging task. Numerous methods have been introduced to solve the problem of multi-modal medical image segmentation in recent years. In this paper, we propose a solution for the task of brain tumor segmentation. To this end, we first introduce a method of enhancing an existing magnetic resonance imaging (MRI) dataset by generating synthetic computed tomography (CT) images. Then, we discuss a process of systematic optimization of a convolutional neural network (CNN) architecture that uses this enhanced dataset, in order to customize it for our task. Using publicly available datasets, we show that the proposed method outperforms similar existing methods.
Collapse
|
52
|
Cong W, De Man B, Wang G. Projection decomposition via univariate optimization for dual-energy CT. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2022; 30:725-736. [PMID: 35634811 PMCID: PMC9427723 DOI: 10.3233/xst-221153] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Dual-energy computed tomography (DECT) acquires two x-ray projection datasets with different x-ray energy spectra, performs material-specific image reconstruction based on the energy-dependent non-linear integral model, and provides more accurate quantification of attenuation coefficients than single energy spectrum CT. In the diagnostic energy range, x-ray energy-dependent attenuation is mainly caused by photoelectric absorption and Compton scattering. Theoretically, these two physical components of the x-ray attenuation mechanism can be determined from two projection datasets with distinct energy spectra. Practically, the solution of the non-linear integral equation is complicated due to spectral uncertainty, detector sensitivity, and data noise. Conventional multivariable optimization methods are prone to local minima. In this paper, we develop a new method for DECT image reconstruction in the projection domain. This method combines an analytic solution of a polynomial equation and a univariate optimization to solve the polychromatic non-linear integral equation. The polynomial equation of an odd order has a unique real solution with sufficient accuracy for image reconstruction, and the univariate optimization can achieve the global optimal solution, allowing accurate and stable projection decomposition for DECT. Numerical and physical phantom experiments are performed to demonstrate the effectiveness of the method in comparison with the state-of-the-art projection decomposition methods. As a result, the univariate optimization method yields a quality improvement of 15% for image reconstruction and substantial reduction of the computational time, as compared to the multivariable optimization methods.
Collapse
Affiliation(s)
- Wenxiang Cong
- Biomedical Imaging Center, Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY, USA
| | - Bruno De Man
- GE Research, One Research Circle, Niskayuna, NY, USA
| | - Ge Wang
- Biomedical Imaging Center, Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY, USA
| |
Collapse
|
53
|
Cui X, Guo Y, Zhang X, Shangguan H, Liu B, Wang A. Artifact-Assisted multi-level and multi-scale feature fusion attention network for low-dose CT denoising. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2022; 30:875-889. [PMID: 35694948 DOI: 10.3233/xst-221149] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
BACKGROUND AND OBJECTIVE Since low-dose computed tomography (LDCT) images typically have higher noise that may affect accuracy of disease diagnosis, the objective of this study is to develop and evaluate a new artifact-assisted feature fusion attention (AAFFA) network to extract and reduce image artifact and noise in LDCT images. METHODS In AAFFA network, a feature fusion attention block is constructed for local multi-scale artifact feature extraction and progressive fusion from coarse to fine. A multi-level fusion architecture based on skip connection and attention modules is also introduced for artifact feature extraction. Specifically, long-range skip connections are used to enhance and fuse artifact features with different depth levels. Then, the fused shallower features enter channel attention for better extraction of artifact features, and the fused deeper features are sent into pixel attention for focusing on the artifact pixel information. Besides, an artifact channel is designed to provide rich artifact features and guide the extraction of noise and artifact features. The AAPM LDCT Challenge dataset is used to train and test the network. The performance is evaluated by using both visual observation and quantitative metrics including peak signal-noise-ratio (PSNR), structural similarity index (SSIM) and visual information fidelity (VIF). RESULTS Using AAFFA network improves the averaged PSNR/SSIM/VIF values of AAPM LDCT images from 43.4961, 0.9595, 0.3926 to 48.2513, 0.9859, 0.4589, respectively. CONCLUSIONS The proposed AAFFA network is able to effectively reduce noise and artifacts while preserving object edges. Assessment of visual quality and quantitative index demonstrates the significant improvement compared with other image denoising methods.
Collapse
Affiliation(s)
- Xueying Cui
- School of Applied Science, Taiyuan University of Science and Technology, Taiyuan, China
| | - Yingting Guo
- School of Applied Science, Taiyuan University of Science and Technology, Taiyuan, China
| | - Xiong Zhang
- School of Electronic Information Engineering, Taiyuan University of Science and Technology, Taiyuan, China
| | - Hong Shangguan
- School of Electronic Information Engineering, Taiyuan University of Science and Technology, Taiyuan, China
| | - Bin Liu
- School of Applied Science, Taiyuan University of Science and Technology, Taiyuan, China
| | - Anhong Wang
- School of Electronic Information Engineering, Taiyuan University of Science and Technology, Taiyuan, China
| |
Collapse
|
54
|
Hu S, Lei B, Wang S, Wang Y, Feng Z, Shen Y. Bidirectional Mapping Generative Adversarial Networks for Brain MR to PET Synthesis. IEEE TRANSACTIONS ON MEDICAL IMAGING 2022; 41:145-157. [PMID: 34428138 DOI: 10.1109/tmi.2021.3107013] [Citation(s) in RCA: 48] [Impact Index Per Article: 16.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Fusing multi-modality medical images, such as magnetic resonance (MR) imaging and positron emission tomography (PET), can provide various anatomical and functional information about the human body. However, PET data is not always available for several reasons, such as high cost, radiation hazard, and other limitations. This paper proposes a 3D end-to-end synthesis network called Bidirectional Mapping Generative Adversarial Networks (BMGAN). Image contexts and latent vectors are effectively used for brain MR-to-PET synthesis. Specifically, a bidirectional mapping mechanism is designed to embed the semantic information of PET images into the high-dimensional latent space. Moreover, the 3D Dense-UNet generator architecture and the hybrid loss functions are further constructed to improve the visual quality of cross-modality synthetic images. The most appealing part is that the proposed method can synthesize perceptually realistic PET images while preserving the diverse brain structures of different subjects. Experimental results demonstrate that the performance of the proposed method outperforms other competitive methods in terms of quantitative measures, qualitative displays, and evaluation metrics for classification.
Collapse
|
55
|
Bera S, Biswas PK. Noise Conscious Training of Non Local Neural Network Powered by Self Attentive Spectral Normalized Markovian Patch GAN for Low Dose CT Denoising. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:3663-3673. [PMID: 34224348 DOI: 10.1109/tmi.2021.3094525] [Citation(s) in RCA: 19] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
The explosive rise of the use of Computer tomography (CT) imaging in medical practice has heightened public concern over the patient's associated radiation dose. On the other hand, reducing the radiation dose leads to increased noise and artifacts, which adversely degrades the scan's interpretability. In recent times, the deep learning-based technique has emerged as a promising method for low dose CT(LDCT) denoising. However, some common bottleneck still exists, which hinders deep learning-based techniques from furnishing the best performance. In this study, we attempted to mitigate these problems with three novel accretions. First, we propose a novel convolutional module as the first attempt to utilize neighborhood similarity of CT images for denoising tasks. Our proposed module assisted in boosting the denoising by a significant margin. Next, we moved towards the problem of non-stationarity of CT noise and introduced a new noise aware mean square error loss for LDCT denoising. The loss mentioned above also assisted to alleviate the laborious effort required while training CT denoising network using image patches. Lastly, we propose a novel discriminator function for CT denoising tasks. The conventional vanilla discriminator tends to overlook the fine structural details and focus on the global agreement. Our proposed discriminator leverage self-attention and pixel-wise GANs for restoring the diagnostic quality of LDCT images. Our method validated on a publicly available dataset of the 2016 NIH-AAPM-Mayo Clinic Low Dose CT Grand Challenge performed remarkably better than the existing state of the art method. The corresponding source code is available at: https://github.com/reach2sbera/ldct_nonlocal.
Collapse
|
56
|
Shi L, Lu Y, Dvornek N, Weyman CA, Miller EJ, Sinusas AJ, Liu C. Automatic Inter-Frame Patient Motion Correction for Dynamic Cardiac PET Using Deep Learning. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:3293-3304. [PMID: 34018932 PMCID: PMC8670362 DOI: 10.1109/tmi.2021.3082578] [Citation(s) in RCA: 19] [Impact Index Per Article: 4.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/30/2023]
Abstract
Patient motion during dynamic PET imaging can induce errors in myocardial blood flow (MBF) estimation. Motion correction for dynamic cardiac PET is challenging because the rapid tracer kinetics of 82Rb leads to substantial tracer distribution change across different dynamic frames over time, which can cause difficulties for image registration-based motion correction, particularly for early dynamic frames. In this paper, we developed an automatic deep learning-based motion correction (DeepMC) method for dynamic cardiac PET. In this study we focused on the detection and correction of inter-frame rigid translational motion caused by voluntary body movement and pattern change of respiratory motion. A bidirectional-3D LSTM network was developed to fully utilize both local and nonlocal temporal information in the 4D dynamic image data for motion detection. The network was trained and evaluated over motion-free patient scans with simulated motion so that the motion ground-truths are available, where one million samples based on 65 patient scans were used in training, and 600 samples based on 20 patient scans were used in evaluation. The proposed method was also evaluated using additional 10 patient datasets with real motion. We demonstrated that the proposed DeepMC obtained superior performance compared to conventional registration-based methods and other convolutional neural networks (CNN), in terms of motion estimation and MBF quantification accuracy. Once trained, DeepMC is much faster than the registration-based methods and can be easily integrated into the clinical workflow. In the future work, additional investigation is needed to evaluate this approach in a clinical context with realistic patient motion.
Collapse
|
57
|
Dong S, Hangel G, Bogner W, Trattnig S, Rossler K, Widhalm G, De Feyter HM, De Graaf RA, Duncan JS. High-Resolution Magnetic Resonance Spectroscopic Imaging using a Multi-Encoder Attention U-Net with Structural and Adversarial Loss. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2021; 2021:2891-2895. [PMID: 34891851 DOI: 10.1109/embc46164.2021.9630146] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Common to most medical imaging techniques, the spatial resolution of Magnetic Resonance Spectroscopic Imaging (MRSI) is ultimately limited by the achievable SNR. This work presents a deep learning method for 1H-MRSI spatial resolution enhancement, based on the observation that multi-parametric MRI images provide relevant spatial priors for MRSI enhancement. A Multi-encoder Attention U-Net (MAU-Net) architecture was constructed to process a MRSI metabolic map and three different MRI modalities through separate encoding paths. Spatial attention modules were incorporated to automatically learn spatial weights that highlight salient features for each MRI modality. MAU-Net was trained based on in vivo brain imaging data from patients with high-grade gliomas, using a combined loss function consisting of pixel, structural and adversarial loss. Experimental results showed that the proposed method is able to reconstruct high-quality metabolic maps with a high-resolution of 64×64 from a low-resolution of 16 × 16, with better performance compared to several baseline methods.
Collapse
|
58
|
Bai T, Wang B, Nguyen D, Wang B, Dong B, Cong W, Kalra MK, Jiang S. Deep Interactive Denoiser (DID) for X-Ray Computed Tomography. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:2965-2975. [PMID: 34329156 DOI: 10.1109/tmi.2021.3101241] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/13/2023]
Abstract
Low-dose computed tomography (LDCT) is desirable for both diagnostic imaging and image-guided interventions. Denoisers are widely used to improve the quality of LDCT. Deep learning (DL)-based denoisers have shown state-of-the-art performance and are becoming mainstream methods. However, there are two challenges to using DL-based denoisers: 1) a trained model typically does not generate different image candidates with different noise-resolution tradeoffs, which are sometimes needed for different clinical tasks; and 2) the model's generalizability might be an issue when the noise level in the testing images differs from that in the training dataset. To address these two challenges, in this work, we introduce a lightweight optimization process that can run on top of any existing DL-based denoiser during the testing phase to generate multiple image candidates with different noise-resolution tradeoffs suitable for different clinical tasks in real time. Consequently, our method allows users to interact with the denoiser to efficiently review various image candidates and quickly pick the desired one; thus, we termed this method deep interactive denoiser (DID). Experimental results demonstrated that DID can deliver multiple image candidates with different noise-resolution tradeoffs and shows great generalizability across various network architectures, as well as training and testing datasets with various noise levels.
Collapse
|
59
|
Lin H, Li Z, Yang Z, Wang Y. Variance-aware attention U-Net for multi-organ segmentation. Med Phys 2021; 48:7864-7876. [PMID: 34716711 DOI: 10.1002/mp.15322] [Citation(s) in RCA: 16] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2021] [Revised: 10/06/2021] [Accepted: 10/23/2021] [Indexed: 01/20/2023] Open
Abstract
PURPOSE With the continuous development of deep learning based medical image segmentation technology, it is expected to attain more robust and accurate performance for more challenging tasks, such as multi-organs, small/irregular areas, and ambiguous boundary issues. METHODS We propose a variance-aware attention U-Net to solve the problem of multi-organ segmentation. Specifically, a simple yet effective variance-based uncertainty mechanism is devised to evaluate the discrimination of each voxel via its prediction probability. The proposed variance uncertainty is further embedded into an attention architecture, which not only aggregates multi-level deep features in a global-level but also enforces the network to pay extra attention to voxels with uncertain predictions during training. RESULTS Extensive experiments on challenging abdominal multi-organ CT dataset show that our proposed method consistently outperforms cutting-edge attention networks with respect to the evaluation metrics of Dice index (DSC), 95% Hausdorff distance (95HD) and average symmetric surface distance (ASSD). CONCLUSIONS The proposed network provides an accurate and robust solution for multi-organ segmentation and has the potential to be used for improving other segmentation applications.
Collapse
Affiliation(s)
- Haoneng Lin
- National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, Guangdong Key Laboratory for Biomedical Measurements and Ultrasound Imaging, School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, China.,Smart Medical Imaging, Learning and Engineering (SMILE) Lab, Shenzhen University, Shenzhen, China
| | - Zongshang Li
- National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, Guangdong Key Laboratory for Biomedical Measurements and Ultrasound Imaging, School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, China.,Smart Medical Imaging, Learning and Engineering (SMILE) Lab, Shenzhen University, Shenzhen, China
| | - Zefan Yang
- National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, Guangdong Key Laboratory for Biomedical Measurements and Ultrasound Imaging, School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, China.,Smart Medical Imaging, Learning and Engineering (SMILE) Lab, Shenzhen University, Shenzhen, China
| | - Yi Wang
- National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, Guangdong Key Laboratory for Biomedical Measurements and Ultrasound Imaging, School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, China.,Smart Medical Imaging, Learning and Engineering (SMILE) Lab, Shenzhen University, Shenzhen, China.,Medical UltraSound Image Computing (MUSIC) Lab, Shenzhen University, Shenzhen, China.,Marshall Laboratory of Biomedical Engineering, Shenzhen University, Shenzhen, China
| |
Collapse
|
60
|
Chen K, Long K, Ren Y, Sun J, Pu X. Lesion-Inspired Denoising Network: Connecting Medical Image Denoising and Lesion Detection. PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA 2021. [DOI: 10.1145/3474085.3475480] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/08/2023]
Affiliation(s)
- Kecheng Chen
- University of Electronic Science and Technology of China, Chengdu, China
| | - Kun Long
- University of Electronic Science and Technology of China, Chengdu, China
| | - Yazhou Ren
- University of Electronic Science and Technology of China, Chengdu, China
| | - Jiayu Sun
- West China Hospital of SiChuan University, Chengdu, China
| | - Xiaorong Pu
- University of Electronic Science and Technology of China, Chengdu, China
| |
Collapse
|
61
|
The feasibility of deep learning-based synthetic contrast-enhanced CT from nonenhanced CT in emergency department patients with acute abdominal pain. Sci Rep 2021; 11:20390. [PMID: 34650183 PMCID: PMC8516935 DOI: 10.1038/s41598-021-99896-4] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/29/2021] [Accepted: 09/30/2021] [Indexed: 12/04/2022] Open
Abstract
Our objective was to investigate the feasibility of deep learning-based synthetic contrast-enhanced CT (DL-SCE-CT) from nonenhanced CT (NECT) in patients who visited the emergency department (ED) with acute abdominal pain (AAP). We trained an algorithm generating DL-SCE-CT using NECT with paired precontrast/postcontrast images. For clinical application, 353 patients from three institutions who visited the ED with AAP were included. Six reviewers (experienced radiologists, ER1-3; training radiologists, TR1-3) made diagnostic and disposition decisions using NECT alone and then with NECT and DL-SCE-CT together. The radiologists’ confidence in decisions was graded using a 5-point scale. The diagnostic accuracy using DL-SCE-CT improved in three radiologists (50%, P = 0.023, 0.012, < 0.001, especially in 2/3 of TRs). The confidence of diagnosis and disposition improved significantly in five radiologists (83.3%, P < 0.001). Particularly, in subgroups with underlying malignancy and miscellaneous medical conditions (MMCs) and in CT-negative cases, more radiologists reported increased confidence in diagnosis (83.3% [5/6], 100.0% [6/6], and 83.3% [5/6], respectively) and disposition (66.7% [4/6], 83.3% [5/6] and 100% [6/6], respectively). In conclusion, DL-SCE-CT enhances the accuracy and confidence of diagnosis and disposition regarding patients with AAP in the ED, especially for less experienced radiologists, in CT-negative cases, and in certain disease subgroups with underlying malignancy and MMCs.
Collapse
|
62
|
Remedios LW, Lingam S, Remedios SW, Gao R, Clark SW, Davis LT, Landman BA. Comparison of convolutional neural networks for detecting large vessel occlusion on computed tomography angiography. Med Phys 2021; 48:6060-6068. [PMID: 34287944 PMCID: PMC8568625 DOI: 10.1002/mp.15122] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/02/2021] [Revised: 07/04/2021] [Accepted: 07/05/2021] [Indexed: 12/19/2022] Open
Abstract
PURPOSE Artificial intelligence diagnosis and triage of large vessel occlusion may quicken clinical response for a subset of time-sensitive acute ischemic stroke patients, improving outcomes. Differences in architectural elements within data-driven convolutional neural network (CNN) models impact performance. Foreknowledge of effective model architectural elements for domain-specific problems can narrow the search for candidate models and inform strategic model design and adaptation to optimize performance on available data. Here, we study CNN architectures with a range of learnable parameters and which span the inclusion of architectural elements, such as parallel processing branches and residual connections with varying methods of recombining residual information. METHODS We compare five CNNs: ResNet-50, DenseNet-121, EfficientNet-B0, PhiNet, and an Inception module-based network, on a computed tomography angiography large vessel occlusion detection task. The models were trained and preliminarily evaluated with 10-fold cross-validation on preprocessed scans (n = 240). An ablation study was performed on PhiNet due to superior cross-validated test performance across accuracy, precision, recall, specificity, and F1 score. The final evaluation of all models was performed on a withheld external validation set (n = 60) and these predictions were subsequently calibrated with sigmoid curves. RESULTS Uncalibrated results on the withheld external validation set show that DenseNet-121 had the best average performance on accuracy, precision, recall, specificity, and F1 score. After calibration DenseNet-121 maintained superior performance on all metrics except recall. CONCLUSIONS The number of learnable parameters in our five models and best-ablated PhiNet directly related to cross-validated test performance-the smaller the model the better. However, this pattern did not hold when looking at generalization on the withheld external validation set. DenseNet-121 generalized the best; we posit this was due to its heavy use of residual connections utilizing concatenation, which causes feature maps from earlier layers to be used deeper in the network, while aiding in gradient flow and regularization.
Collapse
Affiliation(s)
- Lucas W. Remedios
- Department of Computer Science, Vanderbilt University,
Nashville, TN, 37235, USA
| | - Sneha Lingam
- School of Medicine, Vanderbilt University, Nashville, TN,
37240, USA
| | - Samuel W. Remedios
- Department of Computer Science, Johns Hopkins University,
Baltimore, MD, 21218, USA
- Department of Radiology and Imaging Sciences, National
Institutes of Health, Bethesda, MD, 20892, USA
| | - Riqiang Gao
- Department of Computer Science, Vanderbilt University,
Nashville, TN, 37235, USA
| | - Stephen W. Clark
- Department of Neurology, Vanderbilt University Medical
Center, Nashville, TN, 37232, USA
| | - Larry T. Davis
- Department of Radiology and Radiological Sciences,
Vanderbilt University Medical Center, Nashville, TN, 37232, USA
- Department of Neurology, Vanderbilt University Medical
Center, Nashville, TN, 37232, USA
| | - Bennett A. Landman
- Department of Computer Science, Vanderbilt University,
Nashville, TN, 37235, USA
- Department of Electrical Engineering, Vanderbilt
University, Nashville, TN, 37235, USA
- Department of Radiology and Radiological Sciences,
Vanderbilt University Medical Center, Nashville, TN, 37232, USA
| |
Collapse
|
63
|
Kulathilake KASH, Abdullah NA, Bandara AMRR, Lai KW. InNetGAN: Inception Network-Based Generative Adversarial Network for Denoising Low-Dose Computed Tomography. JOURNAL OF HEALTHCARE ENGINEERING 2021; 2021:9975762. [PMID: 34552709 PMCID: PMC8452440 DOI: 10.1155/2021/9975762] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/07/2021] [Revised: 08/18/2021] [Accepted: 08/27/2021] [Indexed: 12/24/2022]
Abstract
Low-dose Computed Tomography (LDCT) has gained a great deal of attention in clinical procedures due to its ability to reduce the patient's risk of exposure to the X-ray radiation. However, reducing the X-ray dose increases the quantum noise and artifacts in the acquired LDCT images. As a result, it produces visually low-quality LDCT images that adversely affect the disease diagnosing and treatment planning in clinical procedures. Deep Learning (DL) has recently become the cutting-edge technology of LDCT denoising due to its high performance and data-driven execution compared to conventional denoising approaches. Although the DL-based models perform fairly well in LDCT noise reduction, some noise components are still retained in denoised LDCT images. One reason for this noise retention is the direct transmission of feature maps through the skip connections of contraction and extraction path-based DL modes. Therefore, in this study, we propose a Generative Adversarial Network with Inception network modules (InNetGAN) as a solution for filtering the noise transmission through skip connections and preserving the texture and fine structure of LDCT images. The proposed Generator is modeled based on the U-net architecture. The skip connections in the U-net architecture are modified with three different inception network modules to filter out the noise in the feature maps passing over them. The quantitative and qualitative experimental results have shown the performance of the InNetGAN model in reducing noise and preserving the subtle structures and texture details in LDCT images compared to the other state-of-the-art denoising algorithms.
Collapse
Affiliation(s)
- K. A. Saneera Hemantha Kulathilake
- Department of Computer System and Technology, Faculty of Computer Science and Information Technology, Universiti Malaya, 50603 Kuala Lumpur, Malaysia
- Department of Computing, Faculty of Applied Sciences, Rajarata University of Sri Lanka, Mihintale, Sri Lanka
| | - Nor Aniza Abdullah
- Department of Computer System and Technology, Faculty of Computer Science and Information Technology, Universiti Malaya, 50603 Kuala Lumpur, Malaysia
| | | | - Khin Wee Lai
- Department of Biomedical Engineering, Faculty of Engineering, Universiti Malaya, 50603 Kuala Lumpur, Malaysia
| |
Collapse
|
64
|
Low-Dose CT Image Denoising with Improving WGAN and Hybrid Loss Function. COMPUTATIONAL AND MATHEMATICAL METHODS IN MEDICINE 2021; 2021:2973108. [PMID: 34484414 PMCID: PMC8416402 DOI: 10.1155/2021/2973108] [Citation(s) in RCA: 12] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 04/22/2021] [Revised: 07/12/2021] [Accepted: 08/12/2021] [Indexed: 11/17/2022]
Abstract
The X-ray radiation from computed tomography (CT) brought us the potential risk. Simply decreasing the dose makes the CT images noisy and diagnostic performance compromised. Here, we develop a novel denoising low-dose CT image method. Our framework is based on an improved generative adversarial network coupling with the hybrid loss function, including the adversarial loss, perceptual loss, sharpness loss, and structural similarity loss. Among the loss function terms, perceptual loss and structural similarity loss are made use of to preserve textural details, and sharpness loss can make reconstruction images clear. The adversarial loss can sharp the boundary regions. The results of experiments show the proposed method can effectively remove noise and artifacts better than the state-of-the-art methods in the aspects of the visual effect, the quantitative measurements, and the texture details.
Collapse
|
65
|
Tian S, Wang M, Yuan F, Dai N, Sun Y, Xie W, Qin J. Efficient Computer-Aided Design of Dental Inlay Restoration: A Deep Adversarial Framework. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:2415-2427. [PMID: 33945473 DOI: 10.1109/tmi.2021.3077334] [Citation(s) in RCA: 17] [Impact Index Per Article: 4.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Restoring the normal masticatory function of broken teeth is a challenging task primarily due to the defect location and size of a patient's teeth. In recent years, although some representative image-to-image transformation methods (e.g. Pix2Pix) can be potentially applicable to restore the missing crown surface, most of them fail to generate dental inlay surface with realistic crown details (e.g. occlusal groove) that are critical to the restoration of defective teeth with varying shapes. In this article, we design a computer-aided Deep Adversarial-driven dental Inlay reStoration (DAIS) framework to automatically reconstruct a realistic surface for a defective tooth. Specifically, DAIS consists of a Wasserstein generative adversarial network (WGAN) with a specially designed loss measurement, and a new local-global discriminator mechanism. The local discriminator focuses on missing regions to ensure the local consistency of a generated occlusal surface, while the global discriminator aims at defective teeth and adjacent teeth to assess if it is coherent as a whole. Experimental results demonstrate that DAIS is highly efficient to deal with a large area of missing teeth in arbitrary shapes and generate realistic occlusal surface completion. Moreover, the designed watertight inlay prostheses have enough anatomical morphology, thus providing higher clinical applicability compared with more state-of-the-art methods.
Collapse
|
66
|
Wang J, Li W, Chen Y, Fang W, Kong W, He Y, Shi G. Weakly supervised anomaly segmentation in retinal OCT images using an adversarial learning approach. BIOMEDICAL OPTICS EXPRESS 2021; 12:4713-4729. [PMID: 34513220 PMCID: PMC8407839 DOI: 10.1364/boe.426803] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/02/2021] [Revised: 06/17/2021] [Accepted: 06/26/2021] [Indexed: 05/09/2023]
Abstract
Lesion detection is a critical component of disease diagnosis, but the manual segmentation of lesions in medical images is time-consuming and experience-demanding. These issues have recently been addressed through deep learning models. However, most of the existing algorithms were developed using supervised training, which requires time-intensive manual labeling and prevents the model from detecting unaware lesions. As such, this study proposes a weakly supervised learning network based on CycleGAN for lesions segmentation in full-width optical coherence tomography (OCT) images. The model was trained to reconstruct underlying normal anatomic structures from abnormal input images, then the lesions can be detected by calculating the difference between the input and output images. A customized network architecture and a multi-scale similarity perceptual reconstruction loss were used to extend the CycleGAN model to transfer between objects exhibiting shape deformations. The proposed technique was validated using an open-source retinal OCT image dataset. Image-level anomaly detection and pixel-level lesion detection results were assessed using area-under-curve (AUC) and the Dice similarity coefficient, producing results of 96.94% and 0.8239, respectively, higher than all comparative methods. The average test time required to generate a single full-width image was 0.039 s, which is shorter than that reported in recent studies. These results indicate that our model can accurately detect and segment retinopathy lesions in real-time, without the need for supervised labeling. And we hope this method will be helpful to accelerate the clinical diagnosis process and reduce the misdiagnosis rate.
Collapse
Affiliation(s)
- Jing Wang
- Jiangsu Key Laboratory of Medical Optics, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Science, Suzhou 215163, China
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei 230026, China
| | - Wanyue Li
- Jiangsu Key Laboratory of Medical Optics, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Science, Suzhou 215163, China
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei 230026, China
| | - Yiwei Chen
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei 230026, China
| | - Wangyi Fang
- Department of Ophthalmology and Vision Science, Eye and ENT Hospital, Fudan University, Shanghai 201112, China
- Key Laboratory of Myopia of State Health Ministry, and Key Laboratory of Visual Impairment and Restoration of Shanghai, Shanghai 200003, China
| | - Wen Kong
- Jiangsu Key Laboratory of Medical Optics, Suzhou Institute of Biomedical Engineering and Technology, Chinese Academy of Science, Suzhou 215163, China
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei 230026, China
| | - Yi He
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei 230026, China
| | - Guohua Shi
- School of Biomedical Engineering (Suzhou), Division of Life Sciences and Medicine, University of Science and Technology of China, Hefei 230026, China
- Center for Excellence in Brain Science and Intelligence Technology, Chinese Academy of Science, Shanghai 200031, China
| |
Collapse
|
67
|
Li W, Chen J, Chen P, Yu L, Cui X, Li Y, Cheng F, Ouyang W. NIA-Network: Towards improving lung CT infection detection for COVID-19 diagnosis. Artif Intell Med 2021; 117:102082. [PMID: 34127245 PMCID: PMC8153959 DOI: 10.1016/j.artmed.2021.102082] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/20/2020] [Revised: 04/12/2021] [Accepted: 04/26/2021] [Indexed: 01/08/2023]
Abstract
During pandemics (e.g., COVID-19) physicians have to focus on diagnosing and treating patients, which often results in that only a limited amount of labeled CT images is available. Although recent semi-supervised learning algorithms may alleviate the problem of annotation scarcity, limited real-world CT images still cause those algorithms producing inaccurate detection results, especially in real-world COVID-19 cases. Existing models often cannot detect the small infected regions in COVID-19 CT images, such a challenge implicitly causes that many patients with minor symptoms are misdiagnosed and develop more severe symptoms, causing a higher mortality. In this paper, we propose a new method to address this challenge. Not only can we detect severe cases, but also detect minor symptoms using real-world COVID-19 CT images in which the source domain only includes limited labeled CT images but the target domain has a lot of unlabeled CT images. Specifically, we adopt Network-in-Network and Instance Normalization to build a new module (we term it NI module) and extract discriminative representations from CT images from both source and target domains. A domain classifier is utilized to implement infected region adaptation from source domain to target domain in an Adversarial Learning manner, and learns domain-invariant region proposal network (RPN) in the Faster R-CNN model. We call our model NIA-Network (Network-in-Network, Instance Normalization and Adversarial Learning), and conduct extensive experiments on two COVID-19 datasets to validate our approach. The experimental results show that our model can effectively detect infected regions with different sizes and achieve the highest diagnostic accuracy compared with existing SOTA methods.
Collapse
Affiliation(s)
- Wei Li
- School of Artificial Intelligence and Computer Science, Jiangnan University, Wuxi, Jiangsu, PR China; Jiangsu Key Laboratory of Media Design and Software Technology, Wuxi, Jiangsu, PR China; Science Center for Future Foods, Jiangnan University, Wuxi, Jiangsu, PR China
| | - Jinlin Chen
- Department of Computing, The Hong Kong Polytechnic University, Hong Kong, China
| | - Ping Chen
- Department of Engineering, University of Massachusetts, Boston, USA.
| | - Lequan Yu
- Department of Statistics and Actuarial Science, The University of Hong Kong, Hong Kong, China
| | - Xiaohui Cui
- School of Cyber Science and Engineering, Wuhan University, Wuhan, Hubei, PR China
| | - Yiwei Li
- School of Computer Science and Information Engineering, Hefei University of Technology, Hefei, PR China
| | - Fang Cheng
- Department of Cancer Center, Union Hospital of Tongji Medical College of Huazhong University of Science and Technology, Wuhan, PR China
| | - Wen Ouyang
- Department of Radiation and Medical Oncology, Zhongnan Hospital, Wuhan University, Wuhan, PR China
| |
Collapse
|
68
|
Lyu Q, Xia D, Liu Y, Yang X, Li R. Pyramidal convolution attention generative adversarial network with data augmentation for image denoising. Soft comput 2021. [DOI: 10.1007/s00500-021-05870-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
69
|
Jolivet F, Lesaint J, Fournier C, Garcin M, Brambilla A. An Efficient One-Step Method for Spectral CT Based on an Approximate Linear Model. IEEE TRANSACTIONS ON RADIATION AND PLASMA MEDICAL SCIENCES 2021. [DOI: 10.1109/trpms.2020.3015598] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
|
70
|
Mühlberg A, Kärgel R, Katzmann A, Durlak F, Allard PE, Faivre JB, Sühling M, Rémy-Jardin M, Taubmann O. Unraveling the interplay of image formation, data representation and learning in CT-based COPD phenotyping automation: The need for a meta-strategy. Med Phys 2021; 48:5179-5191. [PMID: 34129688 DOI: 10.1002/mp.15049] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/23/2020] [Revised: 04/20/2021] [Accepted: 06/01/2021] [Indexed: 12/19/2022] Open
Abstract
PURPOSE In the literature on automated phenotyping of chronic obstructive pulmonary disease (COPD), there is a multitude of isolated classical machine learning and deep learning techniques, mostly investigating individual phenotypes, with small study cohorts and heterogeneous meta-parameters, e.g., different scan protocols or segmented regions. The objective is to compare the impact of different experimental setups, i.e., varying meta-parameters related to image formation and data representation, with the impact of the learning technique for subtyping automation for a variety of phenotypes. The identified associations of these parameters with automation performance and their interactions might be a first step towards a determination of optimal meta-parameters, i.e., a meta-strategy. METHODS A clinical cohort of 981 patients (53.8 ± 15.1 years, 554 male) was examined. The inspiratory CT images were analyzed to automate the diagnosis of 13 COPD phenotypes given by two radiologists. A benchmark feature set that integrates many quantitative criteria was extracted from the lung and trained a variety of learning algorithms on the first 654 patients (two thirds) and the respective algorithm retrospectively assessed the remaining 327 patients (one third). The automation performance was evaluated by the area under the receiver operating characteristic curve (AUC). 1717 experiments were conducted with varying meta-parameters such as reconstruction kernel, segmented regions and input dimensionality, i.e., number of extracted features. The association of the meta-parameters with the automation performance was analyzed by multivariable general linear model decomposition of the automation performance in the contributions of meta-parameters and the learning technique. RESULTS The automation performance varied strongly for varying meta-parameters. For emphysema-predominant phenotypes, an AUC of 93%-95% could be achieved for the best meta-configuration. The airways-predominant phenotypes led to a lower performance of 65%-85%, while smooth kernel configurations on average were unexpectedly superior to those with sharp kernels. The performance impact of meta-parameters, even that of often neglected ones like the missing-data imputation, was in general larger than that of the learning technique. Advanced learning techniques like 3D deep learning or automated machine learning yielded inferior automation performance for non-optimal meta-configurations in comparison to simple techniques with suitable meta-configurations. The best automation performance was achieved by a combination of modern learning techniques and a suitable meta-configuration. CONCLUSIONS Our results indicate that for COPD phenotype automation, study design parameters such as reconstruction kernel and the model input dimensionality should be adapted to the learning technique and may be more important than the technique itself. To achieve optimal automation and prediction results, the interaction between input those meta-parameters and the learning technique should be considered. This might be particularly relevant for the development of specific scan protocols for novel learning algorithms, and towards an understanding of good study design for automated phenotyping.
Collapse
Affiliation(s)
| | - Rainer Kärgel
- CT R&D Image Analytics, Siemens Healthineers, Forchheim, Germany
| | | | - Felix Durlak
- CT R&D Image Analytics, Siemens Healthineers, Forchheim, Germany
| | | | | | - Michael Sühling
- CT R&D Image Analytics, Siemens Healthineers, Forchheim, Germany
| | | | - Oliver Taubmann
- CT R&D Image Analytics, Siemens Healthineers, Forchheim, Germany
| |
Collapse
|
71
|
Kulathilake KASH, Abdullah NA, Sabri AQM, Lai KW. A review on Deep Learning approaches for low-dose Computed Tomography restoration. COMPLEX INTELL SYST 2021; 9:2713-2745. [PMID: 34777967 PMCID: PMC8164834 DOI: 10.1007/s40747-021-00405-x] [Citation(s) in RCA: 41] [Impact Index Per Article: 10.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/25/2020] [Accepted: 05/18/2021] [Indexed: 02/08/2023]
Abstract
Computed Tomography (CT) is a widely use medical image modality in clinical medicine, because it produces excellent visualizations of fine structural details of the human body. In clinical procedures, it is desirable to acquire CT scans by minimizing the X-ray flux to prevent patients from being exposed to high radiation. However, these Low-Dose CT (LDCT) scanning protocols compromise the signal-to-noise ratio of the CT images because of noise and artifacts over the image space. Thus, various restoration methods have been published over the past 3 decades to produce high-quality CT images from these LDCT images. More recently, as opposed to conventional LDCT restoration methods, Deep Learning (DL)-based LDCT restoration approaches have been rather common due to their characteristics of being data-driven, high-performance, and fast execution. Thus, this study aims to elaborate on the role of DL techniques in LDCT restoration and critically review the applications of DL-based approaches for LDCT restoration. To achieve this aim, different aspects of DL-based LDCT restoration applications were analyzed. These include DL architectures, performance gains, functional requirements, and the diversity of objective functions. The outcome of the study highlights the existing limitations and future directions for DL-based LDCT restoration. To the best of our knowledge, there have been no previous reviews, which specifically address this topic.
Collapse
Affiliation(s)
- K. A. Saneera Hemantha Kulathilake
- Department of Computer System and Technology, Faculty of Computer Science and Information Technology, Universiti Malaya, 50603 Kuala Lumpur, Malaysia
| | - Nor Aniza Abdullah
- Department of Computer System and Technology, Faculty of Computer Science and Information Technology, Universiti Malaya, 50603 Kuala Lumpur, Malaysia
| | - Aznul Qalid Md Sabri
- Department of Artificial Intelligence, Faculty of Computer Science and Information Technology, Universiti Malaya, 50603 Kuala Lumpur, Malaysia
| | - Khin Wee Lai
- Department of Biomedical Engineering, Faculty of Engineering, Universiti Malaya, 50603 Kuala Lumpur, Malaysia
| |
Collapse
|
72
|
Shi Z, Li H, Cao Q, Wang Z, Cheng M. A material decomposition method for dual-energy CT via dual interactive Wasserstein generative adversarial networks. Med Phys 2021; 48:2891-2905. [PMID: 33704786 DOI: 10.1002/mp.14828] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/21/2020] [Revised: 02/26/2021] [Accepted: 02/28/2021] [Indexed: 11/09/2022] Open
Abstract
PURPOSE Dual-energy computed tomography (DECT) is highly promising for material characterization and identification, whereas reconstructed material-specific images are affected by magnified noise and beam-hardening artifacts. Although various DECT material decomposition methods have been proposed to solve this problem, the quality of the decomposed images is still unsatisfactory, particularly in the image edges. In this study, a data-driven approach using dual interactive Wasserstein generative adversarial networks (DIWGAN) is developed to improve DECT decomposition accuracy and perform edge-preserving images. METHODS In proposed DIWGAN, two interactive generators are used to synthesize decomposed images of two basis materials by modeling the spatial and spectral correlations from input DECT reconstructed images, and the corresponding discriminators are employed to distinguish the difference between the generated images and labels. The DECT images reconstructed from high- and low-energy bins are sent to two generators separately, and each generator synthesizes one material-specific image, thereby ensuring the specificity of the network modeling. In addition, the information from different energy bins is exploited through the feature sharing of two generators. During decomposition model training, a hybrid loss function including L1 loss, edge loss, and adversarial loss is incorporated to preserve the texture and edges in the generated images. Additionally, a selector is employed to define the generator that should be trained in each iteration, which can ensure the modeling ability of two different generators and improve the material decomposition accuracy. The performance of the proposed method is evaluated using digital phantom, XCAT phantom, and real data from a mouse. RESULTS On the digital phantom, the regions of bone and soft tissue are strictly and accurately separated using the trained decomposition model. The material densities in different bone and soft-tissue regions are near the ground truth, and the error of material densities is lower than 3 mg/ml. The results from XCAT phantom show that the material-specific images generated by directed matrix inversion and iterative decomposition methods have severe noise and artifacts. Regarding to the learning-based methods, the decomposed images of fully convolutional network (FCN) and butterfly network (Butterfly-Net) still contain varying degrees of artifacts, while proposed DIWGAN can yield high quality images. Compared to Butterfly-Net, the root-mean-square error (RMSE) of soft-tissue images generated by the DIWGAN decreased by 0.01 g/ml, whereas the peak-signal-to-noise ratio (PSNR) and structural similarity (SSIM) of the soft-tissue images reached 31.43 dB and 0.9987, respectively. The mass densities of the decomposed materials are nearest to the ground truth when using the DIWGAN method. The noise standard deviation of the decomposition images reduced by 69%, 60%, 33%, and 21% compared with direct matrix inversion, iterative decomposition, FCN, and Butterfly-Net, respectively. Furthermore, the performance of the mouse data indicates the potential of the proposed material decomposition method in real scanned data. CONCLUSIONS A DECT material decomposition method based on deep learning is proposed, and the relationship between reconstructed and material-specific images is mapped by training the DIWGAN model. Results from both the simulation phantoms and real data demonstrate the advantages of this method in suppressing noise and beam-hardening artifacts.
Collapse
Affiliation(s)
- Zaifeng Shi
- School of Microelectronics, Tianjin University, Tianjin, 300072, China.,Tianjin Key Laboratory of Imaging and Sensing Microelectronic Technology, Tianjin, 300072, China
| | - Huilong Li
- School of Microelectronics, Tianjin University, Tianjin, 300072, China
| | - Qingjie Cao
- School of Mathematical Sciences, Tianjin Normal University, Tianjin, 300072, China
| | - Zhongqi Wang
- School of Microelectronics, Tianjin University, Tianjin, 300072, China
| | - Ming Cheng
- School of Microelectronics, Tianjin University, Tianjin, 300072, China
| |
Collapse
|
73
|
Xue Y, Qin W, Luo C, Yang P, Jiang Y, Tsui T, He H, Wang L, Qin J, Xie Y, Niu T. Multi-Material Decomposition for Single Energy CT Using Material Sparsity Constraint. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:1303-1318. [PMID: 33460369 DOI: 10.1109/tmi.2021.3051416] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Multi-material decomposition (MMD) decomposes CT images into basis material images, and is a promising technique in clinical diagnostic CT to identify material compositions within the human body. MMD could be implemented on measurements obtained from spectral CT protocol, although spectral CT data acquisition is not readily available in most clinical environments. MMD methods using single energy CT (SECT), broadly applied in radiological departments of most hospitals, have been proposed in the literature while challenged by the inferior decomposition accuracy and the limited number of material bases due to the constrained material information in the SECT measurement. In this paper, we propose an image-domain SECT MMD method using material sparsity as an assistance under the condition that each voxel of the CT image contains at most two different elemental materials. L0 norm represents the material sparsity constraint (MSC) and is integrated into the decomposition objective function with a least-square data fidelity term, total variation term, and a sum-to-one constraint of material volume fractions. An accelerated primal-dual (APD) algorithm with line-search scheme is applied to solve the problem. The pixelwise direct inversion method with the two-material assumption (TMA) is applied to estimate the initials. We validate the proposed method on phantom and patient data. Compared with the TMA method, the proposed MSC method increases the volume fraction accuracy (VFA) from 92.0% to 98.5% in the phantom study. In the patient study, the calcification area can be clearly visualized in the virtual non-contrast image generated by the proposed method, and has a similar shape to that in the ground-truth contrast-free CT image. The high decomposition image quality from the proposed method substantially facilitates the SECT-based MMD clinical applications.
Collapse
|
74
|
Shin Y, Yang J, Lee YH. Deep Generative Adversarial Networks: Applications in Musculoskeletal Imaging. Radiol Artif Intell 2021; 3:e200157. [PMID: 34136816 PMCID: PMC8204145 DOI: 10.1148/ryai.2021200157] [Citation(s) in RCA: 14] [Impact Index Per Article: 3.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/02/2020] [Revised: 02/10/2021] [Accepted: 02/16/2021] [Indexed: 12/12/2022]
Abstract
In recent years, deep learning techniques have been applied in musculoskeletal radiology to increase the diagnostic potential of acquired images. Generative adversarial networks (GANs), which are deep neural networks that can generate or transform images, have the potential to aid in faster imaging by generating images with a high level of realism across multiple contrast and modalities from existing imaging protocols. This review introduces the key architectures of GANs as well as their technical background and challenges. Key research trends are highlighted, including: (a) reconstruction of high-resolution MRI; (b) image synthesis with different modalities and contrasts; (c) image enhancement that efficiently preserves high-frequency information suitable for human interpretation; (d) pixel-level segmentation with annotation sharing between domains; and (e) applications to different musculoskeletal anatomies. In addition, an overview is provided of the key issues wherein clinical applicability is challenging to capture with conventional performance metrics and expert evaluation. When clinically validated, GANs have the potential to improve musculoskeletal imaging. Keywords: Adults and Pediatrics, Computer Aided Diagnosis (CAD), Computer Applications-General (Informatics), Informatics, Skeletal-Appendicular, Skeletal-Axial, Soft Tissues/Skin © RSNA, 2021.
Collapse
Affiliation(s)
- YiRang Shin
- From the Department of Radiology, Research Institute of Radiological Science, and Center for Clinical Imaging Data Science (CCIDS), Yonsei University College of Medicine, 250 Seongsanno, Seodaemun-gu, Seoul 220-701, Republic of Korea (Y.S., J.Y., Y.H.L.); Systems Molecular Radiology at Yonsei (SysMolRaY), Seoul, Republic of Korea (J.Y.); and Severance Biomedical Science Institute (SBSI), Yonsei University College of Medicine, Seoul, Republic of Korea (J.Y.)
| | - Jaemoon Yang
- From the Department of Radiology, Research Institute of Radiological Science, and Center for Clinical Imaging Data Science (CCIDS), Yonsei University College of Medicine, 250 Seongsanno, Seodaemun-gu, Seoul 220-701, Republic of Korea (Y.S., J.Y., Y.H.L.); Systems Molecular Radiology at Yonsei (SysMolRaY), Seoul, Republic of Korea (J.Y.); and Severance Biomedical Science Institute (SBSI), Yonsei University College of Medicine, Seoul, Republic of Korea (J.Y.)
| | - Young Han Lee
- From the Department of Radiology, Research Institute of Radiological Science, and Center for Clinical Imaging Data Science (CCIDS), Yonsei University College of Medicine, 250 Seongsanno, Seodaemun-gu, Seoul 220-701, Republic of Korea (Y.S., J.Y., Y.H.L.); Systems Molecular Radiology at Yonsei (SysMolRaY), Seoul, Republic of Korea (J.Y.); and Severance Biomedical Science Institute (SBSI), Yonsei University College of Medicine, Seoul, Republic of Korea (J.Y.)
| |
Collapse
|
75
|
Park HY, Bae HJ, Hong GS, Kim M, Yun J, Park S, Chung WJ, Kim N. Realistic High-Resolution Body Computed Tomography Image Synthesis by Using Progressive Growing Generative Adversarial Network: Visual Turing Test. JMIR Med Inform 2021; 9:e23328. [PMID: 33609339 PMCID: PMC8077702 DOI: 10.2196/23328] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2020] [Revised: 11/15/2020] [Accepted: 02/20/2021] [Indexed: 12/14/2022] Open
Abstract
BACKGROUND Generative adversarial network (GAN)-based synthetic images can be viable solutions to current supervised deep learning challenges. However, generating highly realistic images is a prerequisite for these approaches. OBJECTIVE The aim of this study was to investigate and validate the unsupervised synthesis of highly realistic body computed tomography (CT) images by using a progressive growing GAN (PGGAN) trained to learn the probability distribution of normal data. METHODS We trained the PGGAN by using 11,755 body CT scans. Ten radiologists (4 radiologists with <5 years of experience [Group I], 4 radiologists with 5-10 years of experience [Group II], and 2 radiologists with >10 years of experience [Group III]) evaluated the results in a binary approach by using an independent validation set of 300 images (150 real and 150 synthetic) to judge the authenticity of each image. RESULTS The mean accuracy of the 10 readers in the entire image set was higher than random guessing (1781/3000, 59.4% vs 1500/3000, 50.0%, respectively; P<.001). However, in terms of identifying synthetic images as fake, there was no significant difference in the specificity between the visual Turing test and random guessing (779/1500, 51.9% vs 750/1500, 50.0%, respectively; P=.29). The accuracy between the 3 reader groups with different experience levels was not significantly different (Group I, 696/1200, 58.0%; Group II, 726/1200, 60.5%; and Group III, 359/600, 59.8%; P=.36). Interreader agreements were poor (κ=0.11) for the entire image set. In subgroup analysis, the discrepancies between real and synthetic CT images occurred mainly in the thoracoabdominal junction and in the anatomical details. CONCLUSIONS The GAN can synthesize highly realistic high-resolution body CT images that are indistinguishable from real images; however, it has limitations in generating body images of the thoracoabdominal junction and lacks accuracy in the anatomical details.
Collapse
Affiliation(s)
- Ho Young Park
- Department of Radiology and Research Institute of Radiology, University of Ulsan College of Medicine & Asan Medical Center, Seoul, Republic of Korea
| | - Hyun-Jin Bae
- Department of Medicine, University of Ulsan College of Medicine & Asan Medical Center, Seoul, Republic of Korea
| | - Gil-Sun Hong
- Department of Radiology and Research Institute of Radiology, University of Ulsan College of Medicine & Asan Medical Center, Seoul, Republic of Korea
| | - Minjee Kim
- Department of Biomedical Engineering, Asan Medical Institute of Convergence Science and Technology, Asan Medical Center, University of Ulsan College of Medicine, Seoul, Republic of Korea, Seoul, Republic of Korea
| | - JiHye Yun
- Department of Radiology and Research Institute of Radiology, University of Ulsan College of Medicine & Asan Medical Center, Seoul, Republic of Korea
| | - Sungwon Park
- Department of Health Screening and Promotion Center, University of Ulsan College of Medicine & Asan Medical Center, Seoul, Republic of Korea
| | - Won Jung Chung
- Department of Health Screening and Promotion Center, University of Ulsan College of Medicine & Asan Medical Center, Seoul, Republic of Korea
| | - NamKug Kim
- Department of Radiology and Research Institute of Radiology, University of Ulsan College of Medicine & Asan Medical Center, Seoul, Republic of Korea
- Department of Convergence Medicine, University of Ulsan College of Medicine & Asan Medical Center, Seoul, Republic of Korea
| |
Collapse
|
76
|
Bai T, Wang B, Nguyen D, Jiang S. Probabilistic self-learning framework for low-dose CT denoising. Med Phys 2021; 48:2258-2270. [PMID: 33621348 DOI: 10.1002/mp.14796] [Citation(s) in RCA: 15] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/21/2020] [Revised: 02/15/2021] [Accepted: 02/16/2021] [Indexed: 01/03/2023] Open
Abstract
PURPOSE Despite the indispensable role of x-ray computed tomography (CT) in diagnostic medicine, the associated harmful ionizing radiation dose is a major concern, as it may cause genetic diseases and cancer. Decreasing patients' exposure can reduce the radiation dose and hence the related risks, but it would inevitably induce higher quantum noise. Supervised deep learning techniques have been used to train deep neural networks for denoising low-dose CT (LDCT) images, but the success of such strategies requires massive sets of pixel-level paired LDCT and normal-dose CT (NDCT) images, which are rarely available in real clinical practice. Our purpose is to mitigate the data scarcity problem for deep learning-based LDCT denoising. METHODS To solve this problem, we devised a shift-invariant property-based neural network that uses only the LDCT images to characterize both the inherent pixel correlations and the noise distribution, shaping into our probabilistic self-learning (PSL) framework. The AAPM Low-dose CT Challenge dataset was used to train the network. Both simulated datasets and real dataset were employed to test the denoising performance as well as the model generalizability. The performance was compared to a conventional method (total variation (TV)-based), a popular self-learning method (noise2void (N2V)), and a well-known unsupervised learning method (CycleGAN) by using both qualitative visual inspection and quantitative metrics including peak signal-noise-ratio (PSNR), structural similarity index (SSIM) and contrast-to-noise-ratio (CNR). The standard deviations (STD) of selected flat regions were also calculated for comparison. RESULTS The PSL method can improve the averaged PSNR/SSIM values from 27.61/0.5939 (LDCT) to 30.50/0.6797. By contrast, the averaged PSNR/SSIM values were 31.49/0.7284 (TV), 29.43/0.6699 (N2V), and 29.79/0.6992 (CycleGAN). The averaged STDs of selected flat regions were calculated to be 132.3 HU (LDCT), 25.77 HU (TV), 19.95 HU (N2V), 75.06 HU (CycleGAN), 60.62 HU (PSL) and 57.28 HU (NDCT). As for the low-contrast lesion detectability quantification, the CNR were calculated to be 0.202 (LDCT), 0.356 (TV), 0.372 (N2V), 0.383 (CycleGAN), 0.399 (PSL), and 0.359 (NDCT). By visual inspection, we observed that the proposed PSL method can deliver a noise-suppressed and detail-preserved image, while the TV-based method would lead to the blocky artifact, the N2V method would produce over-smoothed structures and CT value biased effect, and the CycleGAN method would generate slightly noisy results with inaccurate CT values. We also verified the generalizability of the PSL method, which exhibited superior denoising performance among various testing datasets with different data distribution shifts. CONCLUSIONS A deep learning-based convolutional neural network can be trained without paired datasets. Qualitatively visual inspection showed the proposed PSL method can achieve superior denoising performance than all the competitors, despite that the employed quantitative metrics in terms of PSNR, SSIM and CNR did not always show consistently better values.
Collapse
Affiliation(s)
- Ti Bai
- Medical Artificial Intelligence and Automation (MAIA) Laboratory, Department of Radiation Oncology, University of Texas Southwestern Medical Centre, Dallas, Texas, 75239, USA
| | - Biling Wang
- Medical Artificial Intelligence and Automation (MAIA) Laboratory, Department of Radiation Oncology, University of Texas Southwestern Medical Centre, Dallas, Texas, 75239, USA
| | - Dan Nguyen
- Medical Artificial Intelligence and Automation (MAIA) Laboratory, Department of Radiation Oncology, University of Texas Southwestern Medical Centre, Dallas, Texas, 75239, USA
| | - Steve Jiang
- Medical Artificial Intelligence and Automation (MAIA) Laboratory, Department of Radiation Oncology, University of Texas Southwestern Medical Centre, Dallas, Texas, 75239, USA
| |
Collapse
|
77
|
Torres-Velázquez M, Chen WJ, Li X, McMillan AB. Application and Construction of Deep Learning Networks in Medical Imaging. IEEE TRANSACTIONS ON RADIATION AND PLASMA MEDICAL SCIENCES 2021; 5:137-159. [PMID: 34017931 PMCID: PMC8132932 DOI: 10.1109/trpms.2020.3030611] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/22/2022]
Abstract
Deep learning (DL) approaches are part of the machine learning (ML) subfield concerned with the development of computational models to train artificial intelligence systems. DL models are characterized by automatically extracting high-level features from the input data to learn the relationship between matching datasets. Thus, its implementation offers an advantage over common ML methods that often require the practitioner to have some domain knowledge of the input data to select the best latent representation. As a result of this advantage, DL has been successfully applied within the medical imaging field to address problems, such as disease classification and tumor segmentation for which it is difficult or impossible to determine which image features are relevant. Therefore, taking into consideration the positive impact of DL on the medical imaging field, this article reviews the key concepts associated with its evolution and implementation. The sections of this review summarize the milestones related to the development of the DL field, followed by a description of the elements of deep neural network and an overview of its application within the medical imaging field. Subsequently, the key steps necessary to implement a supervised DL application are defined, and associated limitations are discussed.
Collapse
Affiliation(s)
- Maribel Torres-Velázquez
- Department of Biomedical Engineering, College of Engineering, University of Wisconsin-Madison, Madison, WI 53705 USA
| | - Wei-Jie Chen
- Department of Electrical and Computer Engineering, College of Engineering, University of Wisconsin-Madison, Madison, WI 53705 USA
| | - Xue Li
- Department of Radiology, University of Wisconsin School of Medicine and Public Health, University of Wisconsin-Madison, Madison, WI 53705 USA
| | - Alan B McMillan
- Department of Radiology, University of Wisconsin School of Medicine and Public Health, University of Wisconsin-Madison, Madison, WI 53705 USA, and also with the Department of Medical Physics, University of Wisconsin-Madison, Madison, WI 53705 USA
| |
Collapse
|
78
|
Zhou W, Du H, Mei W, Fang L. Efficient structurally-strengthened generative adversarial network for MRI reconstruction. Neurocomputing 2021. [DOI: 10.1016/j.neucom.2020.09.008] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/25/2022]
|
79
|
Wang T, Lei Y, Fu Y, Wynne JF, Curran WJ, Liu T, Yang X. A review on medical imaging synthesis using deep learning and its clinical applications. J Appl Clin Med Phys 2021; 22:11-36. [PMID: 33305538 PMCID: PMC7856512 DOI: 10.1002/acm2.13121] [Citation(s) in RCA: 126] [Impact Index Per Article: 31.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/13/2020] [Revised: 11/12/2020] [Accepted: 11/21/2020] [Indexed: 02/06/2023] Open
Abstract
This paper reviewed the deep learning-based studies for medical imaging synthesis and its clinical application. Specifically, we summarized the recent developments of deep learning-based methods in inter- and intra-modality image synthesis by listing and highlighting the proposed methods, study designs, and reported performances with related clinical applications on representative studies. The challenges among the reviewed studies were then summarized with discussion.
Collapse
Affiliation(s)
- Tonghe Wang
- Department of Radiation OncologyEmory UniversityAtlantaGAUSA
- Winship Cancer InstituteEmory UniversityAtlantaGAUSA
| | - Yang Lei
- Department of Radiation OncologyEmory UniversityAtlantaGAUSA
| | - Yabo Fu
- Department of Radiation OncologyEmory UniversityAtlantaGAUSA
| | - Jacob F. Wynne
- Department of Radiation OncologyEmory UniversityAtlantaGAUSA
| | - Walter J. Curran
- Department of Radiation OncologyEmory UniversityAtlantaGAUSA
- Winship Cancer InstituteEmory UniversityAtlantaGAUSA
| | - Tian Liu
- Department of Radiation OncologyEmory UniversityAtlantaGAUSA
- Winship Cancer InstituteEmory UniversityAtlantaGAUSA
| | - Xiaofeng Yang
- Department of Radiation OncologyEmory UniversityAtlantaGAUSA
- Winship Cancer InstituteEmory UniversityAtlantaGAUSA
| |
Collapse
|
80
|
Qiu B, You Y, Huang Z, Meng X, Jiang Z, Zhou C, Liu G, Yang K, Ren Q, Lu Y. N2NSR-OCT: Simultaneous denoising and super-resolution in optical coherence tomography images using semisupervised deep learning. JOURNAL OF BIOPHOTONICS 2021; 14:e202000282. [PMID: 33025760 DOI: 10.1002/jbio.202000282] [Citation(s) in RCA: 22] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/14/2020] [Revised: 09/21/2020] [Accepted: 09/29/2020] [Indexed: 06/11/2023]
Abstract
Optical coherence tomography (OCT) imaging shows a significant potential in clinical routines due to its noninvasive property. However, the quality of OCT images is generally limited by inherent speckle noise of OCT imaging and low sampling rate. To obtain high signal-to-noise ratio (SNR) and high-resolution (HR) OCT images within a short scanning time, we presented a learning-based method to recover high-quality OCT images from noisy and low-resolution OCT images. We proposed a semisupervised learning approach named N2NSR-OCT, to generate denoised and super-resolved OCT images simultaneously using up- and down-sampling networks (U-Net (Semi) and DBPN (Semi)). Additionally, two different super-resolution and denoising models with different upscale factors (2× and 4×) were trained to recover the high-quality OCT image of the corresponding down-sampling rates. The new semisupervised learning approach is able to achieve results comparable with those of supervised learning using up- and down-sampling networks, and can produce better performance than other related state-of-the-art methods in the aspects of maintaining subtle fine retinal structures.
Collapse
Affiliation(s)
- Bin Qiu
- Department of Biomedical Engineering, College of Engineering, Peking University, Beijing, China
- Institute of Biomedical Engineering, Peking University Shenzhen Graduate School, Shenzhen, China
- Institute of Biomedical Engineering, Shenzhen Bay Laboratory, Shenzhen, China
| | - Yunfei You
- Department of Biomedical Engineering, College of Engineering, Peking University, Beijing, China
- Institute of Biomedical Engineering, Peking University Shenzhen Graduate School, Shenzhen, China
- Institute of Biomedical Engineering, Shenzhen Bay Laboratory, Shenzhen, China
| | - Zhiyu Huang
- Department of Biomedical Engineering, College of Engineering, Peking University, Beijing, China
- Institute of Biomedical Engineering, Peking University Shenzhen Graduate School, Shenzhen, China
- Institute of Biomedical Engineering, Shenzhen Bay Laboratory, Shenzhen, China
| | - Xiangxi Meng
- Department of Biomedical Engineering, College of Engineering, Peking University, Beijing, China
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education), Department of Nuclear Medicine, Peking University Cancer Hospital & Institute, Beijing, China
| | - Zhe Jiang
- Department of Biomedical Engineering, College of Engineering, Peking University, Beijing, China
- Institute of Biomedical Engineering, Peking University Shenzhen Graduate School, Shenzhen, China
- Institute of Biomedical Engineering, Shenzhen Bay Laboratory, Shenzhen, China
| | - Chuanqing Zhou
- Institute of Biomedical Engineering, Shenzhen Bay Laboratory, Shenzhen, China
| | - Gangjun Liu
- Institute of Biomedical Engineering, Peking University Shenzhen Graduate School, Shenzhen, China
- Institute of Biomedical Engineering, Shenzhen Bay Laboratory, Shenzhen, China
| | - Kun Yang
- College of Quality and Technical Supervision, Hebei University, Baoding, China
| | - Qiushi Ren
- Department of Biomedical Engineering, College of Engineering, Peking University, Beijing, China
- Institute of Biomedical Engineering, Peking University Shenzhen Graduate School, Shenzhen, China
- Institute of Biomedical Engineering, Shenzhen Bay Laboratory, Shenzhen, China
| | - Yanye Lu
- Department of Biomedical Engineering, College of Engineering, Peking University, Beijing, China
- Institute of Biomedical Engineering, Peking University Shenzhen Graduate School, Shenzhen, China
- Institute of Biomedical Engineering, Shenzhen Bay Laboratory, Shenzhen, China
- Department of Computer Science, Friedrich-Alexander-University Erlangen-Nuremberg, Erlangen, Germany
| |
Collapse
|
81
|
Podgorsak AR, Shiraz Bhurwani MM, Ionita CN. CT artifact correction for sparse and truncated projection data using generative adversarial networks. Med Phys 2020; 48:615-626. [PMID: 32996149 DOI: 10.1002/mp.14504] [Citation(s) in RCA: 13] [Impact Index Per Article: 2.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/21/2020] [Revised: 09/17/2020] [Accepted: 09/18/2020] [Indexed: 12/17/2022] Open
Abstract
PURPOSE Computed tomography image reconstruction using truncated or sparsely acquired projection data to reduce radiation dose, iodine volume, and patient motion artifacts has been widely investigated. To continue these efforts, we investigated the use of machine learning-based reconstruction techniques using deep convolutional generative adversarial networks (DCGANs) and evaluated its effect using standard imaging metrics. METHODS Ten thousand head computed tomography (CT) scans were collected from the 2019 RSNA Intracranial Hemorrhage Detection and Classification Challenge dataset. Sinograms were simulated and then resampled in both a one-third truncated and one-third sparse manner. DCGANs were tasked with correcting the incomplete projection data, either in the sinogram domain where the full sinogram was recovered by the DCGAN and then reconstructed, or the reconstruction domain where the incomplete data were first reconstructed and the sparse or truncation artifacts were corrected by the DCGAN. Seventy-five hundred images were used for network training and 2500 were withheld for network assessment using mean absolute error (MAE), structural similarity index measure (SSIM), and peak signal-to-noise ratio (PSNR) between results of different correction techniques. Image data from a quality-assurance phantom were also resampled in the two manners and corrected and reconstructed for network performance assessment using line profiles across high-contrast features, the modulation transfer function (MTF), noise power spectrum (NPS), and Hounsfield Unit (HU) linearity analysis. RESULTS Better agreement with the fully sampled reconstructions were achieved from sparse acquisition corrected in the sinogram domain and the truncated acquisition corrected in the reconstruction domain. MAE, SSIM, and PSNR showed quantitative improvement from the DCGAN correction techniques. HU linearity of the reconstructions was maintained by the correction techniques for the sparse and truncated acquisitions. MTF curves reached the 10% modulation cutoff frequency at 5.86 lp/cm for the truncated corrected reconstruction compared with 2.98 lp/cm for the truncated uncorrected reconstruction, and 5.36 lp/cm for the sparse corrected reconstruction compared with around 2.91 lp/cm for the sparse uncorrected reconstruction. NPS analyses yielded better agreement across a range of frequencies between the resampled corrected phantom and truth reconstructions. CONCLUSIONS We demonstrated the use of DCGANs for CT-image correction from sparse and truncated simulated projection data, while preserving imaging quality of the fully sampled projection data.
Collapse
Affiliation(s)
- Alexander R Podgorsak
- Canon Stroke and Vascular Research Center, 875 Ellicott Street, Buffalo, NY, 14203, USA.,Medical Physics Program, State University of New York at Buffalo, 955 Main Street, Buffalo, NY, 14203, USA.,Department of Biomedical Engineering, State University of New York at Buffalo, 200 Lee Road, Buffalo, NY, 14228, USA
| | - Mohammad Mahdi Shiraz Bhurwani
- Canon Stroke and Vascular Research Center, 875 Ellicott Street, Buffalo, NY, 14203, USA.,Department of Biomedical Engineering, State University of New York at Buffalo, 200 Lee Road, Buffalo, NY, 14228, USA
| | - Ciprian N Ionita
- Canon Stroke and Vascular Research Center, 875 Ellicott Street, Buffalo, NY, 14203, USA.,Medical Physics Program, State University of New York at Buffalo, 955 Main Street, Buffalo, NY, 14203, USA.,Department of Biomedical Engineering, State University of New York at Buffalo, 200 Lee Road, Buffalo, NY, 14228, USA
| |
Collapse
|
82
|
Zhou W, Du H, Mei W, Fang L. Spatial orthogonal attention generative adversarial network for MRI reconstruction. Med Phys 2020; 48:627-639. [PMID: 33111361 DOI: 10.1002/mp.14509] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/13/2020] [Revised: 07/12/2020] [Accepted: 08/24/2020] [Indexed: 11/08/2022] Open
Abstract
PURPOSE Recent studies have witnessed that self-attention modules can better solve the vision understanding problems by capturing long-range dependencies. However, there are very few works designing a lightweight self-attention module to improve the quality of MRI reconstruction. Furthermore, it can be observed that several important self-attention modules (e.g., the non-local block) cause high computational complexity and need a huge number of GPU memory when the size of the input feature is large. The purpose of this study is to design a lightweight yet effective spatial orthogonal attention module (SOAM) to capture long-range dependencies, and develop a novel spatial orthogonal attention generative adversarial network, termed as SOGAN, to achieve more accurate MRI reconstruction. METHODS We first develop a lightweight SOAM, which can generate two small attention maps to effectively aggregate the long-range contextual information in vertical and horizontal directions, respectively. Then, we embed the proposed SOAMs into the concatenated convolutional autoencoders to form the generator of the proposed SOGAN. RESULTS The experimental results demonstrate that the proposed SOAMs improve the quality of the reconstructed MR images effectively by capturing long-range dependencies. Besides, compared with state-of-the-art deep learning-based CS-MRI methods, the proposed SOGAN reconstructs MR images more accurately, but with fewer model parameters. CONCLUSIONS The proposed SOAM is a lightweight yet effective self-attention module to capture long-range dependencies, thus, can improve the quality of MRI reconstruction to a large extent. Besides, with the help of SOAMs, the proposed SOGAN outperforms the state-of-the-art deep learning-based CS-MRI methods.
Collapse
Affiliation(s)
- Wenzhong Zhou
- School of Information and Electronics, Beijing Institute of Technology, Beijing, 100081, China
| | - Huiqian Du
- School of Information and Electronics, Beijing Institute of Technology, Beijing, 100081, China
| | - Wenbo Mei
- School of Information and Electronics, Beijing Institute of Technology, Beijing, 100081, China
| | - Liping Fang
- School of Mathematics and Statistics, Beijing Institute of Technology, Beijing, 100081, China
| |
Collapse
|
83
|
Thomsen FSL, Delrieux CA, Pisula JI, Fuertes García JM, Lucena M, de Luis García R, Borggrefe J. Noise reduction using novel loss functions to compute tissue mineral density and trabecular bone volume fraction on low resolution QCT. Comput Med Imaging Graph 2020; 86:101816. [PMID: 33221674 DOI: 10.1016/j.compmedimag.2020.101816] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/26/2020] [Revised: 09/22/2020] [Accepted: 10/09/2020] [Indexed: 01/28/2023]
Abstract
Micro-structural parameters of the thoracic or lumbar spine generally carry insufficient accuracy and precision for clinical in vivo studies when assessed on quantitative computed tomography (QCT). We propose a 3D convolutional neural network with specific loss functions for QCT noise reduction to compute micro-structural parameters such as tissue mineral density (TMD) and bone volume ratio (BV/TV) with significantly higher accuracy than using no or standard noise reduction filters. The vertebra-phantom study contained high resolution peripheral and clinical CT scans with simulated in vivo CT noise and nine repetitions of three different tube currents (100, 250 and 360 mAs). Five-fold cross validation was performed on 20466 purely spongy pairs of noisy and ground-truth patches. Comparison of training and test errors revealed high robustness against over-fitting. While not showing effects for the assessment of BMD and voxel-wise densities, the filter improved thoroughly the computation of TMD and BV/TV with respect to the unfiltered data. Root-mean-square and accuracy errors of low resolution TMD and BV/TV decreased to less than 17% of the initial values. Furthermore filtered low resolution scans revealed still more TMD- and BV/TV-relevant information than high resolution CT scans, either unfiltered or filtered with two state-of-the-art standard denoising methods. The proposed architecture is threshold and rotational invariant, applicable on a wide range of image resolutions at once, and likely serves for an accurate computation of further micro-structural parameters. Furthermore, it is less prone for over-fitting than neural networks that compute structural parameters directly. In conclusion, the method is potentially important for the diagnosis of osteoporosis and other bone diseases since it allows to assess relevant 3D micro-structural information from standard low exposure CT protocols such as 100 mAs and 120 kVp.
Collapse
Affiliation(s)
- Felix S L Thomsen
- Departamento de Ingeniería Eléctrica y Computadoras, San Andrés 800, 8000 Bahía Blanca, Argentina.
| | - Claudio A Delrieux
- Departamento de Ingeniería Eléctrica y Computadoras, San Andrés 800, 8000 Bahía Blanca, Argentina.
| | - Juan I Pisula
- Departamento de Ingeniería Eléctrica y Computadoras, San Andrés 800, 8000 Bahía Blanca, Argentina.
| | | | - Manuel Lucena
- Campus Las Lagunillas, Edificio Centros de Investigación (C6), 23071 Jaén, Spain.
| | | | - Jan Borggrefe
- Universitätsinstitut für Radiologie, Neuroradiologie und Nuklearmedizin, Hans-Nolte-Str. 1, 32429 Minden, Germany.
| |
Collapse
|
84
|
Letourneau-Guillon L, Camirand D, Guilbert F, Forghani R. Artificial Intelligence Applications for Workflow, Process Optimization and Predictive Analytics. Neuroimaging Clin N Am 2020; 30:e1-e15. [PMID: 33039002 DOI: 10.1016/j.nic.2020.08.008] [Citation(s) in RCA: 23] [Impact Index Per Article: 4.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/22/2022]
Abstract
There is great potential for artificial intelligence (AI) applications, especially machine learning and natural language processing, in medical imaging. Much attention has been garnered by the image analysis tasks for diagnostic decision support and precision medicine, but there are many other potential applications of AI in radiology and have potential to enhance all levels of the radiology workflow and practice, including workflow optimization and support for interpretation tasks, quality and safety, and operational efficiency. This article reviews the important potential applications of informatics and AI related to process improvement and operations in the radiology department.
Collapse
Affiliation(s)
- Laurent Letourneau-Guillon
- Department of Radiology, Centre Hospitalier de l'Université de Montréal (CHUM), 1051, rue Sanguinet, Montréal, Quebec H2X 0C1, Canada; Centre de Recherche du CHUM (CRCHUM), 900 St Denis St, Montréal, Quebec H2X 0A9, Canada.
| | - David Camirand
- Department of Radiology, Centre Hospitalier de l'Université de Montréal (CHUM), 1051, rue Sanguinet, Montréal, Quebec H2X 0C1, Canada
| | - Francois Guilbert
- Department of Radiology, Centre Hospitalier de l'Université de Montréal (CHUM), 1051, rue Sanguinet, Montréal, Quebec H2X 0C1, Canada; Centre de Recherche du CHUM (CRCHUM), 900 St Denis St, Montréal, Quebec H2X 0A9, Canada
| | - Reza Forghani
- Augmented Intelligence & Precision Health Laboratory (AIPHL), Department of Radiology & Research Institute of the McGill University Health Centre, 5252 Boulevard de Maisonneuve Ouest, Montréal, Quebec H4A 3S5, Canada; Department of Radiology, McGill University, 1650 Cedar Avenue, Montréal, Quebec H3G 1A4, Canada; Segal Cancer Centre, Lady Davis Institute for Medical Research, Jewish General Hospital, 3755 Cote Ste-Catherine Road, Montréal, Quebec H3T 1E2, Canada; Gerald Bronfman Department of Oncology, McGill University, Suite 720, 5100 Maisonneuve Boulevard West, Montréal, Quebec H4A3T2, Canada; Department of Otolaryngology - Head and Neck Surgery, Royal Victoria Hospital, McGill University Health Centre, 1001 boul. Decarie Boulevard, Montréal, Quebec H3A 3J1, Canada; 4intelligent Inc., Cote St-Luc, Quebec H3X 4A6, Canada
| |
Collapse
|
85
|
Edge-guided second-order total generalized variation for Gaussian noise removal from depth map. Sci Rep 2020; 10:16329. [PMID: 33004951 PMCID: PMC7530766 DOI: 10.1038/s41598-020-73342-3] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/25/2020] [Accepted: 09/15/2020] [Indexed: 11/24/2022] Open
Abstract
Total generalized variation models have recently demonstrated high-quality denoising capacity for single image. In this paper, we present an accurate denoising method for depth map. Our method uses a weighted second-order total generalized variational model for Gaussian noise removal. By fusing an edge indicator function into the regularization term of the second-order total generalized variational model to guide the diffusion of gradients, our method aims to use the first or second derivative to enhance the intensity of the diffusion tensor. We use the first-order primal–dual algorithm to minimize the proposed energy function and achieve high-quality denoising and edge preserving result for depth maps with high -intensity noise. Extensive quantitative and qualitative evaluations in comparison to bench-mark datasets show that the proposed method provides significant higher accuracy and visual improvements than many state-of-the-art denoising algorithms.
Collapse
|
86
|
Creating Artificial Images for Radiology Applications Using Generative Adversarial Networks (GANs) - A Systematic Review. Acad Radiol 2020; 27:1175-1185. [PMID: 32035758 DOI: 10.1016/j.acra.2019.12.024] [Citation(s) in RCA: 51] [Impact Index Per Article: 10.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2019] [Revised: 12/24/2019] [Accepted: 12/27/2019] [Indexed: 12/22/2022]
Abstract
RATIONALE AND OBJECTIVES Generative adversarial networks (GANs) are deep learning models aimed at generating fake realistic looking images. These novel models made a great impact on the computer vision field. Our study aims to review the literature on GANs applications in radiology. MATERIALS AND METHODS This systematic review followed the PRISMA guidelines. Electronic datasets were searched for studies describing applications of GANs in radiology. We included studies published up-to September 2019. RESULTS Data were extracted from 33 studies published between 2017 and 2019. Eighteen studies focused on CT images generation, ten on MRI, three on PET/MRI and PET/CT, one on ultrasound and one on X-ray. Applications in radiology included image reconstruction and denoising for dose and scan time reduction (fourteen studies), data augmentation (six studies), transfer between modalities (eight studies) and image segmentation (five studies). All studies reported that generated images improved the performance of the developed algorithms. CONCLUSION GANs are increasingly studied for various radiology applications. They enable the creation of new data, which can be used to improve clinical care, education and research.
Collapse
|
87
|
Sheng K. Artificial intelligence in radiotherapy: a technological review. Front Med 2020; 14:431-449. [PMID: 32728877 DOI: 10.1007/s11684-020-0761-1] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2019] [Accepted: 02/14/2020] [Indexed: 12/19/2022]
Abstract
Radiation therapy (RT) is widely used to treat cancer. Technological advances in RT have occurred in the past 30 years. These advances, such as three-dimensional image guidance, intensity modulation, and robotics, created challenges and opportunities for the next breakthrough, in which artificial intelligence (AI) will possibly play important roles. AI will replace certain repetitive and labor-intensive tasks and improve the accuracy and consistency of others, particularly those with increased complexity because of technological advances. The improvement in efficiency and consistency is important to manage the increasing cancer patient burden to the society. Furthermore, AI may provide new functionalities that facilitate satisfactory RT. The functionalities include superior images for real-time intervention and adaptive and personalized RT. AI may effectively synthesize and analyze big data for such purposes. This review describes the RT workflow and identifies areas, including imaging, treatment planning, quality assurance, and outcome prediction, that benefit from AI. This review primarily focuses on deep-learning techniques, although conventional machine-learning techniques are also mentioned.
Collapse
Affiliation(s)
- Ke Sheng
- Department of Radiation Oncology, University of California, Los Angeles, CA, 90095, USA.
| |
Collapse
|
88
|
Podgorsak AR, Sommer KN, Reddy A, Iyer V, Wilson MF, Rybicki FJ, Mitsouras D, Sharma U, Fujimoto S, Kumamaru KK, Angel E, Ionita CN. Initial evaluation of a convolutional neural network used for noninvasive assessment of coronary artery disease severity from coronary computed tomography angiography data. Med Phys 2020; 47:3996-4004. [DOI: 10.1002/mp.14339] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/13/2020] [Revised: 06/09/2020] [Accepted: 06/10/2020] [Indexed: 12/13/2022] Open
Affiliation(s)
- Alexander R. Podgorsak
- From the Canon Stroke and Vascular Research Center 875 Ellicott Street Buffalo NY 14222USA
| | - Kelsey N. Sommer
- From the Canon Stroke and Vascular Research Center 875 Ellicott Street Buffalo NY 14222USA
| | - Abhinay Reddy
- From the Canon Stroke and Vascular Research Center 875 Ellicott Street Buffalo NY 14222USA
| | - Vijay Iyer
- From the Canon Stroke and Vascular Research Center 875 Ellicott Street Buffalo NY 14222USA
| | - Michael F. Wilson
- From the Canon Stroke and Vascular Research Center 875 Ellicott Street Buffalo NY 14222USA
| | - Frank J. Rybicki
- Department of Radiology University of Cincinnati 234 Goodman Street Cincinnati OH USA
| | - Dimitrios Mitsouras
- San Francisco Department of Radiology and Biomedical Imaging University of California 505 Parnassus Avenue San Francisco CA 94143USA
| | - Umesh Sharma
- From the Canon Stroke and Vascular Research Center 875 Ellicott Street Buffalo NY 14222USA
| | - Shinchiro Fujimoto
- Department of Cardiovascular Medicine Juntendo University 3‐1‐3 Hongo, Bunkyo‐ku Tokyo Japan
| | - Kanako K. Kumamaru
- Department of Radiology Juntendo University 3‐1‐3 Hongo, Bunkyo‐ku Tokyo Japan
| | - Erin Angel
- Canon Medical Systems USA, Inc. 2441 Michelle Drive Tustin CA 92780USA
| | - Ciprian N. Ionita
- From the Canon Stroke and Vascular Research Center 875 Ellicott Street Buffalo NY 14222USA
| |
Collapse
|
89
|
Li M, Hsu W, Xie X, Cong J, Gao W. SACNN: Self-Attention Convolutional Neural Network for Low-Dose CT Denoising With Self-Supervised Perceptual Loss Network. IEEE TRANSACTIONS ON MEDICAL IMAGING 2020; 39:2289-2301. [PMID: 31985412 DOI: 10.1109/tmi.2020.2968472] [Citation(s) in RCA: 111] [Impact Index Per Article: 22.2] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/10/2023]
Abstract
Computed tomography (CT) is a widely used screening and diagnostic tool that allows clinicians to obtain a high-resolution, volumetric image of internal structures in a non-invasive manner. Increasingly, efforts have been made to improve the image quality of low-dose CT (LDCT) to reduce the cumulative radiation exposure of patients undergoing routine screening exams. The resurgence of deep learning has yielded a new approach for noise reduction by training a deep multi-layer convolutional neural networks (CNN) to map the low-dose to normal-dose CT images. However, CNN-based methods heavily rely on convolutional kernels, which use fixed-size filters to process one local neighborhood within the receptive field at a time. As a result, they are not efficient at retrieving structural information across large regions. In this paper, we propose a novel 3D self-attention convolutional neural network for the LDCT denoising problem. Our 3D self-attention module leverages the 3D volume of CT images to capture a wide range of spatial information both within CT slices and between CT slices. With the help of the 3D self-attention module, CNNs are able to leverage pixels with stronger relationships regardless of their distance and achieve better denoising results. In addition, we propose a self-supervised learning scheme to train a domain-specific autoencoder as the perceptual loss function. We combine these two methods and demonstrate their effectiveness on both CNN-based neural networks and WGAN-based neural networks with comprehensive experiments. Tested on the AAPM-Mayo Clinic Low Dose CT Grand Challenge data set, our experiments demonstrate that self-attention (SA) module and autoencoder (AE) perceptual loss function can efficiently enhance traditional CNNs and can achieve comparable or better results than the state-of-the-art methods.
Collapse
|
90
|
Gao Y, Liang Z, Zhang H, Yang J, Ferretti J, Bilfinger T, Yaddanapudi K, Schweitzer M, Bhattacharji P, Moore W. A Task-dependent Investigation on Dose and Texture in CT Image Reconstruction. IEEE TRANSACTIONS ON RADIATION AND PLASMA MEDICAL SCIENCES 2020; 4:441-449. [PMID: 33907724 PMCID: PMC8075295 DOI: 10.1109/trpms.2019.2957459] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/03/2023]
Abstract
Localizing and characterizing clinically-significant lung nodules, a potential precursor to lung cancer, at the lowest achievable radiation dose is demanded to minimize the stochastic radiation effects from x-ray computed tomography (CT). A minimal dose level is heavily dependent on the image reconstruction algorithms and clinical task, in which the tissue texture always plays an important role. This study aims to investigate the dependence through a task-based evaluation at multiple dose levels and variable textures in reconstructions with prospective patient studies. 133 patients with a suspicious pulmonary nodule scheduled for biopsy were recruited and the data was acquired at120kVp with three different dose levels of 100, 40 and 20mAs. Three reconstruction algorithms were implemented: analytical filtered back-projection (FBP) with optimal noise filtering; statistical Markov random field (MRF) model with optimal Huber weighting (MRF-H) for piecewise smooth reconstruction; and tissue-specific texture model (MRF-T) for texture preserved statistical reconstruction. Experienced thoracic radiologists reviewed and scored all images at random, blind to the CT dose and reconstruction algorithms. The radiologists identified the nodules in each image including the 133 biopsy target nodules and 66 other non-target nodules. For target nodule characterization, only MRF-T at 40mAs showed no statistically significant difference from FBP at 100mAs. For localizing both the target nodules and the non-target nodules, some as small as 3mm, MRF-T at 40 and 20mAs levels showed no statistically significant difference from FBP at 100mAs, respectively. MRF-H and FBP at 40 and 20mAs levels performed statistically differently from FBP at 100mAs. This investigation concluded that (1) the textures in the MRF-T reconstructions improves both the tasks of localizing and characterizing nodules at low dose CT and (2) the task of characterizing nodules is more challenging than the task of localizing nodules and needs more dose or enhanced textures from reconstruction.
Collapse
Affiliation(s)
- Yongfeng Gao
- Department of Radiology, Stony Brook University, Stony Brook, NY
11794, USA
| | - Zhengrong Liang
- Departments of Radiology, Biomedical Engineering, Computer Science,
and Electrical Engineering, Stony Brook University, Stony Brook, NY 11794,
USA
| | - Hao Zhang
- Departments of Radiology and Biomedical Engineering, Stony Brook
University, Stony Brook, NY 11794, USA and now with the Department of
Radiation Oncology, Stanford University, Stanford, CA 94035, USA
| | - Jie Yang
- Department of Family, Population and Preventive Medicine, Stony
Brook University, Stony Brook, NY 11794, USA
| | - John Ferretti
- Department of Radiology, Stony Brook University, Stony Brook, NY
11794, USA
| | - Thomas Bilfinger
- Department of Surgery, Stony Brook University, Stony Brook, NY
11794, USA)
| | | | - Mark Schweitzer
- Department of Radiology, Stony Brook University, Stony Brook, NY
11794, USA
| | - Priya Bhattacharji
- Department of Radiology, Stony Brook University, Stony Brook, NY
11794, USA, and now with the Department of Radiology, New York University,
New York, NY 10016, USA
| | - William Moore
- Department of Radiology, Stony Brook University, Stony Brook, NY
11794, USA, and now with the Department of Radiology, New York University,
New York, NY 10016, USA
| |
Collapse
|
91
|
Shan H, Jia X, Yan P, Li Y, Paganetti H, Wang G. Synergizing medical imaging and radiotherapy with deep learning. MACHINE LEARNING-SCIENCE AND TECHNOLOGY 2020. [DOI: 10.1088/2632-2153/ab869f] [Citation(s) in RCA: 18] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/14/2022]
|
92
|
Fan F, Shan H, Kalra MK, Singh R, Qian G, Getzin M, Teng Y, Hahn J, Wang G. Quadratic Autoencoder (Q-AE) for Low-Dose CT Denoising. IEEE TRANSACTIONS ON MEDICAL IMAGING 2020; 39:2035-2050. [PMID: 31902758 PMCID: PMC7376975 DOI: 10.1109/tmi.2019.2963248] [Citation(s) in RCA: 50] [Impact Index Per Article: 10.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/05/2023]
Abstract
Inspired by complexity and diversity of biological neurons, our group proposed quadratic neurons by replacing the inner product in current artificial neurons with a quadratic operation on input data, thereby enhancing the capability of an individual neuron. Along this direction, we are motivated to evaluate the power of quadratic neurons in popular network architectures, simulating human-like learning in the form of "quadratic-neuron-based deep learning". Our prior theoretical studies have shown important merits of quadratic neurons and networks in representation, efficiency, and interpretability. In this paper, we use quadratic neurons to construct an encoder-decoder structure, referred as the quadratic autoencoder, and apply it to low-dose CT denoising. The experimental results on the Mayo low-dose CT dataset demonstrate the utility and robustness of quadratic autoencoder in terms of image denoising and model efficiency. To our best knowledge, this is the first time that the deep learning approach is implemented with a new type of neurons and demonstrates a significant potential in the medical imaging field.
Collapse
Affiliation(s)
- Fenglei Fan
- Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY, USA
| | - Hongming Shan
- Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY, USA
| | - Mannudeep K. Kalra
- Department of Radiology, Massachusetts General Hospital, Boston, MA 02114, USA
| | - Ramandeep Singh
- Department of Radiology, Massachusetts General Hospital, Boston, MA 02114, USA
| | - Guhan Qian
- Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY, USA
| | - Matthew Getzin
- Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY, USA
| | - Yueyang Teng
- Sino-Dutch Biomedical and Information Engineering School, Northeastern University, Shenyang, China, 110169
| | - Juergen Hahn
- Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY, USA
| | - Ge Wang
- Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY, USA
| |
Collapse
|
93
|
A novel super-resolution CT image reconstruction via semi-supervised generative adversarial network. Neural Comput Appl 2020. [DOI: 10.1007/s00521-020-04905-8] [Citation(s) in RCA: 12] [Impact Index Per Article: 2.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/26/2022]
|
94
|
Yang Q, Li N, Zhao Z, Fan X, Chang EIC, Xu Y. MRI Cross-Modality Image-to-Image Translation. Sci Rep 2020; 10:3753. [PMID: 32111966 PMCID: PMC7048849 DOI: 10.1038/s41598-020-60520-6] [Citation(s) in RCA: 40] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/09/2019] [Accepted: 02/12/2020] [Indexed: 11/23/2022] Open
Abstract
We present a cross-modality generation framework that learns to generate translated modalities from given modalities in MR images. Our proposed method performs Image Modality Translation (abbreviated as IMT) by means of a deep learning model that leverages conditional generative adversarial networks (cGANs). Our framework jointly exploits the low-level features (pixel-wise information) and high-level representations (e.g. brain tumors, brain structure like gray matter, etc.) between cross modalities which are important for resolving the challenging complexity in brain structures. Our framework can serve as an auxiliary method in medical use and has great application potential. Based on our proposed framework, we first propose a method for cross-modality registration by fusing the deformation fields to adopt the cross-modality information from translated modalities. Second, we propose an approach for MRI segmentation, translated multichannel segmentation (TMS), where given modalities, along with translated modalities, are segmented by fully convolutional networks (FCN) in a multichannel manner. Both of these two methods successfully adopt the cross-modality information to improve the performance without adding any extra data. Experiments demonstrate that our proposed framework advances the state-of-the-art on five brain MRI datasets. We also observe encouraging results in cross-modality registration and segmentation on some widely adopted brain datasets. Overall, our work can serve as an auxiliary method in medical use and be applied to various tasks in medical fields.
Collapse
Grants
- This work is supported by Microsoft Research under the eHealth program, the National Natural Science Foundation in China under Grant 81771910, the National Science and Technology Major Project of the Ministry of Science and Technology in China under Grant 2017YFC0110903, the Beijing Natural Science Foundation in China under Grant 4152033, the Technology and Innovation Commission of Shenzhen in China under Grant shenfagai2016-627, Beijing Young Talent Project in China, the Fundamental Research Funds for the Central Universities of China under Grant SKLSDE-2017ZX-08 from the State Key Laboratory of Software Development Environment in Beihang University in China, the 111 Project in China under Grant B13003.
- This work is supported by the National Science and Technology Major Project of the Ministry of Science and Technology in China under Grant 2017YFC0110903, Microsoft Research under the eHealth program, the National Natural Science Foundation in China under Grant 81771910, the Beijing Natural Science Foundation in China under Grant 4152033, the Technology and Innovation Commission of Shenzhen in China under Grant shenfagai2016-627, Beijing Young Talent Project in China, the Fundamental Research Funds for the Central Universities of China under Grant SKLSDE-2017ZX-08 from the State Key Laboratory of Software Development Environment in Beihang University in China, the 111 Project in China under Grant B13003.
Collapse
Affiliation(s)
- Qianye Yang
- State Key Laboratory of Software Development Environment and Key Laboratory of Biomechanics and Mechanobiology of Ministry of Education and Research Institute of Beihang University in Shenzhen, Beijing Advanced Innovation Center for Biomedical Engineering, Beihang University, Beijing, 100191, China
| | - Nannan Li
- State Key Laboratory of Software Development Environment and Key Laboratory of Biomechanics and Mechanobiology of Ministry of Education and Research Institute of Beihang University in Shenzhen, Beijing Advanced Innovation Center for Biomedical Engineering, Beihang University, Beijing, 100191, China
- Ping An Technology (Shenzhen) Co., Ltd., Shanghai, 200030, China
| | - Zixu Zhao
- State Key Laboratory of Software Development Environment and Key Laboratory of Biomechanics and Mechanobiology of Ministry of Education and Research Institute of Beihang University in Shenzhen, Beijing Advanced Innovation Center for Biomedical Engineering, Beihang University, Beijing, 100191, China
| | - Xingyu Fan
- Bioengineering College of Chongqing University, Chongqing, 400044, China
| | | | - Yan Xu
- State Key Laboratory of Software Development Environment and Key Laboratory of Biomechanics and Mechanobiology of Ministry of Education and Research Institute of Beihang University in Shenzhen, Beijing Advanced Innovation Center for Biomedical Engineering, Beihang University, Beijing, 100191, China.
- Microsoft Research Asia, Beijing, 100080, China.
| |
Collapse
|
95
|
Guha I, Nadeem SA, You C, Zhang X, Levy SM, Wang G, Torner JC, Saha PK. Deep Learning Based High-Resolution Reconstruction of Trabecular Bone Microstructures from Low-Resolution CT Scans using GAN-CIRCLE. PROCEEDINGS OF SPIE--THE INTERNATIONAL SOCIETY FOR OPTICAL ENGINEERING 2020; 11317:113170U. [PMID: 32201450 PMCID: PMC7085412 DOI: 10.1117/12.2549318] [Citation(s) in RCA: 19] [Impact Index Per Article: 3.8] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/28/2023]
Abstract
Osteoporosis is a common age-related disease characterized by reduced bone density and increased fracture-risk. Microstructural quality of trabecular bone (Tb), commonly found at axial skeletal sites and at the end of long bones, is an important determinant of bone-strength and fracture-risk. High-resolution emerging CT scanners enable in vivo measurement of Tb microstructures at peripheral sites. However, resolution-dependence of microstructural measures and wide resolution-discrepancies among various CT scanners together with rapid upgrades in technology warrant data harmonization in CT-based cross-sectional and longitudinal bone studies. This paper presents a deep learning-based method for high-resolution reconstruction of Tb microstructures from low-resolution CT scans using GAN-CIRCLE. A network was developed and evaluated using post-registered ankle CT scans of nineteen volunteers on both low- and high-resolution CT scanners. 9,000 matching pairs of low- and high-resolution patches of size 64×64 were randomly harvested from ten volunteers for training and validation. Another 5,000 matching pairs of patches from nine other volunteers were used for evaluation. Quantitative comparison shows that predicted high-resolution scans have significantly improved structural similarity index (p < 0.01) with true high-resolution scans as compared to the same metric for low-resolution data. Different Tb microstructural measures such as thickness, spacing, and network area density are also computed from low- and predicted high-resolution images, and compared with the values derived from true high-resolution scans. Thickness and network area measures from predicted images showed higher agreement with true high-resolution CT (CCC = [0.95, 0.91]) derived values than the same measures from low-resolution images (CCC = [0.72, 0.88]).
Collapse
Affiliation(s)
- Indranil Guha
- Department of Electrical and Computer Engineering, College of Engineering, University of Iowa, Iowa City, IA 52242
| | - Syed Ahmed Nadeem
- Department of Electrical and Computer Engineering, College of Engineering, University of Iowa, Iowa City, IA 52242
| | - Chenyu You
- Department of Computer Science, Yale University, New Haven, CT 05620
| | - Xiaoliu Zhang
- Department of Electrical and Computer Engineering, College of Engineering, University of Iowa, Iowa City, IA 52242
| | - Steven M Levy
- Department of Preventive and Community Dentistry, College of Dentistry, University of Iowa, Iowa City, IA 52242
| | - Ge Wang
- Biomedical Imaging Center, BME/CBIS, Rensselaer Polytechnic Institute, Troy, New York, NY 12180
| | - James C Torner
- Department of Epidemiology, University of Iowa, Iowa City, IA 52242
| | - Punam K Saha
- Department of Electrical and Computer Engineering, College of Engineering, University of Iowa, Iowa City, IA 52242
- Department of Radiology, Carver College of Medicine, University of Iowa, Iowa City, IA 52242
| |
Collapse
|
96
|
Qiu B, Huang Z, Liu X, Meng X, You Y, Liu G, Yang K, Maier A, Ren Q, Lu Y. Noise reduction in optical coherence tomography images using a deep neural network with perceptually-sensitive loss function. BIOMEDICAL OPTICS EXPRESS 2020; 11:817-830. [PMID: 32133225 PMCID: PMC7041484 DOI: 10.1364/boe.379551] [Citation(s) in RCA: 58] [Impact Index Per Article: 11.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/04/2019] [Revised: 01/07/2020] [Accepted: 01/08/2020] [Indexed: 05/02/2023]
Abstract
Optical coherence tomography (OCT) is susceptible to the coherent noise, which is the speckle noise that deteriorates contrast and the detail structural information of OCT images, thus imposing significant limitations on the diagnostic capability of OCT. In this paper, we propose a novel OCT image denoising method by using an end-to-end deep learning network with a perceptually-sensitive loss function. The method has been validated on OCT images acquired from healthy volunteers' eyes. The label images for training and evaluating OCT denoising deep learning models are images generated by averaging 50 frames of respective registered B-scans acquired from a region with scans occurring in one direction. The results showed that the new approach can outperform other related denoising methods on the aspects of preserving detail structure information of retinal layers and improving the perceptual metrics in the human visual perception.
Collapse
Affiliation(s)
- Bin Qiu
- Department of Biomedical Engineering, College of Engineering, Peking University, No. 5 Yihe Yuan Road, Haidian District, Beijing 100871, China
| | - Zhiyu Huang
- Department of Biomedical Engineering, College of Engineering, Peking University, No. 5 Yihe Yuan Road, Haidian District, Beijing 100871, China
| | - Xi Liu
- Department of Biomedical Engineering, College of Engineering, Peking University, No. 5 Yihe Yuan Road, Haidian District, Beijing 100871, China
| | - Xiangxi Meng
- Department of Biomedical Engineering, College of Engineering, Peking University, No. 5 Yihe Yuan Road, Haidian District, Beijing 100871, China
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education), Department of Nuclear Medicine, Peking University Cancer Hospital & Institute, Beijing 100142, China
| | - Yunfei You
- Department of Biomedical Engineering, College of Engineering, Peking University, No. 5 Yihe Yuan Road, Haidian District, Beijing 100871, China
| | - Gangjun Liu
- Institute of Biomedical Engineering, Peking University Shenzhen Graduate School, No. 2199 Lishui Road, Nanshan District, Shenzhen 518055, China
- Institute of Biomedical Engineering, Shenzhen Bay Laboratory 5F, No. 9 Duxue Road, Nanshan District, Shenzhen 518071, China
| | - Kun Yang
- College of Quality and Technical Supervision, Hebei University, No. 2666 Qiyidong Road, Baoding 071000, China
| | - Andreas Maier
- Pattern Recognition Lab, Department of Computer Science, Friedrich-Alexander-University Erlangen-Nuremberg, Martensstrasse 3, 91058 Erlangen, Germany
| | - Qiushi Ren
- Department of Biomedical Engineering, College of Engineering, Peking University, No. 5 Yihe Yuan Road, Haidian District, Beijing 100871, China
- Institute of Biomedical Engineering, Peking University Shenzhen Graduate School, No. 2199 Lishui Road, Nanshan District, Shenzhen 518055, China
- Institute of Biomedical Engineering, Shenzhen Bay Laboratory 5F, No. 9 Duxue Road, Nanshan District, Shenzhen 518071, China
| | - Yanye Lu
- Department of Biomedical Engineering, College of Engineering, Peking University, No. 5 Yihe Yuan Road, Haidian District, Beijing 100871, China
- Institute of Biomedical Engineering, Peking University Shenzhen Graduate School, No. 2199 Lishui Road, Nanshan District, Shenzhen 518055, China
- Pattern Recognition Lab, Department of Computer Science, Friedrich-Alexander-University Erlangen-Nuremberg, Martensstrasse 3, 91058 Erlangen, Germany
| |
Collapse
|
97
|
You C, Yang J, Chapiro J, Duncan JS. Unsupervised Wasserstein Distance Guided Domain Adaptation for 3D Multi-domain Liver Segmentation. INTERPRETABLE AND ANNOTATION-EFFICIENT LEARNING FOR MEDICAL IMAGE COMPUTING 2020. [DOI: 10.1007/978-3-030-61166-8_17] [Citation(s) in RCA: 18] [Impact Index Per Article: 3.6] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/02/2022]
|
98
|
Imae T, Kaji S, Kida S, Matsuda K, Takenaka S, Aoki A, Nakamoto T, Ozaki S, Nawa K, Yamashita H, Nakagawa K, Abe O. [Improvement in Image Quality of CBCT during Treatment by Cycle Generative Adversarial Network]. Nihon Hoshasen Gijutsu Gakkai Zasshi 2020; 76:1173-1184. [PMID: 33229847 DOI: 10.6009/jjrt.2020_jsrt_76.11.1173] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
PURPOSE Volumetric modulated arc therapy (VMAT) can acquire projection images during rotational irradiation, and cone-beam computed tomography (CBCT) images during VMAT delivery can be reconstructed. The poor quality of CBCT images prevents accurate recognition of organ position during the treatment. The purpose of this study was to improve the image quality of CBCT during the treatment by cycle generative adversarial network (CycleGAN). METHOD Twenty patients with clinically localized prostate cancer were treated with VMAT, and projection images for intra-treatment CBCT (iCBCT) were acquired. Synthesis of PCT (SynPCT) with improved image quality by CycleGAN requires only unpaired and unaligned iCBCT and planning CT (PCT) images for training. We performed visual and quantitative evaluation to compare iCBCT, SynPCT and PCT deformable image registration (DIR) to confirm the clinical usefulness. RESULT We demonstrated suitable CycleGAN networks and hyperparameters for SynPCT. The image quality of SynPCT improved visually and quantitatively while preserving anatomical structures of the original iCBCT. The undesirable deformation of PCT was reduced when SynPCT was used as its reference instead of iCBCT. CONCLUSION We have performed image synthesis with preservation of organ position by CycleGAN for iCBCT and confirmed the clinical usefulness.
Collapse
Affiliation(s)
| | - Shizuo Kaji
- Department of Radiology, University of Tokyo Hospital
- Institute of Mathematics for Industry, Kyushu University
| | - Satoshi Kida
- Department of Radiology, University of Tokyo Hospital
| | | | | | - Atsushi Aoki
- Department of Radiology, University of Tokyo Hospital
| | | | - Sho Ozaki
- Department of Radiology, University of Tokyo Hospital
| | - Kanabu Nawa
- Department of Radiology, University of Tokyo Hospital
| | | | | | - Osamu Abe
- Department of Radiology, University of Tokyo Hospital
| |
Collapse
|
99
|
|
100
|
You C, Li G, Zhang Y, Zhang X, Shan H, Li M, Ju S, Zhao Z, Zhang Z, Cong W, Vannier MW, Saha PK, Hoffman EA, Wang G. CT Super-Resolution GAN Constrained by the Identical, Residual, and Cycle Learning Ensemble (GAN-CIRCLE). IEEE TRANSACTIONS ON MEDICAL IMAGING 2020; 39:188-203. [PMID: 31217097 PMCID: PMC11662229 DOI: 10.1109/tmi.2019.2922960] [Citation(s) in RCA: 181] [Impact Index Per Article: 36.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/16/2023]
Abstract
In this paper, we present a semi-supervised deep learning approach to accurately recover high-resolution (HR) CT images from low-resolution (LR) counterparts. Specifically, with the generative adversarial network (GAN) as the building block, we enforce the cycle-consistency in terms of the Wasserstein distance to establish a nonlinear end-to-end mapping from noisy LR input images to denoised and deblurred HR outputs. We also include the joint constraints in the loss function to facilitate structural preservation. In this process, we incorporate deep convolutional neural network (CNN), residual learning, and network in network techniques for feature extraction and restoration. In contrast to the current trend of increasing network depth and complexity to boost the imaging performance, we apply a parallel 1×1 CNN to compress the output of the hidden layer and optimize the number of layers and the number of filters for each convolutional layer. The quantitative and qualitative evaluative results demonstrate that our proposed model is accurate, efficient and robust for super-resolution (SR) image restoration from noisy LR input images. In particular, we validate our composite SR networks on three large-scale CT datasets, and obtain promising results as compared to the other state-of-the-art methods.
Collapse
|