1
|
Sun Q, He N, Yang P, Zhao X. Low dose computed tomography reconstruction with momentum-based frequency adjustment network. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2025; 263:108673. [PMID: 40023964 DOI: 10.1016/j.cmpb.2025.108673] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/11/2024] [Revised: 11/29/2024] [Accepted: 02/13/2025] [Indexed: 03/04/2025]
Abstract
BACKGROUND AND OBJECTIVE Recent investigations into Low-Dose Computed Tomography (LDCT) reconstruction methods have brought Model-Based Data-Driven (MBDD) approaches to the forefront. One prominent architecture within MBDD entails the integration of Model-Based Iterative Reconstruction (MBIR) with Deep Learning (DL). While this approach offers the advantage of harnessing information from sinogram and image domains, it also reveals several deficiencies. First and foremost, the efficacy of DL methods within the realm of MBDD necessitates meticulous enhancement, as it directly impacts the computational cost and the quality of reconstructed images. Next, high computational costs and a high number of iterations limit the development of MBDD methods. Last but not least, CT reconstruction is sensitive to pixel accuracy, and the role of loss functions within DL methods is crucial for meeting this requirement. METHODS This paper advances MBDD methods through three principal contributions. Firstly, we introduce an innovative Frequency Adjustment Network (FAN) that effectively adjusts both high and low-frequency components during the inference phase, resulting in substantial enhancements in reconstruction performance. Second, we develop the Momentum-based Frequency Adjustment Network (MFAN), which leverages momentum terms as an extrapolation strategy to facilitate the amplification of changes throughout successive iterations, culminating in a rapid convergence framework. Lastly, we delve into the visual properties of CT images and present a unique loss function named Focal Detail Loss (FDL). The FDL function preserves fine details throughout the training phase, significantly improving reconstruction quality. RESULTS Through a series of experiments validation on the AAPM-Mayo public dataset and real-world piglet datasets, the aforementioned three contributions demonstrated superior performance. MFAN achieved convergence in 10 iterations as an iteration method, faster than other methods. Ablation studies further highlight the advanced performance of each contribution. CONCLUSIONS This paper presents an MBDD-based LDCT reconstruction method using a momentum-based frequency adjustment network with a focal detail loss function. This approach significantly reduces the number of iterations required for convergence while achieving superior reconstruction results in visual and numerical analyses.
Collapse
Affiliation(s)
- Qixiang Sun
- School of Mathematical Sciences, Capital Normal University, Beijing, 100048, China
| | - Ning He
- Smart City College, Beijing Union University, Beijing, 100101, China
| | - Ping Yang
- School of Mathematical Sciences, Capital Normal University, Beijing, 100048, China
| | - Xing Zhao
- School of Mathematical Sciences, Capital Normal University, Beijing, 100048, China.
| |
Collapse
|
2
|
Xu S, Fu J, Sun Y, Cong P, Xiang X. Deep Radon Prior: A fully unsupervised framework for sparse-view CT reconstruction. Comput Biol Med 2025; 189:109853. [PMID: 40056836 DOI: 10.1016/j.compbiomed.2025.109853] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2024] [Revised: 02/08/2025] [Accepted: 02/11/2025] [Indexed: 03/10/2025]
Abstract
BACKGROUND Sparse-view computed tomography (CT) substantially reduces radiation exposure but often introduces severe artifacts that compromise image fidelity. Recent advances in deep learning for solving inverse problems have shown considerable promise in enhancing CT reconstruction; however, most approaches heavily rely on high-quality training datasets and lack interpretability. METHOD To address these challenges, this paper introduces a novel, fully unsupervised deep learning framework that mitigates the dependency on extensive labeled data and improves the interpretability of the reconstruction process. Specifically, we propose the Deep Radon Prior (DRP) framework, inspired by the Deep Image Prior (DIP), which integrates a neural network as an implicit prior into the iterative reconstruction process. This integration facilitates the image domain and the Radon domain gradient feedback and progressively optimizes the neural network through multiple stages, effectively narrowing the solution space in the Radon domain for under-constrained imaging protocols. RESULTS We discuss the convergence properties of DRP and validate our approach experimentally, demonstrating its ability to produce high-fidelity images while significantly reducing artifacts. Results indicate that DRP achieves comparable or superior performance to supervised methods, thereby addressing the inherent challenges of sparse-view CT and substantially enhancing image quality. CONCLUSIONS The introduction of DRP represents a significant advancement in sparse-view CT imaging by leveraging the inherent deep self-correlation of the Radon domain, enabling effective cooperation with neural network manifolds for image reconstruction. This paradigm shift toward fully unsupervised learning offers a scalable and insightful approach to medical imaging, potentially redefining the landscape of CT reconstruction.
Collapse
Affiliation(s)
- Shuo Xu
- Institute of Nuclear and New Energy Technology, Tsinghua University, Beijing, China; Beijing Key Laboratory of Nuclear Detection Technology, Beijing, China; Hefei Meyer Optoelectronic Technology INC, Hefei, Anhui Province, China.
| | - Jintao Fu
- Institute of Nuclear and New Energy Technology, Tsinghua University, Beijing, China; Beijing Key Laboratory of Nuclear Detection Technology, Beijing, China.
| | - Yuewen Sun
- Institute of Nuclear and New Energy Technology, Tsinghua University, Beijing, China; Beijing Key Laboratory of Nuclear Detection Technology, Beijing, China.
| | - Peng Cong
- Institute of Nuclear and New Energy Technology, Tsinghua University, Beijing, China; Beijing Key Laboratory of Nuclear Detection Technology, Beijing, China.
| | - Xincheng Xiang
- Institute of Nuclear and New Energy Technology, Tsinghua University, Beijing, China; Beijing Key Laboratory of Nuclear Detection Technology, Beijing, China.
| |
Collapse
|
3
|
An W, Liu Y, Shang F, Liu H, Jiao L. DEs-Inspired Accelerated Unfolded Linearized ADMM Networks for Inverse Problems. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS 2025; 36:5319-5333. [PMID: 38625778 DOI: 10.1109/tnnls.2024.3382030] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 04/18/2024]
Abstract
Many research works have shown that the traditional alternating direction multiplier methods (ADMMs) can be better understood by continuous-time differential equations (DEs). On the other hand, many unfolded algorithms directly inherit the traditional iterations to build deep networks. Although they achieve superior practical performance and a faster convergence rate than traditional counterparts, there is a lack of clear insight into unfolded network structures. Thus, we attempt to explore the unfolded linearized ADMM (LADMM) from the perspective of DEs, and design more efficient unfolded networks. First, by proposing an unfolded Euler LADMM scheme and inspired by the trapezoid discretization, we design a new more accurate Trapezoid LADMM scheme. For the convenience of implementation, we provide its explicit version via a prediction-correction strategy. Then, to expand the representation space of unfolded networks, we design an accelerated variant of our Euler LADMM scheme, which can be interpreted as second-order DEs with stronger representation capabilities. To fully explore this representation space, we designed an accelerated Trapezoid LADMM scheme. To the best of our knowledge, this is the first work to explore a comprehensive connection with theoretical guarantees between unfolded ADMMs and first- (second-) order DEs. Finally, we instantiate our schemes as (A-)ELADMM and (A-)TLADMM with the proximal operators, and (A-)ELADMM-Net and (A-)TLADMM-Net with convolutional neural networks (CNNs). Extensive inverse problem experiments show that our Trapezoid LADMM schemes perform better than well-known methods.
Collapse
|
4
|
Gan Y, Lan Q, Huang C, Su W, Huang Z. Dense convolution-based attention network for Alzheimer's disease classification. Sci Rep 2025; 15:5693. [PMID: 39962113 PMCID: PMC11832751 DOI: 10.1038/s41598-025-85802-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/09/2024] [Accepted: 01/06/2025] [Indexed: 02/20/2025] Open
Abstract
Recently, deep learning-based medical image classification models have made substantial advancements. However, many existing models prioritize performance at the cost of efficiency, limiting their practicality in clinical use. Traditional Convolutional Neural Network (CNN)-based methods, Transformer-based methods, and hybrid approaches combining these two struggle to balance performance and model complexity. To achieve efficient predictions with a low parameter count, we propose DenseAttentionNetwork (DANet), a lightweight model for Alzheimer's disease detection in 3D MRI images. DANet leverages dense connections and a linear attention mechanism to enhance feature extraction and capture long-range dependencies. Its architecture integrates convolutional layers for localized feature extraction with linear attention for global context, enabling efficient multi-scale feature reuse across the network. By replacing traditional self-attention with a parameter-efficient linear attention mechanism, DANet overcomes some limitations of standard self-attention. Extensive experiments across multi-institutional datasets demonstrate that DANet achieves the best performance in area under the receiver operating characteristic curve (AUC), which underscores the model's robustness and effectiveness in capturing relevant features for Alzheimer's disease detection while also attaining a strong accuracy structure with fewer parameters. Visualizations based on activation maps further verify the model's ability to highlight AD-relevant regions in 3D MRI images, providing clinically interpretable insights into disease progression.
Collapse
Affiliation(s)
- Yingtong Gan
- Key Laboratory of Multimedia Trusted Perception and Efficient Computing, Ministry of Education of China, Xiamen University, Xiamen, 361005, People's Republic of China
- Institute of Artifical Intelligence, Xiamen University, Xiamen, 361005, People's Republic of China
| | - Quan Lan
- Department of Neurology, First Affiliated Hospital of Xiamen University, Xiamen, 361003, People's Republic of China
| | - ChenXi Huang
- Key Laboratory of Multimedia Trusted Perception and Efficient Computing, Ministry of Education of China, Xiamen University, Xiamen, 361005, People's Republic of China.
| | - Weichao Su
- Xiamen Xianyue Hospital, Xianyue Hospital Affiliated with Xiamen Medical College, Fujian Psychiatric Center, Fujian Clinical Research Center for Mental Disorders, Xiamen, 361012, People's Republic of China.
| | - Zhiyuan Huang
- The School of Clinical Medicine, Fujian Medical University, Fuzhou, 350122, People's Republic of China.
- Xiamen Xianyue Hospital, Xianyue Hospital Affiliated with Xiamen Medical College, Fujian Psychiatric Center, Fujian Clinical Research Center for Mental Disorders, Xiamen, 361012, People's Republic of China.
| |
Collapse
|
5
|
Lin Z, Quan G, Qu H, Du Y, Zhao J. LOQUAT: Low-Rank Quaternion Reconstruction for Photon-Counting CT. IEEE TRANSACTIONS ON MEDICAL IMAGING 2025; 44:668-684. [PMID: 39226197 DOI: 10.1109/tmi.2024.3454174] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 09/05/2024]
Abstract
Photon-counting computed tomography (PCCT) may dramatically benefit clinical practice due to its versatility such as dose reduction and material characterization. However, the limited number of photons detected in each individual energy bin can induce severe noise contamination in the reconstructed image. Fortunately, the notable low-rank prior inherent in the PCCT image can guide the reconstruction to a denoised outcome. To fully excavate and leverage the intrinsic low-rankness, we propose a novel reconstruction algorithm based on quaternion representation (QR), called low-rank quaternion reconstruction (LOQUAT). First, we organize a group of nonlocal similar patches into a quaternion matrix. Then, an adjusted weighted Schatten-p norm (AWSN) is introduced and imposed on the matrix to enforce its low-rank nature. Subsequently, we formulate an AWSN-regularized model and devise an alternating direction method of multipliers (ADMM) framework to solve it. Experiments on simulated and real-world data substantiate the superiority of the LOQUAT technique over several state-of-the-art competitors in terms of both visual inspection and quantitative metrics. Moreover, our QR-based method exhibits lower computational complexity than some popular tensor representation (TR) based counterparts. Besides, the global convergence of LOQUAT is theoretically established under a mild condition. These properties bolster the robustness and practicality of LOQUAT, facilitating its application in PCCT clinical scenarios. The source code will be available at https://github.com/linzf23/LOQUAT.
Collapse
|
6
|
Claros-Olivares CC, Clements RG, McIlvain G, Johnson CL, Brockmeier AJ. MRI-based whole-brain elastography and volumetric measurements to predict brain age. Biol Methods Protoc 2024; 10:bpae086. [PMID: 39902188 PMCID: PMC11790219 DOI: 10.1093/biomethods/bpae086] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/15/2024] [Revised: 11/03/2024] [Accepted: 11/12/2024] [Indexed: 02/05/2025] Open
Abstract
Brain age, as a correlate of an individual's chronological age obtained from structural and functional neuroimaging data, enables assessing developmental or neurodegenerative pathology relative to the overall population. Accurately inferring brain age from brain magnetic resonance imaging (MRI) data requires imaging methods sensitive to tissue health and sophisticated statistical models to identify the underlying age-related brain changes. Magnetic resonance elastography (MRE) is a specialized MRI technique which has emerged as a reliable, non-invasive method to measure the brain's mechanical properties, such as the viscoelastic shear stiffness and damping ratio. These mechanical properties have been shown to change across the life span, reflect neurodegenerative diseases, and are associated with individual differences in cognitive function. Here, we aim to develop a machine learning framework to accurately predict a healthy individual's chronological age from maps of brain mechanical properties. This framework can later be applied to understand neurostructural deviations from normal in individuals with neurodevelopmental or neurodegenerative conditions. Using 3D convolutional networks as deep learning models and more traditional statistical models, we relate chronological age as a function of multiple modalities of whole-brain measurements: stiffness, damping ratio, and volume. Evaluations on held-out subjects show that combining stiffness and volume in a multimodal approach achieves the most accurate predictions. Interpretation of the different models highlights important regions that are distinct between the modalities. The results demonstrate the complementary value of MRE measurements in brain age models, which, in future studies, could improve model sensitivity to brain integrity differences in individuals with neuropathology.
Collapse
Affiliation(s)
| | - Rebecca G Clements
- Department of Biomedical Engineering, University of Delaware, Newark, DE 19713, United States
- Department of Biomedical Engineering, Northwestern University, Evanston, IL 60208, United States
- Department of Physical Therapy and Human Movement Sciences, Northwestern University, Chicago, IL 60611, United States
| | - Grace McIlvain
- Department of Biomedical Engineering, University of Delaware, Newark, DE 19713, United States
- Department of Biomedical Engineering, Columbia University, New York, NY 10027, United States
| | - Curtis L Johnson
- Department of Electrical & Computer Engineering, University of Delaware, Newark, DE 19716, United States
- Department of Biomedical Engineering, University of Delaware, Newark, DE 19713, United States
| | - Austin J Brockmeier
- Department of Electrical & Computer Engineering, University of Delaware, Newark, DE 19716, United States
- Department of Computer & Information Sciences, University of Delaware, Newark, DE 19716, United States
| |
Collapse
|
7
|
Kim S, Park H, Park SH. A review of deep learning-based reconstruction methods for accelerated MRI using spatiotemporal and multi-contrast redundancies. Biomed Eng Lett 2024; 14:1221-1242. [PMID: 39465106 PMCID: PMC11502678 DOI: 10.1007/s13534-024-00425-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/01/2024] [Revised: 08/27/2024] [Accepted: 09/06/2024] [Indexed: 10/29/2024] Open
Abstract
Accelerated magnetic resonance imaging (MRI) has played an essential role in reducing data acquisition time for MRI. Acceleration can be achieved by acquiring fewer data points in k-space, which results in various artifacts in the image domain. Conventional reconstruction methods have resolved the artifacts by utilizing multi-coil information, but with limited robustness. Recently, numerous deep learning-based reconstruction methods have been developed, enabling outstanding reconstruction performances with higher acceleration. Advances in hardware and developments of specialized network architectures have produced such achievements. Besides, MRI signals contain various redundant information including multi-coil redundancy, multi-contrast redundancy, and spatiotemporal redundancy. Utilization of the redundant information combined with deep learning approaches allow not only higher acceleration, but also well-preserved details in the reconstructed images. Consequently, this review paper introduces the basic concepts of deep learning and conventional accelerated MRI reconstruction methods, followed by review of recent deep learning-based reconstruction methods that exploit various redundancies. Lastly, the paper concludes by discussing the challenges, limitations, and potential directions of future developments.
Collapse
Affiliation(s)
- Seonghyuk Kim
- School of Electrical Engineering, Korea Advanced Institute of Science and Technology, Daejeon, Republic of Korea
| | - HyunWook Park
- School of Electrical Engineering, Korea Advanced Institute of Science and Technology, Daejeon, Republic of Korea
| | - Sung-Hong Park
- School of Electrical Engineering, Korea Advanced Institute of Science and Technology, Daejeon, Republic of Korea
- Department of Bio and Brain Engineering, Korea Advanced Institute of Science and Technology, 291 Daehak-ro, Yuseong-gu, Daejeon, 34141 Republic of Korea
| |
Collapse
|
8
|
Asgariandehkordi H, Goudarzi S, Sharifzadeh M, Basarab A, Rivaz H. Denoising Plane Wave Ultrasound Images Using Diffusion Probabilistic Models. IEEE TRANSACTIONS ON ULTRASONICS, FERROELECTRICS, AND FREQUENCY CONTROL 2024; 71:1526-1539. [PMID: 39186422 DOI: 10.1109/tuffc.2024.3448209] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/28/2024]
Abstract
Ultrasound plane wave (PW) imaging is a cutting-edge technique that enables high frame-rate imaging. However, one challenge associated with high frame-rate ultrasound imaging is the high noise associated with them, hindering their wider adoption. Therefore, the development of a denoising method becomes imperative to augment the quality of PW images. Drawing inspiration from denoising diffusion probabilistic models (DDPMs), our proposed solution aims to enhance PW image quality. Specifically, the method considers the distinction between low-angle and high-angle compounding PWs as noise and effectively eliminates it by adapting a DDPM to beamformed radio frequency (RF) data. The method underwent training using only 400 simulated images. In addition, our approach employs natural image segmentation masks as intensity maps for the generated images, resulting in accurate denoising for various anatomy shapes. The proposed method was assessed across simulation, phantom, and in vivo images. The results of the evaluations indicate that our approach not only enhances the image quality on simulated data but also demonstrates effectiveness on phantom and in vivo data in terms of image quality. Comparative analysis with other methods underscores the superiority of our proposed method across various evaluation metrics. The source code and trained model will be released along with the dataset at: http://code.sonography.ai.
Collapse
|
9
|
Lu Y, Xu Z, Hyung Choi M, Kim J, Jung SW. Cross-Domain Denoising for Low-Dose Multi-Frame Spiral Computed Tomography. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:3949-3963. [PMID: 38787677 DOI: 10.1109/tmi.2024.3405024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/26/2024]
Abstract
Computed tomography (CT) has been used worldwide as a non-invasive test to assist in diagnosis. However, the ionizing nature of X-ray exposure raises concerns about potential health risks such as cancer. The desire for lower radiation doses has driven researchers to improve reconstruction quality. Although previous studies on low-dose computed tomography (LDCT) denoising have demonstrated the effectiveness of learning-based methods, most were developed on the simulated data. However, the real-world scenario differs significantly from the simulation domain, especially when using the multi-slice spiral scanner geometry. This paper proposes a two-stage method for the commercially available multi-slice spiral CT scanners that better exploits the complete reconstruction pipeline for LDCT denoising across different domains. Our approach makes good use of the high redundancy of multi-slice projections and the volumetric reconstructions while leveraging the over-smoothing issue in conventional cascaded frameworks caused by aggressive denoising. The dedicated design also provides a more explicit interpretation of the data flow. Extensive experiments on various datasets showed that the proposed method could remove up to 70% of noise without compromised spatial resolution, while subjective evaluations by two experienced radiologists further supported its superior performance against state-of-the-art methods in clinical practice. Code is available at https://github.com/YCL92/TMD-LDCT.
Collapse
|
10
|
Mohammadi N, Goswami S, Kabir IE, Khan S, Feng F, McAleavey S, Doyley MM, Cetin M. Integrating Learning-Based Priors With Physics-Based Models in Ultrasound Elasticity Reconstruction. IEEE TRANSACTIONS ON ULTRASONICS, FERROELECTRICS, AND FREQUENCY CONTROL 2024; 71:1406-1419. [PMID: 38913531 DOI: 10.1109/tuffc.2024.3417905] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/26/2024]
Abstract
Ultrasound elastography images, which enable quantitative visualization of tissue stiffness, can be reconstructed by solving an inverse problem. Classical model-based methods are usually formulated in terms of constrained optimization problems. To stabilize the elasticity reconstructions, regularization techniques, such as Tikhonov method, are used with the cost of promoting smoothness and blurriness in the reconstructed images. Thus, incorporating a suitable regularizer is essential for reducing the elasticity reconstruction artifacts, while finding the most suitable one is challenging. In this work, we present a new statistical representation of the physical imaging model, which incorporates effective signal-dependent colored noise modeling. Moreover, we develop a learning-based integrated statistical framework, which combines a physical model with learning-based priors. We use a dataset of simulated phantoms with various elasticity distributions and geometric patterns to train a denoising regularizer as the learning-based prior. We use fixed-point approaches and variants of gradient descent for solving the integrated optimization task following learning-based plug-and-play (PnP) prior and regularization by denoising (RED) paradigms. Finally, we evaluate the performance of the proposed approaches in terms of relative mean square error (RMSE) with nearly 20% improvement for both piecewise smooth simulated phantoms and experimental phantoms compared with the classical model-based methods and 12% improvement for both spatially varying breast-mimicking simulated phantoms and an experimental breast phantom, demonstrating the potential clinical relevance of our work. Moreover, the qualitative comparisons of reconstructed images demonstrate the robust performance of the proposed methods even for complex elasticity structures that might be encountered in clinical settings.
Collapse
|
11
|
Wu W, Pan J, Wang Y, Wang S, Zhang J. Multi-Channel Optimization Generative Model for Stable Ultra-Sparse-View CT Reconstruction. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:3461-3475. [PMID: 38466593 DOI: 10.1109/tmi.2024.3376414] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/13/2024]
Abstract
Score-based generative model (SGM) has risen to prominence in sparse-view CT reconstruction due to its impressive generation capability. The consistency of data is crucial in guiding the reconstruction process in SGM-based reconstruction methods. However, the existing data consistency policy exhibits certain limitations. Firstly, it employs partial data from the reconstructed image of the iteration process for image updates, which leads to secondary artifacts with compromising image quality. Moreover, the updates to the SGM and data consistency are considered as distinct stages, disregarding their interdependent relationship. Additionally, the reference image used to compute gradients in the reconstruction process is derived from the intermediate result rather than ground truth. Motivated by the fact that a typical SGM yields distinct outcomes with different random noise inputs, we propose a Multi-channel Optimization Generative Model (MOGM) for stable ultra-sparse-view CT reconstruction by integrating a novel data consistency term into the stochastic differential equation model. Notably, the unique aspect of this data consistency component is its exclusive reliance on original data for effectively confining generation outcomes. Furthermore, we pioneer an inference strategy that traces back from the current iteration result to ground truth, enhancing reconstruction stability through foundational theoretical support. We also establish a multi-channel optimization reconstruction framework, where conventional iterative techniques are employed to seek the reconstruction solution. Quantitative and qualitative assessments on 23 views datasets from numerical simulation, clinical cardiac and sheep's lung underscore the superiority of MOGM over alternative methods. Reconstructing from just 10 and 7 views, our method consistently demonstrates exceptional performance.
Collapse
|
12
|
Zhang J, Mao H, Wang X, Guo Y, Wu W. Wavelet-Inspired Multi-Channel Score-Based Model for Limited-Angle CT Reconstruction. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:3436-3448. [PMID: 38373130 DOI: 10.1109/tmi.2024.3367167] [Citation(s) in RCA: 6] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/21/2024]
Abstract
Score-based generative model (SGM) has demonstrated great potential in the challenging limited-angle CT (LA-CT) reconstruction. SGM essentially models the probability density of the ground truth data and generates reconstruction results by sampling from it. Nevertheless, direct application of the existing SGM methods to LA-CT suffers multiple limitations. Firstly, the directional distribution of the artifacts attributing to the missing angles is ignored. Secondly, the different distribution properties of the artifacts in different frequency components have not been fully explored. These drawbacks would inevitably degrade the estimation of the probability density and the reconstruction results. After an in-depth analysis of these factors, this paper proposes a Wavelet-Inspired Score-based Model (WISM) for LA-CT reconstruction. Specifically, besides training a typical SGM with the original images, the proposed method additionally performs the wavelet transform and models the probability density in each wavelet component with an extra SGM. The wavelet components preserve the spatial correspondence with the original image while performing frequency decomposition, thereby keeping the directional property of the artifacts for further analysis. On the other hand, different wavelet components possess more specific contents of the original image in different frequency ranges, simplifying the probability density modeling by decomposing the overall density into component-wise ones. The resulting two SGMs in the image-domain and wavelet-domain are integrated into a unified sampling process under the guidance of the observation data, jointly generating high-quality and consistent LA-CT reconstructions. The experimental evaluation on various datasets consistently verifies the superior performance of the proposed method over the competing method.
Collapse
|
13
|
Pan J, Chang D, Wu W, Chen Y, Wang S. Self-supervised tomographic image noise suppression via residual image prior network. Comput Biol Med 2024; 179:108837. [PMID: 38991317 DOI: 10.1016/j.compbiomed.2024.108837] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/10/2024] [Revised: 05/29/2024] [Accepted: 06/29/2024] [Indexed: 07/13/2024]
Abstract
Computed tomography (CT) denoising is a challenging task in medical imaging that has garnered considerable attention. Supervised networks require a lot of noisy-clean image pairs, which are always unavailable in clinical settings. Existing self-supervised algorithms for suppressing noise with paired noisy images have limitations, such as ignoring the residual between similar image pairs during training and insufficiently learning the spectrum information of images. In this study, we propose a Residual Image Prior Network (RIP-Net) to sufficiently model the residual between the paired similar noisy images. Our approach offers new insights into the field by addressing the limitations of existing methods. We first establish a mathematical theorem clarifying the non-equivalence between similar-image-based self-supervised learning and supervised learning. It helps us better understand the strengths and limitations of self-supervised learning. Secondly, we introduce a novel regularization term to model a low-frequency residual image prior. This can improve the accuracy and robustness of our model. Finally, we design a well-structured denoising network capable of exploring spectrum information while simultaneously sensing context messages. The network has dual paths for modeling high and low-frequency compositions in the raw noisy image. Additionally, context perception modules capture local and global interactions to produce high-quality images. The comprehensive experiments on preclinical photon-counting CT, clinical brain CT, and low-dose CT datasets, demonstrate that our RIP-Net is superior to other unsupervised denoising methods.
Collapse
Affiliation(s)
- Jiayi Pan
- School of Biomedical Engineering, Sun Yat-Sen University, Shenzhen, Guangdong, China.
| | - Dingyue Chang
- Institute of Materials, China Academy of Engineering Physics, Mianyang, Sichuan, China
| | - Weiwen Wu
- School of Biomedical Engineering, Sun Yat-Sen University, Shenzhen, Guangdong, China.
| | - Yang Chen
- School of Computer Science and Engineering, Southeast University, Nanjing, Jiangsu, China.
| | - Shaoyu Wang
- School of Information Engineering, Nanchang University, Nanchang, Jiangxi, China.
| |
Collapse
|
14
|
Li Y, Chen G, Wang G, Zhou Z, An S, Dai S, Jin Y, Zhang C, Zhang M, Yu F. Dominating Alzheimer's disease diagnosis with deep learning on sMRI and DTI-MD. Front Neurol 2024; 15:1444795. [PMID: 39211812 PMCID: PMC11358067 DOI: 10.3389/fneur.2024.1444795] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/06/2024] [Accepted: 07/25/2024] [Indexed: 09/04/2024] Open
Abstract
Background Alzheimer's disease (AD) is a progressive and irreversible neurodegenerative disorder that has become one of the major health concerns for the elderly. Computer-aided AD diagnosis can assist doctors in quickly and accurately determining patients' severity and affected regions. Methods In this paper, we propose a method called MADNet for computer-aided AD diagnosis using multimodal datasets. The method selects ResNet-10 as the backbone network, with dual-branch parallel extraction of discriminative features for AD classification. It incorporates long-range dependencies modeling using attention scores in the decision-making layer and fuses the features based on their importance across modalities. To validate the effectiveness of our proposed multimodal classification method, we construct a multimodal dataset based on the publicly available ADNI dataset and a collected XWNI dataset, which includes examples of AD, Mild Cognitive Impairment (MCI), and Cognitively Normal (CN). Results On this dataset, we conduct binary classification experiments of AD vs. CN and MCI vs. CN, and demonstrate that our proposed method outperforms other traditional single-modal deep learning models. Furthermore, this conclusion also confirms the necessity of using multimodal sMRI and DTI data for computer-aided AD diagnosis, as these two modalities complement and convey information to each other. We visualize the feature maps extracted by MADNet using Grad-CAM, generating heatmaps that guide doctors' attention to important regions in patients' sMRI, which play a crucial role in the development of AD, establishing trust between human experts and machine learning models. Conclusion We propose a simple yet effective multimodal deep convolutional neural network model MADNet that outperforms traditional deep learning methods that use a single-modality dataset for AD diagnosis.
Collapse
Affiliation(s)
- Yuxia Li
- Department of Neurology, Tangshan Central Hospital, Hebei, China
| | - Guanqun Chen
- Department of Neurology, Beijing Chao-Yang Hospital, Capital Medical University, Beijing, China
| | - Guoxin Wang
- College of Biomedical Engineering and Instrument Science, Zhejiang University, Hangzhou, China
| | - Zhiyi Zhou
- JD Health International Inc., Beijing, China
| | - Shan An
- JD Health International Inc., Beijing, China
| | - Shipeng Dai
- College of Science, Northeastern University, Shenyang, China
| | - Yuxin Jin
- JD Health International Inc., Beijing, China
| | - Chao Zhang
- JD Health International Inc., Beijing, China
| | - Mingkai Zhang
- Department of Neurology, XuanWu Hospital of Capital Medical University, Beijing, China
| | - Feng Yu
- College of Biomedical Engineering and Instrument Science, Zhejiang University, Hangzhou, China
| |
Collapse
|
15
|
Xie J, Shao HC, Li Y, Zhang Y. Prior frequency guided diffusion model for limited angle (LA)-CBCT reconstruction. Phys Med Biol 2024; 69:135008. [PMID: 38870947 PMCID: PMC11218670 DOI: 10.1088/1361-6560/ad580d] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/01/2024] [Revised: 05/29/2024] [Accepted: 06/13/2024] [Indexed: 06/15/2024]
Abstract
Objective.Cone-beam computed tomography (CBCT) is widely used in image-guided radiotherapy. Reconstructing CBCTs from limited-angle acquisitions (LA-CBCT) is highly desired for improved imaging efficiency, dose reduction, and better mechanical clearance. LA-CBCT reconstruction, however, suffers from severe under-sampling artifacts, making it a highly ill-posed inverse problem. Diffusion models can generate data/images by reversing a data-noising process through learned data distributions; and can be incorporated as a denoiser/regularizer in LA-CBCT reconstruction. In this study, we developed a diffusion model-based framework, prior frequency-guided diffusion model (PFGDM), for robust and structure-preserving LA-CBCT reconstruction.Approach.PFGDM uses a conditioned diffusion model as a regularizer for LA-CBCT reconstruction, and the condition is based on high-frequency information extracted from patient-specific prior CT scans which provides a strong anatomical prior for LA-CBCT reconstruction. Specifically, we developed two variants of PFGDM (PFGDM-A and PFGDM-B) with different conditioning schemes. PFGDM-A applies the high-frequency CT information condition until a pre-optimized iteration step, and drops it afterwards to enable both similar and differing CT/CBCT anatomies to be reconstructed. PFGDM-B, on the other hand, continuously applies the prior CT information condition in every reconstruction step, while with a decaying mechanism, to gradually phase out the reconstruction guidance from the prior CT scans. The two variants of PFGDM were tested and compared with current available LA-CBCT reconstruction solutions, via metrics including peak signal-to-noise ratio (PSNR) and structural similarity index measure (SSIM).Main results.PFGDM outperformed all traditional and diffusion model-based methods. The mean(s.d.) PSNR/SSIM were 27.97(3.10)/0.949(0.027), 26.63(2.79)/0.937(0.029), and 23.81(2.25)/0.896(0.036) for PFGDM-A, and 28.20(1.28)/0.954(0.011), 26.68(1.04)/0.941(0.014), and 23.72(1.19)/0.894(0.034) for PFGDM-B, based on 120°, 90°, and 30° orthogonal-view scan angles respectively. In contrast, the PSNR/SSIM was 19.61(2.47)/0.807(0.048) for 30° for DiffusionMBIR, a diffusion-based method without prior CT conditioning.Significance. PFGDM reconstructs high-quality LA-CBCTs under very-limited gantry angles, allowing faster and more flexible CBCT scans with dose reductions.
Collapse
Affiliation(s)
- Jiacheng Xie
- The Advanced Imaging and Informatics for Radiation Therapy (AIRT) Laboratory, The Medical Artificial Intelligence and Automation (MAIA) Laboratory, Department of Radiation Oncology, University of Texas Southwestern Medical Center, Dallas, TX 75390, United States of America
| | - Hua-Chieh Shao
- The Advanced Imaging and Informatics for Radiation Therapy (AIRT) Laboratory, The Medical Artificial Intelligence and Automation (MAIA) Laboratory, Department of Radiation Oncology, University of Texas Southwestern Medical Center, Dallas, TX 75390, United States of America
| | - Yunxiang Li
- The Advanced Imaging and Informatics for Radiation Therapy (AIRT) Laboratory, The Medical Artificial Intelligence and Automation (MAIA) Laboratory, Department of Radiation Oncology, University of Texas Southwestern Medical Center, Dallas, TX 75390, United States of America
| | - You Zhang
- The Advanced Imaging and Informatics for Radiation Therapy (AIRT) Laboratory, The Medical Artificial Intelligence and Automation (MAIA) Laboratory, Department of Radiation Oncology, University of Texas Southwestern Medical Center, Dallas, TX 75390, United States of America
| |
Collapse
|
16
|
Li X, Jing K, Yang Y, Wang Y, Ma J, Zheng H, Xu Z. Noise-Generating and Imaging Mechanism Inspired Implicit Regularization Learning Network for Low Dose CT Reconstrution. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:1677-1689. [PMID: 38145543 DOI: 10.1109/tmi.2023.3347258] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/27/2023]
Abstract
Low-dose computed tomography (LDCT) helps to reduce radiation risks in CT scanning while maintaining image quality, which involves a consistent pursuit of lower incident rays and higher reconstruction performance. Although deep learning approaches have achieved encouraging success in LDCT reconstruction, most of them treat the task as a general inverse problem in either the image domain or the dual (sinogram and image) domains. Such frameworks have not considered the original noise generation of the projection data and suffer from limited performance improvement for the LDCT task. In this paper, we propose a novel reconstruction model based on noise-generating and imaging mechanism in full-domain, which fully considers the statistical properties of intrinsic noises in LDCT and prior information in sinogram and image domains. To solve the model, we propose an optimization algorithm based on the proximal gradient technique. Specifically, we derive the approximate solutions of the integer programming problem on the projection data theoretically. Instead of hand-crafting the sinogram and image regularizers, we propose to unroll the optimization algorithm to be a deep network. The network implicitly learns the proximal operators of sinogram and image regularizers with two deep neural networks, providing a more interpretable and effective reconstruction procedure. Numerical results demonstrate our proposed method improvements of > 2.9 dB in peak signal to noise ratio, > 1.4% promotion in structural similarity metric, and > 9 HU decrements in root mean square error over current state-of-the-art LDCT methods.
Collapse
|
17
|
Li Y, Feng J, Xiang J, Li Z, Liang D. AIRPORT: A Data Consistency Constrained Deep Temporal Extrapolation Method To Improve Temporal Resolution In Contrast Enhanced CT Imaging. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:1605-1618. [PMID: 38133967 DOI: 10.1109/tmi.2023.3344712] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/24/2023]
Abstract
Typical tomographic image reconstruction methods require that the imaged object is static and stationary during the time window to acquire a minimally complete data set. The violation of this requirement leads to temporal-averaging errors in the reconstructed images. For a fixed gantry rotation speed, to reduce the errors, it is desired to reconstruct images using data acquired over a narrower angular range, i.e., with a higher temporal resolution. However, image reconstruction with a narrower angular range violates the data sufficiency condition, resulting in severe data-insufficiency-induced errors. The purpose of this work is to decouple the trade-off between these two types of errors in contrast-enhanced computed tomography (CT) imaging. We demonstrated that using the developed data consistency constrained deep temporal extrapolation method (AIRPORT), the entire time-varying imaged object can be accurately reconstructed with 40 frames-per-second temporal resolution, the time window needed to acquire a single projection view data using a typical C-arm cone-beam CT system. AIRPORT is applicable to general non-sparse imaging tasks using a single short-scan data acquisition.
Collapse
|
18
|
Maramraju S, Kowalczewski A, Kaza A, Liu X, Singaraju JP, Albert MV, Ma Z, Yang H. AI-organoid integrated systems for biomedical studies and applications. Bioeng Transl Med 2024; 9:e10641. [PMID: 38435826 PMCID: PMC10905559 DOI: 10.1002/btm2.10641] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/29/2023] [Revised: 12/07/2023] [Accepted: 12/13/2023] [Indexed: 03/05/2024] Open
Abstract
In this review, we explore the growing role of artificial intelligence (AI) in advancing the biomedical applications of human pluripotent stem cell (hPSC)-derived organoids. Stem cell-derived organoids, these miniature organ replicas, have become essential tools for disease modeling, drug discovery, and regenerative medicine. However, analyzing the vast and intricate datasets generated from these organoids can be inefficient and error-prone. AI techniques offer a promising solution to efficiently extract insights and make predictions from diverse data types generated from microscopy images, transcriptomics, metabolomics, and proteomics. This review offers a brief overview of organoid characterization and fundamental concepts in AI while focusing on a comprehensive exploration of AI applications in organoid-based disease modeling and drug evaluation. It provides insights into the future possibilities of AI in enhancing the quality control of organoid fabrication, label-free organoid recognition, and three-dimensional image reconstruction of complex organoid structures. This review presents the challenges and potential solutions in AI-organoid integration, focusing on the establishment of reliable AI model decision-making processes and the standardization of organoid research.
Collapse
Affiliation(s)
- Sudhiksha Maramraju
- Department of Biomedical EngineeringUniversity of North TexasDentonTexasUSA
- Texas Academy of Mathematics and ScienceUniversity of North TexasDentonTexasUSA
| | - Andrew Kowalczewski
- Department of Biomedical & Chemical EngineeringSyracuse UniversitySyracuseNew YorkUSA
- BioInspired Institute for Material and Living SystemsSyracuse UniversitySyracuseNew YorkUSA
| | - Anirudh Kaza
- Department of Biomedical EngineeringUniversity of North TexasDentonTexasUSA
- Texas Academy of Mathematics and ScienceUniversity of North TexasDentonTexasUSA
| | - Xiyuan Liu
- Department of Mechanical & Aerospace EngineeringSyracuse UniversitySyracuseNew YorkUSA
| | - Jathin Pranav Singaraju
- Department of Biomedical EngineeringUniversity of North TexasDentonTexasUSA
- Texas Academy of Mathematics and ScienceUniversity of North TexasDentonTexasUSA
| | - Mark V. Albert
- Department of Biomedical EngineeringUniversity of North TexasDentonTexasUSA
- Department of Computer Science and EngineeringUniversity of North TexasDentonTexasUSA
| | - Zhen Ma
- Department of Biomedical & Chemical EngineeringSyracuse UniversitySyracuseNew YorkUSA
- BioInspired Institute for Material and Living SystemsSyracuse UniversitySyracuseNew YorkUSA
| | - Huaxiao Yang
- Department of Biomedical EngineeringUniversity of North TexasDentonTexasUSA
| |
Collapse
|
19
|
Bousse A, Kandarpa VSS, Rit S, Perelli A, Li M, Wang G, Zhou J, Wang G. Systematic Review on Learning-based Spectral CT. IEEE TRANSACTIONS ON RADIATION AND PLASMA MEDICAL SCIENCES 2024; 8:113-137. [PMID: 38476981 PMCID: PMC10927029 DOI: 10.1109/trpms.2023.3314131] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 03/14/2024]
Abstract
Spectral computed tomography (CT) has recently emerged as an advanced version of medical CT and significantly improves conventional (single-energy) CT. Spectral CT has two main forms: dual-energy computed tomography (DECT) and photon-counting computed tomography (PCCT), which offer image improvement, material decomposition, and feature quantification relative to conventional CT. However, the inherent challenges of spectral CT, evidenced by data and image artifacts, remain a bottleneck for clinical applications. To address these problems, machine learning techniques have been widely applied to spectral CT. In this review, we present the state-of-the-art data-driven techniques for spectral CT.
Collapse
Affiliation(s)
- Alexandre Bousse
- LaTIM, Inserm UMR 1101, Université de Bretagne Occidentale, 29238 Brest, France
| | | | - Simon Rit
- Univ Lyon, INSA-Lyon, Université Claude Bernard Lyon 1, UJM-Saint Étienne, CNRS, Inserm, CREATIS UMR 5220, U1294, F-69373, Lyon, France
| | - Alessandro Perelli
- Department of Biomedical Engineering, School of Science and Engineering, University of Dundee, DD1 4HN, UK
| | - Mengzhou Li
- Biomedical Imaging Center, Rensselaer Polytechnic Institute, Troy, New York, USA
| | - Guobao Wang
- Department of Radiology, University of California Davis Health, Sacramento, USA
| | - Jian Zhou
- CTIQ, Canon Medical Research USA, Inc., Vernon Hills, 60061, USA
| | - Ge Wang
- Biomedical Imaging Center, Rensselaer Polytechnic Institute, Troy, New York, USA
| |
Collapse
|
20
|
Rakhmatulin I, Dao MS, Nassibi A, Mandic D. Exploring Convolutional Neural Network Architectures for EEG Feature Extraction. SENSORS (BASEL, SWITZERLAND) 2024; 24:877. [PMID: 38339594 PMCID: PMC10856895 DOI: 10.3390/s24030877] [Citation(s) in RCA: 13] [Impact Index Per Article: 13.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/04/2023] [Revised: 01/12/2024] [Accepted: 01/20/2024] [Indexed: 02/12/2024]
Abstract
The main purpose of this paper is to provide information on how to create a convolutional neural network (CNN) for extracting features from EEG signals. Our task was to understand the primary aspects of creating and fine-tuning CNNs for various application scenarios. We considered the characteristics of EEG signals, coupled with an exploration of various signal processing and data preparation techniques. These techniques include noise reduction, filtering, encoding, decoding, and dimension reduction, among others. In addition, we conduct an in-depth analysis of well-known CNN architectures, categorizing them into four distinct groups: standard implementation, recurrent convolutional, decoder architecture, and combined architecture. This paper further offers a comprehensive evaluation of these architectures, covering accuracy metrics, hyperparameters, and an appendix that contains a table outlining the parameters of commonly used CNN architectures for feature extraction from EEG signals.
Collapse
Affiliation(s)
- Ildar Rakhmatulin
- Department of Electrical and Electronic Engineering, Imperial College London, London SW7 2AZ, UK; (A.N.)
| | - Minh-Son Dao
- National Institute of Information and Communications Technology (NICT), Tokyo 184-0015, Japan
| | - Amir Nassibi
- Department of Electrical and Electronic Engineering, Imperial College London, London SW7 2AZ, UK; (A.N.)
| | - Danilo Mandic
- Department of Electrical and Electronic Engineering, Imperial College London, London SW7 2AZ, UK; (A.N.)
| |
Collapse
|
21
|
Li Z, Wang Y, Zhang J, Wu W, Yu H. Two-and-a-half order score-based model for solving 3D ill-posed inverse problems. Comput Biol Med 2024; 168:107819. [PMID: 38064853 DOI: 10.1016/j.compbiomed.2023.107819] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/28/2023] [Revised: 11/25/2023] [Accepted: 12/03/2023] [Indexed: 01/10/2024]
Abstract
Computed Tomography (CT) and Magnetic Resonance Imaging (MRI) are crucial technologies in the field of medical imaging. Score-based models demonstrated effectiveness in addressing different inverse problems encountered in the field of CT and MRI, such as sparse-view CT and fast MRI reconstruction. However, these models face challenges in achieving accurate three dimensional (3D) volumetric reconstruction. The existing score-based models predominantly concentrate on reconstructing two-dimensional (2D) data distributions, resulting in inconsistencies between adjacent slices in the reconstructed 3D volumetric images. To overcome this limitation, we propose a novel two-and-a-half order score-based model (TOSM). During the training phase, our TOSM learns data distributions in 2D space, simplifying the training process compared to working directly on 3D volumes. However, during the reconstruction phase, the TOSM utilizes complementary scores along three directions (sagittal, coronal, and transaxial) to achieve a more precise reconstruction. The development of TOSM is built on robust theoretical principles, ensuring its reliability and efficacy. Through extensive experimentation on large-scale sparse-view CT and fast MRI datasets, our method achieved state-of-the-art (SOTA) results in solving 3D ill-posed inverse problems, averaging a 1.56 dB peak signal-to-noise ratio (PSNR) improvement over existing sparse-view CT reconstruction methods across 29 views and 0.87 dB PSNR improvement over existing fast MRI reconstruction methods with × 2 acceleration. In summary, TOSM significantly addresses the issue of inconsistency in 3D ill-posed problems by modeling the distribution of 3D data rather than 2D distribution which has achieved remarkable results in both CT and MRI reconstruction tasks.
Collapse
Affiliation(s)
- Zirong Li
- Department of Biomedical Engineering, Sun-Yat-sen University, Shenzhen Campus, Shenzhen, China
| | - Yanyang Wang
- Department of Biomedical Engineering, Sun-Yat-sen University, Shenzhen Campus, Shenzhen, China
| | - Jianjia Zhang
- Department of Biomedical Engineering, Sun-Yat-sen University, Shenzhen Campus, Shenzhen, China
| | - Weiwen Wu
- Department of Biomedical Engineering, Sun-Yat-sen University, Shenzhen Campus, Shenzhen, China.
| | - Hengyong Yu
- Department of Electrical and Computer Engineering, University of Massachusetts Lowell, Lowell, MA, USA.
| |
Collapse
|
22
|
Rahman H, Khan AR, Sadiq T, Farooqi AH, Khan IU, Lim WH. A Systematic Literature Review of 3D Deep Learning Techniques in Computed Tomography Reconstruction. Tomography 2023; 9:2158-2189. [PMID: 38133073 PMCID: PMC10748093 DOI: 10.3390/tomography9060169] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/13/2023] [Revised: 11/27/2023] [Accepted: 12/01/2023] [Indexed: 12/23/2023] Open
Abstract
Computed tomography (CT) is used in a wide range of medical imaging diagnoses. However, the reconstruction of CT images from raw projection data is inherently complex and is subject to artifacts and noise, which compromises image quality and accuracy. In order to address these challenges, deep learning developments have the potential to improve the reconstruction of computed tomography images. In this regard, our research aim is to determine the techniques that are used for 3D deep learning in CT reconstruction and to identify the training and validation datasets that are accessible. This research was performed on five databases. After a careful assessment of each record based on the objective and scope of the study, we selected 60 research articles for this review. This systematic literature review revealed that convolutional neural networks (CNNs), 3D convolutional neural networks (3D CNNs), and deep learning reconstruction (DLR) were the most suitable deep learning algorithms for CT reconstruction. Additionally, two major datasets appropriate for training and developing deep learning systems were identified: 2016 NIH-AAPM-Mayo and MSCT. These datasets are important resources for the creation and assessment of CT reconstruction models. According to the results, 3D deep learning may increase the effectiveness of CT image reconstruction, boost image quality, and lower radiation exposure. By using these deep learning approaches, CT image reconstruction may be made more precise and effective, improving patient outcomes, diagnostic accuracy, and healthcare system productivity.
Collapse
Affiliation(s)
- Hameedur Rahman
- Department of Computer Games Development, Faculty of Computing & AI, Air University, E9, Islamabad 44000, Pakistan;
| | - Abdur Rehman Khan
- Department of Creative Technologies, Faculty of Computing & AI, Air University, E9, Islamabad 44000, Pakistan;
| | - Touseef Sadiq
- Centre for Artificial Intelligence Research, Department of Information and Communication Technology, University of Agder, Jon Lilletuns vei 9, 4879 Grimstad, Norway
| | - Ashfaq Hussain Farooqi
- Department of Computer Science, Faculty of Computing AI, Air University, Islamabad 44000, Pakistan;
| | - Inam Ullah Khan
- Department of Electronic Engineering, School of Engineering & Applied Sciences (SEAS), Isra University, Islamabad Campus, Islamabad 44000, Pakistan;
| | - Wei Hong Lim
- Faculty of Engineering, Technology and Built Environment, UCSI University, Kuala Lumpur 56000, Malaysia;
| |
Collapse
|
23
|
Li T, Xu Y, Wu T, Charlton JR, Bennett KM, Al-Hindawi F. BlobCUT: A Contrastive Learning Method to Support Small Blob Detection in Medical Imaging. Bioengineering (Basel) 2023; 10:1372. [PMID: 38135963 PMCID: PMC10740534 DOI: 10.3390/bioengineering10121372] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/16/2023] [Revised: 11/19/2023] [Accepted: 11/27/2023] [Indexed: 12/24/2023] Open
Abstract
Medical imaging-based biomarkers derived from small objects (e.g., cell nuclei) play a crucial role in medical applications. However, detecting and segmenting small objects (a.k.a. blobs) remains a challenging task. In this research, we propose a novel 3D small blob detector called BlobCUT. BlobCUT is an unpaired image-to-image (I2I) translation model that falls under the Contrastive Unpaired Translation paradigm. It employs a blob synthesis module to generate synthetic 3D blobs with corresponding masks. This is incorporated into the iterative model training as the ground truth. The I2I translation process is designed with two constraints: (1) a convexity consistency constraint that relies on Hessian analysis to preserve the geometric properties and (2) an intensity distribution consistency constraint based on Kullback-Leibler divergence to preserve the intensity distribution of blobs. BlobCUT learns the inherent noise distribution from the target noisy blob images and performs image translation from the noisy domain to the clean domain, effectively functioning as a denoising process to support blob identification. To validate the performance of BlobCUT, we evaluate it on a 3D simulated dataset of blobs and a 3D MRI dataset of mouse kidneys. We conduct a comparative analysis involving six state-of-the-art methods. Our findings reveal that BlobCUT exhibits superior performance and training efficiency, utilizing only 56.6% of the training time required by the state-of-the-art BlobDetGAN. This underscores the effectiveness of BlobCUT in accurately segmenting small blobs while achieving notable gains in training efficiency.
Collapse
Affiliation(s)
- Teng Li
- School of Computing and Augmented Intelligence, Arizona State University, Tempe, AZ 85281, USA; (T.L.); (Y.X.); (F.A.-H.)
| | - Yanzhe Xu
- School of Computing and Augmented Intelligence, Arizona State University, Tempe, AZ 85281, USA; (T.L.); (Y.X.); (F.A.-H.)
| | - Teresa Wu
- School of Computing and Augmented Intelligence, Arizona State University, Tempe, AZ 85281, USA; (T.L.); (Y.X.); (F.A.-H.)
| | - Jennifer R. Charlton
- Division Nephrology, Department of Pediatrics, University of Virginia, Charlottesville, VA 22903, USA;
| | - Kevin M. Bennett
- Department of Radiology, Washington University, St. Louis, MO 63130, USA;
| | - Firas Al-Hindawi
- School of Computing and Augmented Intelligence, Arizona State University, Tempe, AZ 85281, USA; (T.L.); (Y.X.); (F.A.-H.)
| |
Collapse
|
24
|
Hu D, Zhang Y, Li W, Zhang W, Reddy K, Ding Q, Zhang X, Chen Y, Gao H. SEA-Net: Structure-Enhanced Attention Network for Limited-Angle CBCT Reconstruction of Clinical Projection Data. IEEE TRANSACTIONS ON INSTRUMENTATION AND MEASUREMENT 2023; 72:4507613. [PMID: 38957474 PMCID: PMC11218899 DOI: 10.1109/tim.2023.3318712] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/04/2024]
Abstract
This work aims to improve limited-angle (LA) cone beam computed tomography (CBCT) by developing deep learning (DL) methods for real clinical CBCT projection data, which is the first feasibility study of clinical-projection-data-based LA-CBCT, to the best of our knowledge. In radiation therapy (RT), CBCT is routinely used as the on-board imaging modality for patient setup. Compared to diagnostic CT, CBCT has a long acquisition time, e.g., 60 seconds for a full 360° rotation, which is subject to the motion artifact. Therefore, the LA-CBCT, if achievable, is of the great interest for the purpose of RT, for its proportionally reduced scanning time in addition to the radiation dose. However, LA-CBCT suffers from severe wedge artifacts and image distortions. Targeting at real clinical projection data, we have explored various DL methods such as image/data/hybrid-domain methods and finally developed a so-called Structure-Enhanced Attention Network (SEA-Net) method that has the best image quality from clinical projection data among the DL methods we have implemented. Specifically, the proposed SEA-Net employs a specialized structure enhancement sub-network to promote texture preservation. Based on the observation that the distribution of wedge artifacts in reconstruction images is non-uniform, the spatial attention module is utilized to emphasize the relevant regions while ignores the irrelevant ones, which leads to more accurate texture restoration.
Collapse
Affiliation(s)
- Dianlin Hu
- Department of Radiation Oncology, University of Kansas Medical Center, KS 66160, USA
- Jiangsu Provincial Joint International Research Laboratory of Medical Information Processing, the Laboratory of Image Science and Technology, the School of Computer Science and Engineering, and the Key Laboratory of New Generation Artificial Intelligence Technology and Its Interdisciplinary Applications (Southeast University), Ministry of Education, Nanjing 210096, China
| | - Yikun Zhang
- Department of Radiation Oncology, University of Kansas Medical Center, KS 66160, USA
- Jiangsu Provincial Joint International Research Laboratory of Medical Information Processing, the Laboratory of Image Science and Technology, the School of Computer Science and Engineering, and the Key Laboratory of New Generation Artificial Intelligence Technology and Its Interdisciplinary Applications (Southeast University), Ministry of Education, Nanjing 210096, China
| | - Wangyao Li
- Department of Radiation Oncology, University of Kansas Medical Center, KS 66160, USA
| | - Weijie Zhang
- Department of Radiation Oncology, University of Kansas Medical Center, KS 66160, USA
| | - Krishna Reddy
- Department of Radiation Oncology, University of Kansas Medical Center, KS 66160, USA
| | - Qiaoqiao Ding
- Institute of Natural Sciences & School of Mathematical Sciences & MOE-LSC & SJTU-GenSci Joint Lab, Shanghai Jiao Tong University, Shanghai, 200240, China
| | - Xiaoqun Zhang
- Institute of Natural Sciences & School of Mathematical Sciences & MOE-LSC & SJTU-GenSci Joint Lab, Shanghai Jiao Tong University, Shanghai, 200240, China
| | - Yang Chen
- Jiangsu Provincial Joint International Research Laboratory of Medical Information Processing, the Laboratory of Image Science and Technology, the School of Computer Science and Engineering, and the Key Laboratory of New Generation Artificial Intelligence Technology and Its Interdisciplinary Applications (Southeast University), Ministry of Education, Nanjing 210096, China
| | - Hao Gao
- Department of Radiation Oncology, University of Kansas Medical Center, KS 66160, USA
| |
Collapse
|
25
|
Lim H, Dewaraja YK, Fessler JA. SPECT reconstruction with a trained regularizer using CT-side information: Application to 177Lu SPECT imaging. IEEE TRANSACTIONS ON COMPUTATIONAL IMAGING 2023; 9:846-856. [PMID: 38516350 PMCID: PMC10956080 DOI: 10.1109/tci.2023.3318993] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/23/2024]
Abstract
Improving low-count SPECT can shorten scans and support pre-therapy theranostic imaging for dosimetry-based treatment planning, especially with radionuclides like 177Lu known for low photon yields. Conventional methods often underperform in low-count settings, highlighting the need for trained regularization in model-based image reconstruction. This paper introduces a trained regularizer for SPECT reconstruction that leverages segmentation based on CT imaging. The regularizer incorporates CT-side information via a segmentation mask from a pre-trained network (nnUNet). In this proof-of-concept study, we used patient studies with 177Lu DOTATATE to train and tested with phantom and patient datasets, simulating pre-therapy imaging conditions. Our results show that the proposed method outperforms both standard unregularized EM algorithms and conventional regularization with CT-side information. Specifically, our method achieved marked improvements in activity quantification, noise reduction, and root mean square error. The enhanced low-count SPECT approach has promising implications for theranostic imaging, post-therapy imaging, whole body SPECT, and reducing SPECT acquisition times.
Collapse
Affiliation(s)
- Hongki Lim
- Department of Electronic Engineering, Inha University, Incheon, 22212, South Korea
| | - Yuni K Dewaraja
- Department of Electrical Engineering and Computer Science, University of Michigan, Ann Arbor, MI 48109 USA
| | - Jeffrey A Fessler
- Department of Radiology, University of Michigan, Ann Arbor, MI 48109 USA
| |
Collapse
|
26
|
Cheng W, He J, Liu Y, Zhang H, Wang X, Liu Y, Zhang P, Chen H, Gui Z. CAIR: Combining integrated attention with iterative optimization learning for sparse-view CT reconstruction. Comput Biol Med 2023; 163:107161. [PMID: 37311381 DOI: 10.1016/j.compbiomed.2023.107161] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/05/2023] [Revised: 05/21/2023] [Accepted: 06/07/2023] [Indexed: 06/15/2023]
Abstract
Sparse-view CT is an efficient way for low dose scanning but degrades image quality. Inspired by the successful use of non-local attention in natural image denoising and compression artifact removal, we proposed a network combining integrated attention and iterative optimization learning for sparse-view CT reconstruction (CAIR). Specifically, we first unrolled the proximal gradient descent into a deep network and added an enhanced initializer between the gradient term and the approximation term. It can enhance the information flow between different layers, fully preserve the image details, and improve the network convergence speed. Secondly, the integrated attention module was introduced into the reconstruction process as a regularization term. It adaptively fuses the local and non-local features of the image which are used to reconstruct the complex texture and repetitive details of the image, respectively. Note that we innovatively designed a one-shot iteration strategy to simplify the network structure and reduce the reconstruction time while maintaining image quality. Experiments showed that the proposed method is very robust and outperforms state-of-the-art methods in terms of both quantitative and qualitative, greatly improving the preservation of structures and the removal of artifacts.
Collapse
Affiliation(s)
- Weiting Cheng
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, 030051, China
| | - Jichun He
- School of Medical and BioInformation Engineering, Northeastern University, Shenyang, 110000, China
| | - Yi Liu
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, 030051, China
| | - Haowen Zhang
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, 030051, China
| | - Xiang Wang
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, 030051, China
| | - Yuhang Liu
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, 030051, China
| | - Pengcheng Zhang
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, 030051, China
| | - Hao Chen
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, 030051, China
| | - Zhiguo Gui
- State Key Laboratory of Dynamic Testing Technology, North University of China, Taiyuan, 030051, China.
| |
Collapse
|
27
|
Sarvari AVP, Sridevi K. An optimized EBRSA-Bi LSTM model for highly undersampled rapid CT image reconstruction. Biomed Signal Process Control 2023; 83:104637. [PMID: 36776947 PMCID: PMC9904992 DOI: 10.1016/j.bspc.2023.104637] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/02/2022] [Revised: 12/22/2022] [Accepted: 01/25/2023] [Indexed: 02/10/2023]
Abstract
COVID-19 has spread all over the world, causing serious panic around the globe. Chest computed tomography (CT) images are integral in confirming COVID positive patients. Several investigations were conducted to improve or maintain the image reconstruction quality for the sample image reconstruction. Deep learning (DL) methods have recently been proposed to achieve fast reconstruction, but many have focused on a single domain, such as the image domain of k-space. In this research, the highly under-sampled enhanced battle royale self-attention based bi-directional long short-term (EBRSA-bi LSTM) CT image reconstruction model is proposed to reconstruct the image from the under-sampled data. The research is adapted with two phases, namely, pre-processing and reconstruction. The extended cascaded filter (ECF) is proposed for image pre-processing and tends to suppress the noise and enhance the reconstruction accuracy. In the reconstruction model, the battle royale optimization (BrO) is intended to diminish the loss function of the reconstruction network model and weight updation. The proposed model is tested with two datasets, COVID-CT- and SARS-CoV-2 CT. The reconstruction accuracy of the proposed model with two datasets is 93.5 % and 97.7 %, respectively. Also, the image quality assessment parameters such as Peak-Signal to Noise Ratio (PSNR), Root Mean Square Error (RMSE) and Structural Similarity Index metric (SSIM) are evaluated, and it yields an outcome of (45 and 46 dB), (0.0026 and 0.0022) and (0.992, 0.996) with two datasets.
Collapse
Affiliation(s)
- A V P Sarvari
- Department of Electronics and Communication Engineering, GITAM Deemed to be University, Andhra Pradesh 530045, India
| | - K Sridevi
- Department of Electronics and Communication Engineering, GITAM Deemed to be University, Andhra Pradesh 530045, India
| |
Collapse
|
28
|
Song M, Yang Z, Li P, Zhao Z, Liu Y, Yu Y, Wu LA. Single-pixel imaging with high spectral and spatial resolution. APPLIED OPTICS 2023; 62:2610-2616. [PMID: 37132810 DOI: 10.1364/ao.479069] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/04/2023]
Abstract
It has long been a challenge to obtain high spectral and spatial resolution simultaneously for the field of measurement and detection. Here we present a measurement system based on single-pixel imaging with compressive sensing that can realize excellent spectral and spatial resolution at the same time, as well as data compression. Our method can achieve high spectral and spatial resolution, which is different from the mutually restrictive relationship between the two in traditional imaging. In our experiments, 301 spectral channels are obtained in the band of 420-780 nm with a spectral resolution of 1.2 nm and a spatial resolution of 1.11 mrad. A sampling rate of 12.5% for a 64×64p i x e l image is obtained by using compressive sensing, which also reduces the measurement time; thus, high spectral and spatial resolution are realized simultaneously, even at a low sampling rate.
Collapse
|
29
|
Chun IY, Huang Z, Lim H, Fessler JA. Momentum-Net: Fast and Convergent Iterative Neural Network for Inverse Problems. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE 2023; 45:4915-4931. [PMID: 32750839 PMCID: PMC8011286 DOI: 10.1109/tpami.2020.3012955] [Citation(s) in RCA: 27] [Impact Index Per Article: 13.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/25/2023]
Abstract
Iterative neural networks (INN) are rapidly gaining attention for solving inverse problems in imaging, image processing, and computer vision. INNs combine regression NNs and an iterative model-based image reconstruction (MBIR) algorithm, often leading to both good generalization capability and outperforming reconstruction quality over existing MBIR optimization models. This paper proposes the first fast and convergent INN architecture, Momentum-Net, by generalizing a block-wise MBIR algorithm that uses momentum and majorizers with regression NNs. For fast MBIR, Momentum-Net uses momentum terms in extrapolation modules, and noniterative MBIR modules at each iteration by using majorizers, where each iteration of Momentum-Net consists of three core modules: image refining, extrapolation, and MBIR. Momentum-Net guarantees convergence to a fixed-point for general differentiable (non)convex MBIR functions (or data-fit terms) and convex feasible sets, under two asymptomatic conditions. To consider data-fit variations across training and testing samples, we also propose a regularization parameter selection scheme based on the "spectral spread" of majorization matrices. Numerical experiments for light-field photography using a focal stack and sparse-view computational tomography demonstrate that, given identical regression NN architectures, Momentum-Net significantly improves MBIR speed and accuracy over several existing INNs; it significantly improves reconstruction quality compared to a state-of-the-art MBIR method in each application.
Collapse
|
30
|
Du C, Qiao Z. EPRI sparse reconstruction method based on deep learning. Magn Reson Imaging 2023; 97:24-30. [PMID: 36493992 DOI: 10.1016/j.mri.2022.12.008] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/11/2022] [Revised: 11/03/2022] [Accepted: 12/04/2022] [Indexed: 12/12/2022]
Abstract
Electron paramagnetic resonance imaging (EPRI) is an advanced tumor oxygen concentration imaging method. Now, the bottleneck problem of EPRI is that the scanning time is too long. Sparse reconstruction is an effective and fast imaging method, which means reconstructing images from sparse-view projections. However, the EPRI images sparsely reconstructed by the classic filtered back projection (FBP) algorithm often contain severe streak artifacts, which affect subsequent image processing. In this work, we propose a feature pyramid attention-based, residual, dense, deep convolutional network (FRD-Net) to suppress the streak artifacts in the FBP-reconstructed images. This network combines residual connection, attention mechanism, dense connections and introduces perceptual loss. The EPRI image with streak artifacts is used as the input of the network and the output-label is the corresponding high-quality image densely reconstructed by the FBP algorithm. After training, the FRD-Net gets the capability of suppressing streak artifacts. The real data reconstruction experiments show that the FRD-Net can better improve the sparse reconstruction accuracy, compared with three existing representative deep networks.
Collapse
Affiliation(s)
- Congcong Du
- School of Computer and Information Technology, Shanxi University, Taiyuan, Shanxi 030006, China
| | - Zhiwei Qiao
- School of Computer and Information Technology, Shanxi University, Taiyuan, Shanxi 030006, China.
| |
Collapse
|
31
|
Xia W, Shan H, Wang G, Zhang Y. Physics-/Model-Based and Data-Driven Methods for Low-Dose Computed Tomography: A survey. IEEE SIGNAL PROCESSING MAGAZINE 2023; 40:89-100. [PMID: 38404742 PMCID: PMC10883591 DOI: 10.1109/msp.2022.3204407] [Citation(s) in RCA: 8] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/27/2024]
Abstract
Since 2016, deep learning (DL) has advanced tomographic imaging with remarkable successes, especially in low-dose computed tomography (LDCT) imaging. Despite being driven by big data, the LDCT denoising and pure end-to-end reconstruction networks often suffer from the black box nature and major issues such as instabilities, which is a major barrier to apply deep learning methods in low-dose CT applications. An emerging trend is to integrate imaging physics and model into deep networks, enabling a hybridization of physics/model-based and data-driven elements. In this paper, we systematically review the physics/model-based data-driven methods for LDCT, summarize the loss functions and training strategies, evaluate the performance of different methods, and discuss relevant issues and future directions.
Collapse
Affiliation(s)
- Wenjun Xia
- School of Cyber Science and Engineering, Sichuan University, Chengdu 610065, China
| | - Hongming Shan
- Institute of Science and Technology for Brain-Inspired Intelligence and MOE Frontiers Center for Brain Science, Fudan University, Shanghai, and also with Shanghai Center for Brain Science and Brain-Inspired Technology, Shanghai 200433, China
| | - Ge Wang
- Department of Biomedical Engineering, Rensselaer Polytechnic Institute, Troy, NY 12180 USA
| | - Yi Zhang
- School of Cyber Science and Engineering, Sichuan University, Chengdu 610065, China
| |
Collapse
|
32
|
Clements RG, Claros-Olivares CC, McIlvain G, Brockmeier AJ, Johnson CL. Mechanical Property Based Brain Age Prediction using Convolutional Neural Networks. BIORXIV : THE PREPRINT SERVER FOR BIOLOGY 2023:2023.02.12.528186. [PMID: 36824781 PMCID: PMC9948973 DOI: 10.1101/2023.02.12.528186] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/15/2023]
Abstract
Brain age is a quantitative estimate to explain an individual's structural and functional brain measurements relative to the overall population and is particularly valuable in describing differences related to developmental or neurodegenerative pathology. Accurately inferring brain age from brain imaging data requires sophisticated models that capture the underlying age-related brain changes. Magnetic resonance elastography (MRE) is a phase contrast MRI technology that uses external palpations to measure brain mechanical properties. Mechanical property measures of viscoelastic shear stiffness and damping ratio have been found to change across the entire life span and to reflect brain health due to neurodegenerative diseases and even individual differences in cognitive function. Here we develop and train a multi-modal 3D convolutional neural network (CNN) to model the relationship between age and whole brain mechanical properties. After training, the network maps the measurements and other inputs to a brain age prediction. We found high performance using the 3D maps of various mechanical properties to predict brain age. Stiffness maps alone were able to predict ages of the test group subjects with a mean absolute error (MAE) of 3.76 years, which is comparable to single inputs of damping ratio (MAE: 3.82) and outperforms single input of volume (MAE: 4.60). Combining stiffness and volume in a multimodal approach performed the best, with an MAE of 3.60 years, whereas including damping ratio worsened model performance. Our results reflect previous MRE literature that had demonstrated that stiffness is more strongly related to chronological age than damping ratio. This machine learning model provides the first prediction of brain age from brain biomechanical data-an advancement towards sensitively describing brain integrity differences in individuals with neuropathology.
Collapse
|
33
|
Luckett PH, Lee JJ, Park KY, Raut RV, Meeker KL, Gordon EM, Snyder AZ, Ances BM, Leuthardt EC, Shimony JS. Resting state network mapping in individuals using deep learning. Front Neurol 2023; 13:1055437. [PMID: 36712434 PMCID: PMC9878609 DOI: 10.3389/fneur.2022.1055437] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2022] [Accepted: 12/28/2022] [Indexed: 01/14/2023] Open
Abstract
Introduction Resting state functional MRI (RS-fMRI) is currently used in numerous clinical and research settings. The localization of resting state networks (RSNs) has been utilized in applications ranging from group analysis of neurodegenerative diseases to individual network mapping for pre-surgical planning of tumor resections. Reproducibility of these results has been shown to require a substantial amount of high-quality data, which is not often available in clinical or research settings. Methods In this work, we report voxelwise mapping of a standard set of RSNs using a novel deep 3D convolutional neural network (3DCNN). The 3DCNN was trained on publicly available functional MRI data acquired in n = 2010 healthy participants. After training, maps that represent the probability of a voxel belonging to a particular RSN were generated for each participant, and then used to calculate mean and standard deviation (STD) probability maps, which are made publicly available. Further, we compared our results to previously published resting state and task-based functional mappings. Results Our results indicate this method can be applied in individual subjects and is highly resistant to both noisy data and fewer RS-fMRI time points than are typically acquired. Further, our results show core regions within each network that exhibit high average probability and low STD. Discussion The 3DCNN algorithm can generate individual RSN localization maps, which are necessary for clinical applications. The similarity between 3DCNN mapping results and task-based fMRI responses supports the association of specific functional tasks with RSNs.
Collapse
Affiliation(s)
- Patrick H. Luckett
- Division of Neurotechnology, Department of Neurological Surgery, Washington University School of Medicine, St. Louis, MO, United States
| | - John J. Lee
- Mallinckrodt Institute of Radiology, Washington University School of Medicine, St. Louis, MO, United States
| | - Ki Yun Park
- Division of Neurotechnology, Department of Neurological Surgery, Washington University School of Medicine, St. Louis, MO, United States
| | - Ryan V. Raut
- Department of Physiology and Biophysics, University of Washington, Seattle, WA, United States
- MindScope Program, Allen Institute, Seattle, WA, United States
| | - Karin L. Meeker
- Department of Neurology, Washington University School of Medicine, St. Louis, MO, United States
| | - Evan M. Gordon
- Department of Neurology, Washington University School of Medicine, St. Louis, MO, United States
| | - Abraham Z. Snyder
- Mallinckrodt Institute of Radiology, Washington University School of Medicine, St. Louis, MO, United States
- Department of Neurology, Washington University School of Medicine, St. Louis, MO, United States
| | - Beau M. Ances
- Department of Neurology, Washington University School of Medicine, St. Louis, MO, United States
| | - Eric C. Leuthardt
- Division of Neurotechnology, Department of Neurological Surgery, Washington University School of Medicine, St. Louis, MO, United States
- Department of Neuroscience, Washington University School of Medicine, St. Louis, MO, United States
- Department of Biomedical Engineering, Washington University in St. Louis, St. Louis, MO, United States
- Department of Mechanical Engineering and Materials Science, Washington University in St. Louis, St. Louis, MO, United States
- Center for Innovation in Neuroscience and Technology, Division of Neurotechnology, Washington University School of Medicine, St. Louis, MO, United States
- Brain Laser Center, Washington University School of Medicine, St. Louis, MO, United States
- National Center for Adaptive Neurotechnologies, Albany, NY, United States
| | - Joshua S. Shimony
- Mallinckrodt Institute of Radiology, Washington University School of Medicine, St. Louis, MO, United States
| |
Collapse
|
34
|
Evangelista D, Morotti E, Loli Piccolomini E. RISING: A new framework for model-based few-view CT image reconstruction with deep learning. Comput Med Imaging Graph 2023; 103:102156. [PMID: 36528018 DOI: 10.1016/j.compmedimag.2022.102156] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/28/2022] [Revised: 11/10/2022] [Accepted: 11/27/2022] [Indexed: 12/03/2022]
Abstract
Medical image reconstruction from low-dose tomographic data is an active research field, recently revolutionized by the advent of deep learning. In fact, deep learning typically yields superior results than classical optimization approaches, but unstable results have been reported both numerically and theoretically in the literature. This paper proposes RISING, a new framework for sparse-view tomographic image reconstruction combining an early-stopped Rapid Iterative Solver with a subsequent Iteration Network-based Gaining step. In our two-step approach, the first phase executes very few iterations of a regularized model-based algorithm, whereas the second step completes the missing iterations by means of a convolutional neural network. The proposed method is ground-truth free; it exploits the computational speed and flexibility of a data-driven approach, but it also imposes sparsity constraints to the solution as in the model-based setting. Experiments performed both on a digitally created and on a real abdomen data set confirm the numerical and visual accuracy of the reconstructed RISING images in short computational times. These features make the framework promising to be used on real systems for clinical exams.
Collapse
Affiliation(s)
| | - Elena Morotti
- Department of Political and Social Sciences, University of Bologna, Italy.
| | | |
Collapse
|
35
|
Wu S, Nakao M, Imanishi K, Nakamura M, Mizowaki T, Matsuda T. Computed Tomography slice interpolation in the longitudinal direction based on deep learning techniques: To reduce slice thickness or slice increment without dose increase. PLoS One 2022; 17:e0279005. [PMID: 36520814 PMCID: PMC9754169 DOI: 10.1371/journal.pone.0279005] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/28/2021] [Accepted: 11/29/2022] [Indexed: 12/23/2022] Open
Abstract
Large slice thickness or slice increment causes information insufficiency of Computed Tomography (CT) data in the longitudinal direction, which degrades the quality of CT-based diagnosis. Traditional approaches such as high-resolution computed tomography (HRCT) and linear interpolation can solve this problem. However, HRCT suffers from dose increase, and linear interpolation causes artifacts. In this study, we propose a deep-learning-based approach to reconstruct densely sliced CT from sparsely sliced CT data without any dose increase. The proposed method reconstructs CT images from neighboring slices using a U-net architecture. To prevent multiple reconstructed slices from influencing one another, we propose a parallel architecture in which multiple U-net architectures work independently. Moreover, for a specific organ (i.e., the liver), we propose a range-clip technique to improve reconstruction quality, which enhances the learning of CT values within this organ by enlarging the range of the training data. CT data from 130 patients were collected, with 80% used for training and the remaining 20% used for testing. Experiments showed that our parallel U-net architecture reduced the mean absolute error of CT values in the reconstructed slices by 22.05%, and also reduced the incidence of artifacts around the boundaries of target organs, compared with linear interpolation. Further improvements of 15.12%, 11.04%, 10.94%, and 10.63% were achieved for the liver, left kidney, right kidney, and stomach, respectively, using the proposed range-clip algorithm. Also, we compared the proposed architecture with original U-net method, and the experimental results demonstrated the superiority of our approach.
Collapse
Affiliation(s)
- Shuqiong Wu
- The Institute of Scientific and Industrial Research, Osaka University, Ibaraki, Osaka, Japan
- * E-mail:
| | - Megumi Nakao
- Graduate School of Informatics, Kyoto University, Kyoto, Japan
| | | | - Mitsuhiro Nakamura
- Division of Medical Physics, Department of Information Technology and Medical Engineering, Human Health Sciences, Graduate School of Medicine, Kyoto University, Kyoto, Japan
- Department of Radiation Oncology and Image-applied Therapy, Graduate School of Medicine, Kyoto University, Kyoto, Japan
| | - Takashi Mizowaki
- Department of Radiation Oncology and Image-applied Therapy, Graduate School of Medicine, Kyoto University, Kyoto, Japan
| | - Tetsuya Matsuda
- Graduate School of Informatics, Kyoto University, Kyoto, Japan
| |
Collapse
|
36
|
Jia Y, McMichael N, Mokarzel P, Thompson B, Si D, Humphries T. Superiorization-inspired unrolled SART algorithm with U-Net generated perturbations for sparse-view and limited-angle CT reconstruction. Phys Med Biol 2022; 67. [PMID: 36541524 DOI: 10.1088/1361-6560/aca513] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/02/2022] [Accepted: 11/22/2022] [Indexed: 11/23/2022]
Abstract
Objective.Unrolled algorithms are a promising approach for reconstruction of CT images in challenging scenarios, such as low-dose, sparse-view and limited-angle imaging. In an unrolled algorithm, a fixed number of iterations of a reconstruction method are unrolled into multiple layers of a neural network, and interspersed with trainable layers. The entire network is then trained end-to-end in a supervised fashion, to learn an appropriate regularizer from training data. In this paper we propose a novel unrolled algorithm, and compare its performance with several other approaches on sparse-view and limited-angle CT.Approach.The proposed algorithm is inspired by the superiorization methodology, an optimization heuristic in which iterates of a feasibility-seeking method are perturbed between iterations, typically using descent directions of a model-based penalty function. Our algorithm instead uses a modified U-net architecture to introduce the perturbations, allowing a network to learn beneficial perturbations to the image at various stages of the reconstruction, based on the training data.Main Results.In several numerical experiments modeling sparse-view and limited angle CT scenarios, the algorithm provides excellent results. In particular, it outperforms several competing unrolled methods in limited-angle scenarios, while providing comparable or better performance on sparse-view scenarios.Significance.This work represents a first step towards exploiting the power of deep learning within the superiorization methodology. Additionally, it studies the effect of network architecture on the performance of unrolled methods, as well as the effectiveness of the unrolled approach on both limited-angle CT, where previous studies have primarily focused on the sparse-view and low-dose cases.
Collapse
Affiliation(s)
- Yiran Jia
- School of STEM, University of Washington Bothell, Bothell, WA 98011, United States of America
| | - Noah McMichael
- School of STEM, University of Washington Bothell, Bothell, WA 98011, United States of America
| | - Pedro Mokarzel
- School of STEM, University of Washington Bothell, Bothell, WA 98011, United States of America
| | - Brandon Thompson
- School of STEM, University of Washington Bothell, Bothell, WA 98011, United States of America
| | - Dong Si
- School of STEM, University of Washington Bothell, Bothell, WA 98011, United States of America
| | - Thomas Humphries
- School of STEM, University of Washington Bothell, Bothell, WA 98011, United States of America
| |
Collapse
|
37
|
Sundell VM, Mäkelä T, Vitikainen AM, Kaasalainen T. Convolutional neural network -based phantom image scoring for mammography quality control. BMC Med Imaging 2022; 22:216. [PMID: 36476319 PMCID: PMC9727908 DOI: 10.1186/s12880-022-00944-w] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/21/2022] [Accepted: 11/28/2022] [Indexed: 12/13/2022] Open
Abstract
BACKGROUND Visual evaluation of phantom images is an important, but time-consuming part of mammography quality control (QC). Consistent scoring of phantom images over the device's lifetime is highly desirable. Recently, convolutional neural networks (CNNs) have been applied to a wide range of image classification problems, performing with a high accuracy. The purpose of this study was to automate mammography QC phantom scoring task by training CNN models to mimic a human reviewer. METHODS Eight CNN variations consisting of three to ten convolutional layers were trained for detecting targets (fibres, microcalcifications and masses) in American College of Radiology (ACR) accreditation phantom images and the results were compared with human scoring. Regular and artificially degraded/improved QC phantom images from eight mammography devices were visually evaluated by one reviewer. These images were used in training the CNN models. A separate test set consisted of daily QC images from the eight devices and separately acquired images with varying dose levels. These were scored by four reviewers and considered the ground truth for CNN performance testing. RESULTS Although hyper-parameter search space was limited, an optimal network depth after which additional layers resulted in decreased accuracy was identified. The highest scoring accuracy (95%) was achieved with the CNN consisting of six convolutional layers. The highest deviation between the CNN and the reviewers was found at lowest dose levels. No significant difference emerged between the visual reviews and CNN results except in case of smallest masses. CONCLUSION A CNN-based automatic mammography QC phantom scoring system can score phantom images in a good agreement with human reviewers, and can therefore be of benefit in mammography QC.
Collapse
Affiliation(s)
- Veli-Matti Sundell
- grid.7737.40000 0004 0410 2071Department of Physics, University of Helsinki, P.O. Box 64, 00014 Helsinki, Finland ,grid.7737.40000 0004 0410 2071HUS Diagnostic Center, Radiology, University of Helsinki and Helsinki University Hospital, P.O. Box 340, Haartmaninkatu 4, 00290 Helsinki, Finland
| | - Teemu Mäkelä
- grid.7737.40000 0004 0410 2071Department of Physics, University of Helsinki, P.O. Box 64, 00014 Helsinki, Finland ,grid.7737.40000 0004 0410 2071HUS Diagnostic Center, Radiology, University of Helsinki and Helsinki University Hospital, P.O. Box 340, Haartmaninkatu 4, 00290 Helsinki, Finland
| | - Anne-Mari Vitikainen
- grid.7737.40000 0004 0410 2071HUS Diagnostic Center, Radiology, University of Helsinki and Helsinki University Hospital, P.O. Box 340, Haartmaninkatu 4, 00290 Helsinki, Finland
| | - Touko Kaasalainen
- grid.7737.40000 0004 0410 2071HUS Diagnostic Center, Radiology, University of Helsinki and Helsinki University Hospital, P.O. Box 340, Haartmaninkatu 4, 00290 Helsinki, Finland
| |
Collapse
|
38
|
Huang Z, Liu Z, He P, Ren Y, Li S, Lei Y, Luo D, Liang D, Shao D, Hu Z, Zhang N. Segmentation-guided Denoising Network for Low-dose CT Imaging. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 227:107199. [PMID: 36334524 DOI: 10.1016/j.cmpb.2022.107199] [Citation(s) in RCA: 6] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 07/06/2022] [Revised: 10/06/2022] [Accepted: 10/21/2022] [Indexed: 05/15/2023]
Abstract
BACKGROUND To reduce radiation exposure and improve diagnosis in low-dose computed tomography, several deep learning (DL)-based image denoising methods have been proposed to suppress noise and artifacts over the past few years. However, most of them seek an objective data distribution approximating the gold standard and neglect structural semantic preservation. Moreover, the numerical response in CT images presents substantial regional anatomical differences among tissues in terms of X-ray absorbency. METHODS In this paper, we introduce structural semantic information for low-dose CT imaging. First, the regional segmentation prior to low-dose CT can guide the denoising process. Second the structural semantical results can be considered as evaluation metrics on the estimated normal-dose CT images. Then, a semantic feature transform is engaged to combine the semantic and image features on a semantic fusion module. In addition, the structural semantic loss function is introduced to measure the segmentation difference. RESULTS Experiments are conducted on clinical abdomen data obtained from a clinical hospital, and the semantic labels consist of subcutaneous fat, muscle and visceral fat associated with body physical evaluation. Compared with other DL-based methods, the proposed method achieves better performance on quantitative metrics and better semantic evaluation. CONCLUSIONS The quantitative experimental results demonstrate the promising performance of the proposed methods in noise reduction and structural semantic preservation. While, the proposed method may suffer from several limitations on abnormalities, unknown noise and different manufacturers. In the future, the proposed method will be further explored, and wider applications in PET/CT and PET/MR will be sought.
Collapse
Affiliation(s)
- Zhenxing Huang
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
| | - Zhou Liu
- Department of Radiology, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital & Shenzhen Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Shenzhen, 518116, China
| | - Pin He
- Department of Radiology, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital & Shenzhen Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Shenzhen, 518116, China
| | - Ya Ren
- Department of Radiology, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital & Shenzhen Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Shenzhen, 518116, China
| | - Shuluan Li
- Department of Medical Oncology, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital & Shenzhen Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Shenzhen, 518116, China
| | - Yuanyuan Lei
- Department of Pathology, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital & Shenzhen Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Shenzhen, 518116, China
| | - Dehong Luo
- Department of Radiology, National Cancer Center/National Clinical Research Center for Cancer/Cancer Hospital & Shenzhen Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Shenzhen, 518116, China
| | - Dong Liang
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
| | - Dan Shao
- Department of Nuclear Medicine, Guangdong Provincial People's Hospital, Guangdong Academy of Medical Sciences, Guangzhou 510080, China
| | - Zhanli Hu
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China
| | - Na Zhang
- Lauterbur Research Center for Biomedical Imaging, Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen 518055, China.
| |
Collapse
|
39
|
Yurt M, Dalmaz O, Dar S, Ozbey M, Tinaz B, Oguz K, Cukur T. Semi-Supervised Learning of MRI Synthesis Without Fully-Sampled Ground Truths. IEEE TRANSACTIONS ON MEDICAL IMAGING 2022; 41:3895-3906. [PMID: 35969576 DOI: 10.1109/tmi.2022.3199155] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Learning-based translation between MRI contrasts involves supervised deep models trained using high-quality source- and target-contrast images derived from fully-sampled acquisitions, which might be difficult to collect under limitations on scan costs or time. To facilitate curation of training sets, here we introduce the first semi-supervised model for MRI contrast translation (ssGAN) that can be trained directly using undersampled k-space data. To enable semi-supervised learning on undersampled data, ssGAN introduces novel multi-coil losses in image, k-space, and adversarial domains. The multi-coil losses are selectively enforced on acquired k-space samples unlike traditional losses in single-coil synthesis models. Comprehensive experiments on retrospectively undersampled multi-contrast brain MRI datasets are provided. Our results demonstrate that ssGAN yields on par performance to a supervised model, while outperforming single-coil models trained on coil-combined magnitude images. It also outperforms cascaded reconstruction-synthesis models where a supervised synthesis model is trained following self-supervised reconstruction of undersampled data. Thus, ssGAN holds great promise to improve the feasibility of learning-based multi-contrast MRI synthesis.
Collapse
|
40
|
Kim B, Shim H, Baek J. A streak artifact reduction algorithm in sparse-view CT using a self-supervised neural representation. Med Phys 2022; 49:7497-7515. [PMID: 35880806 DOI: 10.1002/mp.15885] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/11/2022] [Revised: 07/13/2022] [Accepted: 07/17/2022] [Indexed: 12/27/2022] Open
Abstract
PURPOSE Sparse-view computed tomography (CT) has been attracting attention for its reduced radiation dose and scanning time. However, analytical image reconstruction methods suffer from streak artifacts due to insufficient projection views. Recently, various deep learning-based methods have been developed to solve this ill-posed inverse problem. Despite their promising results, they are easily overfitted to the training data, showing limited generalizability to unseen systems and patients. In this work, we propose a novel streak artifact reduction algorithm that provides a system- and patient-specific solution. METHODS Motivated by the fact that streak artifacts are deterministic errors, we regenerate the same artifacts from a prior CT image under the same system geometry. This prior image need not be perfect but should contain patient-specific information and be consistent with full-view projection data for accurate regeneration of the artifacts. To this end, we use a coordinate-based neural representation that often causes image blur but can greatly suppress the streak artifacts while having multiview consistency. By employing techniques in neural radiance fields originally proposed for scene representations, the neural representation is optimized to the measured sparse-view projection data via self-supervised learning. Then, we subtract the regenerated artifacts from the analytically reconstructed original image to obtain the final corrected image. RESULTS To validate the proposed method, we used simulated data of extended cardiac-torso phantoms and the 2016 NIH-AAPM-Mayo Clinic Low-Dose CT Grand Challenge and experimental data of physical pediatric and head phantoms. The performance of the proposed method was compared with a total variation-based iterative reconstruction method, naive application of the neural representation, and a convolutional neural network-based method. In visual inspection, it was observed that the small anatomical features were best preserved by the proposed method. The proposed method also achieved the best scores in the visual information fidelity, modulation transfer function, and lung nodule segmentation. CONCLUSIONS The results on both simulated and experimental data suggest that the proposed method can effectively reduce the streak artifacts while preserving small anatomical structures that are easily blurred or replaced with misleading features by the existing methods. Since the proposed method does not require any additional training datasets, it would be useful in clinical practice where the large datasets cannot be collected.
Collapse
Affiliation(s)
- Byeongjoon Kim
- School of Integrated Technology, Yonsei University, Incheon, South Korea
| | - Hyunjung Shim
- School of Integrated Technology, Yonsei University, Incheon, South Korea
| | - Jongduk Baek
- School of Integrated Technology, Yonsei University, Incheon, South Korea
| |
Collapse
|
41
|
Goudarzi S, Basarab A, Rivaz H. Inverse Problem of Ultrasound Beamforming With Denoising-Based Regularized Solutions. IEEE TRANSACTIONS ON ULTRASONICS, FERROELECTRICS, AND FREQUENCY CONTROL 2022; 69:2906-2916. [PMID: 35969567 DOI: 10.1109/tuffc.2022.3198874] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
During the past few years, inverse problem formulations of ultrasound beamforming have attracted growing interest. They usually pose beamforming as a minimization problem of a fidelity term resulting from the measurement model plus a regularization term that enforces a certain class on the resulting image. Here, we take advantage of alternating direction method of multipliers to propose a flexible framework in which each term is optimized separately. Furthermore, the proposed beamforming formulation is extended to replace the regularization term with a denoising algorithm, based on the recent approaches called plug-and-play (PnP) and regularization by denoising (RED). Such regularizations are shown in this work to better preserve speckle texture, an important feature in ultrasound imaging, than sparsity-based approaches previously proposed in the literature. The efficiency of the proposed methods is evaluated on simulations, real phantoms, and in vivo data available from a plane-wave imaging challenge in medical ultrasound. Furthermore, a comprehensive comparison with existing ultrasound beamforming methods is also provided. These results show that the RED algorithm gives the best image quality in terms of contrast index while preserving the speckle statistics.
Collapse
|
42
|
Minnema J, Ernst A, van Eijnatten M, Pauwels R, Forouzanfar T, Batenburg KJ, Wolff J. A review on the application of deep learning for CT reconstruction, bone segmentation and surgical planning in oral and maxillofacial surgery. Dentomaxillofac Radiol 2022; 51:20210437. [PMID: 35532946 PMCID: PMC9522976 DOI: 10.1259/dmfr.20210437] [Citation(s) in RCA: 10] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/12/2021] [Revised: 04/21/2022] [Accepted: 04/25/2022] [Indexed: 12/11/2022] Open
Abstract
Computer-assisted surgery (CAS) allows clinicians to personalize treatments and surgical interventions and has therefore become an increasingly popular treatment modality in maxillofacial surgery. The current maxillofacial CAS consists of three main steps: (1) CT image reconstruction, (2) bone segmentation, and (3) surgical planning. However, each of these three steps can introduce errors that can heavily affect the treatment outcome. As a consequence, tedious and time-consuming manual post-processing is often necessary to ensure that each step is performed adequately. One way to overcome this issue is by developing and implementing neural networks (NNs) within the maxillofacial CAS workflow. These learning algorithms can be trained to perform specific tasks without the need for explicitly defined rules. In recent years, an extremely large number of novel NN approaches have been proposed for a wide variety of applications, which makes it a difficult task to keep up with all relevant developments. This study therefore aimed to summarize and review all relevant NN approaches applied for CT image reconstruction, bone segmentation, and surgical planning. After full text screening, 76 publications were identified: 32 focusing on CT image reconstruction, 33 focusing on bone segmentation and 11 focusing on surgical planning. Generally, convolutional NNs were most widely used in the identified studies, although the multilayer perceptron was most commonly applied in surgical planning tasks. Moreover, the drawbacks of current approaches and promising research avenues are discussed.
Collapse
Affiliation(s)
- Jordi Minnema
- Department of Oral and Maxillofacial Surgery/Pathology, Amsterdam UMC and Academic Centre for Dentistry Amsterdam (ACTA), Vrije Universiteit Amsterdam, 3D Innovationlab, Amsterdam Movement Sciences, Amsterdam, The Netherlands
| | - Anne Ernst
- Institute for Medical Systems Biology, University Hospital Hamburg-Eppendorf, Hamburg, Germany
| | - Maureen van Eijnatten
- Department of Oral and Maxillofacial Surgery/Pathology, Amsterdam UMC and Academic Centre for Dentistry Amsterdam (ACTA), Vrije Universiteit Amsterdam, 3D Innovationlab, Amsterdam Movement Sciences, Amsterdam, The Netherlands
| | - Ruben Pauwels
- Aarhus Institute of Advanced Studies, Aarhus University, Aarhus, Denmark
| | - Tymour Forouzanfar
- Department of Oral and Maxillofacial Surgery/Pathology, Amsterdam UMC and Academic Centre for Dentistry Amsterdam (ACTA), Vrije Universiteit Amsterdam, 3D Innovationlab, Amsterdam Movement Sciences, Amsterdam, The Netherlands
| | - Kees Joost Batenburg
- Department of Oral and Maxillofacial Surgery/Pathology, Amsterdam UMC and Academic Centre for Dentistry Amsterdam (ACTA), Vrije Universiteit Amsterdam, 3D Innovationlab, Amsterdam Movement Sciences, Amsterdam, The Netherlands
| | - Jan Wolff
- Department of Dentistry and Oral Health, Aarhus University, Vennelyst Boulevard, Aarhus, Denmark
| |
Collapse
|
43
|
Shao W, Leung KH, Xu J, Coughlin JM, Pomper MG, Du Y. Generation of Digital Brain Phantom for Machine Learning Application of Dopamine Transporter Radionuclide Imaging. Diagnostics (Basel) 2022; 12:1945. [PMID: 36010295 PMCID: PMC9406894 DOI: 10.3390/diagnostics12081945] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/18/2022] [Revised: 08/03/2022] [Accepted: 08/11/2022] [Indexed: 11/16/2022] Open
Abstract
While machine learning (ML) methods may significantly improve image quality for SPECT imaging for the diagnosis and monitoring of Parkinson's disease (PD), they require a large amount of data for training. It is often difficult to collect a large population of patient data to support the ML research, and the ground truth of lesion is also unknown. This paper leverages a generative adversarial network (GAN) to generate digital brain phantoms for training ML-based PD SPECT algorithms. A total of 594 PET 3D brain models from 155 patients (113 male and 42 female) were reviewed and 1597 2D slices containing the full or a portion of the striatum were selected. Corresponding attenuation maps were also generated based on these images. The data were then used to develop a GAN for generating 2D brain phantoms, where each phantom consisted of a radioactivity image and the corresponding attenuation map. Statistical methods including histogram, Fréchet distance, and structural similarity were used to evaluate the generator based on 10,000 generated phantoms. When the generated phantoms and training dataset were both passed to the discriminator, similar normal distributions were obtained, which indicated the discriminator was unable to distinguish the generated phantoms from the training datasets. The generated digital phantoms can be used for 2D SPECT simulation and serve as the ground truth to develop ML-based reconstruction algorithms. The cumulated experience from this work also laid the foundation for building a 3D GAN for the same application.
Collapse
Affiliation(s)
- Wenyi Shao
- The Russell H. Morgan Department of Radiology and Radiational Science, Johns Hopkins University School of Medicine, Baltimore, MD 21287, USA
| | - Kevin H. Leung
- The Russell H. Morgan Department of Radiology and Radiational Science, Johns Hopkins University School of Medicine, Baltimore, MD 21287, USA
- Department of Biomedical Engineering, Johns Hopkins University School of Medicine, Baltimore, MD 21287, USA
| | - Jingyan Xu
- The Russell H. Morgan Department of Radiology and Radiational Science, Johns Hopkins University School of Medicine, Baltimore, MD 21287, USA
| | - Jennifer M. Coughlin
- The Russell H. Morgan Department of Radiology and Radiational Science, Johns Hopkins University School of Medicine, Baltimore, MD 21287, USA
- Department of Psychiatry and Behavioral Sciences, Johns Hopkins University School of Medicine, Baltimore, MD 21287, USA
| | - Martin G. Pomper
- The Russell H. Morgan Department of Radiology and Radiational Science, Johns Hopkins University School of Medicine, Baltimore, MD 21287, USA
- Department of Biomedical Engineering, Johns Hopkins University School of Medicine, Baltimore, MD 21287, USA
- Department of Psychiatry and Behavioral Sciences, Johns Hopkins University School of Medicine, Baltimore, MD 21287, USA
| | - Yong Du
- The Russell H. Morgan Department of Radiology and Radiational Science, Johns Hopkins University School of Medicine, Baltimore, MD 21287, USA
| |
Collapse
|
44
|
Kandarpa VSS, Perelli A, Bousse A, Visvikis D. LRR-CED: low-resolution reconstruction-aware convolutional encoder–decoder network for direct sparse-view CT image reconstruction. Phys Med Biol 2022; 67. [DOI: 10.1088/1361-6560/ac7bce] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2021] [Accepted: 06/23/2022] [Indexed: 11/12/2022]
Abstract
Abstract
Objective. Sparse-view computed tomography (CT) reconstruction has been at the forefront of research in medical imaging. Reducing the total x-ray radiation dose to the patient while preserving the reconstruction accuracy is a big challenge. The sparse-view approach is based on reducing the number of rotation angles, which leads to poor quality reconstructed images as it introduces several artifacts. These artifacts are more clearly visible in traditional reconstruction methods like the filtered-backprojection (FBP) algorithm. Approach. Over the years, several model-based iterative and more recently deep learning-based methods have been proposed to improve sparse-view CT reconstruction. Many deep learning-based methods improve FBP-reconstructed images as a post-processing step. In this work, we propose a direct deep learning-based reconstruction that exploits the information from low-dimensional scout images, to learn the projection-to-image mapping. This is done by concatenating FBP scout images at multiple resolutions in the decoder part of a convolutional encoder–decoder (CED). Main results. This approach is investigated on two different networks, based on Dense Blocks and U-Net to show that a direct mapping can be learned from a sinogram to an image. The results are compared to two post-processing deep learning methods (FBP-ConvNet and DD-Net) and an iterative method that uses a total variation (TV) regularization. Significance. This work presents a novel method that uses information from both sinogram and low-resolution scout images for sparse-view CT image reconstruction. We also generalize this idea by demonstrating results with two different neural networks. This work is in the direction of exploring deep learning across the various stages of the image reconstruction pipeline involving data correction, domain transfer and image improvement.
Collapse
|
45
|
Zhao Y, Wang Y, Zhang J, Liu X, Li Y, Guo S, Yang X, Hong S. Surgical GAN: Towards real-time path planning for passive flexible tools in endovascular surgeries. Neurocomputing 2022. [DOI: 10.1016/j.neucom.2022.05.044] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022]
|
46
|
Shao W, Zhou B. Dielectric Breast Phantoms by Generative Adversarial Network. IEEE TRANSACTIONS ON ANTENNAS AND PROPAGATION 2022; 70:6256-6264. [PMID: 36969506 PMCID: PMC10038476 DOI: 10.1109/tap.2021.3121149] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/18/2023]
Abstract
In order to conduct the research of machine-learning (ML) based microwave breast imaging (MBI), a large number of digital dielectric breast phantoms that can be used as training data (ground truth) are required but are difficult to be achieved from practice. Although a few dielectric breast phantoms have been developed for research purpose, the number and the diversity are limited and is far inadequate to develop a robust ML algorithm for MBI. This paper presents a neural network method to generate 2D virtual breast phantoms that are similar to the real ones, which can be used to develop ML-based MBI in the future. The generated phantoms are similar but are different from those used in training. Each phantom consists of several images with each representing the distribution of a dielectric parameter in the breast map. Statistical analysis was performed over 10,000 generated phantoms to investigate the performance of the generative network. With the generative network, one may generate unlimited number of breast images with more variations, so the ML-based MBI will be more ready to deploy.
Collapse
Affiliation(s)
- Wenyi Shao
- Russell H. Morgan Department of Radiology and Radiological Science, Johns Hopkins University School of Medicine, Baltimore, MD 21287, USA
| | | |
Collapse
|
47
|
Development of an Artificial Neural Network for the Detection of Supporting Hindlimb Lameness: A Pilot Study in Working Dogs. Animals (Basel) 2022; 12:ani12141755. [PMID: 35883302 PMCID: PMC9311578 DOI: 10.3390/ani12141755] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/23/2022] [Revised: 07/04/2022] [Accepted: 07/05/2022] [Indexed: 11/17/2022] Open
Abstract
Subjective lameness assessment has been a controversial subject given the lack of agreement between observers; this has prompted the development of kinetic and kinematic devices in order to obtain an objective evaluation of locomotor system in dogs. After proper training, neural networks are potentially capable of making a non-human diagnosis of canine lameness. The purpose of this study was to investigate whether artificial neural networks could be used to determine canine hindlimb lameness by computational means only. The outcome of this study could potentially assess the efficacy of certain treatments against diseases that cause lameness. With this aim, input data were obtained from an inertial sensor positioned on the rump. Data from dogs with unilateral hindlimb lameness and sound dogs were used to obtain differences between both groups at walk. The artificial neural network, after necessary adjustments, was integrated into a web management tool, and the preliminary results discriminating between lame and sound dogs are promising. The analysis of spatial data with artificial neural networks was summarized and developed into a web app that has proven to be a useful tool to discriminate between sound and lame dogs. Additionally, this environment allows veterinary clinicians to adequately follow the treatment of lame canine patients.
Collapse
|
48
|
Rusanov B, Hassan GM, Reynolds M, Sabet M, Kendrick J, Farzad PR, Ebert M. Deep learning methods for enhancing cone-beam CT image quality towards adaptive radiation therapy: A systematic review. Med Phys 2022; 49:6019-6054. [PMID: 35789489 PMCID: PMC9543319 DOI: 10.1002/mp.15840] [Citation(s) in RCA: 29] [Impact Index Per Article: 9.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/18/2021] [Revised: 05/21/2022] [Accepted: 06/16/2022] [Indexed: 11/11/2022] Open
Abstract
The use of deep learning (DL) to improve cone-beam CT (CBCT) image quality has gained popularity as computational resources and algorithmic sophistication have advanced in tandem. CBCT imaging has the potential to facilitate online adaptive radiation therapy (ART) by utilizing up-to-date patient anatomy to modify treatment parameters before irradiation. Poor CBCT image quality has been an impediment to realizing ART due to the increased scatter conditions inherent to cone-beam acquisitions. Given the recent interest in DL applications in radiation oncology, and specifically DL for CBCT correction, we provide a systematic theoretical and literature review for future stakeholders. The review encompasses DL approaches for synthetic CT generation, as well as projection domain methods employed in the CBCT correction literature. We review trends pertaining to publications from January 2018 to April 2022 and condense their major findings - with emphasis on study design and deep learning techniques. Clinically relevant endpoints relating to image quality and dosimetric accuracy are summarised, highlighting gaps in the literature. Finally, we make recommendations for both clinicians and DL practitioners based on literature trends and the current DL state of the art methods utilized in radiation oncology. This article is protected by copyright. All rights reserved.
Collapse
Affiliation(s)
- Branimir Rusanov
- School of Physics, Mathematics and Computing, The University of Western Australia, Perth, Western Australia, 6009, Australia.,Department of Radiation Oncology, Sir Chairles Gairdner Hospital, Perth, Western Australia, 6009, Australia
| | - Ghulam Mubashar Hassan
- School of Physics, Mathematics and Computing, The University of Western Australia, Perth, Western Australia, 6009, Australia
| | - Mark Reynolds
- School of Physics, Mathematics and Computing, The University of Western Australia, Perth, Western Australia, 6009, Australia
| | - Mahsheed Sabet
- School of Physics, Mathematics and Computing, The University of Western Australia, Perth, Western Australia, 6009, Australia.,Department of Radiation Oncology, Sir Chairles Gairdner Hospital, Perth, Western Australia, 6009, Australia
| | - Jake Kendrick
- School of Physics, Mathematics and Computing, The University of Western Australia, Perth, Western Australia, 6009, Australia.,Department of Radiation Oncology, Sir Chairles Gairdner Hospital, Perth, Western Australia, 6009, Australia
| | - Pejman Rowshan Farzad
- School of Physics, Mathematics and Computing, The University of Western Australia, Perth, Western Australia, 6009, Australia.,Department of Radiation Oncology, Sir Chairles Gairdner Hospital, Perth, Western Australia, 6009, Australia
| | - Martin Ebert
- School of Physics, Mathematics and Computing, The University of Western Australia, Perth, Western Australia, 6009, Australia.,Department of Radiation Oncology, Sir Chairles Gairdner Hospital, Perth, Western Australia, 6009, Australia
| |
Collapse
|
49
|
Hu D, Zhang Y, Liu J, Luo S, Chen Y. DIOR: Deep Iterative Optimization-Based Residual-Learning for Limited-Angle CT Reconstruction. IEEE TRANSACTIONS ON MEDICAL IMAGING 2022; 41:1778-1790. [PMID: 35100109 DOI: 10.1109/tmi.2022.3148110] [Citation(s) in RCA: 9] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Limited-angle CT is a challenging problem in real applications. Incomplete projection data will lead to severe artifacts and distortions in reconstruction images. To tackle this problem, we propose a novel reconstruction framework termed Deep Iterative Optimization-based Residual-learning (DIOR) for limited-angle CT. Instead of directly deploying the regularization term on image space, the DIOR combines iterative optimization and deep learning based on the residual domain, significantly improving the convergence property and generalization ability. Specifically, the asymmetric convolutional modules are adopted to strengthen the feature extraction capacity in smooth regions for deep priors. Besides, in our DIOR method, the information contained in low-frequency and high-frequency components is also evaluated by perceptual loss to improve the performance in tissue preservation. Both simulated and clinical datasets are performed to validate the performance of DIOR. Compared with existing competitive algorithms, quantitative and qualitative results show that the proposed method brings a promising improvement in artifact removal, detail restoration and edge preservation.
Collapse
|
50
|
Jung C, Abuhamad M, Mohaisen D, Han K, Nyang D. WBC image classification and generative models based on convolutional neural network. BMC Med Imaging 2022; 22:94. [PMID: 35596153 PMCID: PMC9121596 DOI: 10.1186/s12880-022-00818-1] [Citation(s) in RCA: 12] [Impact Index Per Article: 4.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/03/2021] [Accepted: 05/06/2022] [Indexed: 11/25/2022] Open
Abstract
Background Computer-aided methods for analyzing white blood cells (WBC) are popular due to the complexity of the manual alternatives. Recent works have shown highly accurate segmentation and detection of white blood cells from microscopic blood images. However, the classification of the observed cells is still a challenge, in part due to the distribution of the five types that affect the condition of the immune system. Methods (i) This work proposes W-Net, a CNN-based method for WBC classification. We evaluate W-Net on a real-world large-scale dataset that includes 6562 real images of the five WBC types. (ii) For further benefits, we generate synthetic WBC images using Generative Adversarial Network to be used for education and research purposes through sharing. Results (i) W-Net achieves an average accuracy of 97%. In comparison to state-of-the-art methods in the field of WBC classification, we show that W-Net outperforms other CNN- and RNN-based model architectures. Moreover, we show the benefits of using pre-trained W-Net in a transfer learning context when fine-tuned to specific task or accommodating another dataset. (ii) The synthetic WBC images are confirmed by experiments and a domain expert to have a high degree of similarity to the original images. The pre-trained W-Net and the generated WBC dataset are available for the community to facilitate reproducibility and follow up research work. Conclusion This work proposed W-Net, a CNN-based architecture with a small number of layers, to accurately classify the five WBC types. We evaluated W-Net on a real-world large-scale dataset and addressed several challenges such as the transfer learning property and the class imbalance. W-Net achieved an average classification accuracy of 97%. We synthesized a dataset of new WBC image samples using DCGAN, which we released to the public for education and research purposes.
Collapse
Affiliation(s)
- Changhun Jung
- Department of Cyber Security, Ewha Womans University, 52, Ewhayeodae-gil, Seodaemun-gu, Seoul, 03760, Republic of Korea
| | - Mohammed Abuhamad
- Department of Computer Science, Loyola University Chicago, 1032 W Sheridan Rd, Chicago, 60660, USA
| | - David Mohaisen
- Department of Computer Science, University of Central Florida, 4000 Central Florida Blvd, Orlando, FL, 32816, USA
| | - Kyungja Han
- Department of Laboratory Medicine and College of Medicine, The Catholic University of Korea Seoul St. Mary's Hospital, 222, Banpo-daero, Seocho-gu, Seoul, 06591, Republic of Korea
| | - DaeHun Nyang
- Department of Cyber Security, Ewha Womans University, 52, Ewhayeodae-gil, Seodaemun-gu, Seoul, 03760, Republic of Korea.
| |
Collapse
|