1
|
Rabe M, Kurz C, Thummerer A, Landry G. Artificial intelligence for treatment delivery: image-guided radiotherapy. Strahlenther Onkol 2025; 201:283-297. [PMID: 39138806 DOI: 10.1007/s00066-024-02277-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/01/2024] [Accepted: 07/07/2024] [Indexed: 08/15/2024]
Abstract
Radiation therapy (RT) is a highly digitized field relying heavily on computational methods and, as such, has a high affinity for the automation potential afforded by modern artificial intelligence (AI). This is particularly relevant where imaging is concerned and is especially so during image-guided RT (IGRT). With the advent of online adaptive RT (ART) workflows at magnetic resonance (MR) linear accelerators (linacs) and at cone-beam computed tomography (CBCT) linacs, the need for automation is further increased. AI as applied to modern IGRT is thus one area of RT where we can expect important developments in the near future. In this review article, after outlining modern IGRT and online ART workflows, we cover the role of AI in CBCT and MRI correction for dose calculation, auto-segmentation on IGRT imaging, motion management, and response assessment based on in-room imaging.
Collapse
Affiliation(s)
- Moritz Rabe
- Department of Radiation Oncology, LMU University Hospital, LMU Munich, Marchioninistraße 15, 81377, Munich, Bavaria, Germany
| | - Christopher Kurz
- Department of Radiation Oncology, LMU University Hospital, LMU Munich, Marchioninistraße 15, 81377, Munich, Bavaria, Germany
| | - Adrian Thummerer
- Department of Radiation Oncology, LMU University Hospital, LMU Munich, Marchioninistraße 15, 81377, Munich, Bavaria, Germany
| | - Guillaume Landry
- Department of Radiation Oncology, LMU University Hospital, LMU Munich, Marchioninistraße 15, 81377, Munich, Bavaria, Germany.
- German Cancer Consortium (DKTK), partner site Munich, a partnership between the DKFZ and the LMU University Hospital Munich, Marchioninistraße 15, 81377, Munich, Bavaria, Germany.
- Bavarian Cancer Research Center (BZKF), Marchioninistraße 15, 81377, Munich, Bavaria, Germany.
| |
Collapse
|
2
|
Geng H, Fan J, Yang S, Chen S, Xiao D, Ai D, Fu T, Song H, Yuan K, Duan F, Wang Y, Yang J. DSC-Recon: Dual-Stage Complementary 4-D Organ Reconstruction From X-Ray Image Sequence for Intraoperative Fusion. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:3909-3923. [PMID: 38805326 DOI: 10.1109/tmi.2024.3406876] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/30/2024]
Abstract
Accurately reconstructing 4D critical organs contributes to the visual guidance in X-ray image-guided interventional operation. Current methods estimate intraoperative dynamic meshes by refining a static initial organ mesh from the semantic information in the single-frame X-ray images. However, these methods fall short of reconstructing an accurate and smooth organ sequence due to the distinct respiratory patterns between the initial mesh and X-ray image. To overcome this limitation, we propose a novel dual-stage complementary 4D organ reconstruction (DSC-Recon) model for recovering dynamic organ meshes by utilizing the preoperative and intraoperative data with different respiratory patterns. DSC-Recon is structured as a dual-stage framework: 1) The first stage focuses on addressing a flexible interpolation network applicable to multiple respiratory patterns, which could generate dynamic shape sequences between any pair of preoperative 3D meshes segmented from CT scans. 2) In the second stage, we present a deformation network to take the generated dynamic shape sequence as the initial prior and explore the discriminate feature (i.e., target organ areas and meaningful motion information) in the intraoperative X-ray images, predicting the deformed mesh by introducing a designed feature mapping pipeline integrated into the initialized shape refinement process. Experiments on simulated and clinical datasets demonstrate the superiority of our method over state-of-the-art methods in both quantitative and qualitative aspects.
Collapse
|
3
|
Sun X, Li Y, Li Y, Wang S, Qin Y, Chen P. Reconstruction method suitable for fast CT imaging. OPTICS EXPRESS 2024; 32:17072-17087. [PMID: 38858899 DOI: 10.1364/oe.522097] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/21/2024] [Accepted: 04/09/2024] [Indexed: 06/12/2024]
Abstract
Reconstructing computed tomography (CT) images from an extremely limited set of projections is crucial in practical applications. As the available projections significantly decrease, traditional reconstruction and model-based iterative reconstruction methods become constrained. This work aims to seek a reconstruction method applicable to fast CT imaging when available projections are highly sparse. To minimize the time and cost associated with projections acquisition, we propose a deep learning model, X-CTReNet, which parameterizes a nonlinear mapping function from orthogonal projections to CT volumes for 3D reconstruction. The proposed model demonstrates effective capability in inferring CT volumes from two-view projections compared to baseline methods, highlighting the significant potential for drastically reducing projection acquisition in fast CT imaging.
Collapse
|
4
|
Chen C, Fu Z, Ye S, Zhao C, Golovko V, Ye S, Bai Z. Study on high-precision three-dimensional reconstruction of pulmonary lesions and surrounding blood vessels based on CT images. OPTICS EXPRESS 2024; 32:1371-1390. [PMID: 38297691 DOI: 10.1364/oe.510398] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/27/2023] [Accepted: 12/15/2023] [Indexed: 02/02/2024]
Abstract
The adoption of computerized tomography (CT) technology has significantly elevated the role of pulmonary CT imaging in diagnosing and treating pulmonary diseases. However, challenges persist due to the complex relationship between lesions within pulmonary tissue and the surrounding blood vessels. These challenges involve achieving precise three-dimensional reconstruction while maintaining accurate relative positioning of these elements. To effectively address this issue, this study employs a semi-automatic precise labeling process for the target region. This procedure ensures a high level of consistency in the relative positions of lesions and the surrounding blood vessels. Additionally, a morphological gradient interpolation algorithm, combined with Gaussian filtering, is applied to facilitate high-precision three-dimensional reconstruction of both lesions and blood vessels. Furthermore, this technique enables post-reconstruction slicing at any layer, facilitating intuitive exploration of the correlation between blood vessels and lesion layers. Moreover, the study utilizes physiological knowledge to simulate real-world blood vessel intersections, determining the range of blood vessel branch angles and achieving seamless continuity at internal blood vessel branch points. The experimental results achieved a satisfactory reconstruction with an average Hausdorff distance of 1.5 mm and an average Dice coefficient of 92%, obtained by comparing the reconstructed shape with the original shape,the approach also achieves a high level of accuracy in three-dimensional reconstruction and visualization. In conclusion, this study is a valuable source of technical support for the diagnosis and treatment of pulmonary diseases and holds promising potential for widespread adoption in clinical practice.
Collapse
|
5
|
Fum WKS, Md Shah MN, Raja Aman RRA, Abd Kadir KA, Wen DW, Leong S, Tan LK. Generation of fluoroscopy-alike radiographs as alternative datasets for deep learning in interventional radiology. Phys Eng Sci Med 2023; 46:1535-1552. [PMID: 37695509 DOI: 10.1007/s13246-023-01317-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/12/2023] [Accepted: 08/03/2023] [Indexed: 09/12/2023]
Abstract
In fluoroscopy-guided interventions (FGIs), obtaining large quantities of labelled data for deep learning (DL) can be difficult. Synthetic labelled data can serve as an alternative, generated via pseudo 2D projections of CT volumetric data. However, contrasted vessels have low visibility in simple 2D projections of contrasted CT data. To overcome this, we propose an alternative method to generate fluoroscopy-like radiographs from contrasted head CT Angiography (CTA) volumetric data. The technique involves segmentation of brain tissue, bone, and contrasted vessels from CTA volumetric data, followed by an algorithm to adjust HU values, and finally, a standard ray-based projection is applied to generate the 2D image. The resulting synthetic images were compared to clinical fluoroscopy images for perceptual similarity and subject contrast measurements. Good perceptual similarity was demonstrated on vessel-enhanced synthetic images as compared to the clinical fluoroscopic images. Statistical tests of equivalence show that enhanced synthetic and clinical images have statistically equivalent mean subject contrast within 25% bounds. Furthermore, validation experiments confirmed that the proposed method for generating synthetic images improved the performance of DL models in certain regression tasks, such as localizing anatomical landmarks in clinical fluoroscopy images. Through enhanced pseudo 2D projection of CTA volume data, synthetic images with similar features to real clinical fluoroscopic images can be generated. The use of synthetic images as an alternative source for DL datasets represents a potential solution to the application of DL in FGIs procedures.
Collapse
Affiliation(s)
- Wilbur K S Fum
- Department of Biomedical Imaging, Faculty of Medicine, Universiti Malaya, 50603, Kuala Lumpur, Malaysia
- Division of Radiological Sciences, Singapore General Hospital, Outram Road, Singapore, 169608, Singapore
| | - Mohammad Nazri Md Shah
- Department of Biomedical Imaging, Faculty of Medicine, Universiti Malaya, 50603, Kuala Lumpur, Malaysia
| | | | - Khairul Azmi Abd Kadir
- Department of Biomedical Imaging, Faculty of Medicine, Universiti Malaya, 50603, Kuala Lumpur, Malaysia
| | - David Wei Wen
- Department of Vascular and Interventional Radiology, Singapore General Hospital, Outram Road, Singapore, 169608, Singapore
| | - Sum Leong
- Department of Vascular and Interventional Radiology, Singapore General Hospital, Outram Road, Singapore, 169608, Singapore
| | - Li Kuo Tan
- Department of Biomedical Imaging, Faculty of Medicine, Universiti Malaya, 50603, Kuala Lumpur, Malaysia.
| |
Collapse
|
6
|
Sarmah M, Neelima A, Singh HR. Survey of methods and principles in three-dimensional reconstruction from two-dimensional medical images. Vis Comput Ind Biomed Art 2023; 6:15. [PMID: 37495817 PMCID: PMC10371974 DOI: 10.1186/s42492-023-00142-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2023] [Accepted: 06/27/2023] [Indexed: 07/28/2023] Open
Abstract
Three-dimensional (3D) reconstruction of human organs has gained attention in recent years due to advances in the Internet and graphics processing units. In the coming years, most patient care will shift toward this new paradigm. However, development of fast and accurate 3D models from medical images or a set of medical scans remains a daunting task due to the number of pre-processing steps involved, most of which are dependent on human expertise. In this review, a survey of pre-processing steps was conducted, and reconstruction techniques for several organs in medical diagnosis were studied. Various methods and principles related to 3D reconstruction were highlighted. The usefulness of 3D reconstruction of organs in medical diagnosis was also highlighted.
Collapse
Affiliation(s)
- Mriganka Sarmah
- Department of Computer Science and Engineering, National Institute of Technology, Nagaland, 797103, India.
| | - Arambam Neelima
- Department of Computer Science and Engineering, National Institute of Technology, Nagaland, 797103, India
| | - Heisnam Rohen Singh
- Department of Information Technology, Nagaland University, Nagaland, 797112, India
| |
Collapse
|
7
|
Jadhav S, Deng G, Zawin M, Kaufman AE. COVID-view: Diagnosis of COVID-19 using Chest CT. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2022; 28:227-237. [PMID: 34587075 PMCID: PMC8981756 DOI: 10.1109/tvcg.2021.3114851] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/21/2021] [Revised: 06/13/2021] [Accepted: 08/08/2021] [Indexed: 05/02/2023]
Abstract
Significant work has been done towards deep learning (DL) models for automatic lung and lesion segmentation and classification of COVID-19 on chest CT data. However, comprehensive visualization systems focused on supporting the dual visual+DL diagnosis of COVID-19 are non-existent. We present COVID-view, a visualization application specially tailored for radiologists to diagnose COVID-19 from chest CT data. The system incorporates a complete pipeline of automatic lungs segmentation, localization/isolation of lung abnormalities, followed by visualization, visual and DL analysis, and measurement/quantification tools. Our system combines the traditional 2D workflow of radiologists with newer 2D and 3D visualization techniques with DL support for a more comprehensive diagnosis. COVID-view incorporates a novel DL model for classifying the patients into positive/negative COVID-19 cases, which acts as a reading aid for the radiologist using COVID-view and provides the attention heatmap as an explainable DL for the model output. We designed and evaluated COVID-view through suggestions, close feedback and conducting case studies of real-world patient data by expert radiologists who have substantial experience diagnosing chest CT scans for COVID-19, pulmonary embolism, and other forms of lung infections. We present requirements and task analysis for the diagnosis of COVID-19 that motivate our design choices and results in a practical system which is capable of handling real-world patient cases.
Collapse
Affiliation(s)
| | - Gaofeng Deng
- Department of Computer ScienceStony Brook UniversityUSA
| | - Marlene Zawin
- Department of RadiologyStony Brook University HospitalUSA
| | | |
Collapse
|
8
|
Wang Z, Nakao M, Nakamura M, Matsuda T. Shape Reconstruction for Abdominal Organs based on a Graph Convolutional Network. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2021; 2021:2960-2963. [PMID: 34891866 DOI: 10.1109/embc46164.2021.9630826] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Computed tomography and magnetic resonance imaging produce high-resolution images; however, during surgery or radiotherapy, only low-resolution cone-beam CT and low-dimensional X-ray images can be obtained. Furthermore, because the duodenum and stomach are filled with air, even in high-resolution CT images, it is hard to accurately segment their contours. In this paper, we propose a method that is based on a graph convolutional network (GCN) to reconstruct organs that are hard to detect in medical images. The method uses surrounding detectable-organ features to determine the shape and location of the target organ and learns mesh deformation parameters, which are applied to a target organ template. The role of the template is to establish an initial topological structure for the target organ. We conducted experiments with both single and multiple organ meshes to verify the performance of our proposed method.
Collapse
|
9
|
Li X, Wang S, Niu X, Wang L, Chen P. 3D M-Net: Object-Specific 3D Segmentation Network Based on a Single Projection. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2021; 2021:5852595. [PMID: 34335721 PMCID: PMC8292052 DOI: 10.1155/2021/5852595] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/02/2021] [Accepted: 06/28/2021] [Indexed: 11/30/2022]
Abstract
The internal assembly correctness of industrial products directly affects their performance and service life. Industrial products are usually protected by opaque housing, so most internal detection methods are based on X-rays. Since the dense structural features of industrial products, it is challenging to detect the occluded parts only from projections. Limited by the data acquisition and reconstruction speeds, CT-based detection methods do not achieve real-time detection. To solve the above problems, we design an end-to-end single-projection 3D segmentation network. For a specific product, the network adopts a single projection as input to segment product components and output 3D segmentation results. In this study, the feasibility of the network was verified against data containing several typical assembly errors. The qualitative and quantitative results reveal that the segmentation results can meet industrial assembly real-time detection requirements and exhibit high robustness to noise and component occlusion.
Collapse
Affiliation(s)
- Xuan Li
- State Key Lab for Electronic Testing Technology, North University of China, Taiyuan 030051, China
| | - Sukai Wang
- State Key Lab for Electronic Testing Technology, North University of China, Taiyuan 030051, China
| | - Xiaodong Niu
- State Key Lab for Electronic Testing Technology, North University of China, Taiyuan 030051, China
| | - Liming Wang
- State Key Lab for Electronic Testing Technology, North University of China, Taiyuan 030051, China
| | - Ping Chen
- State Key Lab for Electronic Testing Technology, North University of China, Taiyuan 030051, China
| |
Collapse
|
10
|
Wang Y, Yan G, Zhu H, Buch S, Wang Y, Haacke EM, Hua J, Zhong Z. VC-Net: Deep Volume-Composition Networks for Segmentation and Visualization of Highly Sparse and Noisy Image Data. IEEE TRANSACTIONS ON VISUALIZATION AND COMPUTER GRAPHICS 2021; 27:1301-1311. [PMID: 33048701 DOI: 10.1109/tvcg.2020.3030374] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
The fundamental motivation of the proposed work is to present a new visualization-guided computing paradigm to combine direct 3D volume processing and volume rendered clues for effective 3D exploration. For example, extracting and visualizing microstructures in-vivo have been a long-standing challenging problem. However, due to the high sparseness and noisiness in cerebrovasculature data as well as highly complex geometry and topology variations of micro vessels, it is still extremely challenging to extract the complete 3D vessel structure and visualize it in 3D with high fidelity. In this paper, we present an end-to-end deep learning method, VC-Net, for robust extraction of 3D microvascular structure through embedding the image composition, generated by maximum intensity projection (MIP), into the 3D volumetric image learning process to enhance the overall performance. The core novelty is to automatically leverage the volume visualization technique (e.g., MIP - a volume rendering scheme for 3D volume images) to enhance the 3D data exploration at the deep learning level. The MIP embedding features can enhance the local vessel signal (through canceling out the noise) and adapt to the geometric variability and scalability of vessels, which is of great importance in microvascular tracking. A multi-stream convolutional neural network (CNN) framework is proposed to effectively learn the 3D volume and 2D MIP feature vectors, respectively, and then explore their inter-dependencies in a joint volume-composition embedding space by unprojecting the 2D feature vectors into the 3D volume embedding space. It is noted that the proposed framework can better capture the small/micro vessels and improve the vessel connectivity. To our knowledge, this is the first time that a deep learning framework is proposed to construct a joint convolutional embedding space, where the computed vessel probabilities from volume rendering based 2D projection and 3D volume can be explored and integrated synergistically. Experimental results are evaluated and compared with the traditional 3D vessel segmentation methods and the state-of-the-art in deep learning, by using extensive public and real patient (micro- )cerebrovascular image datasets. The application of this accurate segmentation and visualization of sparse and complicated 3D microvascular structure facilitated by our method demonstrates the potential in a powerful MR arteriogram and venogram diagnosis of vascular disease.
Collapse
|
11
|
Statistical deformation reconstruction using multi-organ shape features for pancreatic cancer localization. Med Image Anal 2021; 67:101829. [DOI: 10.1016/j.media.2020.101829] [Citation(s) in RCA: 9] [Impact Index Per Article: 2.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2020] [Revised: 08/12/2020] [Accepted: 09/12/2020] [Indexed: 11/20/2022]
|
12
|
Lei Y, Tian Z, Wang T, Higgins K, Bradley JD, Curran WJ, Liu T, Yang X. Deep learning-based real-time volumetric imaging for lung stereotactic body radiation therapy: a proof of concept study. Phys Med Biol 2020; 65:235003. [PMID: 33080578 PMCID: PMC11756341 DOI: 10.1088/1361-6560/abc303] [Citation(s) in RCA: 22] [Impact Index Per Article: 4.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/25/2022]
Abstract
Due to the inter- and intra- variation of respiratory motion, it is highly desired to provide real-time volumetric images during the treatment delivery of lung stereotactic body radiation therapy (SBRT) for accurate and active motion management. In this proof-of-concept study, we propose a novel generative adversarial network integrated with perceptual supervision to derive instantaneous volumetric images from a single 2D projection. Our proposed network, named TransNet, consists of three modules, i.e. encoding, transformation and decoding modules. Rather than only using image distance loss between the generated 3D images and the ground truth 3D CT images to supervise the network, perceptual loss in feature space is integrated into loss function to force the TransNet to yield accurate lung boundary. Adversarial supervision is also used to improve the realism of generated 3D images. We conducted a simulation study on 20 patient cases, who had received lung SBRT treatments in our institution and undergone 4D-CT simulation, and evaluated the efficacy and robustness of our method for four different projection angles, i.e. 0°, 30°, 60° and 90°. For each 3D CT image set of a breathing phase, we simulated its 2D projections at these angles. For each projection angle, a patient's 3D CT images of 9 phases and the corresponding 2D projection data were used to train our network for that specific patient, with the remaining phase used for testing. The mean absolute error of the 3D images obtained by our method are 99.3 ± 14.1 HU. The peak signal-to-noise ratio and structural similarity index metric within the tumor region of interest are 15.4 ± 2.5 dB and 0.839 ± 0.090, respectively. The center of mass distance between the manual tumor contours on the 3D images obtained by our method and the manual tumor contours on the corresponding 3D phase CT images are within 2.6 mm, with a mean value of 1.26 mm averaged over all the cases. Our method has also been validated in a simulated challenging scenario with increased respiratory motion amplitude and tumor shrinkage, and achieved acceptable results. Our experimental results demonstrate the feasibility and efficacy of our 2D-to-3D method for lung cancer patients, which provides a potential solution for in-treatment real-time on-board volumetric imaging for tumor tracking and dose delivery verification to ensure the effectiveness of lung SBRT treatment.
Collapse
Affiliation(s)
- Yang Lei
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA, 30322, United States of America
- Co-first author
| | - Zhen Tian
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA, 30322, United States of America
- Co-first author
| | - Tonghe Wang
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA, 30322, United States of America
| | - Kristin Higgins
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA, 30322, United States of America
| | - Jeffrey D Bradley
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA, 30322, United States of America
| | - Walter J Curran
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA, 30322, United States of America
| | - Tian Liu
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA, 30322, United States of America
| | - Xiaofeng Yang
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA, 30322, United States of America
| |
Collapse
|
13
|
Tong F, Nakao M, Wu S, Nakamura M, Matsuda T. X-ray2Shape: Reconstruction of 3D Liver Shape from a Single 2D Projection Image. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2020; 2020:1608-1611. [PMID: 33018302 DOI: 10.1109/embc44109.2020.9176655] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
Computed tomography (CT) and magnetic resonance imaging (MRI) scanners measure three-dimensional (3D) images of patients. However, only low-dimensional local two-dimensional (2D) images may be obtained during surgery or radiotherapy. Although computer vision techniques have shown that 3D shapes can be estimated from multiple 2D images, shape reconstruction from a single 2D image such as an endoscopic image or an X-ray image remains a challenge. In this study, we propose X-ray2Shape, which permits a deep learning-based 3D organ mesh to be reconstructed from a single 2D projection image. The method learns the mesh deformation from a mean template and deep features computed from the individual projection images. Experiments with organ meshes and digitally reconstructed radiograph (DRR) images of abdominal regions were performed to confirm the estimation performance of the methods.
Collapse
|
14
|
Choudhary A, Tong L, Zhu Y, Wang MD. Advancing Medical Imaging Informatics by Deep Learning-Based Domain Adaptation. Yearb Med Inform 2020; 29:129-138. [PMID: 32823306 PMCID: PMC7442502 DOI: 10.1055/s-0040-1702009] [Citation(s) in RCA: 25] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/08/2023] Open
Abstract
INTRODUCTION There has been a rapid development of deep learning (DL) models for medical imaging. However, DL requires a large labeled dataset for training the models. Getting large-scale labeled data remains a challenge, and multi-center datasets suffer from heterogeneity due to patient diversity and varying imaging protocols. Domain adaptation (DA) has been developed to transfer the knowledge from a labeled data domain to a related but unlabeled domain in either image space or feature space. DA is a type of transfer learning (TL) that can improve the performance of models when applied to multiple different datasets. OBJECTIVE In this survey, we review the state-of-the-art DL-based DA methods for medical imaging. We aim to summarize recent advances, highlighting the motivation, challenges, and opportunities, and to discuss promising directions for future work in DA for medical imaging. METHODS We surveyed peer-reviewed publications from leading biomedical journals and conferences between 2017-2020, that reported the use of DA in medical imaging applications, grouping them by methodology, image modality, and learning scenarios. RESULTS We mainly focused on pathology and radiology as application areas. Among various DA approaches, we discussed domain transformation (DT) and latent feature-space transformation (LFST). We highlighted the role of unsupervised DA in image segmentation and described opportunities for future development. CONCLUSION DA has emerged as a promising solution to deal with the lack of annotated training data. Using adversarial techniques, unsupervised DA has achieved good performance, especially for segmentation tasks. Opportunities include domain transferability, multi-modal DA, and applications that benefit from synthetic data.
Collapse
Affiliation(s)
- Anirudh Choudhary
- Department of Computational Science and Engineering, Georgia Institute of Technology, GA, USA
| | - Li Tong
- Department of Biomedical Engineering, Georgia Institute of Technology and Emory University, GA, USA
| | - Yuanda Zhu
- School of Electrical and Computer Engineering, Georgia Institute of Technology, GA, USA
| | - May D. Wang
- Department of Computational Science and Engineering, Georgia Institute of Technology, GA, USA
- Department of Biomedical Engineering, Georgia Institute of Technology and Emory University, GA, USA
- School of Electrical and Computer Engineering, Georgia Institute of Technology, GA, USA
| |
Collapse
|
15
|
Virgolin M, Wang Z, Alderliesten T, Bosman PAN. Machine learning for the prediction of pseudorealistic pediatric abdominal phantoms for radiation dose reconstruction. J Med Imaging (Bellingham) 2020; 7:046501. [PMID: 32743017 PMCID: PMC7390892 DOI: 10.1117/1.jmi.7.4.046501] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/07/2019] [Accepted: 07/15/2020] [Indexed: 11/14/2022] Open
Abstract
Purpose: Current phantoms used for the dose reconstruction of long-term childhood cancer survivors lack individualization. We design a method to predict highly individualized abdominal three-dimensional (3-D) phantoms automatically. Approach: We train machine learning (ML) models to map (2-D) patient features to 3-D organ-at-risk (OAR) metrics upon a database of 60 pediatric abdominal computed tomographies with liver and spleen segmentations. Next, we use the models in an automatic pipeline that outputs a personalized phantom given the patient's features, by assembling 3-D imaging from the database. A step to improve phantom realism (i.e., avoid OAR overlap) is included. We compare five ML algorithms, in terms of predicting OAR left-right (LR), anterior-posterior (AP), inferior-superior (IS) positions, and surface Dice-Sørensen coefficient (sDSC). Furthermore, two existing human-designed phantom construction criteria and two additional control methods are investigated for comparison. Results: Different ML algorithms result in similar test mean absolute errors: ∼ 8 mm for liver LR, IS, and spleen AP, IS; ∼ 5 mm for liver AP and spleen LR; ∼ 80 % for abdomen sDSC; and ∼ 60 % to 65% for liver and spleen sDSC. One ML algorithm (GP-GOMEA) significantly performs the best for 6/9 metrics. The control methods and the human-designed criteria in particular perform generally worse, sometimes substantially ( + 5 - mm error for spleen IS, - 10 % sDSC for liver). The automatic step to improve realism generally results in limited metric accuracy loss, but fails in one case (out of 60). Conclusion: Our ML-based pipeline leads to phantoms that are significantly and substantially more individualized than currently used human-designed criteria.
Collapse
Affiliation(s)
- Marco Virgolin
- Centrum Wiskunde and Informatica, Life Sciences and Health Group, Amsterdam, The Netherlands
| | - Ziyuan Wang
- Amsterdam UMC, University of Amsterdam, Department of Radiation Oncology, Amsterdam, The Netherlands
| | - Tanja Alderliesten
- Amsterdam UMC, University of Amsterdam, Department of Radiation Oncology, Amsterdam, The Netherlands
- Leiden University Medical Center, Department of Radiation Oncology, Leiden, The Netherlands
| | - Peter A N Bosman
- Centrum Wiskunde and Informatica, Life Sciences and Health Group, Amsterdam, The Netherlands
- Delft University of Technology, Algorithmics Group, Delft, The Netherlands
| |
Collapse
|
16
|
Spectral Processing for Denoising and Compression of 3D Meshes Using Dynamic Orthogonal Iterations. J Imaging 2020; 6:jimaging6060055. [PMID: 34460601 PMCID: PMC8321066 DOI: 10.3390/jimaging6060055] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.2] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/14/2020] [Revised: 06/18/2020] [Accepted: 06/23/2020] [Indexed: 11/17/2022] Open
Abstract
Recently, spectral methods have been extensively used in the processing of 3D meshes. They usually take advantage of some unique properties that the eigenvalues and the eigenvectors of the decomposed Laplacian matrix have. However, despite their superior behavior and performance, they suffer from computational complexity, especially while the number of vertices of the model increases. In this work, we suggest the use of a fast and efficient spectral processing approach applied to dense static and dynamic 3D meshes, which can be ideally suited for real-time denoising and compression applications. To increase the computational efficiency of the method, we exploit potential spectral coherence between adjacent parts of a mesh and then we apply an orthogonal iteration approach for the tracking of the graph Laplacian eigenspaces. Additionally, we present a dynamic version that automatically identifies the optimal subspace size that satisfies a given reconstruction quality threshold. In this way, we overcome the problem of the perceptual distortions, due to the fixed number of subspace sizes that is used for all the separated parts individually. Extensive simulations carried out using different 3D models in different use cases (i.e., compression and denoising), showed that the proposed approach is very fast, especially in comparison with the SVD based spectral processing approaches, while at the same time the quality of the reconstructed models is of similar or even better reconstruction quality. The experimental analysis also showed that the proposed approach could also be used by other denoising methods as a preprocessing step, in order to optimize the reconstruction quality of their results and decrease their computational complexity since they need fewer iterations to converge.
Collapse
|