1
|
Siewerdsen JH. Image quality models for 2D and 3D x-ray imaging systems: A perspective vignette. Med Phys 2022. [PMID: 36542332 DOI: 10.1002/mp.16051] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/06/2022] [Revised: 10/12/2022] [Accepted: 10/12/2022] [Indexed: 12/24/2022] Open
Abstract
Image quality models based on cascaded systems analysis and task-based imaging performance were an important aspect of the emergence of 2D and 3D digital x-ray systems over the last 25 years. This perspective vignette offers cursory review of such developments and personal insights that may not be obvious within previously published scientific literature. The vignette traces such models to the mid-1990s, when flat-panel x-ray detectors were emerging as a new base technology for digital radiography and benefited from the rigorous, objective characterization of imaging performance gained from such models. The connection of models for spatial resolution and noise to spatial-frequency-dependent descriptors of imaging task provided a useful framework for system optimization that helped to accelerate the development of new technologies to first clinical use. Extension of the models to new technologies and applications is also described, including dual-energy imaging, photon-counting detectors, phase contrast imaging, tomosynthesis, cone-beam CT, 3D image reconstruction, and image registration.
Collapse
Affiliation(s)
- Jeffrey H Siewerdsen
- Departments of Imaging Physics, Neurosurgery, and Radiation Physics, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA.,Director of Surgical Data Science, Institute for Data Science in Oncology, The University of Texas MD Anderson Cancer Center, Houston, Texas, USA
| |
Collapse
|
2
|
Huang Y, Jones CK, Zhang X, Johnston A, Waktola S, Aygun N, Witham TF, Bydon A, Theodore N, Helm PA, Siewerdsen JH, Uneri A. Multi-perspective region-based CNNs for vertebrae labeling in intraoperative long-length images. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 227:107222. [PMID: 36370597 DOI: 10.1016/j.cmpb.2022.107222] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/17/2022] [Revised: 10/31/2022] [Accepted: 11/02/2022] [Indexed: 06/16/2023]
Abstract
PURPOSE Effective aggregation of intraoperative x-ray images that capture the patient anatomy from multiple view-angles has the potential to enable and improve automated image analysis that can be readily performed during surgery. We present multi-perspective region-based neural networks that leverage knowledge of the imaging geometry for automatic vertebrae labeling in Long-Film images - a novel tomographic imaging modality with an extended field-of-view for spine imaging. METHOD A multi-perspective network architecture was designed to exploit small view-angle disparities produced by a multi-slot collimator and consolidate information from overlapping image regions. A second network incorporates large view-angle disparities to jointly perform labeling on images from multiple views (viz., AP and lateral). A recurrent module incorporates contextual information and enforce anatomical order for the detected vertebrae. The three modules are combined to form the multi-view multi-slot (MVMS) network for labeling vertebrae using images from all available perspectives. The network was trained on images synthesized from 297 CT images and tested on 50 AP and 50 lateral Long-Film images acquired from 13 cadaveric specimens. Labeling performance of the multi-perspective networks was evaluated with respect to the number of vertebrae appearances and presence of surgical instrumentation. RESULTS The MVMS network achieved an F1 score of >96% and an average vertebral localization error of 3.3 mm, with 88.3% labeling accuracy on both AP and lateral images - (15.5% and 35.0% higher than conventional Faster R-CNN on AP and lateral views, respectively). Aggregation of multiple appearances of the same vertebra using the multi-slot network significantly improved the labeling accuracy (p < 0.05). Using the multi-view network, labeling accuracy on the more challenging lateral views was improved to the same level as that of the AP views. The approach demonstrated robustness to the presence of surgical instrumentation, commonly encountered in intraoperative images, and achieved comparable performance in images with and without instrumentation (88.9% vs. 91.2% labeling accuracy). CONCLUSION The MVMS network demonstrated effective multi-perspective aggregation, providing means for accurate, automated vertebrae labeling during spine surgery. The algorithms may be generalized to other imaging tasks and modalities that involve multiple views with view-angle disparities (e.g., bi-plane radiography). Predicted labels can help avoid adverse events during surgery (e.g., wrong-level surgery), establish correspondence with labels in preoperative modalities to facilitate image registration, and enable automated measurement of spinal alignment metrics for intraoperative assessment of spinal curvature.
Collapse
Affiliation(s)
- Y Huang
- Department of Biomedical Engineering, Johns Hopkins University, Baltimore MD, United States
| | - C K Jones
- Department of Computer Science, Johns Hopkins University, Baltimore MD, United States
| | - X Zhang
- Department of Biomedical Engineering, Johns Hopkins University, Baltimore MD, United States
| | - A Johnston
- Department of Biomedical Engineering, Johns Hopkins University, Baltimore MD, United States
| | - S Waktola
- Department of Biomedical Engineering, Johns Hopkins University, Baltimore MD, United States
| | - N Aygun
- Department of Radiology, Johns Hopkins Medicine, Baltimore MD, United States
| | - T F Witham
- Department of Neurosurgery, Johns Hopkins Medicine, Baltimore MD, United States
| | - A Bydon
- Department of Neurosurgery, Johns Hopkins Medicine, Baltimore MD, United States
| | - N Theodore
- Department of Neurosurgery, Johns Hopkins Medicine, Baltimore MD, United States
| | - P A Helm
- Medtronic, Littleton MA, United States
| | - J H Siewerdsen
- Department of Biomedical Engineering, Johns Hopkins University, Baltimore MD, United States; Department of Computer Science, Johns Hopkins University, Baltimore MD, United States; Department of Radiology, Johns Hopkins Medicine, Baltimore MD, United States; Department of Neurosurgery, Johns Hopkins Medicine, Baltimore MD, United States; Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston TX, United States
| | - A Uneri
- Department of Biomedical Engineering, Johns Hopkins University, Baltimore MD, United States.
| |
Collapse
|
3
|
Ketcha MD, De Silva T, Han R, Uneri A, Vogt S, Kleinszig G, Siewerdsen JH. Learning-based deformable image registration: effect of statistical mismatch between train and test images. J Med Imaging (Bellingham) 2019; 6:044008. [PMID: 31853461 DOI: 10.1117/1.jmi.6.4.044008] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/25/2019] [Accepted: 11/18/2019] [Indexed: 01/11/2023] Open
Abstract
Convolutional neural networks (CNNs) offer a promising means to achieve fast deformable image registration with accuracy comparable to conventional, physics-based methods. A persistent question with CNN methods, however, is whether they will be able to generalize to data outside of the training set. We investigated this question of mismatch between train and test data with respect to first- and second-order image statistics (e.g., spatial resolution, image noise, and power spectrum). A UNet-based architecture was built and trained on simulated CT images for various conditions of image noise (dose), spatial resolution, and deformation magnitude. Target registration error was measured as a function of the difference in statistical properties between the test and training data. Generally, registration error is minimized when the training data exactly match the statistics of the test data; however, networks trained with data exhibiting a diversity in statistical characteristics generalized well across the range of statistical conditions considered. Furthermore, networks trained on simulated image content with first- and second-order statistics selected to match that of real anatomical data were shown to provide reasonable registration performance on real anatomical content, offering potential new means for data augmentation. Characterizing the behavior of a CNN in the presence of statistical mismatch is an important step in understanding how these networks behave when deployed on new, unobserved data. Such characterization can inform decisions on whether retraining is necessary and can guide the data collection and/or augmentation process for training.
Collapse
Affiliation(s)
- Michael D Ketcha
- Johns Hopkins University, Department of Biomedical Engineering, Baltimore, Maryland, United States
| | - Tharindu De Silva
- Johns Hopkins University, Department of Biomedical Engineering, Baltimore, Maryland, United States
| | - Runze Han
- Johns Hopkins University, Department of Biomedical Engineering, Baltimore, Maryland, United States
| | - Ali Uneri
- Johns Hopkins University, Department of Biomedical Engineering, Baltimore, Maryland, United States
| | | | | | - Jeffrey H Siewerdsen
- Johns Hopkins University, Department of Biomedical Engineering, Baltimore, Maryland, United States
| |
Collapse
|