1
|
Hein J, Cavalcanti N, Suter D, Zingg L, Carrillo F, Calvet L, Farshad M, Navab N, Pollefeys M, Fürnstahl P. Next-generation surgical navigation: Marker-less multi-view 6DoF pose estimation of surgical instruments. Med Image Anal 2025; 103:103613. [PMID: 40381257 DOI: 10.1016/j.media.2025.103613] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/22/2023] [Revised: 04/10/2025] [Accepted: 04/18/2025] [Indexed: 05/20/2025]
Abstract
State-of-the-art research of traditional computer vision is increasingly leveraged in the surgical domain. A particular focus in computer-assisted surgery is to replace marker-based tracking systems for instrument localization with pure image-based 6DoF pose estimation using deep-learning methods. However, state-of-the-art single-view pose estimation methods do not yet meet the accuracy required for surgical navigation. In this context, we investigate the benefits of multi-view setups for highly accurate and occlusion-robust 6DoF pose estimation of surgical instruments and derive recommendations for an ideal camera system that addresses the challenges in the operating room. Our contributions are threefold. First, we present a multi-view RGB-D video dataset of ex-vivo spine surgeries, captured with static and head-mounted cameras and including rich annotations for surgeon, instruments, and patient anatomy. Second, we perform an extensive evaluation of three state-of-the-art single-view and multi-view pose estimation methods, analyzing the impact of camera quantities and positioning, limited real-world data, and static, hybrid, or fully mobile camera setups on the pose accuracy, occlusion robustness, and generalizability. Third, we design a multi-camera system for marker-less surgical instrument tracking, achieving an average position error of 1.01mm and orientation error of 0.89° for a surgical drill, and 2.79mm and 3.33° for a screwdriver under optimal conditions. Our results demonstrate that marker-less tracking of surgical instruments is becoming a feasible alternative to existing marker-based systems.
Collapse
Affiliation(s)
- Jonas Hein
- Research in Orthopedic Computer Science, Balgrist University Hospital, University of Zurich, Zurich, Switzerland; Computer Vision and Geometry Group, ETH Zurich, Zurich, Switzerland.
| | - Nicola Cavalcanti
- Research in Orthopedic Computer Science, Balgrist University Hospital, University of Zurich, Zurich, Switzerland
| | - Daniel Suter
- Balgrist University Hospital, University of Zurich, Zurich, Switzerland
| | - Lukas Zingg
- Balgrist University Hospital, University of Zurich, Zurich, Switzerland
| | - Fabio Carrillo
- Research in Orthopedic Computer Science, Balgrist University Hospital, University of Zurich, Zurich, Switzerland; OR-X Translational Center for Surgery, Balgrist University Hospital, University of Zurich, Zurich, Switzerland
| | - Lilian Calvet
- OR-X Translational Center for Surgery, Balgrist University Hospital, University of Zurich, Zurich, Switzerland
| | - Mazda Farshad
- Balgrist University Hospital, University of Zurich, Zurich, Switzerland
| | - Nassir Navab
- Computer Aided Medical Procedures, Technical University Munich, Munich, Germany
| | - Marc Pollefeys
- Computer Vision and Geometry Group, ETH Zurich, Zurich, Switzerland
| | - Philipp Fürnstahl
- Research in Orthopedic Computer Science, Balgrist University Hospital, University of Zurich, Zurich, Switzerland; OR-X Translational Center for Surgery, Balgrist University Hospital, University of Zurich, Zurich, Switzerland
| |
Collapse
|
2
|
Spinos D, Martinos A, Petsiou DP, Mistry N, Garas G. Artificial Intelligence in Temporal Bone Imaging: A Systematic Review. Laryngoscope 2025; 135:973-981. [PMID: 39352072 DOI: 10.1002/lary.31809] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/08/2024] [Revised: 08/03/2024] [Accepted: 09/17/2024] [Indexed: 10/03/2024]
Abstract
OBJECTIVE The human temporal bone comprises more than 30 identifiable anatomical components. With the demand for precise image interpretation in this complex region, the utilization of artificial intelligence (AI) applications is steadily increasing. This systematic review aims to highlight the current role of AI in temporal bone imaging. DATA SOURCES A Systematic Review of English Publications searching MEDLINE (PubMed), COCHRANE Library, and EMBASE. REVIEW METHODS The search algorithm employed consisted of key items such as 'artificial intelligence,' 'machine learning,' 'deep learning,' 'neural network,' 'temporal bone,' and 'vestibular schwannoma.' Additionally, manual retrieval was conducted to capture any studies potentially missed in our initial search. All abstracts and full texts were screened based on our inclusion and exclusion criteria. RESULTS A total of 72 studies were included. 95.8% were retrospective and 88.9% were based on internal databases. Approximately two-thirds involved an AI-to-human comparison. Computed tomography (CT) was the imaging modality in 54.2% of the studies, with vestibular schwannoma (VS) being the most frequent study item (37.5%). Fifty-eight out of 72 articles employed neural networks, with 72.2% using various types of convolutional neural network models. Quality assessment of the included publications yielded a mean score of 13.6 ± 2.5 on a 20-point scale based on the CONSORT-AI extension. CONCLUSION Current research data highlight AI's potential in enhancing diagnostic accuracy with faster results and decreased performance errors compared to those of clinicians, thus improving patient care. However, the shortcomings of the existing research, often marked by heterogeneity and variable quality, underscore the need for more standardized methodological approaches to ensure the consistency and reliability of future data. LEVEL OF EVIDENCE NA Laryngoscope, 135:973-981, 2025.
Collapse
Affiliation(s)
- Dimitrios Spinos
- South Warwickshire NHS Foundation Trust, Warwick, UK
- University Hospitals Birmingham NHS Foundation Trust, Birmingham, UK
| | - Anastasios Martinos
- National and Kapodistrian University of Athens School of Medicine, Athens, Greece
| | | | - Nina Mistry
- Gloucestershire Hospitals NHS Foundation Trust, ENT, Head and Neck Surgery, Gloucester, UK
| | - George Garas
- Surgical Innovation Centre, Department of Surgery and Cancer, Imperial College London, St. Mary's Hospital, London, UK
- Athens Medical Center, Marousi & Psychiko Clinic, Athens, Greece
| |
Collapse
|
3
|
Kiran U, Bhat SN, Anitha H, Naik RR. Feature-based multimodal registration framework for vertebral pose estimation. EUROPEAN SPINE JOURNAL : OFFICIAL PUBLICATION OF THE EUROPEAN SPINE SOCIETY, THE EUROPEAN SPINAL DEFORMITY SOCIETY, AND THE EUROPEAN SECTION OF THE CERVICAL SPINE RESEARCH SOCIETY 2024; 33:2251-2260. [PMID: 38104308 DOI: 10.1007/s00586-023-08054-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/21/2023] [Revised: 08/21/2023] [Accepted: 11/12/2023] [Indexed: 12/19/2023]
Abstract
PURPOSE The reliable estimation of the vertebral body posture helps to aid a safe and effective spine surgery. The proposed work aims to present an MR to X-ray image registration to assess the 3D pose of the vertebral body during spine surgery. The 3D assessment of vertebral pose assists in analyzing the position and orientation of the vertebral body to provide information during various clinical diagnosis conditions such as curvature estimation and pedicle screw insertion surgery. METHODS The proposed feature-based registration framework extracted vertebral end plates to avoid the mismatch between the intensities of MR and X-ray images. Using the projection matrix, the segmented MRI is forward projected and then registered to the X-ray image using binary image matching similarity and the CMA-ES optimizer. RESULTS The proposed method estimated the vertebral pose by registering the simulated X-ray onto pre-operative MRI. To evaluate the efficacy of the proposed approach, a certain number of experiments are carried out on the simulated dataset. CONCLUSION The proposed method is a fast and accurate registration method that can provide 3D information about the vertebral body. This 3D information is useful to improve accuracy during various clinical diagnoses.
Collapse
Affiliation(s)
- Usha Kiran
- Department of Electronics and Communication Engineering, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India
| | - Shyamasunder N Bhat
- Department of Orthopaedics, Kasturba Medical College, Manipal, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India.
| | - H Anitha
- Department of Electronics and Communication Engineering, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India.
| | - Roshan Ramakrishna Naik
- Department of Electronics and Communication Engineering, St. Joseph Engineering College, Vamanjoor, Mangalore, Karnataka, 575028, India
| |
Collapse
|
4
|
Killeen BD, Chaudhary S, Osgood G, Unberath M. Take a shot! Natural language control of intelligent robotic X-ray systems in surgery. Int J Comput Assist Radiol Surg 2024; 19:1165-1173. [PMID: 38619790 PMCID: PMC11178437 DOI: 10.1007/s11548-024-03120-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/04/2024] [Accepted: 03/22/2024] [Indexed: 04/16/2024]
Abstract
PURPOSE The expanding capabilities of surgical systems bring with them increasing complexity in the interfaces that humans use to control them. Robotic C-arm X-ray imaging systems, for instance, often require manipulation of independent axes via joysticks, while higher-level control options hide inside device-specific menus. The complexity of these interfaces hinder "ready-to-hand" use of high-level functions. Natural language offers a flexible, familiar interface for surgeons to express their desired outcome rather than remembering the steps necessary to achieve it, enabling direct access to task-aware, patient-specific C-arm functionality. METHODS We present an English language voice interface for controlling a robotic X-ray imaging system with task-aware functions for pelvic trauma surgery. Our fully integrated system uses a large language model (LLM) to convert natural spoken commands into machine-readable instructions, enabling low-level commands like "Tilt back a bit," to increase the angular tilt or patient-specific directions like, "Go to the obturator oblique view of the right ramus," based on automated image analysis. RESULTS We evaluate our system with 212 prompts provided by an attending physician, in which the system performed satisfactory actions 97% of the time. To test the fully integrated system, we conduct a real-time study in which an attending physician placed orthopedic hardware along desired trajectories through an anthropomorphic phantom, interacting solely with an X-ray system via voice. CONCLUSION Voice interfaces offer a convenient, flexible way for surgeons to manipulate C-arms based on desired outcomes rather than device-specific processes. As LLMs grow increasingly capable, so too will their applications in supporting higher-level interactions with surgical assistance systems.
Collapse
Affiliation(s)
- Benjamin D Killeen
- Laboratory for Computational Sensing and Robotics, Johns Hopkins University, Baltimore, MD, 21218, USA.
| | - Shreayan Chaudhary
- Laboratory for Computational Sensing and Robotics, Johns Hopkins University, Baltimore, MD, 21218, USA
| | - Greg Osgood
- Department of Orthopaedic Surgery, Johns Hopkins University, Baltimore, MD, 212187, USA
| | - Mathias Unberath
- Laboratory for Computational Sensing and Robotics, Johns Hopkins University, Baltimore, MD, 21218, USA
| |
Collapse
|
5
|
Long Z, Chi Y, Yu X, Jiang Z, Yang D. ArthroNavi framework: stereo endoscope-guided instrument localization for arthroscopic minimally invasive surgeries. JOURNAL OF BIOMEDICAL OPTICS 2023; 28:106002. [PMID: 37841507 PMCID: PMC10576396 DOI: 10.1117/1.jbo.28.10.106002] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/05/2023] [Revised: 09/24/2023] [Accepted: 09/29/2023] [Indexed: 10/17/2023]
Abstract
SIGNIFICANCE As an example of a minimally invasive arthroscopic surgical procedure, arthroscopic osteochondral autograft transplantation (OAT) is a common option for repairing focal cartilage defects in the knee joints. Arthroscopic OAT offers considerable benefits to patients, such as less post-operative pain and shorter hospital stays. However, performing OAT arthroscopically is an extremely demanding task because the osteochondral graft harvester must remain perpendicular to the cartilage surface to avoid differences in angulation. AIM We present a practical ArthroNavi framework for instrument pose localization by combining a self-developed stereo endoscopy with electromagnetic computation, which equips surgeons with surgical navigation assistance that eases the operational constraints of arthroscopic OAT surgery. APPROACH A prototype of a stereo endoscope specifically fit for a texture-less scene is introduced extensively. Then, the proposed framework employs the semi-global matching algorithm integrating the matching cubes method for real-time processing of the 3D point cloud. To address issues regarding initialization and occlusion, a displaying method based on patient tracking coordinates is proposed for intra-operative robust navigation. A geometrical constraint method that utilizes the 3D point cloud is used to compute a pose for the instrument. Finally, a hemisphere tabulation method is presented for pose accuracy evaluation. RESULTS Experimental results show that our endoscope achieves 3D shape measurement with an accuracy of < 730 μ m . The mean error of pose localization is 15.4 deg (range of 10.3 deg to 21.3 deg; standard deviation of 3.08 deg) in our ArthroNavi method, which is within the same order of magnitude as that achieved by experienced surgeons using a freehand technique. CONCLUSIONS The effectiveness of the proposed ArthroNavi has been validated on a phantom femur. The potential contribution of this framework may provide a new computer-aided option for arthroscopic OAT surgery.
Collapse
Affiliation(s)
- Zhongjie Long
- Beijing Information Science & Technology University, School of Electromechanical Engineering, Beijing, China
| | - Yongting Chi
- Beijing Information Science & Technology University, School of Electromechanical Engineering, Beijing, China
| | - Xiaotong Yu
- Guang’anmen Hospital, China Academy of Chinese Medical Sciences, Beijing, China
| | - Zhouxiang Jiang
- Beijing Information Science & Technology University, School of Electromechanical Engineering, Beijing, China
| | - Dejin Yang
- Beijing Jishuitan Hospital, Capital Medical School, 4th Clinical College of Peking University, Department of Orthopedics, Beijing, China
| |
Collapse
|
6
|
Kiran U, Ramakrishna Naik R, Bhat SN, H A. Evaluating similarity measure for multimodal 3D to 2D registration. Biomed Phys Eng Express 2023; 9:055015. [PMID: 37487480 DOI: 10.1088/2057-1976/ace9e1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/26/2023] [Accepted: 07/24/2023] [Indexed: 07/26/2023]
Abstract
The 3D to 2D registration technique in spine surgery is vital to aid surgeons in avoiding the wrong site surgery by estimating the vertebral pose. The vertebral poses are estimated by generating the spatial correspondence relationship between pre-operative MR with intra-operative x-ray images, then evaluated using a similarity measure. Different similarity measures are used in 3D to 2D registration techniques to assess the spatial correspondence between the pre-operative and intra-operative images. However, to evaluate the registration performance of the similarity measures, the proposed framework employs three different similarity measures: Binary Image Matching, Dice Coefficients, and Normalized Cross-correlation technique to compare the images based on pixel positions. The registration accuracy of the proposed similarity measures is compared based on the mean Target Registration Error, mean Iteration Times, and success rate. In the absence of simulated test images, the experiment is conducted on the simulated AP and Lateral test images. The experiment conducted on the simulated test images shows that all three similarity measures work well for the feature based 3D to 2D registration in that BIM gives better results. The experiment also indicates high registration accuracy when the initial displacements are varied up to ±20 mm and ±100of the translational and rotational parameters, respectively, for three similarity measures.
Collapse
Affiliation(s)
- Usha Kiran
- Department of Electronics and Communication Engineering, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India
| | - Roshan Ramakrishna Naik
- Department of Electronics and Communication Engineering, St. Joseph Engineering College, Vamanjoor, Mangalore, Karnataka, 575028, India
| | - Shyamasunder N Bhat
- Department of Orthopaedics, Kasturba Medical College, Manipal, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India
| | - Anitha H
- Department of Electronics and Communication Engineering, Manipal Institute of Technology, Manipal Academy of Higher Education, Manipal, Karnataka, 576104, India
| |
Collapse
|
7
|
Killeen BD, Gao C, Oguine KJ, Darcy S, Armand M, Taylor RH, Osgood G, Unberath M. An autonomous X-ray image acquisition and interpretation system for assisting percutaneous pelvic fracture fixation. Int J Comput Assist Radiol Surg 2023; 18:1201-1208. [PMID: 37213057 PMCID: PMC11002911 DOI: 10.1007/s11548-023-02941-y] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/22/2023] [Accepted: 04/25/2023] [Indexed: 05/23/2023]
Abstract
PURPOSE Percutaneous fracture fixation involves multiple X-ray acquisitions to determine adequate tool trajectories in bony anatomy. In order to reduce time spent adjusting the X-ray imager's gantry, avoid excess acquisitions, and anticipate inadequate trajectories before penetrating bone, we propose an autonomous system for intra-operative feedback that combines robotic X-ray imaging and machine learning for automated image acquisition and interpretation, respectively. METHODS Our approach reconstructs an appropriate trajectory in a two-image sequence, where the optimal second viewpoint is determined based on analysis of the first image. A deep neural network is responsible for detecting the tool and corridor, here a K-wire and the superior pubic ramus, respectively, in these radiographs. The reconstructed corridor and K-wire pose are compared to determine likelihood of cortical breach, and both are visualized for the clinician in a mixed reality environment that is spatially registered to the patient and delivered by an optical see-through head-mounted display. RESULTS We assess the upper bounds on system performance through in silico evaluation across 11 CTs with fractures present, in which the corridor and K-wire are adequately reconstructed. In post hoc analysis of radiographs across 3 cadaveric specimens, our system determines the appropriate trajectory to within 2.8 ± 1.3 mm and 2.7 ± 1.8[Formula: see text]. CONCLUSION An expert user study with an anthropomorphic phantom demonstrates how our autonomous, integrated system requires fewer images and lower movement to guide and confirm adequate placement compared to current clinical practice. Code and data are available.
Collapse
Affiliation(s)
| | - Cong Gao
- Johns Hopkins University, Baltimore, 21210, MD, USA
| | | | - Sean Darcy
- Johns Hopkins University, Baltimore, 21210, MD, USA
| | - Mehran Armand
- Johns Hopkins University, Baltimore, 21210, MD, USA
- Department of Orthopaedic Surgery, Johns Hopkins University, Baltimore, USA
| | | | - Greg Osgood
- Department of Orthopaedic Surgery, Johns Hopkins University, Baltimore, USA
| | | |
Collapse
|
8
|
Long Z, Chi Y, Yang D, Jiang Z, Bai L. Hemisphere Tabulation Method: An Ingenious Approach for Pose Evaluation of Instruments Using the Electromagnetic-Based Stereo Imaging Method. MICROMACHINES 2023; 14:446. [PMID: 36838146 PMCID: PMC9964370 DOI: 10.3390/mi14020446] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 12/26/2022] [Revised: 02/07/2023] [Accepted: 02/13/2023] [Indexed: 06/18/2023]
Abstract
Drilling of a bone surface often occurs in clinical orthopaedic surgery. The position and orientation of the instrument are the most important factors in this process. Theoretically, some mechanical components may assist in orienting an instrument to certain bone shapes, such as the knee joint and caput femoris. However, the mechanical assisting component does not seem to work in some confined spaces where the bone shape is a free-form surface. In this paper, we propose an ingenious hemisphere tabulation method (HTM) for assessing the pose accuracy of an instrument. The acquisition and assessment of HTM is conducted based on an electromagnetic-based stereo imaging method using a custom-made optical measurement unit, and the operation steps of HTM are described in detail. Experimental results based on 50 tests show that the HTM can identify ideal poses and the evaluated pose of an instrument location on a hemisphere model. The mean error of pose localisation is 7.24 deg, with a range of 1.35 to 15.84 and a standard of 3.66 deg, which is more accurate than our previous method.
Collapse
Affiliation(s)
- Zhongjie Long
- School of Electromechanical Engineering, Beijing Information Science & Technology University, Beijing 100192, China
- Key Laboratory of Modern Measurement & Control Technology, Ministry of Education, Beijing Information Science & Technology University, Beijing 100192, China
| | - Yongting Chi
- School of Electromechanical Engineering, Beijing Information Science & Technology University, Beijing 100192, China
| | - Dejin Yang
- Department of Orthopedics, Beijing Jishuitan Hospital, 4th Clinical College of Peking University, Beijing 100035, China
| | - Zhouxiang Jiang
- School of Electromechanical Engineering, Beijing Information Science & Technology University, Beijing 100192, China
- Key Laboratory of Modern Measurement & Control Technology, Ministry of Education, Beijing Information Science & Technology University, Beijing 100192, China
| | - Long Bai
- School of Electromechanical Engineering, Beijing Information Science & Technology University, Beijing 100192, China
- Key Laboratory of Modern Measurement & Control Technology, Ministry of Education, Beijing Information Science & Technology University, Beijing 100192, China
| |
Collapse
|
9
|
Wagner MG, Periyasamy S, Kutlu AZ, Pieper AA, Swietlik JF, Ziemlewicz TJ, Hall TL, Xu Z, Speidel MA, Jr FTL, Laeseke PF. An X-Ray C-Arm Guided Automatic Targeting System for Histotripsy. IEEE Trans Biomed Eng 2023; 70:592-602. [PMID: 35984807 PMCID: PMC9929026 DOI: 10.1109/tbme.2022.3198600] [Citation(s) in RCA: 9] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/10/2022]
Abstract
OBJECTIVE Histotripsy is an emerging noninvasive, nonionizing and nonthermal focal cancer therapy that is highly precise and can create a treatment zone of virtually any size and shape. Current histotripsy systems rely on ultrasound imaging to target lesions. However, deep or isoechoic targets obstructed by bowel gas or bone can often not be treated safely using ultrasound imaging alone. This work presents an alternative x-ray C-arm based targeting approach and a fully automated robotic targeting system. METHODS The approach uses conventional cone beam CT (CBCT) images to localize the target lesion and 2D fluoroscopy to determine the 3D position and orientation of the histotripsy transducer relative to the C-arm. The proposed pose estimation uses a digital model and deep learning-based feature segmentation to estimate the transducer focal point relative to the CBCT coordinate system. Additionally, the integrated robotic arm was calibrated to the C-arm by estimating the transducer pose for four preprogrammed transducer orientations and positions. The calibrated system can then automatically position the transducer such that the focal point aligns with any target selected in a CBCT image. RESULTS The accuracy of the proposed targeting approach was evaluated in phantom studies, where the selected target location was compared to the center of the spherical ablation zones in post-treatment CBCTs. The mean and standard deviation of the Euclidean distance was 1.4 ±0.5 mm. The mean absolute error of the predicted treatment radius was 0.5 ±0.5 mm. CONCLUSION CBCT-based histotripsy targeting enables accurate and fully automated treatment without ultrasound guidance. SIGNIFICANCE The proposed approach could considerably decrease operator dependency and enable treatment of tumors not visible under ultrasound.
Collapse
|
10
|
Machine Learning in the Management of Lateral Skull Base Tumors: A Systematic Review. JOURNAL OF OTORHINOLARYNGOLOGY, HEARING AND BALANCE MEDICINE 2022. [DOI: 10.3390/ohbm3040007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
The application of machine learning (ML) techniques to otolaryngology remains a topic of interest and prevalence in the literature, though no previous articles have summarized the current state of ML application to management and the diagnosis of lateral skull base (LSB) tumors. Subsequently, we present a systematic overview of previous applications of ML techniques to the management of LSB tumors. Independent searches were conducted on PubMed and Web of Science between August 2020 and February 2021 to identify the literature pertaining to the use of ML techniques in LSB tumor surgery written in the English language. All articles were assessed in regard to their application task, ML methodology, and their outcomes. A total of 32 articles were examined. The number of articles involving applications of ML techniques to LSB tumor surgeries has significantly increased since the first article relevant to this field was published in 1994. The most commonly employed ML category was tree-based algorithms. Most articles were included in the category of surgical management (13; 40.6%), followed by those in disease classification (8; 25%). Overall, the application of ML techniques to the management of LSB tumor has evolved rapidly over the past two decades, and the anticipated growth in the future could significantly augment the surgical outcomes and management of LSB tumors.
Collapse
|
11
|
Surgical Tool Datasets for Machine Learning Research: A Survey. Int J Comput Vis 2022. [DOI: 10.1007/s11263-022-01640-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/05/2022]
Abstract
AbstractThis paper is a comprehensive survey of datasets for surgical tool detection and related surgical data science and machine learning techniques and algorithms. The survey offers a high level perspective of current research in this area, analyses the taxonomy of approaches adopted by researchers using surgical tool datasets, and addresses key areas of research, such as the datasets used, evaluation metrics applied and deep learning techniques utilised. Our presentation and taxonomy provides a framework that facilitates greater understanding of current work, and highlights the challenges and opportunities for further innovative and useful research.
Collapse
|
12
|
Kausch L, Thomas S, Kunze H, Norajitra T, Klein A, Ayala L, El Barbari J, Mandelka E, Privalov M, Vetter S, Mahnken A, Maier-Hein L, Maier-Hein K. C-arm positioning for standard projections during spinal implant placement. Med Image Anal 2022; 81:102557. [DOI: 10.1016/j.media.2022.102557] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/16/2022] [Revised: 06/09/2022] [Accepted: 07/22/2022] [Indexed: 10/16/2022]
|
13
|
Leveraging spatial uncertainty for online error compensation in EMT. Int J Comput Assist Radiol Surg 2020; 15:1043-1051. [PMID: 32440957 PMCID: PMC7303086 DOI: 10.1007/s11548-020-02189-w] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/16/2019] [Accepted: 04/23/2020] [Indexed: 01/26/2023]
Abstract
Purpose Electromagnetic tracking (EMT) can potentially complement fluoroscopic navigation, reducing radiation exposure in a hybrid setting. Due to the susceptibility to external distortions, systematic error in EMT needs to be compensated algorithmically. Compensation algorithms for EMT in guidewire procedures are only practical in an online setting. Methods We collect positional data and train a symmetric artificial neural network (ANN) architecture for compensating navigation error. The results are evaluated in both online and offline scenarios and are compared to polynomial fits. We assess spatial uncertainty of the compensation proposed by the ANN. Simulations based on real data show how this uncertainty measure can be utilized to improve accuracy and limit radiation exposure in hybrid navigation. Results ANNs compensate unseen distortions by more than 70%, outperforming polynomial regression. Working on known distortions, ANNs outperform polynomials as well. We empirically demonstrate a linear relationship between tracking accuracy and model uncertainty. The effectiveness of hybrid tracking is shown in a simulation experiment. Conclusion ANNs are suitable for EMT error compensation and can generalize across unseen distortions. Model uncertainty needs to be assessed when spatial error compensation algorithms are developed, so that training data collection can be optimized. Finally, we find that error compensation in EMT reduces the need for X-ray images in hybrid navigation.
Collapse
|
14
|
Vercauteren T, Unberath M, Padoy N, Navab N. CAI4CAI: The Rise of Contextual Artificial Intelligence in Computer Assisted Interventions. PROCEEDINGS OF THE IEEE. INSTITUTE OF ELECTRICAL AND ELECTRONICS ENGINEERS 2020; 108:198-214. [PMID: 31920208 PMCID: PMC6952279 DOI: 10.1109/jproc.2019.2946993] [Citation(s) in RCA: 55] [Impact Index Per Article: 11.0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/01/2019] [Revised: 09/12/2019] [Accepted: 10/04/2019] [Indexed: 05/10/2023]
Abstract
Data-driven computational approaches have evolved to enable extraction of information from medical images with a reliability, accuracy and speed which is already transforming their interpretation and exploitation in clinical practice. While similar benefits are longed for in the field of interventional imaging, this ambition is challenged by a much higher heterogeneity. Clinical workflows within interventional suites and operating theatres are extremely complex and typically rely on poorly integrated intra-operative devices, sensors, and support infrastructures. Taking stock of some of the most exciting developments in machine learning and artificial intelligence for computer assisted interventions, we highlight the crucial need to take context and human factors into account in order to address these challenges. Contextual artificial intelligence for computer assisted intervention, or CAI4CAI, arises as an emerging opportunity feeding into the broader field of surgical data science. Central challenges being addressed in CAI4CAI include how to integrate the ensemble of prior knowledge and instantaneous sensory information from experts, sensors and actuators; how to create and communicate a faithful and actionable shared representation of the surgery among a mixed human-AI actor team; how to design interventional systems and associated cognitive shared control schemes for online uncertainty-aware collaborative decision making ultimately producing more precise and reliable interventions.
Collapse
Affiliation(s)
- Tom Vercauteren
- School of Biomedical Engineering & Imaging SciencesKing’s College LondonLondonWC2R 2LSU.K.
| | - Mathias Unberath
- Department of Computer ScienceJohns Hopkins UniversityBaltimoreMD21218USA
| | - Nicolas Padoy
- ICube institute, CNRS, IHU Strasbourg, University of Strasbourg67081StrasbourgFrance
| | - Nassir Navab
- Fakultät für InformatikTechnische Universität München80333MunichGermany
| |
Collapse
|
15
|
Unberath M, Zaech JN, Gao C, Bier B, Goldmann F, Lee SC, Fotouhi J, Taylor R, Armand M, Navab N. Enabling machine learning in X-ray-based procedures via realistic simulation of image formation. Int J Comput Assist Radiol Surg 2019; 14:1517-1528. [PMID: 31187399 PMCID: PMC7297499 DOI: 10.1007/s11548-019-02011-2] [Citation(s) in RCA: 34] [Impact Index Per Article: 5.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/28/2019] [Accepted: 06/03/2019] [Indexed: 12/19/2022]
Abstract
PURPOSE Machine learning-based approaches now outperform competing methods in most disciplines relevant to diagnostic radiology. Image-guided procedures, however, have not yet benefited substantially from the advent of deep learning, in particular because images for procedural guidance are not archived and thus unavailable for learning, and even if they were available, annotations would be a severe challenge due to the vast amounts of data. In silico simulation of X-ray images from 3D CT is an interesting alternative to using true clinical radiographs since labeling is comparably easy and potentially readily available. METHODS We extend our framework for fast and realistic simulation of fluoroscopy from high-resolution CT, called DeepDRR, with tool modeling capabilities. The framework is publicly available, open source, and tightly integrated with the software platforms native to deep learning, i.e., Python, PyTorch, and PyCuda. DeepDRR relies on machine learning for material decomposition and scatter estimation in 3D and 2D, respectively, but uses analytic forward projection and noise injection to ensure acceptable computation times. On two X-ray image analysis tasks, namely (1) anatomical landmark detection and (2) segmentation and localization of robot end-effectors, we demonstrate that convolutional neural networks (ConvNets) trained on DeepDRRs generalize well to real data without re-training or domain adaptation. To this end, we use the exact same training protocol to train ConvNets on naïve and DeepDRRs and compare their performance on data of cadaveric specimens acquired using a clinical C-arm X-ray system. RESULTS Our findings are consistent across both considered tasks. All ConvNets performed similarly well when evaluated on the respective synthetic testing set. However, when applied to real radiographs of cadaveric anatomy, ConvNets trained on DeepDRRs significantly outperformed ConvNets trained on naïve DRRs ([Formula: see text]). CONCLUSION Our findings for both tasks are positive and promising. Combined with complementary approaches, such as image style transfer, the proposed framework for fast and realistic simulation of fluoroscopy from CT contributes to promoting the implementation of machine learning in X-ray-guided procedures. This paradigm shift has the potential to revolutionize intra-operative image analysis to simplify surgical workflows.
Collapse
Affiliation(s)
- Mathias Unberath
- Department of Computer Science, Johns Hopkins University, Baltimore, MD, USA.
- Laboratory for Computational Sensing + Robotics, Johns Hopkins University, Baltimore, MD, USA.
- Computer Aided Medical Procedures, Johns Hopkins University, Baltimore, MD, USA.
| | - Jan-Nico Zaech
- Laboratory for Computational Sensing + Robotics, Johns Hopkins University, Baltimore, MD, USA
- Computer Aided Medical Procedures, Johns Hopkins University, Baltimore, MD, USA
| | - Cong Gao
- Department of Computer Science, Johns Hopkins University, Baltimore, MD, USA
- Laboratory for Computational Sensing + Robotics, Johns Hopkins University, Baltimore, MD, USA
| | - Bastian Bier
- Laboratory for Computational Sensing + Robotics, Johns Hopkins University, Baltimore, MD, USA
- Computer Aided Medical Procedures, Johns Hopkins University, Baltimore, MD, USA
| | - Florian Goldmann
- Laboratory for Computational Sensing + Robotics, Johns Hopkins University, Baltimore, MD, USA
- Computer Aided Medical Procedures, Johns Hopkins University, Baltimore, MD, USA
| | - Sing Chun Lee
- Department of Computer Science, Johns Hopkins University, Baltimore, MD, USA
- Laboratory for Computational Sensing + Robotics, Johns Hopkins University, Baltimore, MD, USA
- Computer Aided Medical Procedures, Johns Hopkins University, Baltimore, MD, USA
| | - Javad Fotouhi
- Department of Computer Science, Johns Hopkins University, Baltimore, MD, USA
- Laboratory for Computational Sensing + Robotics, Johns Hopkins University, Baltimore, MD, USA
- Computer Aided Medical Procedures, Johns Hopkins University, Baltimore, MD, USA
| | - Russell Taylor
- Department of Computer Science, Johns Hopkins University, Baltimore, MD, USA
- Laboratory for Computational Sensing + Robotics, Johns Hopkins University, Baltimore, MD, USA
| | - Mehran Armand
- Laboratory for Computational Sensing + Robotics, Johns Hopkins University, Baltimore, MD, USA
- Johns Hopkins University Applied Physics Laboratory, Laurel, MD, USA
| | - Nassir Navab
- Department of Computer Science, Johns Hopkins University, Baltimore, MD, USA
- Laboratory for Computational Sensing + Robotics, Johns Hopkins University, Baltimore, MD, USA
- Computer Aided Medical Procedures, Johns Hopkins University, Baltimore, MD, USA
| |
Collapse
|