1
|
Guo J, Wang J, Wei R, Kang D, Dou Q, Liu YH. UC-NeRF: Uncertainty-Aware Conditional Neural Radiance Fields From Endoscopic Sparse Views. IEEE TRANSACTIONS ON MEDICAL IMAGING 2025; 44:1284-1296. [PMID: 39531569 DOI: 10.1109/tmi.2024.3496558] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/16/2024]
Abstract
Visualizing surgical scenes is crucial for revealing internal anatomical structures during minimally invasive procedures. Novel View Synthesis is a vital technique that offers geometry and appearance reconstruction, enhancing understanding, planning, and decision-making in surgical scenes. Despite the impressive achievements of Neural Radiance Field (NeRF), its direct application to surgical scenes produces unsatisfying results due to two challenges: endoscopic sparse views and significant photometric inconsistencies. In this paper, we propose uncertainty-aware conditional NeRF for novel view synthesis to tackle the severe shape-radiance ambiguity from sparse surgical views. The core of UC-NeRF is to incorporate the multi-view uncertainty estimation to condition the neural radiance field for modeling the severe photometric inconsistencies adaptively. Specifically, our UC-NeRF first builds a consistency learner in the form of multi-view stereo network, to establish the geometric correspondence from sparse views and generate uncertainty estimation and feature priors. In neural rendering, we design a base-adaptive NeRF network to exploit the uncertainty estimation for explicitly handling the photometric inconsistencies. Furthermore, an uncertainty-guided geometry distillation is employed to enhance geometry learning. Experiments on the SCARED and Hamlyn datasets demonstrate our superior performance in rendering appearance and geometry, consistently outperforming the current state-of-the-art approaches. Our code will be released at https://github.com/wrld/UC-NeRF.
Collapse
|
2
|
Ye Z, Shao C, Zhu K. Unsupervised neural network-based image stitching method for bladder endoscopy. PLoS One 2025; 20:e0311637. [PMID: 39964991 PMCID: PMC11835325 DOI: 10.1371/journal.pone.0311637] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/11/2024] [Accepted: 01/22/2025] [Indexed: 02/20/2025] Open
Abstract
Bladder endoscopy enables the observation of intravesical lesion characteristics, making it an essential tool in urology. Image stitching techniques are commonly employed to expand the field of view of bladder endoscopy. Traditional image stitching methods rely on feature matching. In recent years, deep-learning techniques have garnered significant attention in the field of computer vision. However, the commonly employed supervised learning approaches often require a substantial amount of labeled data, which can be challenging to acquire, especially in the context of medical data. To address this limitation, this study proposes an unsupervised neural network-based image stitching method for bladder endoscopy, which eliminates the need for labeled datasets. The method comprises two modules: an unsupervised alignment network and an unsupervised fusion network. In the unsupervised alignment network, we employed feature convolution, regression networks, and linear transformations to align images. In the unsupervised fusion network, we achieved image fusion from features to pixel by simultaneously eliminating artifacts and enhancing the resolution. Experiments demonstrated our method's consistent stitching success rate of 98.11% and robust image stitching accuracy at various resolutions. Our method eliminates sutures and flocculent debris from cystoscopy images, presenting good image smoothness while preserving rich textural features. Moreover, our method could successfully stitch challenging scenes such as dim and blurry scenes. Our application of unsupervised deep learning methods in the field of cystoscopy image stitching was successfully validated, laying the foundation for real-time panoramic stitching of bladder endoscopic video images. This advancement provides opportunities for the future development of computer-vision-assisted diagnostic systems for bladder cavities.
Collapse
Affiliation(s)
- Zixing Ye
- Department of Urology, Peking Union Medical College Hospital, Beijing, China
| | - Chenyu Shao
- National Elite Institute of Engineering, Northwestern Polytechnical University, Xi’an, China
| | - Kelei Zhu
- School of Software, Northwestern Polytechnical University, Xi’an, China
| |
Collapse
|
3
|
Whitley P, Creasey C, Clarkson MJ, Thompson S. A Serious Game to Study Reduced Field of View in Keyhole Surgery: Development and Experimental Study. JMIR Serious Games 2025; 13:e56269. [PMID: 39933172 PMCID: PMC11862761 DOI: 10.2196/56269] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/11/2024] [Revised: 09/30/2024] [Accepted: 01/17/2025] [Indexed: 02/13/2025] Open
Abstract
BACKGROUND During keyhole surgery, the surgeon is required to perform highly demanding tasks while only being able to see part of the patient's anatomy. This limited field of view is widely cited as a key limitation of the procedure, and many computational methods have been proposed to overcome it. However, the precise effects of a limited field of view on task performance remain unknown due to the lack of tools to study these effects effectively. OBJECTIVE This paper describes our work on developing a serious game with 2 objectives: (1) to create an engaging game that communicates some of the challenges of keyhole surgery, and (2) to test the effect of a limited field of view on task performance. The development of a serious game that can be played by a wide range of participants will enable us to gather quantitative data on the effects of the reduced field of view on task performance. These data can inform the future development of technologies to help surgeons reduce the impact of a limited field of view on clinical outcomes for patients. The game is open source and may be adapted and used by other researchers to study related problems. METHODS We implemented an open-source serious game in JavaScript, inspired by the surgical task of selectively cauterizing blood vessels during twin-to-twin transfusion surgery. During the game, the player is required to identify and cut the correct blood vessel under different fields of view and varying levels of vascular complexity. We conducted a quantitative analysis of task performance time under different conditions and a formative analysis of the game using participant questionnaires. RESULTS We recruited 25 players to test the game and recorded their task performance time, accuracy, and qualitative metrics. Reducing the field of view resulted in participants taking significantly longer (P<.001) to perform otherwise identical tasks (mean 6.4 seconds, 95% CI 5.0-7.8 seconds vs mean 13.6 seconds, 95% CI 10.3-16.9 seconds). Participants found the game engaging and agreed that it enhanced their understanding of the limited field of view during keyhole surgery. CONCLUSIONS We recruited 25 players to test the game and recorded their task performance time, accuracy, and qualitative metrics. Reducing the field of view resulted in participants taking statistically significantly longer (16.4 vs 9.8 seconds; P=.05) to perform otherwise identical tasks. Participants found the game engaging and agreed that it enhanced their understanding of the limited field of view during keyhole surgery.
Collapse
Affiliation(s)
- Phoebe Whitley
- Department of Medical Physics and Biomedical Engineering, Faculty of Engineering Sciences, University College London, London, United Kingdom
| | - Connor Creasey
- Department of Medical Physics and Biomedical Engineering, Faculty of Engineering Sciences, University College London, London, United Kingdom
| | - Matthew J Clarkson
- UCL Hawkes Institute, Faculty of Engineering Sciences, University College London, London, United Kingdom
| | - Stephen Thompson
- Advanced Research Computing, University College London, London, United Kingdom
| |
Collapse
|
4
|
Li M, Varble NA, Gurram S, Long D, Valera V, Gopal N, Bakhutashvili I, Reed S, Pritchard WF, Karanian JW, Xu S, Wood BJ. Bladder image stitching algorithm for navigation and referencing using a standard cystoscope. Sci Rep 2024; 14:29168. [PMID: 39587232 PMCID: PMC11589604 DOI: 10.1038/s41598-024-80284-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/07/2024] [Accepted: 11/18/2024] [Indexed: 11/27/2024] Open
Abstract
To aid in the diagnosis, monitoring, and surveillance of bladder carcinoma, this study aimed to develop and test an algorithm that creates a referenceable bladder map rendered from standard cystoscopy videos without the need for specialized equipment. A vision-based algorithm was developed to generate 2D bladder maps from individual video frames, by sequentially stitching image frames based on matching surface features, and subsequently localize and track frames during reevaluation. The algorithm was developed and calibrated in a 2D model and 3D anthropomorphic bladder phantom. The performance was evaluated in vivo in swine and with retrospective clinical cystoscopy video. Results showed that the algorithm was capable of capturing and stitching intravesical images with different sweeping patterns. Between 93% and 99% of frames had sufficient features for bladder map generation. Upon reevaluation, the cystoscope accurately localized a frame within 4.5 s. In swine, a virtual mucosal surface map was generated that matched the explant anatomy. A surface map could be generated based on archived patient cystoscopy images. This tool could aid recording and referencing pathologic findings and biopsy or treatment locations for subsequent procedures and may have utility in patients with metachronous bladder cancer and in low-resource settings.
Collapse
Affiliation(s)
- Ming Li
- Center for Interventional Oncology, National Institutes of Health, Bethesda, MD, USA.
| | - Nicole A Varble
- Center for Interventional Oncology, National Institutes of Health, Bethesda, MD, USA
- Philips Healthcare, Cambridge, MA, USA
| | - Sandeep Gurram
- Urologic Oncology Branch, National Cancer Institute, National Institute of Health, Bethesda, MD, USA
| | - Dilara Long
- Center for Interventional Oncology, National Institutes of Health, Bethesda, MD, USA
| | - Vladimir Valera
- Urologic Oncology Branch, National Cancer Institute, National Institute of Health, Bethesda, MD, USA
| | - Nikhil Gopal
- Urologic Oncology Branch, National Cancer Institute, National Institute of Health, Bethesda, MD, USA
| | - Ivane Bakhutashvili
- Center for Interventional Oncology, National Institutes of Health, Bethesda, MD, USA
| | - Sheridan Reed
- Center for Interventional Oncology, National Institutes of Health, Bethesda, MD, USA
| | - William F Pritchard
- Center for Interventional Oncology, National Institutes of Health, Bethesda, MD, USA
| | - John W Karanian
- Center for Interventional Oncology, National Institutes of Health, Bethesda, MD, USA
| | - Sheng Xu
- Center for Interventional Oncology, National Institutes of Health, Bethesda, MD, USA
| | - Bradford J Wood
- Center for Interventional Oncology, National Institutes of Health, Bethesda, MD, USA.
- Urologic Oncology Branch, National Cancer Institute, National Institute of Health, Bethesda, MD, USA.
- Radiology & Imaging Sciences, Clinical Center, National Institutes of Health, Bethesda, MD, USA.
- National Institute of Biomedical Imaging and Bioengineering, Bethesda, MD, USA.
| |
Collapse
|
5
|
Göbel B, Reiterer A, Möller K. Image-Based 3D Reconstruction in Laparoscopy: A Review Focusing on the Quantitative Evaluation by Applying the Reconstruction Error. J Imaging 2024; 10:180. [PMID: 39194969 DOI: 10.3390/jimaging10080180] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/28/2024] [Revised: 07/16/2024] [Accepted: 07/18/2024] [Indexed: 08/29/2024] Open
Abstract
Image-based 3D reconstruction enables laparoscopic applications as image-guided navigation and (autonomous) robot-assisted interventions, which require a high accuracy. The review's purpose is to present the accuracy of different techniques to label the most promising. A systematic literature search with PubMed and google scholar from 2015 to 2023 was applied by following the framework of "Review articles: purpose, process, and structure". Articles were considered when presenting a quantitative evaluation (root mean squared error and mean absolute error) of the reconstruction error (Euclidean distance between real and reconstructed surface). The search provides 995 articles, which were reduced to 48 articles after applying exclusion criteria. From these, a reconstruction error data set could be generated for the techniques of stereo vision, Shape-from-Motion, Simultaneous Localization and Mapping, deep-learning, and structured light. The reconstruction error varies from below one millimeter to higher than ten millimeters-with deep-learning and Simultaneous Localization and Mapping delivering the best results under intraoperative conditions. The high variance emerges from different experimental conditions. In conclusion, submillimeter accuracy is challenging, but promising image-based 3D reconstruction techniques could be identified. For future research, we recommend computing the reconstruction error for comparison purposes and use ex/in vivo organs as reference objects for realistic experiments.
Collapse
Affiliation(s)
- Birthe Göbel
- Department of Sustainable Systems Engineering-INATECH, University of Freiburg, Emmy-Noether-Street 2, 79110 Freiburg im Breisgau, Germany
- KARL STORZ SE & Co. KG, Dr.-Karl-Storz-Street 34, 78532 Tuttlingen, Germany
| | - Alexander Reiterer
- Department of Sustainable Systems Engineering-INATECH, University of Freiburg, Emmy-Noether-Street 2, 79110 Freiburg im Breisgau, Germany
- Fraunhofer Institute for Physical Measurement Techniques IPM, 79110 Freiburg im Breisgau, Germany
| | - Knut Möller
- Institute of Technical Medicine-ITeM, Furtwangen University (HFU), 78054 Villingen-Schwenningen, Germany
- Mechanical Engineering, University of Canterbury, Christchurch 8140, New Zealand
| |
Collapse
|
6
|
Schmidt A, Mohareri O, DiMaio SP, Salcudean SE. Surgical Tattoos in Infrared: A Dataset for Quantifying Tissue Tracking and Mapping. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:2634-2645. [PMID: 38437151 DOI: 10.1109/tmi.2024.3372828] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 03/06/2024]
Abstract
Quantifying performance of methods for tracking and mapping tissue in endoscopic environments is essential for enabling image guidance and automation of medical interventions and surgery. Datasets developed so far either use rigid environments, visible markers, or require annotators to label salient points in videos after collection. These are respectively: not general, visible to algorithms, or costly and error-prone. We introduce a novel labeling methodology along with a dataset that uses said methodology, Surgical Tattoos in Infrared (STIR). STIR has labels that are persistent but invisible to visible spectrum algorithms. This is done by labelling tissue points with IR-fluorescent dye, indocyanine green (ICG), and then collecting visible light video clips. STIR comprises hundreds of stereo video clips in both in vivo and ex vivo scenes with start and end points labelled in the IR spectrum. With over 3,000 labelled points, STIR will help to quantify and enable better analysis of tracking and mapping methods. After introducing STIR, we analyze multiple different frame-based tracking methods on STIR using both 3D and 2D endpoint error and accuracy metrics. STIR is available at https://dx.doi.org/10.21227/w8g4-g548.
Collapse
|
7
|
Yang Z, Dai J, Pan J. 3D reconstruction from endoscopy images: A survey. Comput Biol Med 2024; 175:108546. [PMID: 38704902 DOI: 10.1016/j.compbiomed.2024.108546] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/15/2023] [Revised: 01/05/2024] [Accepted: 04/28/2024] [Indexed: 05/07/2024]
Abstract
Three-dimensional reconstruction of images acquired through endoscopes is playing a vital role in an increasing number of medical applications. Endoscopes used in the clinic are commonly classified as monocular endoscopes and binocular endoscopes. We have reviewed the classification of methods for depth estimation according to the type of endoscope. Basically, depth estimation relies on feature matching of images and multi-view geometry theory. However, these traditional techniques have many problems in the endoscopic environment. With the increasing development of deep learning techniques, there is a growing number of works based on learning methods to address challenges such as inconsistent illumination and texture sparsity. We have reviewed over 170 papers published in the 10 years from 2013 to 2023. The commonly used public datasets and performance metrics are summarized. We also give a taxonomy of methods and analyze the advantages and drawbacks of algorithms. Summary tables and result atlas are listed to facilitate the comparison of qualitative and quantitative performance of different methods in each category. In addition, we summarize commonly used scene representation methods in endoscopy and speculate on the prospects of deep estimation research in medical applications. We also compare the robustness performance, processing time, and scene representation of the methods to facilitate doctors and researchers in selecting appropriate methods based on surgical applications.
Collapse
Affiliation(s)
- Zhuoyue Yang
- State Key Laboratory of Virtual Reality Technology and Systems, Beihang University, 37 Xueyuan Road, Haidian District, Beijing, 100191, China; Peng Cheng Lab, 2 Xingke 1st Street, Nanshan District, Shenzhen, Guangdong Province, 518000, China
| | - Ju Dai
- Peng Cheng Lab, 2 Xingke 1st Street, Nanshan District, Shenzhen, Guangdong Province, 518000, China
| | - Junjun Pan
- State Key Laboratory of Virtual Reality Technology and Systems, Beihang University, 37 Xueyuan Road, Haidian District, Beijing, 100191, China; Peng Cheng Lab, 2 Xingke 1st Street, Nanshan District, Shenzhen, Guangdong Province, 518000, China.
| |
Collapse
|
8
|
Xu Y, Zhang P, Wang L, Li Y, Luo B, Yu Y, Chen R. Performance evaluation and future prospects of capsule robot localization technology. GEO-SPATIAL INFORMATION SCIENCE 2024:1-31. [DOI: 10.1080/10095020.2024.2354239] [Citation(s) in RCA: 1] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Received: 08/13/2023] [Accepted: 05/07/2024] [Indexed: 01/04/2025]
Affiliation(s)
- Yan Xu
- State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan, China
| | - Peng Zhang
- State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan, China
- Institute of Medical Informatics, Renmin Hospital of Wuhan University, Wuhan, China
| | - Lei Wang
- State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan, China
| | - You Li
- State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan, China
- Institute of Medical Informatics, Renmin Hospital of Wuhan University, Wuhan, China
| | - Bin Luo
- State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan, China
- Institute of Medical Informatics, Renmin Hospital of Wuhan University, Wuhan, China
| | - Yue Yu
- State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan, China
| | - Ruizhi Chen
- State Key Laboratory of Information Engineering in Surveying, Mapping and Remote Sensing, Wuhan University, Wuhan, China
- Institute of Medical Informatics, Renmin Hospital of Wuhan University, Wuhan, China
| |
Collapse
|
9
|
Schmidt A, Mohareri O, DiMaio S, Yip MC, Salcudean SE. Tracking and mapping in medical computer vision: A review. Med Image Anal 2024; 94:103131. [PMID: 38442528 DOI: 10.1016/j.media.2024.103131] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/16/2023] [Revised: 02/08/2024] [Accepted: 02/29/2024] [Indexed: 03/07/2024]
Abstract
As computer vision algorithms increase in capability, their applications in clinical systems will become more pervasive. These applications include: diagnostics, such as colonoscopy and bronchoscopy; guiding biopsies, minimally invasive interventions, and surgery; automating instrument motion; and providing image guidance using pre-operative scans. Many of these applications depend on the specific visual nature of medical scenes and require designing algorithms to perform in this environment. In this review, we provide an update to the field of camera-based tracking and scene mapping in surgery and diagnostics in medical computer vision. We begin with describing our review process, which results in a final list of 515 papers that we cover. We then give a high-level summary of the state of the art and provide relevant background for those who need tracking and mapping for their clinical applications. After which, we review datasets provided in the field and the clinical needs that motivate their design. Then, we delve into the algorithmic side, and summarize recent developments. This summary should be especially useful for algorithm designers and to those looking to understand the capability of off-the-shelf methods. We maintain focus on algorithms for deformable environments while also reviewing the essential building blocks in rigid tracking and mapping since there is a large amount of crossover in methods. With the field summarized, we discuss the current state of the tracking and mapping methods along with needs for future algorithms, needs for quantification, and the viability of clinical applications. We then provide some research directions and questions. We conclude that new methods need to be designed or combined to support clinical applications in deformable environments, and more focus needs to be put into collecting datasets for training and evaluation.
Collapse
Affiliation(s)
- Adam Schmidt
- Department of Electrical and Computer Engineering, University of British Columbia, 2329 West Mall, Vancouver V6T 1Z4, BC, Canada.
| | - Omid Mohareri
- Advanced Research, Intuitive Surgical, 1020 Kifer Rd, Sunnyvale, CA 94086, USA
| | - Simon DiMaio
- Advanced Research, Intuitive Surgical, 1020 Kifer Rd, Sunnyvale, CA 94086, USA
| | - Michael C Yip
- Department of Electrical and Computer Engineering, University of California San Diego, 9500 Gilman Dr, La Jolla, CA 92093, USA
| | - Septimiu E Salcudean
- Department of Electrical and Computer Engineering, University of British Columbia, 2329 West Mall, Vancouver V6T 1Z4, BC, Canada
| |
Collapse
|
10
|
Yuan W, Poosa SRP, Dirks RF. Comparative Analysis of Color Space and Channel, Detector, and Descriptor for Feature-Based Image Registration. J Imaging 2024; 10:105. [PMID: 38786559 PMCID: PMC11122496 DOI: 10.3390/jimaging10050105] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2024] [Revised: 04/26/2024] [Accepted: 04/26/2024] [Indexed: 05/25/2024] Open
Abstract
The current study aimed to quantify the value of color spaces and channels as a potential superior replacement for standard grayscale images, as well as the relative performance of open-source detectors and descriptors for general feature-based image registration purposes, based on a large benchmark dataset. The public dataset UDIS-D, with 1106 diverse image pairs, was selected. In total, 21 color spaces or channels including RGB, XYZ, Y'CrCb, HLS, L*a*b* and their corresponding channels in addition to grayscale, nine feature detectors including AKAZE, BRISK, CSE, FAST, HL, KAZE, ORB, SIFT, and TBMR, and 11 feature descriptors including AKAZE, BB, BRIEF, BRISK, DAISY, FREAK, KAZE, LATCH, ORB, SIFT, and VGG were evaluated according to reprojection error (RE), root mean square error (RMSE), structural similarity index measure (SSIM), registration failure rate, and feature number, based on 1,950,984 image registrations. No meaningful benefits from color space or channel were observed, although XYZ, RGB color space and L* color channel were able to outperform grayscale by a very minor margin. Per the dataset, the best-performing color space or channel, detector, and descriptor were XYZ/RGB, SIFT/FAST, and AKAZE. The most robust color space or channel, detector, and descriptor were L*a*b*, TBMR, and VGG. The color channel, detector, and descriptor with the most initial detector features and final homography features were Z/L*, FAST, and KAZE. In terms of the best overall unfailing combinations, XYZ/RGB+SIFT/FAST+VGG/SIFT seemed to provide the highest image registration quality, while Z+FAST+VGG provided the most image features.
Collapse
Affiliation(s)
- Wenan Yuan
- Independent Researcher, Oak Brook, IL 60523, USA; (S.R.P.P.); (R.F.D.)
| | | | | |
Collapse
|
11
|
Barbour MC, Amin SN, Friedman SD, Perez FA, Bly RA, Johnson KE, Parikh SR, Richardson CM, Dahl JP, Aliseda A. Surface Reconstruction of the Pediatric Larynx via Structure from Motion Photogrammetry: A Pilot Study. Otolaryngol Head Neck Surg 2024; 170:1195-1199. [PMID: 38168480 PMCID: PMC10960702 DOI: 10.1002/ohn.635] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/23/2023] [Revised: 11/10/2023] [Accepted: 12/07/2023] [Indexed: 01/05/2024]
Abstract
Endoscopy is the gold standard for characterizing pediatric airway disorders, however, it is limited for quantitative analysis due to lack of three-dimensional (3D) vision and poor stereotactic depth perception. We utilize structure from motion (SfM) photogrammetry, to reconstruct 3D surfaces of pathologic and healthy pediatric larynges from monocular two-dimensional (2D) endoscopy. Models of pediatric subglottic stenosis were 3D printed and airway endoscopies were simulated. 3D surfaces were successfully reconstructed from endoscopic videos of all models using an SfM analysis toolkit. Average subglottic surface error between SfM reconstructed surfaces and 3D printed models was 0.65 mm as measured by Modified Hausdorff Distance. Average volumetric similarity between SfM surfaces and printed models was 0.82 as measured by Jaccard Index. SfM can be used to accurately reconstruct 3D surface renderings of the larynx from 2D endoscopy video. This technique has immense potential for use in quantitative analysis of airway geometry and virtual surgical planning.
Collapse
Affiliation(s)
- Michael C Barbour
- Department of Mechanical Engineering, University of Washington, Seattle, Washington, USA
| | - Shaunak N Amin
- Department of Otolaryngology-Head and Neck Surgery, University of Washington, Seattle, Washington, USA
| | - Seth D Friedman
- Center for Respiratory Biology and Therapeutics, Seattle Children's Hospital, Seattle, Washington, USA
| | - Francisco A Perez
- Department of Pediatric Radiology, Seattle Children's Hospital, Seattle, Washington, USA
| | - Randall A Bly
- Department of Otolaryngology-Head and Neck Surgery, University of Washington, Seattle, Washington, USA
- Division of Pediatric Otolaryngology-Head and Neck Surgery, Seattle Children's Hospital, Seattle, Washington, USA
| | - Kaalan E Johnson
- Department of Otolaryngology-Head and Neck Surgery, University of Washington, Seattle, Washington, USA
- Division of Pediatric Otolaryngology-Head and Neck Surgery, Seattle Children's Hospital, Seattle, Washington, USA
| | - Sanjay R Parikh
- Department of Otolaryngology-Head and Neck Surgery, University of Washington, Seattle, Washington, USA
- Division of Pediatric Otolaryngology-Head and Neck Surgery, Seattle Children's Hospital, Seattle, Washington, USA
| | - Clare M Richardson
- Division of Pediatric Otolaryngology-Head and Neck Surgery, Seattle Children's Hospital, Seattle, Washington, USA
- Division of Pediatric Otolaryngology-Head and Neck Surgery, Phoenix Children's Hospital, Phoenix, Arizona, USA
| | - John P Dahl
- Department of Otolaryngology-Head and Neck Surgery, University of Washington, Seattle, Washington, USA
- Division of Pediatric Otolaryngology-Head and Neck Surgery, Seattle Children's Hospital, Seattle, Washington, USA
| | - Alberto Aliseda
- Department of Mechanical Engineering, University of Washington, Seattle, Washington, USA
| |
Collapse
|
12
|
Ma T, Meng B, Yang J, Gou N, Shi W. A half jaw panoramic stitching method of intraoral endoscopy images based on dental arch arrangement. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2024; 21:494-522. [PMID: 38303432 DOI: 10.3934/mbe.2024022] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/03/2024]
Abstract
To address the challenges of repetitive and low-texture features in intraoral endoscopic images, a novel methodology for stitching panoramic half jaw images of the oral cavity is proposed. Initially, an enhanced self-attention mechanism guided by Time-Weighting concepts is employed to augment the clustering potential of feature points, thereby increasing the number of matched features. Subsequently, a combination of the Sinkhorn algorithm and Random Sample Consensus (RANSAC) is utilized to maximize the count of matched feature pairs, accurately remove outliers and minimize error. Last, to address the unique spatial alignment among intraoral endoscopic images, a wavelet transform and weighted fusion algorithm based on dental arch arrangement in intraoral endoscopic images have been developed, specifically for use in the fusion stage of intraoral endoscopic images. This enables the local oral images to be precisely positioned along the dental arch, and seamless stitching is achieved through wavelet transformation and a gradual weighted fusion technique. Experimental results demonstrate that this method yields promising outcomes in panoramic stitching tasks for intraoral endoscopic images, achieving a matching accuracy of 84.6% and a recall rate of 78.4% in a dataset with an average overlap of 35%. A novel solution for panoramic stitching of intraoral endoscopic images is provided by this method.
Collapse
Affiliation(s)
- Tian Ma
- College of Computer Science and Technology, Xi'an University of Science and Technology, Xi'an Shaanxi 710054, China
| | - Boyang Meng
- College of Computer Science and Technology, Xi'an University of Science and Technology, Xi'an Shaanxi 710054, China
| | - Jiayi Yang
- College of Computer Science and Technology, Xi'an University of Science and Technology, Xi'an Shaanxi 710054, China
| | - Nana Gou
- College of Computer Science and Technology, Xi'an University of Science and Technology, Xi'an Shaanxi 710054, China
| | - Weilu Shi
- College of Computer Science and Technology, Xi'an University of Science and Technology, Xi'an Shaanxi 710054, China
| |
Collapse
|
13
|
Groenhuis V, de Groot AG, Cornel EB, Stramigioli S, Siepel FJ. 3-D and 2-D reconstruction of bladders for the assessment of inter-session detection of tissue changes: a proof of concept. Int J Comput Assist Radiol Surg 2023; 18:1915-1924. [PMID: 37085675 PMCID: PMC10497453 DOI: 10.1007/s11548-023-02900-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/22/2022] [Accepted: 03/31/2023] [Indexed: 04/23/2023]
Abstract
PURPOSE Abnormalities in the bladder wall require careful investigation regarding type, spatial position and invasiveness. Construction of a 3-D model of the bladder is helpful to ensure adequate coverage of the scanning procedure, quantitative comparison of bladder wall textures between successive sessions and finding back previously discovered abnormalities. METHODS Videos of both an in vivo bladder and a textured bladder phantom were acquired. Structure-from-motion and bundle adjustment algorithms were used to construct a 3-D point cloud, approximate it by a surface mesh, texture it with the back-projected camera frames and draw the corresponding 2-D atlas. Reconstructions of successive sessions were compared; those of the bladder phantom were co-registered, transformed using 3-D thin plate splines and post-processed to highlight significant changes in texture. RESULTS The reconstruction algorithms of the presented workflow were able to construct 3-D models and corresponding 2-D atlas of both the in vivo bladder and the bladder phantom. For the in vivo bladder the portion of the reconstructed surface area was 58% and 79% for the pre- and post-operative scan, respectively. For the bladder phantom the full surface was reconstructed and the mean reprojection error was 0.081 mm (range 0-0.79 mm). In inter-session comparison the changes in texture were correctly indicated for all six locations. CONCLUSION The proposed proof of concept was able to perform 3-D and 2-D reconstruction of an in vivo bladder wall based on a set of monocular images. In a phantom study the computer vision algorithms were also effective in co-registering reconstructions of successive sessions and highlighting texture changes between sessions. These techniques may be useful for detecting, monitoring and revisiting suspicious lesions.
Collapse
Affiliation(s)
- Vincent Groenhuis
- Robotics and Mechatronics, University of Twente, Drienerlolaan 5, 7522 NB Enschede, The Netherlands
| | - Antonius G. de Groot
- Robotics and Mechatronics, University of Twente, Drienerlolaan 5, 7522 NB Enschede, The Netherlands
| | - Erik B. Cornel
- Department of Urology, Ziekenhuisgroep Twente (ZGT), Zilvermeeuw 1, 7609 PP Almelo, The Netherlands
| | - Stefano Stramigioli
- Robotics and Mechatronics, University of Twente, Drienerlolaan 5, 7522 NB Enschede, The Netherlands
| | - Françoise J. Siepel
- Robotics and Mechatronics, University of Twente, Drienerlolaan 5, 7522 NB Enschede, The Netherlands
| |
Collapse
|
14
|
Oliva Maza L, Steidle F, Klodmann J, Strobl K, Triebel R. An ORB-SLAM3-based Approach for Surgical Navigation in Ureteroscopy. COMPUTER METHODS IN BIOMECHANICS AND BIOMEDICAL ENGINEERING: IMAGING & VISUALIZATION 2022. [DOI: 10.1080/21681163.2022.2156392] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/24/2022]
Affiliation(s)
- Laura Oliva Maza
- PEK (Perzeption und Kognition), Deutsches Zentrum für Luft- und Raumfahrt (DLR), Weßling, Germany
| | - Florian Steidle
- PEK (Perzeption und Kognition), Deutsches Zentrum für Luft- und Raumfahrt (DLR), Weßling, Germany
| | - Julian Klodmann
- PEK (Perzeption und Kognition), Deutsches Zentrum für Luft- und Raumfahrt (DLR), Weßling, Germany
- ARR (Analyse und Regelung komplexer Robotersysteme), Deutsches Zentrum für Luft- und Raumfahrt (DLR), Weßling, Germany
| | - Klaus Strobl
- PEK (Perzeption und Kognition), Deutsches Zentrum für Luft- und Raumfahrt (DLR), Weßling, Germany
| | - Rudolph Triebel
- PEK (Perzeption und Kognition), Deutsches Zentrum für Luft- und Raumfahrt (DLR), Weßling, Germany
| |
Collapse
|
15
|
Fan J, Feng Y, Mo J, Wang S, Liang Q. Texture-less surface reconstruction using shape-based image augmentation. Comput Biol Med 2022; 150:106114. [PMID: 36179513 DOI: 10.1016/j.compbiomed.2022.106114] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/05/2022] [Revised: 09/09/2022] [Accepted: 09/17/2022] [Indexed: 11/24/2022]
Abstract
The development of intelligent Robot-Assisted Minimally Invasive Surgery demands geometric reconstruction from endoscopic images. However, images of human tissue surfaces are commonly texture-less. Obtaining the dense depth map of a texture-less scene is very difficult because traditional feature-based 3D reconstruction methods cannot detect enough features to build dense correspondences for depth computation. Given this problem, this study proposes a novel reconstruction method based on our shape-based image augmentation method. The main contribution of this method is the provision of a novel means to resolve the texture-less problem mainly on the input data level. In our method, we first calculate two shape gradient maps using Shape-From-Shading (SFS) method and we build Fast Point Feature Histogram (FPFH) 3D descriptor map according to the shape. Second, a series of augmented images can be computed by combining shape gradient maps, FPFH map, and the original image with different weights. Finally, we detect features on the new augmented images. Based on feature calculated sparse depth information and SFS calculated dense shape information, we further integrate a rectified dense depth map. Experiments show that our method can reconstruct texture-less surfaces with good accuracy.
Collapse
Affiliation(s)
- Jiacheng Fan
- School of Mechanical Engineering, Shanghai Jiao Tong University, 800 Dongchuan Road, Shanghai, CN, 200240, China.
| | - Yuan Feng
- School of Mechanical Engineering, Shanghai Jiao Tong University, 800 Dongchuan Road, Shanghai, CN, 200240, China
| | - Jinqiu Mo
- School of Mechanical Engineering, Shanghai Jiao Tong University, 800 Dongchuan Road, Shanghai, CN, 200240, China
| | - Shigang Wang
- School of Mechanical Engineering, Shanghai Jiao Tong University, 800 Dongchuan Road, Shanghai, CN, 200240, China
| | - Qinghua Liang
- School of Mechanical Engineering, Shanghai Jiao Tong University, 800 Dongchuan Road, Shanghai, CN, 200240, China.
| |
Collapse
|
16
|
Zenteno O, Trinh DH, Treuillet S, Lucas Y, Bazin T, Lamarque D, Daul C. Optical biopsy mapping on endoscopic image mosaics with a marker-free probe. Comput Biol Med 2022; 143:105234. [PMID: 35093845 DOI: 10.1016/j.compbiomed.2022.105234] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/19/2021] [Revised: 12/25/2021] [Accepted: 01/10/2022] [Indexed: 12/24/2022]
Abstract
Gastric cancer is the second leading cause of cancer-related deaths worldwide. Early diagnosis significantly increases the chances of survival; therefore, improved assisted exploration and screening techniques are necessary. Previously, we made use of an augmented multi-spectral endoscope by inserting an optical probe into the instrumentation channel. However, the limited field of view and the lack of markings left by optical biopsies on the tissue complicate the navigation and revisit of the suspect areas probed in-vivo. In this contribution two innovative tools are introduced to significantly increase the traceability and monitoring of patients in clinical practice: (i) video mosaicing to build a more comprehensive and panoramic view of large gastric areas; (ii) optical biopsy targeting and registration with the endoscopic images. The proposed optical flow-based mosaicing technique selects images that minimize texture discontinuities and is robust despite the lack of texture and illumination variations. The optical biopsy targeting is based on automatic tracking of a free-marker probe in the endoscopic view using deep learning to dynamically estimate its pose during exploration. The accuracy of pose estimation is sufficient to ensure a precise overlapping of the standard white-light color image and the hyperspectral probe image, assuming that the small target area of the organ is almost flat. This allows the mapping of all spatio-temporally tracked biopsy sites onto the panoramic mosaic. Experimental validations are carried out from videos acquired on patients in hospital. The proposed technique is purely software-based and therefore easily integrable into clinical practice. It is also generic and compatible to any imaging modality that connects to a fiberscope.
Collapse
Affiliation(s)
- Omar Zenteno
- Laboratoire PRISME, Université d'Orléans, Orléans, France
| | - Dinh-Hoan Trinh
- CRAN, UMR 7039 CNRS and Université de Lorraine, Vandœuvre-lès-Nancy, France
| | | | - Yves Lucas
- Laboratoire PRISME, Université d'Orléans, Orléans, France
| | - Thomas Bazin
- Service d'Hépato-gastroentérologie et oncologie digestive, Hôpital Ambroise Paré, Boulogne-Billancourt, France
| | - Dominique Lamarque
- Service d'Hépato-gastroentérologie et oncologie digestive, Hôpital Ambroise Paré, Boulogne-Billancourt, France
| | - Christian Daul
- CRAN, UMR 7039 CNRS and Université de Lorraine, Vandœuvre-lès-Nancy, France.
| |
Collapse
|
17
|
AIM in Endoscopy Procedures. Artif Intell Med 2022. [DOI: 10.1007/978-3-030-64573-1_164] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/07/2023]
|
18
|
Zhang Z, Liu Y, Tian J, Liu S, Yang B, Xiang L, Yin L, Zheng W. Study on Reconstruction and Feature Tracking of Silicone Heart 3D Surface. SENSORS 2021; 21:s21227570. [PMID: 34833646 PMCID: PMC8619637 DOI: 10.3390/s21227570] [Citation(s) in RCA: 12] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 09/29/2021] [Revised: 10/27/2021] [Accepted: 11/11/2021] [Indexed: 12/03/2022]
Abstract
At present, feature-based 3D reconstruction and tracking technology is widely applied in the medical field. In minimally invasive surgery, the surgeon can achieve three-dimensional reconstruction through the images obtained by the endoscope in the human body, restore the three-dimensional scene of the area to be operated on, and track the motion of the soft tissue surface. This enables doctors to have a clearer understanding of the location depth of the surgical area, greatly reducing the negative impact of 2D image defects and ensuring smooth operation. In this study, firstly, the 3D coordinates of each feature point are calculated by using the parameters of the parallel binocular endoscope and the spatial geometric constraints. At the same time, the discrete feature points are divided into multiple triangles using the Delaunay triangulation method. Then, the 3D coordinates of feature points and the division results of each triangle are combined to complete the 3D surface reconstruction. Combined with the feature matching method based on convolutional neural network, feature tracking is realized by calculating the three-dimensional coordinate changes of the same feature point in different frames. Finally, experiments are carried out on the endoscope image to complete the 3D surface reconstruction and feature tracking.
Collapse
Affiliation(s)
- Ziyan Zhang
- School of Innovation and Entrepreneurship, Xi’an Fanyi University, Xi’an 710105, China;
| | - Yan Liu
- School of Automation, University of Electronic Science and Technology of China, Chengdu 610054, China; (Y.L.); (J.T.); (B.Y.); (L.X.); (W.Z.)
| | - Jiawei Tian
- School of Automation, University of Electronic Science and Technology of China, Chengdu 610054, China; (Y.L.); (J.T.); (B.Y.); (L.X.); (W.Z.)
| | - Shan Liu
- School of Automation, University of Electronic Science and Technology of China, Chengdu 610054, China; (Y.L.); (J.T.); (B.Y.); (L.X.); (W.Z.)
- Correspondence:
| | - Bo Yang
- School of Automation, University of Electronic Science and Technology of China, Chengdu 610054, China; (Y.L.); (J.T.); (B.Y.); (L.X.); (W.Z.)
| | - Longhai Xiang
- School of Automation, University of Electronic Science and Technology of China, Chengdu 610054, China; (Y.L.); (J.T.); (B.Y.); (L.X.); (W.Z.)
| | - Lirong Yin
- Department of Geography and Anthropology, Louisiana State University, Baton Rouge, LA 70803, USA;
| | - Wenfeng Zheng
- School of Automation, University of Electronic Science and Technology of China, Chengdu 610054, China; (Y.L.); (J.T.); (B.Y.); (L.X.); (W.Z.)
| |
Collapse
|
19
|
Li L, Bano S, Deprest J, David A, Stoyanov D, Vasconcelos F. Globally Optimal Fetoscopic Mosaicking Based on Pose Graph Optimisation With Affine Constraints. IEEE Robot Autom Lett 2021. [DOI: 10.1109/lra.2021.3100938] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
20
|
Gong C, Brunton SL, Schowengerdt BT, Seibel EJ. Intensity-Mosaic: automatic panorama mosaicking of disordered images with insufficient features. J Med Imaging (Bellingham) 2021; 8:054002. [PMID: 34604440 PMCID: PMC8479456 DOI: 10.1117/1.jmi.8.5.054002] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/19/2021] [Accepted: 09/13/2021] [Indexed: 11/14/2022] Open
Abstract
Purpose: Handling low-quality and few-feature medical images is a challenging task in automatic panorama mosaicking. Current mosaicking methods for disordered input images are based on feature point matching, whereas in this case intensity-based registration achieves better performance than feature-point registration methods. We propose a mosaicking method that enables the use of mutual information (MI) registration for mosaicking randomly ordered input images with insufficient features. Approach: Dimensionality reduction is used to map disordered input images into a low dimensional space. Based on the low dimensional representation, the image global correspondence can be recognized efficiently. For adjacent image pairs, we optimize the MI metric for registration. The panorama is then created after image blending. We demonstrate our method on relatively lower-cost handheld devices that acquire images from the retina in vivo, kidney ex vivo, and bladder phantom, all of which contain sparse features. Results: Our method is compared with three baselines: AutoStitch, "dimension reduction + SIFT," and "MI-Only." Our method compared to the first two feature-point based methods exhibits 1.25 (ex vivo microscope dataset) to two times (in vivo retina dataset) rate of mosaic completion, and MI-Only has the lowest complete rate among three datasets. When comparing the subsequent complete mosaics, our target registration errors can be 2.2 and 3.8 times reduced when using the microscopy and bladder phantom datasets. Conclusions: Using dimensional reduction increases the success rate of detecting adjacent images, which makes MI-based registration feasible and narrows the search range of MI optimization. To the best of our knowledge, this is the first mosaicking method that allows automatic stitching of disordered images with intensity-based alignment, which provides more robust and accurate results when there are insufficient features for classic mosaicking methods.
Collapse
Affiliation(s)
- Chen Gong
- University of Washington, Department of Mechanical Engineering, Seattle, Washington, United States
| | - Steven L. Brunton
- University of Washington, Department of Mechanical Engineering, Seattle, Washington, United States
| | | | - Eric J. Seibel
- University of Washington, Department of Mechanical Engineering, Seattle, Washington, United States
| |
Collapse
|
21
|
Zhou H, Jayender J. Real-Time Nonrigid Mosaicking of Laparoscopy Images. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:1726-1736. [PMID: 33690113 PMCID: PMC8169627 DOI: 10.1109/tmi.2021.3065030] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
The ability to extend the field of view of laparoscopy images can help the surgeons to obtain a better understanding of the anatomical context. However, due to tissue deformation, complex camera motion and significant three-dimensional (3D) anatomical surface, image pixels may have non-rigid deformation and traditional mosaicking methods cannot work robustly for laparoscopy images in real-time. To solve this problem, a novel two-dimensional (2D) non-rigid simultaneous localization and mapping (SLAM) system is proposed in this paper, which is able to compensate for the deformation of pixels and perform image mosaicking in real-time. The key algorithm of this 2D non-rigid SLAM system is the expectation maximization and dual quaternion (EMDQ) algorithm, which can generate smooth and dense deformation field from sparse and noisy image feature matches in real-time. An uncertainty-based loop closing method has been proposed to reduce the accumulative errors. To achieve real-time performance, both CPU and GPU parallel computation technologies are used for dense mosaicking of all pixels. Experimental results on in vivo and synthetic data demonstrate the feasibility and accuracy of our non-rigid mosaicking method.
Collapse
|
22
|
Afifi A, Takada C, Yoshimura Y, Nakaguchi T. Real-Time Expanded Field-of-View for Minimally Invasive Surgery Using Multi-Camera Visual Simultaneous Localization and Mapping. SENSORS 2021; 21:s21062106. [PMID: 33802766 PMCID: PMC8002421 DOI: 10.3390/s21062106] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 01/14/2021] [Revised: 03/12/2021] [Accepted: 03/13/2021] [Indexed: 01/23/2023]
Abstract
Minimally invasive surgery is widely used because of its tremendous benefits to the patient. However, there are some challenges that surgeons face in this type of surgery, the most important of which is the narrow field of view. Therefore, we propose an approach to expand the field of view for minimally invasive surgery to enhance surgeons’ experience. It combines multiple views in real-time to produce a dynamic expanded view. The proposed approach extends the monocular Oriented features from an accelerated segment test and Rotated Binary robust independent elementary features—Simultaneous Localization And Mapping (ORB-SLAM) to work with a multi-camera setup. The ORB-SLAM’s three parallel threads, namely tracking, mapping and loop closing, are performed for each camera and new threads are added to calculate the relative cameras’ pose and to construct the expanded view. A new algorithm for estimating the optimal inter-camera correspondence matrix from a set of corresponding 3D map points is presented. This optimal transformation is then used to produce the final view. The proposed approach was evaluated using both human models and in vivo data. The evaluation results of the proposed correspondence matrix estimation algorithm prove its ability to reduce the error and to produce an accurate transformation. The results also show that when other approaches fail, the proposed approach can produce an expanded view. In this work, a real-time dynamic field-of-view expansion approach that can work in all situations regardless of images’ overlap is proposed. It outperforms the previous approaches and can also work at 21 fps.
Collapse
Affiliation(s)
- Ahmed Afifi
- Department of Computer Science, College of Computer Science and Information Technology, King Faisal University, P.O. Box 400, Al-Ahsa 31982, Saudi Arabia
- Faculty of Computers and Information, Menoufia University, Menoufia 32511, Egypt
- Correspondence: (A.A.); (T.N.)
| | - Chisato Takada
- Graduate School of Science and Engineering, Chiba University, Chiba 263-8522, Japan;
| | - Yuichiro Yoshimura
- Center for Frontier Medical Engineering, Chiba University, Chiba 263-8522, Japan;
| | - Toshiya Nakaguchi
- Center for Frontier Medical Engineering, Chiba University, Chiba 263-8522, Japan;
- Correspondence: (A.A.); (T.N.)
| |
Collapse
|
23
|
Widya AR, Monno Y, Okutomi M, Suzuki S, Gotoda T, Miki K. Stomach 3D Reconstruction Using Virtual Chromoendoscopic Images. IEEE JOURNAL OF TRANSLATIONAL ENGINEERING IN HEALTH AND MEDICINE-JTEHM 2021; 9:1700211. [PMID: 33796417 PMCID: PMC8009143 DOI: 10.1109/jtehm.2021.3062226] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/16/2020] [Revised: 01/19/2021] [Accepted: 02/15/2021] [Indexed: 12/23/2022]
Abstract
Gastric endoscopy is a golden standard in the clinical process that enables medical practitioners to diagnose various lesions inside a patient’s stomach. If a lesion is found, a success in identifying the location of the found lesion relative to the global view of the stomach will lead to better decision making for the next clinical treatment. Our previous research showed that the lesion localization could be achieved by reconstructing the whole stomach shape from chromoendoscopic indigo carmine (IC) dye-sprayed images using a structure-from-motion (SfM) pipeline. However, spraying the IC dye to the whole stomach requires additional time, which is not desirable for both patients and practitioners. Our objective is to propose an alternative way to achieve whole stomach 3D reconstruction without the need of the IC dye. We generate virtual IC-sprayed (VIC) images based on image-to-image style translation trained on unpaired real no-IC and IC-sprayed images, where we have investigated the effect of input and output color channel selection for generating the VIC images. We validate our reconstruction results by comparing them with the results using real IC-sprayed images and confirm that the obtained stomach 3D structures are comparable to each other. We also propose a local reconstruction technique to obtain a more detailed surface and texture around an interesting region. The proposed method achieves the whole stomach reconstruction without the need of real IC dye using SfM. We have found that translating no-IC green-channel images to IC-sprayed red-channel images gives the best SfM reconstruction result. Clinical impact We offer a method of the frame localization and local 3D reconstruction of a found gastric lesion using standard endoscopy images, leading to better clinical decision.
Collapse
Affiliation(s)
- Aji Resindra Widya
- Department of Systems and Control EngineeringSchool of EngineeringTokyo Institute of TechnologyTokyo152-8550Japan
| | - Yusuke Monno
- Department of Systems and Control EngineeringSchool of EngineeringTokyo Institute of TechnologyTokyo152-8550Japan
| | - Masatoshi Okutomi
- Department of Systems and Control EngineeringSchool of EngineeringTokyo Institute of TechnologyTokyo152-8550Japan
| | - Sho Suzuki
- Division of Gastroenterology and HepatologyDepartment of MedicineNihon University School of MedicineTokyo101-8309Japan
| | - Takuji Gotoda
- Division of Gastroenterology and HepatologyDepartment of MedicineNihon University School of MedicineTokyo101-8309Japan
| | - Kenji Miki
- Department of Internal MedicineTsujinaka Hospital KashiwanohaKashiwa277-0871Japan
| |
Collapse
|
24
|
Gong L, Zheng J, Ping Z, Wang Y, Wang S, Zuo S. Robust Mosaicing of Endomicroscopic Videos via Context-Weighted Correlation Ratio. IEEE Trans Biomed Eng 2021; 68:579-591. [PMID: 32746056 DOI: 10.1109/tbme.2020.3007768] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/07/2022]
Abstract
Probe-based confocal laser endomicroscopy (pCLE) is a promising imaging tool that provides in situ and in vivo optical imaging to perform real-time pathological assessments. However, due to limited field of view, it is difficult for clinicians to get a full understanding of the scanned tissues. In this paper, we develop a novel mosaicing framework to assemble all frame sequences into a full view image. First, a hybrid rigid registration that combines feature matching and template matching is presented to achieve a global alignment of all frames. Then, the parametric free-form deformation (FFD) model with a multiresolution architecture is implemented to accommodate non-rigid tissue distortions. More importantly, we devise a robust similarity metric called context-weighted correlation ratio (CWCR) to promote registration accuracy, where spatial and geometric contexts are incorporated into the estimation of functional intensity dependence. Experiments on both robotic setup and manual manipulation have demonstrated that the proposed scheme significantly precedes some state-of-the-art mosaicing schemes in the presence of intensity fluctuations, insufficient overlap and tissue distortions. Moreover, the comparisons of the proposed CWCR metric and two other metrics have validated the effectiveness of the context-weighted strategy in quantifying the differences between two frames. Benefiting from more rational and delicate mosaics, the proposed scheme is more suitable to instruct diagnosis and treatment during optical biopsies.
Collapse
|
25
|
Marzullo A, Moccia S, Calimeri F, De Momi E. AIM in Endoscopy Procedures. Artif Intell Med 2021. [DOI: 10.1007/978-3-030-58080-3_164-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/07/2022]
|
26
|
Zhou Y, Eimen RL, Seibel EJ, Bowden AK. Cost-Efficient Video Synthesis and Evaluation for Development of Virtual 3D Endoscopy. IEEE JOURNAL OF TRANSLATIONAL ENGINEERING IN HEALTH AND MEDICINE 2021; 9:1800711. [PMID: 34950539 PMCID: PMC8673697 DOI: 10.1109/jtehm.2021.3132193] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 06/16/2021] [Revised: 10/11/2021] [Accepted: 11/13/2021] [Indexed: 11/06/2022]
Affiliation(s)
- Yaxuan Zhou
- Department of Electrical and Computer EngineeringUniversity of Washington Seattle WA 98195 USA
- Human Photonics LaboratoryDepartment of Mechanical EngineeringUniversity of Washington Seattle WA 98195 USA
| | - Rachel L Eimen
- Department of Biomedical EngineeringVanderbilt University Nashville TN 37232 USA
| | - Eric J Seibel
- Human Photonics LaboratoryDepartment of Mechanical EngineeringUniversity of Washington Seattle WA 98195 USA
| | - Audrey K Bowden
- Department of Biomedical EngineeringVanderbilt University Nashville TN 37232 USA
- Department of Electrical Engineering and Computer ScienceVanderbilt University Nashville TN 37232 USA
| |
Collapse
|
27
|
He Q, Bano S, Ahmad OF, Yang B, Chen X, Valdastri P, Lovat LB, Stoyanov D, Zuo S. Deep learning-based anatomical site classification for upper gastrointestinal endoscopy. Int J Comput Assist Radiol Surg 2020; 15:1085-1094. [PMID: 32377939 PMCID: PMC7316667 DOI: 10.1007/s11548-020-02148-5] [Citation(s) in RCA: 17] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [Key Words] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/19/2019] [Accepted: 03/31/2020] [Indexed: 02/07/2023]
Abstract
Purpose Upper gastrointestinal (GI) endoscopic image documentation has provided an efficient, low-cost solution to address quality control for endoscopic reporting. The problem is, however, challenging for computer-assisted techniques, because different sites have similar appearances. Additionally, across different patients, site appearance variation may be large and inconsistent. Therefore, according to the British and modified Japanese guidelines, we propose a set of oesophagogastroduodenoscopy (EGD) images to be routinely captured and evaluate its efficiency for deep learning-based classification methods. Methods A novel EGD image dataset standardising upper GI endoscopy to several steps is established following landmarks proposed in guidelines and annotated by an expert clinician. To demonstrate the discrimination of proposed landmarks that enable the generation of an automated endoscopic report, we train several deep learning-based classification models utilising the well-annotated images. Results We report results for a clinical dataset composed of 211 patients (comprising a total of 3704 EGD images) acquired during routine upper GI endoscopic examinations. We find close agreement between predicted labels using our method and the ground truth labelled by human experts. We observe the limitation of current static image classification scheme for EGD image classification. Conclusion Our study presents a framework for developing automated EGD reports using deep learning. We demonstrate that our method is feasible to address EGD image classification and can lead towards improved performance and additionally qualitatively demonstrate its performance on our dataset.
Collapse
Affiliation(s)
- Qi He
- Key Laboratory of Mechanism Theory and Equipment Design of Ministry of Education, Tianjin University, Tianjin, China
| | - Sophia Bano
- Wellcome/EPSRC Centre for Interventional and Surgical Sciences (WEISS), University College London, London, UK
| | - Omer F Ahmad
- Wellcome/EPSRC Centre for Interventional and Surgical Sciences (WEISS), University College London, London, UK
| | - Bo Yang
- General Hospital, Tianjin Medical University, Tianjin, China
| | - Xin Chen
- General Hospital, Tianjin Medical University, Tianjin, China
| | - Pietro Valdastri
- School of Electronic and Electrical Engineering, University of Leeds, Leeds, UK
| | - Laurence B Lovat
- Wellcome/EPSRC Centre for Interventional and Surgical Sciences (WEISS), University College London, London, UK
| | - Danail Stoyanov
- Wellcome/EPSRC Centre for Interventional and Surgical Sciences (WEISS), University College London, London, UK
| | - Siyang Zuo
- Key Laboratory of Mechanism Theory and Equipment Design of Ministry of Education, Tianjin University, Tianjin, China.
| |
Collapse
|
28
|
Perperidis A, Dhaliwal K, McLaughlin S, Vercauteren T. Image computing for fibre-bundle endomicroscopy: A review. Med Image Anal 2020; 62:101620. [PMID: 32279053 PMCID: PMC7611433 DOI: 10.1016/j.media.2019.101620] [Citation(s) in RCA: 29] [Impact Index Per Article: 5.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/28/2018] [Accepted: 11/18/2019] [Indexed: 12/12/2022]
Abstract
Endomicroscopy is an emerging imaging modality, that facilitates the acquisition of in vivo, in situ optical biopsies, assisting diagnostic and potentially therapeutic interventions. While there is a diverse and constantly expanding range of commercial and experimental optical biopsy platforms available, fibre-bundle endomicroscopy is currently the most widely used platform and is approved for clinical use in a range of clinical indications. Miniaturised, flexible fibre-bundles, guided through the working channel of endoscopes, needles and catheters, enable high-resolution imaging across a variety of organ systems. Yet, the nature of image acquisition though a fibre-bundle gives rise to several inherent characteristics and limitations necessitating novel and effective image pre- and post-processing algorithms, ranging from image formation, enhancement and mosaicing to pathology detection and quantification. This paper introduces the underlying technology and most prevalent clinical applications of fibre-bundle endomicroscopy, and provides a comprehensive, up-to-date, review of relevant image reconstruction, analysis and understanding/inference methodologies. Furthermore, current limitations as well as future challenges and opportunities in fibre-bundle endomicroscopy computing are identified and discussed.
Collapse
Affiliation(s)
- Antonios Perperidis
- Institute of Sensors, Signals and Systems (ISSS), Heriot Watt University, EH14 4AS, UK; EPSRC IRC "Hub" in Optical Molecular Sensing & Imaging, MRC Centre for Inflammation Research, Queen's Medical Research Institute (QMRI), University of Edinburgh, EH16 4TJ, UK.
| | - Kevin Dhaliwal
- EPSRC IRC "Hub" in Optical Molecular Sensing & Imaging, MRC Centre for Inflammation Research, Queen's Medical Research Institute (QMRI), University of Edinburgh, EH16 4TJ, UK.
| | - Stephen McLaughlin
- Institute of Sensors, Signals and Systems (ISSS), Heriot Watt University, EH14 4AS, UK.
| | - Tom Vercauteren
- School of Biomedical Engineering and Imaging Sciences, King's College London, WC2R 2LS, UK.
| |
Collapse
|
29
|
Qiu L, Ren H. Endoscope navigation with SLAM-based registration to computed tomography for transoral surgery. INTERNATIONAL JOURNAL OF INTELLIGENT ROBOTICS AND APPLICATIONS 2020. [DOI: 10.1007/s41315-020-00127-2] [Citation(s) in RCA: 4] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 02/05/2023]
|
30
|
Perrot R, Bourdon P, Helbert D. Confidence-based dynamic optimization model for biomedical image mosaicking. JOURNAL OF THE OPTICAL SOCIETY OF AMERICA. A, OPTICS, IMAGE SCIENCE, AND VISION 2019; 36:C28-C39. [PMID: 31873691 DOI: 10.1364/josaa.36.000c28] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 06/18/2019] [Accepted: 09/25/2019] [Indexed: 06/10/2023]
Abstract
Biomedical image mosaicking is a trending topic. It consists of computing a single large image from multiple observations and becomes a challenging task when said observations barely overlap or are subject to illumination changes, poor resolution, blur, or either highly textured or predominantly homogeneous content. Because such challenges are common in biomedical images, classical keypoint/feature-based methods perform poorly. In this paper, we propose a new framework based on pairwise template matching coupled with a constrained, confidence-driven global optimization strategy to solve the issue of microscopic biomedical image mosaicking. First we provide experimental results that show significant improvement on a subjective level. Then we describe a new validation strategy for objective assessment, which shows improvement as well.
Collapse
|
31
|
Attar R, Xie X, Wang Z, Yue S. 2D reconstruction of small intestine's interior wall. Comput Biol Med 2018; 105:54-63. [PMID: 30583250 DOI: 10.1016/j.compbiomed.2018.12.001] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/06/2018] [Revised: 12/02/2018] [Accepted: 12/02/2018] [Indexed: 10/27/2022]
Abstract
Examining and interpreting of a large number of wireless endoscopic images from the gastrointestinal tract is a tiresome task for physicians. A practical solution is to automatically construct a two dimensional representation of the gastrointestinal tract for easy inspection. However, little has been done on wireless endoscopic image stitching, let alone systematic investigation. The proposed new wireless endoscopic image stitching method consists of two main steps to improve the accuracy and efficiency of image registration. First, the keypoints are extracted by Principle Component Analysis and Scale Invariant Feature Transform (PCA-SIFT) algorithm and refined with Maximum Likelihood Estimation SAmple Consensus (MLESAC) outlier removal to find the most reliable keypoints. Second, the optimal transformation parameters obtained from first step are fed to the Normalised Mutual Information (NMI) algorithm as an initial solution. With modified Marquardt-Levenberg search strategy in a multiscale framework, the NMI can find the optimal transformation parameters in the shortest time. The proposed methodology has been tested on two different datasets - one with real wireless endoscopic images and another with images obtained from Micro-Ball (a new wireless cubic endoscopy system with six image sensors). The results have demonstrated the accuracy and robustness of the proposed methodology both visually and quantitatively - registration residual error of 0.93±0.33 pixels on 2500 real endoscopy image pairs and residual error accumulation of 16.59 pixels and without affecting the visual registration quality on stitching 152 images of Micro-Ball.
Collapse
Affiliation(s)
- Rahman Attar
- School of Computing, University of Leeds, Leeds, UK; School of Computer Science, University of Lincoln, Lincoln, UK
| | - Xiang Xie
- Institute of Microelectronics, Tsinghua University, Beijing, China
| | - Zhihua Wang
- Institute of Microelectronics, Tsinghua University, Beijing, China
| | - Shigang Yue
- School of Computer Science, University of Lincoln, Lincoln, UK.
| |
Collapse
|
32
|
Long M, Li Z, Xie X, Li G, Wang Z. Adaptive Image Enhancement Based on Guide Image and Fraction-Power Transformation for Wireless Capsule Endoscopy. IEEE TRANSACTIONS ON BIOMEDICAL CIRCUITS AND SYSTEMS 2018; 12:993-1003. [PMID: 30346276 DOI: 10.1109/tbcas.2018.2869530] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/08/2023]
Abstract
Good image quality of the wireless capsule endoscopy (WCE) is the key for doctors to diagnose gastrointestinal (GI) tract diseases. However, the poor illumination, limited performance of the camera in WCE, and complex environment in the GI tract usually result in low-quality endoscopic images. Existing image enhancement methods only use the information of the image itself or multiple images of the same scene to accomplish the enhancement. In this paper, we propose an adaptive image enhancement method based on guide image and fraction-power transformation. First, intensities of endoscopic images are analyzed to assess the illumination conditions. Second, images captured under poor illumination conditions are enhanced by a brand-new image enhancement method called adaptive guide image based enhancement (AGIE). AGIE enhances low-quality images by using the information of a good quality image of the similar scene. Otherwise, images are enhanced by the proposed adaptive fraction-power transformation. Experimental results show that the proposed method improves the average intensity of endoscopic images by 64.20% and the average local entropy by 31.25%, which outperforms the state-of-art methods.
Collapse
|
33
|
Takada C, Afifi A, Suzuki T, Nakaguchi T. An enhanced hybrid tracking-mosaicking approach for surgical view expansion. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2018; 2017:3692-3695. [PMID: 29060700 DOI: 10.1109/embc.2017.8037659] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.1] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
Abstract
The aim of this work is to overcome the narrow surgical field of view problem in minimally invasive surgery. We achieve this by combining multiple views of the camera-retractable trocar which can obtain surgical viewpoints different from laparoscopic view. However, the accuracy and time are essential factors in this process. Therefore, we tend to improve the accuracy of a hybrid tracking-mosaicking approach which can combine several views at high speed. Two improvements are presented and analyzed here. The first improvement utilizes two sharping methodologies to enhance the image quality. This enhancement, in turn, improves the interest point extraction process and increases the number of extracted points. In the second enhancement, the tracking accuracy is improved by applying a filtering methodology to select the set of valid flow vectors only. This process reduces the tracking error which may accumulate during tracking. The experimental evaluation, shows that these improvements enhance the final mosaicking accuracy and allows us to construct a more accurate expanded view.
Collapse
|
34
|
Péntek Q, Hein S, Miernik A, Reiterer A. Image-based 3D surface approximation of the bladder using structure-from-motion for enhanced cystoscopy based on phantom data. ACTA ACUST UNITED AC 2018; 63:461-466. [DOI: 10.1515/bmt-2016-0185] [Citation(s) in RCA: 13] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/14/2016] [Accepted: 05/16/2017] [Indexed: 11/15/2022]
Abstract
Abstract
Bladder cancer is likely to recur after resection. For this reason, bladder cancer survivors often undergo follow-up cystoscopy for years after treatment to look for bladder cancer recurrence. 3D modeling of the bladder could provide more reliable cystoscopic documentation by giving an overall picture of the organ and tumor positions. However, 3D reconstruction of the urinary bladder based on endoscopic images is challenging. This is due to the small field of view of the endoscope, considerable image distortion, and occlusion by urea, blood or particles. In this paper, we will demonstrate a method for the conversion of uncalibrated, monocular, endoscopic videos of the bladder into a 3D model using structure-from-motion (SfM). First of all, frames are extracted from video sequences. Distortions are then corrected in a calibration procedure. Finally, the 3D reconstruction algorithm generates a sparse surface approximation of the bladder lining based on the corrected frames. This method was tested using an endoscopic video of a phantom that mimics the rich structure of the bladder. The reconstructed 3D model covered a large part of the object, with an average reprojection error of 1.15 pixels and a relative accuracy of 99.4%.
Collapse
|
35
|
Abstract
Bronchoscopy enables many minimally invasive chest procedures for diseases such as lung cancer and asthma. Guided by the bronchoscope's video stream, a physician can navigate the complex three-dimensional (3-D) airway tree to collect tissue samples or administer a disease treatment. Unfortunately, physicians currently discard procedural video because of the overwhelming amount of data generated. Hence, they must rely on memory and anecdotal snapshots to document a procedure. We propose a robust automatic method for summarizing an endobronchial video stream. Inspired by the multimedia concept of the video summary and by research in other endoscopy domains, our method consists of three main steps: 1) shot segmentation, 2) motion analysis, and 3) keyframe selection. Overall, the method derives a true hierarchical decomposition, consisting of a shot set and constituent keyframe set, for a given procedural video. No other method to our knowledge gives such a structured summary for the raw, unscripted, unedited videos arising in endoscopy. Results show that our method more efficiently covers the observed endobronchial regions than other keyframe-selection approaches and is robust to parameter variations. Over a wide range of video sequences, our method required on average only 6.5% of available video frames to achieve a video coverage = 92.7%. We also demonstrate how the derived video summary facilitates direct fusion with a patient's 3-D chest computed-tomography scan in a system under development, thereby enabling efficient video browsing and retrieval through the complex airway tree.
Collapse
|
36
|
Chen L, Tang W, John NW, Wan TR, Zhang JJ. SLAM-based dense surface reconstruction in monocular Minimally Invasive Surgery and its application to Augmented Reality. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2018; 158:135-146. [PMID: 29544779 DOI: 10.1016/j.cmpb.2018.02.006] [Citation(s) in RCA: 47] [Impact Index Per Article: 6.7] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 02/14/2017] [Revised: 01/03/2018] [Accepted: 02/02/2018] [Indexed: 06/08/2023]
Abstract
BACKGROUND AND OBJECTIVE While Minimally Invasive Surgery (MIS) offers considerable benefits to patients, it also imposes big challenges on a surgeon's performance due to well-known issues and restrictions associated with the field of view (FOV), hand-eye misalignment and disorientation, as well as the lack of stereoscopic depth perception in monocular endoscopy. Augmented Reality (AR) technology can help to overcome these limitations by augmenting the real scene with annotations, labels, tumour measurements or even a 3D reconstruction of anatomy structures at the target surgical locations. However, previous research attempts of using AR technology in monocular MIS surgical scenes have been mainly focused on the information overlay without addressing correct spatial calibrations, which could lead to incorrect localization of annotations and labels, and inaccurate depth cues and tumour measurements. In this paper, we present a novel intra-operative dense surface reconstruction framework that is capable of providing geometry information from only monocular MIS videos for geometry-aware AR applications such as site measurements and depth cues. We address a number of compelling issues in augmenting a scene for a monocular MIS environment, such as drifting and inaccurate planar mapping. METHODS A state-of-the-art Simultaneous Localization And Mapping (SLAM) algorithm used in robotics has been extended to deal with monocular MIS surgical scenes for reliable endoscopic camera tracking and salient point mapping. A robust global 3D surface reconstruction framework has been developed for building a dense surface using only unorganized sparse point clouds extracted from the SLAM. The 3D surface reconstruction framework employs the Moving Least Squares (MLS) smoothing algorithm and the Poisson surface reconstruction framework for real time processing of the point clouds data set. Finally, the 3D geometric information of the surgical scene allows better understanding and accurate placement AR augmentations based on a robust 3D calibration. RESULTS We demonstrate the clinical relevance of our proposed system through two examples: (a) measurement of the surface; (b) depth cues in monocular endoscopy. The performance and accuracy evaluations of the proposed framework consist of two steps. First, we have created a computer-generated endoscopy simulation video to quantify the accuracy of the camera tracking by comparing the results of the video camera tracking with the recorded ground-truth camera trajectories. The accuracy of the surface reconstruction is assessed by evaluating the Root Mean Square Distance (RMSD) of surface vertices of the reconstructed mesh with that of the ground truth 3D models. An error of 1.24 mm for the camera trajectories has been obtained and the RMSD for surface reconstruction is 2.54 mm, which compare favourably with previous approaches. Second, in vivo laparoscopic videos are used to examine the quality of accurate AR based annotation and measurement, and the creation of depth cues. These results show the potential promise of our geometry-aware AR technology to be used in MIS surgical scenes. CONCLUSIONS The results show that the new framework is robust and accurate in dealing with challenging situations such as the rapid endoscopy camera movements in monocular MIS scenes. Both camera tracking and surface reconstruction based on a sparse point cloud are effective and operated in real-time. This demonstrates the potential of our algorithm for accurate AR localization and depth augmentation with geometric cues and correct surface measurements in MIS with monocular endoscopes.
Collapse
|
37
|
Moccia S, Vanone GO, Momi ED, Laborai A, Guastini L, Peretti G, Mattos LS. Learning-based classification of informative laryngoscopic frames. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2018; 158:21-30. [PMID: 29544787 DOI: 10.1016/j.cmpb.2018.01.030] [Citation(s) in RCA: 24] [Impact Index Per Article: 3.4] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/27/2017] [Revised: 12/18/2017] [Accepted: 01/29/2018] [Indexed: 06/08/2023]
Abstract
BACKGROUND AND OBJECTIVE Early-stage diagnosis of laryngeal cancer is of primary importance to reduce patient morbidity. Narrow-band imaging (NBI) endoscopy is commonly used for screening purposes, reducing the risks linked to a biopsy but at the cost of some drawbacks, such as large amount of data to review to make the diagnosis. The purpose of this paper is to present a strategy to perform automatic selection of informative endoscopic video frames, which can reduce the amount of data to process and potentially increase diagnosis performance. METHODS A new method to classify NBI endoscopic frames based on intensity, keypoint and image spatial content features is proposed. Support vector machines with the radial basis function and the one-versus-one scheme are used to classify frames as informative, blurred, with saliva or specular reflections, or underexposed. RESULTS When tested on a balanced set of 720 images from 18 different laryngoscopic videos, a classification recall of 91% was achieved for informative frames, significantly overcoming three state of the art methods (Wilcoxon rank-signed test, significance level = 0.05). CONCLUSIONS Due to the high performance in identifying informative frames, the approach is a valuable tool to perform informative frame selection, which can be potentially applied in different fields, such us computer-assisted diagnosis and endoscopic view expansion.
Collapse
Affiliation(s)
- Sara Moccia
- Department of Electronics, Information, and Bioengineering, Politecnico di Milano, Milan, Italy; Department of Advanced Robotics, Istituto Italiano di Tecnologia, Genoa, Italy.
| | - Gabriele O Vanone
- Department of Electronics, Information, and Bioengineering, Politecnico di Milano, Milan, Italy
| | - Elena De Momi
- Department of Electronics, Information, and Bioengineering, Politecnico di Milano, Milan, Italy
| | - Andrea Laborai
- Department of Otorhinolaryngology, Head and Neck Surgery, University of Genoa, Genoa, Italy
| | - Luca Guastini
- Department of Otorhinolaryngology, Head and Neck Surgery, University of Genoa, Genoa, Italy
| | - Giorgio Peretti
- Department of Otorhinolaryngology, Head and Neck Surgery, University of Genoa, Genoa, Italy
| | - Leonardo S Mattos
- Department of Advanced Robotics, Istituto Italiano di Tecnologia, Genoa, Italy
| |
Collapse
|
38
|
Peter L, Tella-Amo M, Shakir DI, Attilakos G, Wimalasundera R, Deprest J, Ourselin S, Vercauteren T. Retrieval and registration of long-range overlapping frames for scalable mosaicking of in vivo fetoscopy. Int J Comput Assist Radiol Surg 2018; 13:713-720. [PMID: 29546573 PMCID: PMC5953985 DOI: 10.1007/s11548-018-1728-4] [Citation(s) in RCA: 12] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/29/2018] [Accepted: 02/28/2018] [Indexed: 11/17/2022]
Abstract
Purpose The standard clinical treatment of Twin-to-Twin transfusion syndrome consists in the photo-coagulation of undesired anastomoses located on the placenta which are responsible to a blood transfer between the two twins. While being the standard of care procedure, fetoscopy suffers from a limited field-of-view of the placenta resulting in missed anastomoses. To facilitate the task of the clinician, building a global map of the placenta providing a larger overview of the vascular network is highly desired. Methods To overcome the challenging visual conditions inherent to in vivo sequences (low contrast, obstructions or presence of artifacts, among others), we propose the following contributions: (1) robust pairwise registration is achieved by aligning the orientation of the image gradients, and (2) difficulties regarding long-range consistency (e.g. due to the presence of outliers) is tackled via a bag-of-word strategy, which identifies overlapping frames of the sequence to be registered regardless of their respective location in time. Results In addition to visual difficulties, in vivo sequences are characterised by the intrinsic absence of gold standard. We present mosaics motivating qualitatively our methodological choices and demonstrating their promising aspect. We also demonstrate semi-quantitatively, via visual inspection of registration results, the efficacy of our registration approach in comparison with two standard baselines. Conclusion This paper proposes the first approach for the construction of mosaics of placenta in in vivo fetoscopy sequences. Robustness to visual challenges during registration and long-range temporal consistency are proposed, offering first positive results on in vivo data for which standard mosaicking techniques are not applicable. Electronic supplementary material The online version of this article (10.1007/s11548-018-1728-4) contains supplementary material, which is available to authorized users.
Collapse
Affiliation(s)
- Loïc Peter
- Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, London, UK.
| | - Marcel Tella-Amo
- Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, London, UK
| | - Dzhoshkun Ismail Shakir
- Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, London, UK
| | | | | | - Jan Deprest
- Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, London, UK.,Department of Development and Regeneration, Cluster Woman and Child, Centre for Surgical Technologies, KU Leuven, Leuven, Belgium
| | - Sébastien Ourselin
- Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, London, UK
| | - Tom Vercauteren
- Wellcome/EPSRC Centre for Interventional and Surgical Sciences, University College London, London, UK.,Department of Development and Regeneration, Cluster Woman and Child, Centre for Surgical Technologies, KU Leuven, Leuven, Belgium
| |
Collapse
|
39
|
Marmol A, Peynot T, Eriksson A, Jaiprakash A, Roberts J, Crawford R. Evaluation of Keypoint Detectors and Descriptors in Arthroscopic Images for Feature-Based Matching Applications. IEEE Robot Autom Lett 2017. [DOI: 10.1109/lra.2017.2714150] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
|
40
|
Shi C, Luo X, Qi P, Li T, Song S, Najdovski Z, Fukuda T, Ren H. Shape Sensing Techniques for Continuum Robots in Minimally Invasive Surgery: A Survey. IEEE Trans Biomed Eng 2017; 64:1665-1678. [DOI: 10.1109/tbme.2016.2622361] [Citation(s) in RCA: 161] [Impact Index Per Article: 20.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/07/2022]
|
41
|
Workflow and simulation of image-to-physical registration of holes inside spongy bone. Int J Comput Assist Radiol Surg 2017; 12:1425-1437. [DOI: 10.1007/s11548-017-1594-5] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.4] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/17/2016] [Accepted: 04/24/2017] [Indexed: 10/19/2022]
|
42
|
The status of augmented reality in laparoscopic surgery as of 2016. Med Image Anal 2017; 37:66-90. [DOI: 10.1016/j.media.2017.01.007] [Citation(s) in RCA: 183] [Impact Index Per Article: 22.9] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/10/2016] [Revised: 01/16/2017] [Accepted: 01/23/2017] [Indexed: 12/27/2022]
|
43
|
Kriegmair MC, Bergen T, Ritter M, Mandel P, Michel MS, Wittenberg T, Bolenz C. Digital Mapping of the Urinary Bladder: Potential for Standardized Cystoscopy Reports. Urology 2017; 104:235-241. [PMID: 28214573 DOI: 10.1016/j.urology.2017.02.019] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.9] [Reference Citation Analysis] [Abstract] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/06/2016] [Revised: 01/16/2017] [Accepted: 02/09/2017] [Indexed: 11/27/2022]
Abstract
OBJECTIVE To develop a standardized digital reporting tool for cystoscopy of the urinary bladder using panoramic imaging. MATERIALS AND METHODS An image processing and stitching software (Endorama) was developed to generate panoramic images from cystoscopy data. In a processing phase, algorithms were modulated and refined by reference to cystoscopy sequences (n = 30). Subsequently, standard systematic cystoscopies (n = 12) were recorded in patients undergoing transurethral resection of a bladder tumor to create panoramic images. RESULTS All sequences were applicable for the development and refinements of the software. Processing increasingly allowed the creation of images illustrating large parts of the bladder and relevant anatomic landmarks in different locations. The pathway covered by the endoscope during the intervention was illustrated as a route in the respective digital image. During the application phase, panoramic images were successfully created in 10 out of 12 cases. The resolution of the images was 4096 × 2048 pixels and the images required a median digital memory of 3.9 MB (3.4-5.7). The panoramic images illustrated 22 relevant findings of which 7 were papillary tumors. CONCLUSION High-quality digital panoramic maps of the urinary bladder were created using specifically processed data of videocystoscopy. In this preliminary series, relevant findings were illustrated in the respective image. Our tool may help improve standardization of cystoscopy reports and reduce interobserver variability.
Collapse
Affiliation(s)
| | - Tobias Bergen
- Image Processing and Medical Engineering Department, Fraunhofer Institute for Integrated Circuits IIS, Erlangen, Germany
| | - Manuel Ritter
- Department of Urology, University Medical Center Mannheim, Mannheim, Germany
| | - Philipp Mandel
- Department of Urology, University Hospital Hamburg-Eppendorf, Hamburg, Germany
| | - Maurice S Michel
- Department of Urology, University Medical Center Mannheim, Mannheim, Germany
| | - Thomas Wittenberg
- Image Processing and Medical Engineering Department, Fraunhofer Institute for Integrated Circuits IIS, Erlangen, Germany
| | | |
Collapse
|
44
|
Armin MA, Chetty G, De Visser H, Dumas C, Grimpen F, Salvado O. Automated visibility map of the internal colon surface from colonoscopy video. Int J Comput Assist Radiol Surg 2016; 11:1599-610. [PMID: 27492067 DOI: 10.1007/s11548-016-1462-8] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.1] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/27/2016] [Accepted: 07/19/2016] [Indexed: 01/14/2023]
Abstract
PURPOSE Optical colonoscopy is a prominent procedure by which clinicians examine the surface of the colon for cancerous polyps using a flexible colonoscope. One of the main concerns regarding the quality of the colonoscopy is to ensure that the whole colonic surface has been inspected for abnormalities. In this paper, we aim at estimating areas that have not been covered thoroughly by providing a map from the internal colon surface. METHODS Camera parameters were estimated using optical flow between consecutive colonoscopy frames. A cylinder model was fitted to the colon structure using 3D pseudo stereo vision and projected into each frame. A circumferential band from the cylinder was extracted to unroll the internal colon surface (band image). By registering these band images, drift in estimating camera motion could be reduced, and a visibility map of the colon surface could be generated, revealing uncovered areas by the colonoscope. Hidden areas behind haustral folds were ignored in this study. The method was validated on simulated and actual colonoscopy videos. The realistic simulated videos were generated using a colonoscopy simulator with known ground truth, and the actual colonoscopy videos were manually assessed by a clinical expert. RESULTS The proposed method obtained a sensitivity and precision of 98 and 96 % for detecting the number of uncovered areas on simulated data, whereas validation on real videos showed a sensitivity and precision of 96 and 78 %, respectively. Error in camera motion drift could be reduced by almost 50 % using results from band image registration. CONCLUSION Using a simple cylindrical model for the colon and reducing drift by registering band images allows for the generation of visibility maps. The current results also suggest that the provided feedback through the visibility map could enhance clinicians' awareness of uncovered areas, which in return could reduce the probability of missing polyps.
Collapse
Affiliation(s)
- Mohammad Ali Armin
- HCT, University of Canberra, Bruce, Canberra, ACT, Australia.
- CSIRO Biomedical Informatics, The Australian e-Health Research Centre, Level 5, UQ Health Sciences, Brisbane, QLD, 4029, Australia.
| | - Girija Chetty
- HCT, University of Canberra, Bruce, Canberra, ACT, Australia
| | - Hans De Visser
- CSIRO Biomedical Informatics, The Australian e-Health Research Centre, Level 5, UQ Health Sciences, Brisbane, QLD, 4029, Australia
| | - Cedric Dumas
- CSIRO Biomedical Informatics, The Australian e-Health Research Centre, Level 5, UQ Health Sciences, Brisbane, QLD, 4029, Australia
| | - Florian Grimpen
- Department of Gastroenterology and Hepatology, Royal Brisbane and Women's Hospital, Herston, Brisbane, QLD, Australia
| | - Olivier Salvado
- CSIRO Biomedical Informatics, The Australian e-Health Research Centre, Level 5, UQ Health Sciences, Brisbane, QLD, 4029, Australia.
| |
Collapse
|
45
|
Moccia S, Penza V, Vanone GO, De Momi E, Mattos LS. Automatic workflow for narrow-band laryngeal video stitching. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2016; 2016:1188-1191. [PMID: 28268537 DOI: 10.1109/embc.2016.7590917] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/06/2023]
Abstract
In narrow band (NB) laryngeal endoscopy, the clinician usually positions the endoscope near the tissue for a correct inspection of possible vascular pattern alterations, indicative of laryngeal malignancies. The video is usually reviewed many times to refine the diagnosis, resulting in loss of time since the salient frames of the video are mixed with blurred, noisy, and redundant frames caused by the endoscope movements. The aim of this work is to provide to the clinician a unique larynx panorama, obtained through an automatic frame selection strategy to discard non-informative frames. Anisotropic diffusion filtering was exploited to lower the noise level while encouraging the selection of meaningful image features, and a feature-based stitching approach was carried out to generate the panorama. The frame selection strategy, tested on on six pathological NB endoscopic videos, was compared with standard strategies, as uniform and random sampling, showing higher performance of the subsequent stitching procedure, both visually, in terms of vascular structure preservation, and numerically, through a blur estimation metric.
Collapse
|
46
|
Penza V, Ortiz J, Mattos LS, Forgione A, De Momi E. Dense soft tissue 3D reconstruction refined with super-pixel segmentation for robotic abdominal surgery. Int J Comput Assist Radiol Surg 2015; 11:197-206. [DOI: 10.1007/s11548-015-1276-0] [Citation(s) in RCA: 21] [Impact Index Per Article: 2.1] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/12/2015] [Accepted: 07/27/2015] [Indexed: 11/30/2022]
|
47
|
Comparative study on surface reconstruction accuracy of stereo imaging devices for microsurgery. Int J Comput Assist Radiol Surg 2015; 11:145-56. [PMID: 26100121 DOI: 10.1007/s11548-015-1240-z] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/12/2015] [Accepted: 06/04/2015] [Indexed: 10/23/2022]
Abstract
PURPOSE Processing stereoscopic image data is an emerging field. Especially in microsurgery that requires sub-millimeter accuracy, application of stereo-based methods on endoscopic and microscopic scenarios is of major interest. In this context, direct comparison of stereo-based surface reconstruction applied to several camera settings is presented. METHODS A method for stereo matching is proposed and validated on in-vitro data. Demonstrating suitability for surgical scenarios, this method is applied to two custom-made stereo cameras, a miniaturized, bendable surgical endoscope and an operating microscope. Reconstruction accuracy is assessed on a custom-made reference sample. Subsequent to its fabrication, a coordinate measuring arm is used to acquire ground truth. Next, the sample is positioned by a robot at varying distances to each camera. Surface estimation is performed, while the specimen is localized based on. markers. Finally, the error between estimated surface and ground truth is computed. RESULTS Sample measurement with the coordinate measuring arm yields reliable ground truth data with a root-mean-square error of 11.2 μm. Overall surface reconstruction with analyzed cameras is quantified by a root-mean-square error of less than 0.18 mm. Microscope setting with the highest magnification yields the most accurate measurement, while the maximum deviation does not exceed 0.5 mm. Custom-made stereo cameras perform similar but with outliers of increased magnitude. Miniaturized, bendable surgical endoscope produces the maximum error of approximately 1.2 mm. CONCLUSIONS Reconstruction results reveal that microscopic imaging outperforms investigated chip-on-the-tip solutions, i.e., at higher magnification. Nonetheless, custom-made cameras are suitable for application in microsurgery. Although reconstruction with the miniaturized endoscope is more inaccurate, it provides a good trade-off between accuracy, outer dimensions and accessibility to hard-to-reach surgical sites.
Collapse
|