1
|
Chen YW, Zhang J, Wang P, Hu ZY, Zhong KH. Convolutional-de-convolutional neural networks for recognition of surgical workflow. Front Comput Neurosci 2022; 16:998096. [PMID: 36157842 PMCID: PMC9491113 DOI: 10.3389/fncom.2022.998096] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/19/2022] [Accepted: 08/18/2022] [Indexed: 11/13/2022] Open
Abstract
Computer-assisted surgery (CAS) has occupied an important position in modern surgery, further stimulating the progress of methodology and technology. In recent years, a large number of computer vision-based methods have been widely used in surgical workflow recognition tasks. For training the models, a lot of annotated data are necessary. However, the annotation of surgical data requires expert knowledge and thus becomes difficult and time-consuming. In this paper, we focus on the problem of data deficiency and propose a knowledge transfer learning method based on artificial neural network to compensate a small amount of labeled training data. To solve this problem, we propose an unsupervised method for pre-training a Convolutional-De-Convolutional (CDC) neural network for sequencing surgical workflow frames, which performs neural convolution in space (for semantic abstraction) and neural de-convolution in time (for frame level resolution) simultaneously. Specifically, through neural convolution transfer learning, we only fine-tuned the CDC neural network to classify the surgical phase. We performed some experiments for validating the model, and it showed that the proposed model can effectively extract the surgical feature and determine the surgical phase. The accuracy (Acc), recall, precision (Pres) of our model reached 91.4, 78.9, and 82.5%, respectively.
Collapse
Affiliation(s)
- Yu-wen Chen
- Chongqing Institute of Green and Intelligent Technology, Chinese Academy of Sciences, Chongqing, China
| | - Ju Zhang
- Chongqing Institute of Green and Intelligent Technology, Chinese Academy of Sciences, Chongqing, China
| | - Peng Wang
- Southwest Hospital, Third Military Medical University, Chongqing, China
| | - Zheng-yu Hu
- Chongqing Institute of Green and Intelligent Technology, Chinese Academy of Sciences, Chongqing, China
| | - Kun-hua Zhong
- Chongqing Institute of Green and Intelligent Technology, Chinese Academy of Sciences, Chongqing, China
- *Correspondence: Kun-hua Zhong,
| |
Collapse
|
2
|
Oregi I, Pérez A, Del Ser J, Lozano JA. An active adaptation strategy for streaming time series classification based on elastic similarity measures. Neural Comput Appl 2022. [DOI: 10.1007/s00521-022-07358-3] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/30/2022]
|
3
|
Zhou XH, Xie XL, Feng ZQ, Hou ZG, Bian GB, Li RQ, Ni ZL, Liu SQ, Zhou YJ. A Multilayer and Multimodal-Fusion Architecture for Simultaneous Recognition of Endovascular Manipulations and Assessment of Technical Skills. IEEE TRANSACTIONS ON CYBERNETICS 2022; 52:2565-2577. [PMID: 32697730 DOI: 10.1109/tcyb.2020.3004653] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/11/2023]
Abstract
The clinical success of the percutaneous coronary intervention (PCI) is highly dependent on endovascular manipulation skills and dexterous manipulation strategies of interventionalists. However, the analysis of endovascular manipulations and related discussion for technical skill assessment are limited. In this study, a multilayer and multimodal-fusion architecture is proposed to recognize six typical endovascular manipulations. The synchronously acquired multimodal motion signals from ten subjects are used as the inputs of the architecture independently. Six classification-based and two rule-based fusion algorithms are evaluated for performance comparisons. The recognition metrics under the determined architecture are further used to assess technical skills. The experimental results indicate that the proposed architecture can achieve the overall accuracy of 96.41%, much higher than that of a single-layer recognition architecture (92.85%). In addition, the multimodal fusion brings significant performance improvement in comparison with single-modal schemes. Furthermore, the K -means-based skill assessment can obtain an accuracy of 95% to cluster the attempts made by different skill-level groups. These hopeful results indicate the great possibility of the architecture to facilitate clinical skill assessment and skill learning.
Collapse
|
4
|
Wang Y, Dai J, Morgan TN, Elsaied M, Garbens A, Qu X, Steinberg R, Gahan J, Larson EC. Evaluating robotic-assisted surgery training videos with multi-task convolutional neural networks. J Robot Surg 2021; 16:917-925. [PMID: 34709538 DOI: 10.1007/s11701-021-01316-2] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/02/2021] [Accepted: 10/03/2021] [Indexed: 10/20/2022]
Abstract
We seek to understand if an automated algorithm can replace human scoring of surgical trainees performing the urethrovesical anastomosis in radical prostatectomy with synthetic tissue. Specifically, we investigate neural networks for predicting the surgical proficiency score (GEARS score) from video clips. We evaluate videos of surgeons performing the urethral anastomosis using synthetic tissue. The algorithm tracks surgical instrument locations from video, saving the positions of key points on the instruments over time. These positional features are used to train a multi-task convolutional network to infer each sub-category of the GEARS score to determine the proficiency level of trainees. Experimental results demonstrate that the proposed method achieves good performance with scores matching manual inspection in 86.1% of all GEARS sub-categories. Furthermore, the model can detect the difference between proficiency (novice to expert) in 83.3% of videos. Evaluation of GEARS sub-categories with artificial neural networks is possible for novice and intermediate surgeons, but additional research is needed to understand if expert surgeons can be evaluated with a similar automated system.
Collapse
Affiliation(s)
- Yihao Wang
- Department of Computer Science, Southern Methodist University, Dallas, USA
| | - Jessica Dai
- Department of Urology, University of Texas Southwestern Medical Center, Dallas, USA
| | - Tara N Morgan
- Department of Urology, University of Texas Southwestern Medical Center, Dallas, USA
| | - Mohamed Elsaied
- Department of Computer Science, Southern Methodist University, Dallas, USA
| | - Alaina Garbens
- Department of Urology, University of Texas Southwestern Medical Center, Dallas, USA
| | - Xingming Qu
- Department of Computer Science, Southern Methodist University, Dallas, USA
| | - Ryan Steinberg
- Department of Urology, University of Texas Southwestern Medical Center, Dallas, USA
| | - Jeffrey Gahan
- Department of Urology, University of Texas Southwestern Medical Center, Dallas, USA
| | - Eric C Larson
- Department of Computer Science, Southern Methodist University, Dallas, USA.
| |
Collapse
|
5
|
Bonatti J, Wallner S, Winkler B, Grabenwöger M. Robotic totally endoscopic coronary artery bypass grafting: current status and future prospects. Expert Rev Med Devices 2020; 17:33-40. [PMID: 31829047 DOI: 10.1080/17434440.2020.1704252] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.6] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/20/2022]
Abstract
Introduction: Totally endoscopic coronary artery bypass grafting (TECAB) can only be performed in a reproducible manner using robotic technology. This operation has been developed for more than 20 years seeing three generations of surgical robots. TECAB can be carried out beating heart but also on the arrested heart. Single and multiple grafts can be placed and TECAB can be combined with percutaneous coronary intervention in hybrid procedures.Areas covered: This review outlines indications for the procedure, the surgical technique, and the postoperative care. Intra- and postoperative results as available in the literature are reported. Further areas focus on technological development, training methods, learning curves as well as on cost. Finally, we give an outlook on the potential future of this operation.Expert opinion: Robotic TECAB represents a complex, sophisticated but safe, and over-the-years grown procedure. Even though results seem to be in line with conventional coronary surgery worldwide adoption still has been slow probably due to procedure times, costs and learning curves. Main advantages of TECAB are minimized surgical trauma and subsequent reduction of postoperative healing time. With the current introduction of new robotic devices, a new era of procedure development is on its way.
Collapse
Affiliation(s)
- Johannes Bonatti
- Department of Cardio-Vascular Surgery, Vienna North Hospital and Karl Landsteiner Institute for Cardio-Vascular Research, Vienna, Austria
| | - Stephanie Wallner
- Department of Cardio-Vascular Surgery, Vienna North Hospital and Karl Landsteiner Institute for Cardio-Vascular Research, Vienna, Austria
| | - Bernhard Winkler
- Department of Cardio-Vascular Surgery, Vienna North Hospital and Karl Landsteiner Institute for Cardio-Vascular Research, Vienna, Austria.,Center for Biomedical Research, Medical University of Vienna, Vienna, Austria
| | - Martin Grabenwöger
- Department of Cardio-Vascular Surgery, Vienna North Hospital and Karl Landsteiner Institute for Cardio-Vascular Research, Vienna, Austria.,Medical Faculty, Sigmund Freud University, Vienna, Austria
| |
Collapse
|
6
|
A Robotic Recording and Playback Platform for Training Surgeons and Learning Autonomous Behaviors Using the da Vinci Surgical System. ROBOTICS 2019. [DOI: 10.3390/robotics8010009] [Citation(s) in RCA: 10] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/16/2022] Open
Abstract
This paper describes a recording and playback system developed using a da Vinci Standard Surgical System and research kit. The system records stereo laparoscopic videos, robot arm joint angles, and surgeon–console interactions in a synchronized manner. A user can then, on-demand and at adjustable speeds, watch stereo videos and feel recorded movements on the hand controllers of entire procedures or sub procedures. Currently, there is no reported comprehensive ability to capture expert surgeon movements and insights and reproduce them on hardware directly. This system has important applications in several areas: (1) training of surgeons, (2) collection of learning data for the development of advanced control algorithms and intelligent autonomous behaviors, and (3) use as a “black box” for retrospective error analysis. We show a prototype of such an immersive system on a clinically-relevant platform along with its recording and playback fidelity. Lastly, we convey possible research avenues to create better systems for training and assisting robotic surgeons.
Collapse
|
7
|
Wang Z, Fey AM. SATR-DL: Improving Surgical Skill Assessment And Task Recognition In Robot-Assisted Surgery With Deep Neural Networks. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2018; 2018:1793-1796. [PMID: 30440742 DOI: 10.1109/embc.2018.8512575] [Citation(s) in RCA: 9] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Abstract
PURPOSE This paper focuses on an automated analysis of surgical motion profiles for objective skill assessment and task recognition in robot-assisted surgery. Existing techniques heavily rely on conventional statistic measures or shallow modelings based on hand-engineered features and gesture segmentation. Such developments require significant expert knowledge, are prone to errors, and are less efficient in online adaptive training systems. METHODS In this work, we present an efficient analytic framework with a parallel deep learning architecture, SATR-DL, to assess trainee expertise and recognize surgical training activity. Through an end-to-end learning technique, abstract information of spatial representations and temporal dynamics is jointly obtained directly from raw motion sequences. RESULTS By leveraging a shared highlevel representation learning, the resulting model is successful in the recognition of trainee skills and surgical tasks, suturing, needle-passing, and knot-tying. Meanwhile, we explore the use of ensemble in classification at the trial level, where the SATR-DL outperforms state-of-the-art performance by achieving accuracies of 0.960 and 1.000 in skill assessment and task recognition, respectively. CONCLUSION This study highlights the potential of SATR-DL to provide improvements for an efficient data-driven assessment in intelligent robotic surgery.
Collapse
|
8
|
Forestier G, Petitjean F, Senin P, Despinoy F, Huaulmé A, Fawaz HI, Weber J, Idoumghar L, Muller PA, Jannin P. Surgical motion analysis using discriminative interpretable patterns. Artif Intell Med 2018; 91:3-11. [PMID: 30172445 DOI: 10.1016/j.artmed.2018.08.002] [Citation(s) in RCA: 34] [Impact Index Per Article: 4.9] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/08/2017] [Revised: 07/06/2018] [Accepted: 08/13/2018] [Indexed: 11/29/2022]
Abstract
OBJECTIVE The analysis of surgical motion has received a growing interest with the development of devices allowing their automatic capture. In this context, the use of advanced surgical training systems makes an automated assessment of surgical trainee possible. Automatic and quantitative evaluation of surgical skills is a very important step in improving surgical patient care. MATERIAL AND METHOD In this paper, we present an approach for the discovery and ranking of discriminative and interpretable patterns of surgical practice from recordings of surgical motions. A pattern is defined as a series of actions or events in the kinematic data that together are distinctive of a specific gesture or skill level. Our approach is based on the decomposition of continuous kinematic data into a set of overlapping gestures represented by strings (bag of words) for which we compute comparative numerical statistic (tf-idf) enabling the discriminative gesture discovery via its relative occurrence frequency. RESULTS We carried out experiments on three surgical motion datasets. The results show that the patterns identified by the proposed method can be used to accurately classify individual gestures, skill levels and surgical interfaces. We also present how the patterns provide a detailed feedback on the trainee skill assessment. CONCLUSIONS The proposed approach is an interesting addition to existing learning tools for surgery as it provides a way to obtain a feedback on which parts of an exercise have been used to classify the attempt as correct or incorrect.
Collapse
Affiliation(s)
- Germain Forestier
- IRIMAS, Université de Haute-Alsace, Mulhouse, France; Faculty of Information Technology, Monash University, Melbourne, Australia.
| | - François Petitjean
- Faculty of Information Technology, Monash University, Melbourne, Australia.
| | - Pavel Senin
- Los Alamos National Laboratory, University Of Hawai'i at Mānoa, United States.
| | - Fabien Despinoy
- Univ Rennes, Inserm, LTSI - UMR_S 1099, F35000 Rennes, France.
| | - Arnaud Huaulmé
- Univ Rennes, Inserm, LTSI - UMR_S 1099, F35000 Rennes, France.
| | | | | | | | | | - Pierre Jannin
- Univ Rennes, Inserm, LTSI - UMR_S 1099, F35000 Rennes, France.
| |
Collapse
|