1
|
Xiao X, Zhang J, Shao Y, Liu J, Shi K, He C, Kong D. Deep Learning-Based Medical Ultrasound Image and Video Segmentation Methods: Overview, Frontiers, and Challenges. SENSORS (BASEL, SWITZERLAND) 2025; 25:2361. [PMID: 40285051 PMCID: PMC12031589 DOI: 10.3390/s25082361] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 03/13/2025] [Revised: 04/03/2025] [Accepted: 04/05/2025] [Indexed: 04/29/2025]
Abstract
The intricate imaging structures, artifacts, and noise present in ultrasound images and videos pose significant challenges for accurate segmentation. Deep learning has recently emerged as a prominent field, playing a crucial role in medical image processing. This paper reviews ultrasound image and video segmentation methods based on deep learning techniques, summarizing the latest developments in this field, such as diffusion and segment anything models as well as classical methods. These methods are classified into four main categories based on the characteristics of the segmentation methods. Each category is outlined and evaluated in the corresponding section. We provide a comprehensive overview of deep learning-based ultrasound image segmentation methods, evaluation metrics, and common ultrasound datasets, hoping to explain the advantages and disadvantages of each method, summarize its achievements, and discuss challenges and future trends.
Collapse
Affiliation(s)
- Xiaolong Xiao
- College of Mathematical Medicine, Zhejiang Normal University, Jinhua 321004, China; (X.X.); (Y.S.); (J.L.); (K.S.); (C.H.); (D.K.)
- School of Computer Science and Technology (School of Artificial Intelligence), Zhejiang Normal University, Jinhua 321004, China
| | - Jianfeng Zhang
- College of Mathematical Medicine, Zhejiang Normal University, Jinhua 321004, China; (X.X.); (Y.S.); (J.L.); (K.S.); (C.H.); (D.K.)
- Puyang Institute of Big Data and Artificial Intelligence, Puyang 457006, China
| | - Yuan Shao
- College of Mathematical Medicine, Zhejiang Normal University, Jinhua 321004, China; (X.X.); (Y.S.); (J.L.); (K.S.); (C.H.); (D.K.)
- School of Computer Science and Technology (School of Artificial Intelligence), Zhejiang Normal University, Jinhua 321004, China
| | - Jialong Liu
- College of Mathematical Medicine, Zhejiang Normal University, Jinhua 321004, China; (X.X.); (Y.S.); (J.L.); (K.S.); (C.H.); (D.K.)
- School of Computer Science and Technology (School of Artificial Intelligence), Zhejiang Normal University, Jinhua 321004, China
| | - Kaibing Shi
- College of Mathematical Medicine, Zhejiang Normal University, Jinhua 321004, China; (X.X.); (Y.S.); (J.L.); (K.S.); (C.H.); (D.K.)
| | - Chunlei He
- College of Mathematical Medicine, Zhejiang Normal University, Jinhua 321004, China; (X.X.); (Y.S.); (J.L.); (K.S.); (C.H.); (D.K.)
| | - Dexing Kong
- College of Mathematical Medicine, Zhejiang Normal University, Jinhua 321004, China; (X.X.); (Y.S.); (J.L.); (K.S.); (C.H.); (D.K.)
- School of Mathematical Sciences, Zhejiang University, Hangzhou 310027, China
| |
Collapse
|
2
|
Liu Y, Yuan D, Xu Z, Zhan Y, Zhang H, Lu J, Lukasiewicz T. Pixel level deep reinforcement learning for accurate and robust medical image segmentation. Sci Rep 2025; 15:8213. [PMID: 40064951 PMCID: PMC11894052 DOI: 10.1038/s41598-025-92117-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/14/2025] [Accepted: 02/25/2025] [Indexed: 03/14/2025] Open
Abstract
Existing deep learning methods have achieved significant success in medical image segmentation. However, this success largely relies on stacking advanced modules and architectures, which has created a path dependency. This path dependency is unsustainable, as it leads to increasingly larger model parameters and higher deployment costs. To break this path dependency, we introduce deep reinforcement learning to enhance segmentation performance. However, current deep reinforcement learning methods face challenges such as high training cost, independent iterative processes, and high uncertainty of segmentation masks. Consequently, we propose a Pixel-level Deep Reinforcement Learning model with pixel-by-pixel Mask Generation (PixelDRL-MG) for more accurate and robust medical image segmentation. PixelDRL-MG adopts a dynamic iterative update policy, directly segmenting the regions of interest without requiring user interaction or coarse segmentation masks. We propose a Pixel-level Asynchronous Advantage Actor-Critic (PA3C) strategy to treat each pixel as an agent whose state (foreground or background) is iteratively updated through direct actions. Our experiments on two commonly used medical image segmentation datasets demonstrate that PixelDRL-MG achieves more superior segmentation performances than the state-of-the-art segmentation baselines (especially in boundaries) using significantly fewer model parameters. We also conducted detailed ablation studies to enhance understanding and facilitate practical application. Additionally, PixelDRL-MG performs well in low-resource settings (i.e., 50-shot or 100-shot), making it an ideal choice for real-world scenarios.
Collapse
Affiliation(s)
- Yunxin Liu
- State Key Laboratory of Reliability and Intelligence of Electrical Equipment, School of Health Sciences and Biomedical Engineering, Hebei University of Technology, Tianjin, China
- Tianjin Key Laboratory of Bioelectromagnetic Technology and Intelligent Health, Hebei University of Technology, Tianjin, China
| | - Di Yuan
- State Key Laboratory of Reliability and Intelligence of Electrical Equipment, School of Health Sciences and Biomedical Engineering, Hebei University of Technology, Tianjin, China
- Tianjin Key Laboratory of Bioelectromagnetic Technology and Intelligent Health, Hebei University of Technology, Tianjin, China
| | - Zhenghua Xu
- State Key Laboratory of Reliability and Intelligence of Electrical Equipment, School of Health Sciences and Biomedical Engineering, Hebei University of Technology, Tianjin, China.
- Tianjin Key Laboratory of Bioelectromagnetic Technology and Intelligent Health, Hebei University of Technology, Tianjin, China.
| | - Yuefu Zhan
- The Third People's Hospital of Longgang District Shenzhen, Shenzhen, China.
- The Seventh People's Hospital of Chongqing, No. 1, Village 1, Lijiatuo Labor Union, Banan District, Chongqing, China.
- Longgang Institute of Medical Imaging, Shantou University Medical College, Shenzhen, China.
- Hainan Women and Children's Medical Center, Hainan, China.
| | - Hongwei Zhang
- BigBear (Tianjin) Medical Technology Co., Ltd, Tianjin, China
| | - Jun Lu
- BigBear (Tianjin) Medical Technology Co., Ltd, Tianjin, China
| | - Thomas Lukasiewicz
- Institute of Logic and Computation, Vienna University of Technology, Vienna, Austria
- Department of Computer Science, University of Oxford, Oxford, United Kingdom
| |
Collapse
|
3
|
Xu Z, Liu Y, Xu G, Lukasiewicz T. Self-Supervised Medical Image Segmentation Using Deep Reinforced Adaptive Masking. IEEE TRANSACTIONS ON MEDICAL IMAGING 2025; 44:180-193. [PMID: 39088493 DOI: 10.1109/tmi.2024.3436608] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 08/03/2024]
Abstract
Self-supervised learning aims to learn transferable representations from unlabeled data for downstream tasks. Inspired by masked language modeling in natural language processing, masked image modeling (MIM) has achieved certain success in the field of computer vision, but its effectiveness in medical images remains unsatisfactory. This is mainly due to the high redundancy and small discriminative regions in medical images compared to natural images. Therefore, this paper proposes an adaptive hard masking (AHM) approach based on deep reinforcement learning to expand the application of MIM in medical images. Unlike predefined random masks, AHM uses an asynchronous advantage actor-critic (A3C) model to predict reconstruction loss for each patch, enabling the model to learn where masking is valuable. By optimizing the non-differentiable sampling process using reinforcement learning, AHM enhances the understanding of key regions, thereby improving downstream task performance. Experimental results on two medical image datasets demonstrate that AHM outperforms state-of-the-art methods. Additional experiments under various settings validate the effectiveness of AHM in constructing masked images.
Collapse
|
4
|
Karunanayake N, Moodleah S, Makhanov SS. Edge-Driven Multi-Agent Reinforcement Learning: A Novel Approach to Ultrasound Breast Tumor Segmentation. Diagnostics (Basel) 2023; 13:3611. [PMID: 38132195 PMCID: PMC10742763 DOI: 10.3390/diagnostics13243611] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/20/2023] [Revised: 11/05/2023] [Accepted: 11/20/2023] [Indexed: 12/23/2023] Open
Abstract
A segmentation model of the ultrasound (US) images of breast tumors based on virtual agents trained using reinforcement learning (RL) is proposed. The agents, living in the edge map, are able to avoid false boundaries, connect broken parts, and finally, accurately delineate the contour of the tumor. The agents move similarly to robots navigating in the unknown environment with the goal of maximizing the rewards. The individual agent does not know the goal of the entire population. However, since the robots communicate, the model is able to understand the global information and fit the irregular boundaries of complicated objects. Combining the RL with a neural network makes it possible to automatically learn and select the local features. In particular, the agents handle the edge leaks and artifacts typical for the US images. The proposed model outperforms 13 state-of-the-art algorithms, including selected deep learning models and their modifications.
Collapse
Affiliation(s)
- Nalan Karunanayake
- Sirindhorn International Institute of Technology, Thammasat University, Pathum Thani 12120, Thailand;
| | - Samart Moodleah
- King Mongkut’s Institute of Technology Ladkrabang, Bangkok 10520, Thailand;
| | - Stanislav S. Makhanov
- Sirindhorn International Institute of Technology, Thammasat University, Pathum Thani 12120, Thailand;
| |
Collapse
|
5
|
Hu M, Zhang J, Matkovic L, Liu T, Yang X. Reinforcement learning in medical image analysis: Concepts, applications, challenges, and future directions. J Appl Clin Med Phys 2023; 24:e13898. [PMID: 36626026 PMCID: PMC9924115 DOI: 10.1002/acm2.13898] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/04/2022] [Revised: 12/14/2022] [Accepted: 12/23/2022] [Indexed: 01/11/2023] Open
Abstract
MOTIVATION Medical image analysis involves a series of tasks used to assist physicians in qualitative and quantitative analyses of lesions or anatomical structures which can significantly improve the accuracy and reliability of medical diagnoses and prognoses. Traditionally, these tedious tasks were finished by experienced physicians or medical physicists and were marred with two major problems, low efficiency and bias. In the past decade, many machine learning methods have been applied to accelerate and automate the image analysis process. Compared to the enormous deployments of supervised and unsupervised learning models, attempts to use reinforcement learning in medical image analysis are still scarce. We hope that this review article could serve as the stepping stone for related research in the future. SIGNIFICANCE We found that although reinforcement learning has gradually gained momentum in recent years, many researchers in the medical analysis field still find it hard to understand and deploy in clinical settings. One possible cause is a lack of well-organized review articles intended for readers without professional computer science backgrounds. Rather than to provide a comprehensive list of all reinforcement learning models applied in medical image analysis, the aim of this review is to help the readers formulate and solve their medical image analysis research through the lens of reinforcement learning. APPROACH & RESULTS We selected published articles from Google Scholar and PubMed. Considering the scarcity of related articles, we also included some outstanding newest preprints. The papers were carefully reviewed and categorized according to the type of image analysis task. In this article, we first reviewed the basic concepts and popular models of reinforcement learning. Then, we explored the applications of reinforcement learning models in medical image analysis. Finally, we concluded the article by discussing the reviewed reinforcement learning approaches' limitations and possible future improvements.
Collapse
Affiliation(s)
- Mingzhe Hu
- Department of Radiation OncologySchool of MedicineEmory UniversityAtlantaGeorgiaUSA,Department of Computer Science and InformaticsEmory UniversityAtlantaGeorgiaUSA
| | - Jiahan Zhang
- Department of Radiation OncologySchool of MedicineEmory UniversityAtlantaGeorgiaUSA
| | - Luke Matkovic
- Department of Radiation OncologySchool of MedicineEmory UniversityAtlantaGeorgiaUSA
| | - Tian Liu
- Department of Radiation OncologySchool of MedicineEmory UniversityAtlantaGeorgiaUSA
| | - Xiaofeng Yang
- Department of Radiation OncologySchool of MedicineEmory UniversityAtlantaGeorgiaUSA,Department of Computer Science and InformaticsEmory UniversityAtlantaGeorgiaUSA
| |
Collapse
|
6
|
Xu L, Zhu S, Wen N. Deep reinforcement learning and its applications in medical imaging and radiation therapy: a survey. Phys Med Biol 2022; 67. [PMID: 36270582 DOI: 10.1088/1361-6560/ac9cb3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/16/2022] [Accepted: 10/21/2022] [Indexed: 11/07/2022]
Abstract
Reinforcement learning takes sequential decision-making approaches by learning the policy through trial and error based on interaction with the environment. Combining deep learning and reinforcement learning can empower the agent to learn the interactions and the distribution of rewards from state-action pairs to achieve effective and efficient solutions in more complex and dynamic environments. Deep reinforcement learning (DRL) has demonstrated astonishing performance in surpassing the human-level performance in the game domain and many other simulated environments. This paper introduces the basics of reinforcement learning and reviews various categories of DRL algorithms and DRL models developed for medical image analysis and radiation treatment planning optimization. We will also discuss the current challenges of DRL and approaches proposed to make DRL more generalizable and robust in a real-world environment. DRL algorithms, by fostering the designs of the reward function, agents interactions and environment models, can resolve the challenges from scarce and heterogeneous annotated medical image data, which has been a major obstacle to implementing deep learning models in the clinic. DRL is an active research area with enormous potential to improve deep learning applications in medical imaging and radiation therapy planning.
Collapse
Affiliation(s)
- Lanyu Xu
- Department of Computer Science and Engineering, Oakland University, Rochester, MI, United States of America
| | - Simeng Zhu
- Department of Radiation Oncology, Henry Ford Health Systems, Detroit, MI, United States of America
| | - Ning Wen
- Department of Radiology/The Institute for Medical Imaging Technology, Ruijin Hospital, Shanghai Jiaotong University School of Medicine, Shanghai, People's Republic of China.,The Global Institute of Future Technology, Shanghai Jiaotong University, Shanghai, People's Republic of China
| |
Collapse
|
7
|
Zhou SK, Le HN, Luu K, V Nguyen H, Ayache N. Deep reinforcement learning in medical imaging: A literature review. Med Image Anal 2021; 73:102193. [PMID: 34371440 DOI: 10.1016/j.media.2021.102193] [Citation(s) in RCA: 57] [Impact Index Per Article: 14.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/27/2021] [Revised: 05/22/2021] [Accepted: 07/20/2021] [Indexed: 12/29/2022]
Abstract
Deep reinforcement learning (DRL) augments the reinforcement learning framework, which learns a sequence of actions that maximizes the expected reward, with the representative power of deep neural networks. Recent works have demonstrated the great potential of DRL in medicine and healthcare. This paper presents a literature review of DRL in medical imaging. We start with a comprehensive tutorial of DRL, including the latest model-free and model-based algorithms. We then cover existing DRL applications for medical imaging, which are roughly divided into three main categories: (i) parametric medical image analysis tasks including landmark detection, object/lesion detection, registration, and view plane localization; (ii) solving optimization tasks including hyperparameter tuning, selecting augmentation strategies, and neural architecture search; and (iii) miscellaneous applications including surgical gesture segmentation, personalized mobile health intervention, and computational model personalization. The paper concludes with discussions of future perspectives.
Collapse
Affiliation(s)
- S Kevin Zhou
- Medical Imaging, Robotics, and Analytic Computing Laboratory and Enigineering (MIRACLE) Center, School of Biomedical Engineering & Suzhou Institute for Advanced Research, University of Science and Technology of China; Key Lab of Intelligent Information Processing of Chinese Academy of Sciences (CAS), Institute of Computing Technology, CAS, China.
| | | | - Khoa Luu
- CSCE Department, University of Arkansas, US
| | | | | |
Collapse
|
8
|
Xiong J, Po LM, Cheung KW, Xian P, Zhao Y, Rehman YAU, Zhang Y. Edge-Sensitive Left Ventricle Segmentation Using Deep Reinforcement Learning. SENSORS (BASEL, SWITZERLAND) 2021; 21:2375. [PMID: 33805558 PMCID: PMC8037138 DOI: 10.3390/s21072375] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/21/2021] [Revised: 03/25/2021] [Accepted: 03/26/2021] [Indexed: 12/18/2022]
Abstract
Deep reinforcement learning (DRL) has been utilized in numerous computer vision tasks, such as object detection, autonomous driving, etc. However, relatively few DRL methods have been proposed in the area of image segmentation, particularly in left ventricle segmentation. Reinforcement learning-based methods in earlier works often rely on learning proper thresholds to perform segmentation, and the segmentation results are inaccurate due to the sensitivity of the threshold. To tackle this problem, a novel DRL agent is designed to imitate the human process to perform LV segmentation. For this purpose, we formulate the segmentation problem as a Markov decision process and innovatively optimize it through DRL. The proposed DRL agent consists of two neural networks, i.e., First-P-Net and Next-P-Net. The First-P-Net locates the initial edge point, and the Next-P-Net locates the remaining edge points successively and ultimately obtains a closed segmentation result. The experimental results show that the proposed model has outperformed the previous reinforcement learning methods and achieved comparable performances compared with deep learning baselines on two widely used LV endocardium segmentation datasets, namely Automated Cardiac Diagnosis Challenge (ACDC) 2017 dataset, and Sunnybrook 2009 dataset. Moreover, the proposed model achieves higher F-measure accuracy compared with deep learning methods when training with a very limited number of samples.
Collapse
Affiliation(s)
- Jingjing Xiong
- Department of Electrical Engineering, City University of Hong Kong, Tat Chee Avenue, Kowloon, Hong Kong, China; (L.-M.P.); (P.X.); (Y.Z.); (Y.Z.)
| | - Lai-Man Po
- Department of Electrical Engineering, City University of Hong Kong, Tat Chee Avenue, Kowloon, Hong Kong, China; (L.-M.P.); (P.X.); (Y.Z.); (Y.Z.)
| | - Kwok Wai Cheung
- School of Communication, The Hang Seng University of Hong Kong, Hang Shin Link, Siu Lek Yuen, Shatin, Hong Kong, China;
| | - Pengfei Xian
- Department of Electrical Engineering, City University of Hong Kong, Tat Chee Avenue, Kowloon, Hong Kong, China; (L.-M.P.); (P.X.); (Y.Z.); (Y.Z.)
| | - Yuzhi Zhao
- Department of Electrical Engineering, City University of Hong Kong, Tat Chee Avenue, Kowloon, Hong Kong, China; (L.-M.P.); (P.X.); (Y.Z.); (Y.Z.)
| | - Yasar Abbas Ur Rehman
- TCL Corporate Research (HK) Co., Ltd., 22 Science Park East Avenue, Shatin, Hong Kong, China;
| | - Yujia Zhang
- Department of Electrical Engineering, City University of Hong Kong, Tat Chee Avenue, Kowloon, Hong Kong, China; (L.-M.P.); (P.X.); (Y.Z.); (Y.Z.)
| |
Collapse
|
9
|
Habijan M, Babin D, Galić I, Leventić H, Romić K, Velicki L, Pižurica A. Overview of the Whole Heart and Heart Chamber Segmentation Methods. Cardiovasc Eng Technol 2020; 11:725-747. [DOI: 10.1007/s13239-020-00494-8] [Citation(s) in RCA: 6] [Impact Index Per Article: 1.2] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 07/02/2020] [Accepted: 10/06/2020] [Indexed: 12/13/2022]
|
10
|
Wang R, Pan W, Jin L, Li Y, Geng Y, Gao C, Chen G, Wang H, Ma D, Liao S. Artificial intelligence in reproductive medicine. Reproduction 2019; 158:R139-R154. [PMID: 30970326 PMCID: PMC6733338 DOI: 10.1530/rep-18-0523] [Citation(s) in RCA: 89] [Impact Index Per Article: 14.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/03/2018] [Accepted: 04/10/2019] [Indexed: 12/16/2022]
Abstract
Artificial intelligence (AI) has experienced rapid growth over the past few years, moving from the experimental to the implementation phase in various fields, including medicine. Advances in learning algorithms and theories, the availability of large datasets and improvements in computing power have contributed to breakthroughs in current AI applications. Machine learning (ML), a subset of AI, allows computers to detect patterns from large complex datasets automatically and uses these patterns to make predictions. AI is proving to be increasingly applicable to healthcare, and multiple machine learning techniques have been used to improve the performance of assisted reproductive technology (ART). Despite various challenges, the integration of AI and reproductive medicine is bound to give an essential direction to medical development in the future. In this review, we discuss the basic aspects of AI and machine learning, and we address the applications, potential limitations and challenges of AI. We also highlight the prospects and future directions in the context of reproductive medicine.
Collapse
Affiliation(s)
- Renjie Wang
- Department of Obstetrics and Gynecology, Cancer Biology Research Center, Tongji Hospital, Tongji Medical College of HUST, Wuhan, Hubei, People’s Republic of China
| | - Wei Pan
- School of Economics and Management, Wuhan University, Wuhan, Hubei, People’s Republic of China
| | - Lei Jin
- Department of Obstetrics and Gynecology, Cancer Biology Research Center, Tongji Hospital, Tongji Medical College of HUST, Wuhan, Hubei, People’s Republic of China
| | - Yuehan Li
- Department of Obstetrics and Gynecology, Cancer Biology Research Center, Tongji Hospital, Tongji Medical College of HUST, Wuhan, Hubei, People’s Republic of China
| | - Yudi Geng
- Department of Obstetrics and Gynecology, Cancer Biology Research Center, Tongji Hospital, Tongji Medical College of HUST, Wuhan, Hubei, People’s Republic of China
| | - Chun Gao
- Department of Obstetrics and Gynecology, Cancer Biology Research Center, Tongji Hospital, Tongji Medical College of HUST, Wuhan, Hubei, People’s Republic of China
| | - Gang Chen
- Department of Obstetrics and Gynecology, Cancer Biology Research Center, Tongji Hospital, Tongji Medical College of HUST, Wuhan, Hubei, People’s Republic of China
| | - Hui Wang
- Department of Obstetrics and Gynecology, Cancer Biology Research Center, Tongji Hospital, Tongji Medical College of HUST, Wuhan, Hubei, People’s Republic of China
| | - Ding Ma
- Department of Obstetrics and Gynecology, Cancer Biology Research Center, Tongji Hospital, Tongji Medical College of HUST, Wuhan, Hubei, People’s Republic of China
| | - Shujie Liao
- Department of Obstetrics and Gynecology, Cancer Biology Research Center, Tongji Hospital, Tongji Medical College of HUST, Wuhan, Hubei, People’s Republic of China
- Correspondence should be addressed to S Liao;
| |
Collapse
|
11
|
Kucur ŞS, Márquez-Neila P, Abegg M, Sznitman R. Patient-attentive sequential strategy for perimetry-based visual field acquisition. Med Image Anal 2019; 54:179-192. [PMID: 30933865 DOI: 10.1016/j.media.2019.03.002] [Citation(s) in RCA: 5] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/22/2018] [Revised: 03/08/2019] [Accepted: 03/14/2019] [Indexed: 11/28/2022]
Abstract
Perimetry is a non-invasive clinical psychometric examination used for diagnosing ophthalmic and neurological conditions. At its core, perimetry relies on a subject pressing a button whenever they see a visual stimulus within their field of view. This sequential process then yields a 2D visual field image that is critical for clinical use. Perimetry is painfully slow however, with examinations lasting 7-8 minutes per eye. Maintaining high levels of concentration during that time is exhausting for the patient and negatively affects the acquired visual field. We introduce PASS, a novel perimetry testing strategy, based on reinforcement learning, that requires fewer locations in order to effectively estimate 2D visual fields. PASS uses a selection policy that determines what locations should be tested in order to reconstruct the complete visual field as accurately as possible, and then separately reconstructs the visual field from sparse observations. Furthermore, PASS is patient-specific and non-greedy. It adaptively selects what locations to query based on the patient's answers to previous queries, and the locations are jointly selected to maximize the quality of the final reconstruction. In our experiments, we show that PASS outperforms state-of-the-art methods, leading to more accurate reconstructions while reducing between 30% and 70% the duration of the patient examination.
Collapse
Affiliation(s)
- Şerife Seda Kucur
- ARTORG Center for Biomedical Engineering Research, University of Bern, Bern, Switzerland.
| | - Pablo Márquez-Neila
- ARTORG Center for Biomedical Engineering Research, University of Bern, Bern, Switzerland
| | - Mathias Abegg
- Department of Ophthalmology, Bern University Hospital, Inselspital, Bern, Switzerland
| | - Raphael Sznitman
- ARTORG Center for Biomedical Engineering Research, University of Bern, Bern, Switzerland
| |
Collapse
|
12
|
Fujii K, Gras G, Salerno A, Yang GZ. Gaze gesture based human robot interaction for laparoscopic surgery. Med Image Anal 2017; 44:196-214. [PMID: 29277075 DOI: 10.1016/j.media.2017.11.011] [Citation(s) in RCA: 24] [Impact Index Per Article: 3.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/12/2016] [Revised: 11/22/2017] [Accepted: 11/23/2017] [Indexed: 02/07/2023]
Abstract
While minimally invasive surgery offers great benefits in terms of reduced patient trauma, bleeding, as well as faster recovery time, it still presents surgeons with major ergonomic challenges. Laparoscopic surgery requires the surgeon to bimanually control surgical instruments during the operation. A dedicated assistant is thus required to manoeuvre the camera, which is often difficult to synchronise with the surgeon's movements. This article introduces a robotic system in which a rigid endoscope held by a robotic arm is controlled via the surgeon's eye movement, thus forgoing the need for a camera assistant. Gaze gestures detected via a series of eye movements are used to convey the surgeon's intention to initiate gaze contingent camera control. Hidden Markov Models (HMMs) are used for real-time gaze gesture recognition, allowing the robotic camera to pan, tilt, and zoom, whilst immune to aberrant or unintentional eye movements. A novel online calibration method for the gaze tracker is proposed, which overcomes calibration drift and simplifies its clinical application. This robotic system has been validated by comprehensive user trials and a detailed analysis performed on usability metrics to assess the performance of the system. The results demonstrate that the surgeons can perform their tasks quicker and more efficiently when compared to the use of a camera assistant or foot switches.
Collapse
Affiliation(s)
- Kenko Fujii
- The Hamlyn Centre for Robotic Surgery, Imperial College London, UK
| | - Gauthier Gras
- The Hamlyn Centre for Robotic Surgery, Imperial College London, UK
| | - Antonino Salerno
- The Hamlyn Centre for Robotic Surgery, Imperial College London, UK
| | - Guang-Zhong Yang
- The Hamlyn Centre for Robotic Surgery, Imperial College London, UK.
| |
Collapse
|
13
|
Peng P, Lekadir K, Gooya A, Shao L, Petersen SE, Frangi AF. A review of heart chamber segmentation for structural and functional analysis using cardiac magnetic resonance imaging. MAGMA (NEW YORK, N.Y.) 2016; 29:155-95. [PMID: 26811173 PMCID: PMC4830888 DOI: 10.1007/s10334-015-0521-4] [Citation(s) in RCA: 128] [Impact Index Per Article: 14.2] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 07/09/2015] [Revised: 12/01/2015] [Accepted: 12/17/2015] [Indexed: 01/19/2023]
Abstract
Cardiovascular magnetic resonance (CMR) has become a key imaging modality in clinical cardiology practice due to its unique capabilities for non-invasive imaging of the cardiac chambers and great vessels. A wide range of CMR sequences have been developed to assess various aspects of cardiac structure and function, and significant advances have also been made in terms of imaging quality and acquisition times. A lot of research has been dedicated to the development of global and regional quantitative CMR indices that help the distinction between health and pathology. The goal of this review paper is to discuss the structural and functional CMR indices that have been proposed thus far for clinical assessment of the cardiac chambers. We include indices definitions, the requirements for the calculations, exemplar applications in cardiovascular diseases, and the corresponding normal ranges. Furthermore, we review the most recent state-of-the art techniques for the automatic segmentation of the cardiac boundaries, which are necessary for the calculation of the CMR indices. Finally, we provide a detailed discussion of the existing literature and of the future challenges that need to be addressed to enable a more robust and comprehensive assessment of the cardiac chambers in clinical practice.
Collapse
Affiliation(s)
- Peng Peng
- Department of Electronic and Electrical Engineering, The University of Sheffield, Sheffield, S1 3JD, UK
| | | | - Ali Gooya
- Department of Electronic and Electrical Engineering, The University of Sheffield, Sheffield, S1 3JD, UK
| | - Ling Shao
- Department of Computer Science and Digital Technologies, Northumbria University, Newcastle upon Tyne, NE1 8ST, UK
| | - Steffen E Petersen
- Centre Lead for Advanced Cardiovascular Imaging, William Harvey Research Institute, Queen Mary University of London, London, EC1M 6BQ, UK
| | - Alejandro F Frangi
- Department of Electronic and Electrical Engineering, The University of Sheffield, Sheffield, S1 3JD, UK.
| |
Collapse
|
14
|
Albà X, Pereañez M, Hoogendoorn C, Swift AJ, Wild JM, Frangi AF, Lekadir K. An Algorithm for the Segmentation of Highly Abnormal Hearts Using a Generic Statistical Shape Model. IEEE TRANSACTIONS ON MEDICAL IMAGING 2016; 35:845-859. [PMID: 26552082 DOI: 10.1109/tmi.2015.2497906] [Citation(s) in RCA: 15] [Impact Index Per Article: 1.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/05/2023]
Abstract
Statistical shape models (SSMs) have been widely employed in cardiac image segmentation. However, in conditions that induce severe shape abnormality and remodeling, such as in the case of pulmonary hypertension (PH) or hypertrophic cardiomyopathy (HCM), a single SSM is rarely capable of capturing the anatomical variability in the extremes of the distribution. This work presents a new algorithm for the segmentation of severely abnormal hearts. The algorithm is highly flexible, as it does not require a priori knowledge of the involved pathology or any specific parameter tuning to be applied to the cardiac image under analysis. The fundamental idea is to approximate the gross effect of the abnormality with a virtual remodeling transformation between the patient-specific geometry and the average shape of the reference model (e.g., average normal morphology). To define this mapping, a set of landmark points are automatically identified during boundary point search, by estimating the reliability of the candidate points. With the obtained transformation, the feature points extracted from the patient image volume are then projected onto the space of the reference SSM, where the model is used to effectively constrain and guide the segmentation process. The extracted shape in the reference space is finally propagated back to the original image of the abnormal heart to obtain the final segmentation. Detailed validation with patients diagnosed with PH and HCM shows the robustness and flexibility of the technique for the segmentation of highly abnormal hearts of different pathologies.
Collapse
|
15
|
Albà X, Figueras I Ventura RM, Lekadir K, Tobon-Gomez C, Hoogendoorn C, Frangi AF. Automatic cardiac LV segmentation in MRI using modified graph cuts with smoothness and interslice constraints. Magn Reson Med 2013; 72:1775-84. [PMID: 24347347 DOI: 10.1002/mrm.25079] [Citation(s) in RCA: 32] [Impact Index Per Article: 2.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/10/2013] [Revised: 11/11/2013] [Accepted: 11/19/2013] [Indexed: 11/06/2022]
Abstract
PURPOSE Magnetic resonance imaging (MRI), specifically late-enhanced MRI, is the standard clinical imaging protocol to assess cardiac viability. Segmentation of myocardial walls is a prerequisite for this assessment. Automatic and robust multisequence segmentation is required to support processing massive quantities of data. METHODS A generic rule-based framework to automatically segment the left ventricle myocardium is presented here. We use intensity information, and include shape and interslice smoothness constraints, providing robustness to subject- and study-specific changes. Our automatic initialization considers the geometrical and appearance properties of the left ventricle, as well as interslice information. The segmentation algorithm uses a decoupled, modified graph cut approach with control points, providing a good balance between flexibility and robustness. RESULTS The method was evaluated on late-enhanced MRI images from a 20-patient in-house database, and on cine-MRI images from a 15-patient open access database, both using as reference manually delineated contours. Segmentation agreement, measured using the Dice coefficient, was 0.81±0.05 and 0.92±0.04 for late-enhanced MRI and cine-MRI, respectively. The method was also compared favorably to a three-dimensional Active Shape Model approach. CONCLUSION The experimental validation with two magnetic resonance sequences demonstrates increased accuracy and versatility.
Collapse
Affiliation(s)
- Xènia Albà
- Center for Computational Imaging & Simulation Technologies in Biomedicine, Universitat Pompeu Fabra, Barcelona, Spain; Networking Research Center on Bioengineering, Biomaterials and Nanomedicine (CIBER-BBN), Barcelona, Spain
| | | | | | | | | | | |
Collapse
|