1
|
Kumari V, Katiyar A, Bhagawati M, Maindarkar M, Gupta S, Paul S, Chhabra T, Boi A, Tiwari E, Rathore V, Singh IM, Al-Maini M, Anand V, Saba L, Suri JS. Transformer and Attention-Based Architectures for Segmentation of Coronary Arterial Walls in Intravascular Ultrasound: A Narrative Review. Diagnostics (Basel) 2025; 15:848. [PMID: 40218198 PMCID: PMC11988294 DOI: 10.3390/diagnostics15070848] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/05/2025] [Revised: 03/08/2025] [Accepted: 03/20/2025] [Indexed: 04/14/2025] Open
Abstract
Background: The leading global cause of death is coronary artery disease (CAD), necessitating early and precise diagnosis. Intravascular ultrasound (IVUS) is a sophisticated imaging technique that provides detailed visualization of coronary arteries. However, the methods for segmenting walls in the IVUS scan into internal wall structures and quantifying plaque are still evolving. This study explores the use of transformers and attention-based models to improve diagnostic accuracy for wall segmentation in IVUS scans. Thus, the objective is to explore the application of transformer models for wall segmentation in IVUS scans to assess their inherent biases in artificial intelligence systems for improving diagnostic accuracy. Methods: By employing the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) framework, we pinpointed and examined the top strategies for coronary wall segmentation using transformer-based techniques, assessing their traits, scientific soundness, and clinical relevancy. Coronary artery wall thickness is determined by using the boundaries (inner: lumen-intima and outer: media-adventitia) through cross-sectional IVUS scans. Additionally, it is the first to investigate biases in deep learning (DL) systems that are associated with IVUS scan wall segmentation. Finally, the study incorporates explainable AI (XAI) concepts into the DL structure for IVUS scan wall segmentation. Findings: Because of its capacity to automatically extract features at numerous scales in encoders, rebuild segmented pictures via decoders, and fuse variations through skip connections, the UNet and transformer-based model stands out as an efficient technique for segmenting coronary walls in IVUS scans. Conclusions: The investigation underscores a deficiency in incentives for embracing XAI and pruned AI (PAI) models, with no UNet systems attaining a bias-free configuration. Shifting from theoretical study to practical usage is crucial to bolstering clinical evaluation and deployment.
Collapse
Affiliation(s)
- Vandana Kumari
- School of Computer Science and Engineering, Galgotias University, Greater Noida 201310, India; (V.K.); (A.K.)
| | - Alok Katiyar
- School of Computer Science and Engineering, Galgotias University, Greater Noida 201310, India; (V.K.); (A.K.)
| | - Mrinalini Bhagawati
- Department of Biomedical Engineering, North Eastern Hill University, Shillong 793022, India; (M.B.); (S.P.)
| | - Mahesh Maindarkar
- School of Bioengineering Research and Sciences, MIT Art, Design and Technology University, Pune 412021, India;
| | - Siddharth Gupta
- Department of Computer Science and Engineering, Bharati Vidyapeeth’s College of Engineering, New Delhi 110063, India;
- Stroke Monitoring and Diagnostic Division, AtheroPoint™, Roseville, CA 95661, USA; (V.R.); (I.M.S.); (V.A.)
| | - Sudip Paul
- Department of Biomedical Engineering, North Eastern Hill University, Shillong 793022, India; (M.B.); (S.P.)
| | - Tisha Chhabra
- Department of Information Technology, Bharati Vidyapeeth’s College of Engineering, New Delhi 110063, India;
| | - Alberto Boi
- Department of Cardiology, University of Cagliari, 09124 Cagliari, Italy; (A.B.); (L.S.)
| | - Ekta Tiwari
- Department of Computer Science, Visvesvaraya National Institute of Technology (VNIT), Nagpur 440010, India;
| | - Vijay Rathore
- Stroke Monitoring and Diagnostic Division, AtheroPoint™, Roseville, CA 95661, USA; (V.R.); (I.M.S.); (V.A.)
| | - Inder M. Singh
- Stroke Monitoring and Diagnostic Division, AtheroPoint™, Roseville, CA 95661, USA; (V.R.); (I.M.S.); (V.A.)
| | - Mustafa Al-Maini
- Allergy, Clinical Immunology and Rheumatology Institute, Toronto, ON M5G 1N8, Canada;
| | - Vinod Anand
- Stroke Monitoring and Diagnostic Division, AtheroPoint™, Roseville, CA 95661, USA; (V.R.); (I.M.S.); (V.A.)
| | - Luca Saba
- Department of Cardiology, University of Cagliari, 09124 Cagliari, Italy; (A.B.); (L.S.)
| | - Jasjit S. Suri
- Stroke Monitoring and Diagnostic Division, AtheroPoint™, Roseville, CA 95661, USA; (V.R.); (I.M.S.); (V.A.)
- Department of Electrical and Computer Engineering, Idaho State University, Pocatello, ID 83209, USA
- Department of Computer Engineering, Graphic Era Deemed to be University, Dehradun 248002, India
- Symbiosis Institute of Technology, Nagpur Campus, Symbiosis International (Deemed University), Pune 440008, India
- University Centre for Research & Development, Chandigarh University, Mohali 140413, India
| |
Collapse
|
2
|
Zuo J, Simpson DG, O'Brien WD, McFarlin BL, Han A. Automated Field of Interest Determination for Quantitative Ultrasound Analyses of Cervical Tissues: Toward Real-time Clinical Translation in Spontaneous Preterm Birth Risk Assessment. ULTRASOUND IN MEDICINE & BIOLOGY 2024; 50:1861-1867. [PMID: 39271408 PMCID: PMC11490401 DOI: 10.1016/j.ultrasmedbio.2024.08.011] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 05/15/2024] [Revised: 07/30/2024] [Accepted: 08/12/2024] [Indexed: 09/15/2024]
Abstract
OBJECTIVE Quantitative ultrasound (QUS) analysis of the human cervix is valuable for predicting spontaneous preterm birth risk. However, this approach currently requires an offline processing step wherein a medically trained analyst manually draws a free-hand field of interest (Manual FOI) for QUS computation. This offline step hinders the clinical adoption of QUS. To address this challenge, we developed a method to determine automatically the cervical FOI (Auto FOI). This study's objective is to evaluate the agreement between QUS results obtained from the Auto and Manual FOIs and assess the feasibility of using the Auto FOI to replace the Manual FOI for cervical QUS computation. METHODS The auto FOI method was developed and evaluated using cervical ultrasound data from 527 pregnant women, using Manual FOIs as the reference. A deep learning model was developed using the cervical B-mode image as the input to determine automatically the FOI. RESULTS Quantitative comparison between the Auto and Manual FOIs yielded a high pixel accuracy of 97% and a Dice coefficient of 87%. Further, the Auto FOI yielded QUS biomarker values that were highly correlated with those obtained from the Manual FOIs. For example, the Pearson correlation coefficient was 0.87 between attenuation coefficient values obtained using Auto and Manual FOIs. Further, Bland-Altman analyses showed negligible bias between QUS biomarker values computed using the Auto and Manual FOIs. CONCLUSION The results support the feasibility of using Auto FOIs to replace Manual FOIs in QUS computation, an important step toward the clinical adoption of QUS technology.
Collapse
Affiliation(s)
- Jingyi Zuo
- Department of Biomedical Engineering and Mechanics, Virginia Polytechnic Institute and State University, Blacksburg, VA, USA
| | - Douglas G Simpson
- Department of Statistics, University of Illinois Urbana-Champaign, Champaign, IL, USA
| | - William D O'Brien
- Bioacoustics Research Laboratory, Department of Electrical and Computer Engineering, University of Illinois Urbana-Champaign, Urbana, IL, USA
| | - Barbara L McFarlin
- Department of Human Development Nursing Sciences, UIC College of Nursing, University of Illinois Chicago, Chicago, IL, USA
| | - Aiguo Han
- Department of Biomedical Engineering and Mechanics, Virginia Polytechnic Institute and State University, Blacksburg, VA, USA.
| |
Collapse
|
3
|
Awais M, Al Taie M, O’Connor CS, Castelo AH, Acidi B, Tran Cao HS, Brock KK. Enhancing Surgical Guidance: Deep Learning-Based Liver Vessel Segmentation in Real-Time Ultrasound Video Frames. Cancers (Basel) 2024; 16:3674. [PMID: 39518111 PMCID: PMC11545685 DOI: 10.3390/cancers16213674] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/02/2024] [Revised: 10/21/2024] [Accepted: 10/25/2024] [Indexed: 11/16/2024] Open
Abstract
BACKGROUND/OBJECTIVES In the field of surgical medicine, the planning and execution of liver resection procedures present formidable challenges, primarily attributable to the intricate and highly individualized nature of liver vascular anatomy. In the current surgical milieu, intraoperative ultrasonography (IOUS) has become indispensable; however, traditional 2D ultrasound imaging's interpretability is hindered by noise and speckle artifacts. Accurate identification of critical structures for preservation during hepatectomy requires advanced surgical skills. METHODS An AI-based model that can help detect and recognize vessels including the inferior vena cava (IVC); the right (RHV), middle (MHV), and left (LVH) hepatic veins; the portal vein (PV) and its major first and second order branches the left portal vein (LPV), right portal vein (RPV), and right anterior (RAPV) and posterior (RPPV) portal veins, for real-time IOUS navigation can be of immense value in liver surgery. This research aims to advance the capabilities of IOUS-guided interventions by applying an innovative AI-based approach named the "2D-weigthed U-Net model" for the segmentation of multiple blood vessels in real-time IOUS video frames. RESULTS Our proposed deep learning (DL) model achieved a mean Dice score of 0.92 for IVC, 0.90 for RHV, 0.89 for MHV, 0.86 for LHV, 0.95 for PV, 0.93 for LPV, 0.84 for RPV, 0.85 for RAPV, and 0.96 for RPPV. CONCLUSION In the future, this research will be extended for real-time multi-label segmentation of extended vasculature in the liver, followed by the translation of our model into the surgical suite.
Collapse
Affiliation(s)
- Muhammad Awais
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA; (M.A.)
| | - Mais Al Taie
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA; (M.A.)
| | - Caleb S. O’Connor
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA; (M.A.)
| | - Austin H. Castelo
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA; (M.A.)
| | - Belkacem Acidi
- Department of Surgical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA (H.S.T.C.)
| | - Hop S. Tran Cao
- Department of Surgical Oncology, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA (H.S.T.C.)
| | - Kristy K. Brock
- Department of Imaging Physics, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA; (M.A.)
- Department the of Radiation Physics, The University of Texas MD Anderson Cancer Center, Houston, TX 77030, USA
| |
Collapse
|
4
|
Brooks JA, Kallenbach M, Radu IP, Berzigotti A, Dietrich CF, Kather JN, Luedde T, Seraphin TP. Artificial Intelligence for Contrast-Enhanced Ultrasound of the Liver: A Systematic Review. Digestion 2024:1-18. [PMID: 39312896 DOI: 10.1159/000541540] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 06/04/2024] [Accepted: 09/18/2024] [Indexed: 09/25/2024]
Abstract
INTRODUCTION The research field of artificial intelligence (AI) in medicine and especially in gastroenterology is rapidly progressing with the first AI tools entering routine clinical practice, for example, in colorectal cancer screening. Contrast-enhanced ultrasound (CEUS) is a highly reliable, low-risk, and low-cost diagnostic modality for the examination of the liver. However, doctors need many years of training and experience to master this technique and, despite all efforts to standardize CEUS, it is often believed to contain significant interrater variability. As has been shown for endoscopy, AI holds promise to support examiners at all training levels in their decision-making and efficiency. METHODS In this systematic review, we analyzed and compared original research studies applying AI methods to CEUS examinations of the liver published between January 2010 and February 2024. We performed a structured literature search on PubMed, Web of Science, and IEEE. Two independent reviewers screened the articles and subsequently extracted relevant methodological features, e.g., cohort size, validation process, machine learning algorithm used, and indicative performance measures from the included articles. RESULTS We included 41 studies with most applying AI methods for classification tasks related to focal liver lesions. These included distinguishing benign versus malignant or classifying the entity itself, while a few studies tried to classify tumor grading, microvascular invasion status, or response to transcatheter arterial chemoembolization directly from CEUS. Some articles tried to segment or detect focal liver lesions, while others aimed to predict survival and recurrence after ablation. The majority (25/41) of studies used hand-picked and/or annotated images as data input to their models. We observed mostly good to high reported model performances with accuracies ranging between 58.6% and 98.9%, while noticing a general lack of external validation. CONCLUSION Even though multiple proof-of-concept studies for the application of AI methods to CEUS examinations of the liver exist and report high performance, more prospective, externally validated, and multicenter research is needed to bring such algorithms from desk to bedside.
Collapse
Affiliation(s)
- James A Brooks
- Department of Gastroenterology, Hepatology and Infectious Diseases, University Hospital Dusseldorf, Medical Faculty at Heinrich-Heine-University, Dusseldorf, Germany
| | - Michael Kallenbach
- Department of Gastroenterology, Hepatology and Infectious Diseases, University Hospital Dusseldorf, Medical Faculty at Heinrich-Heine-University, Dusseldorf, Germany
| | - Iuliana-Pompilia Radu
- Department for Visceral Surgery and Medicine, Inselspital, University of Bern, Bern, Switzerland
| | - Annalisa Berzigotti
- Department for Visceral Surgery and Medicine, Inselspital, University of Bern, Bern, Switzerland
| | - Christoph F Dietrich
- Department Allgemeine Innere Medizin (DAIM), Kliniken Hirslanden Beau Site, Salem and Permanence, Bern, Switzerland
| | - Jakob N Kather
- Else Kroener Fresenius Center for Digital Health, Medical Faculty Carl Gustav Carus, Technical University Dresden, Dresden, Germany
| | - Tom Luedde
- Department of Gastroenterology, Hepatology and Infectious Diseases, University Hospital Dusseldorf, Medical Faculty at Heinrich-Heine-University, Dusseldorf, Germany
| | - Tobias P Seraphin
- Department of Gastroenterology, Hepatology and Infectious Diseases, University Hospital Dusseldorf, Medical Faculty at Heinrich-Heine-University, Dusseldorf, Germany
| |
Collapse
|
5
|
Wu J, Liu F, Sun W, Liu Z, Hou H, Jiang R, Hu H, Ren P, Zhang R, Zhang X. Boundary-aware convolutional attention network for liver segmentation in ultrasound images. Sci Rep 2024; 14:21529. [PMID: 39278955 PMCID: PMC11403006 DOI: 10.1038/s41598-024-70527-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/15/2024] [Accepted: 08/19/2024] [Indexed: 09/18/2024] Open
Abstract
Liver ultrasound is widely used in clinical practice due to its advantages of non-invasiveness, non-radiation, and real-time imaging. Accurate segmentation of the liver region in ultrasound images is essential for accelerating the auxiliary diagnosis of liver-related diseases. This paper proposes BACANet, a deep learning algorithm designed for real-time liver ultrasound segmentation. Our approach utilizes a lightweight network backbone for liver feature extraction and incorporates a convolutional attention mechanism to enhance the network's ability to capture global contextual information. To improve early localization of liver boundaries, we developed a selective large kernel convolution module for boundary feature extraction and introduced explicit liver boundary supervision. Additionally, we designed an enhanced attention gate to efficiently convey liver body and boundary features to the decoder to enhance the feature representation capability. Experimental results across multiple datasets demonstrate that BACANet effectively completes the task of liver ultrasound segmentation, achieving a balance between inference speed and segmentation accuracy. On a public dataset, BACANet achieved a DSC of 0.921 and an IOU of 0.854. On a private test dataset, BACANet achieved a DSC of 0.950 and an IOU of 0.907, with an inference time of approximately 0.32 s per image on a CPU processor.
Collapse
Affiliation(s)
- Jiawei Wu
- School of Medical Informatics and Engineering, Xuzhou Medical University, Xuzhou, 221000, China
| | - Fulong Liu
- School of Medical Informatics and Engineering, Xuzhou Medical University, Xuzhou, 221000, China
| | - Weiqin Sun
- School of Medical Informatics and Engineering, Xuzhou Medical University, Xuzhou, 221000, China
| | - Zhipeng Liu
- Department of Information, Taizhou People's Hospital Affiliated to Nanjing Medical University, Taizhou, 225300, China
| | - Hui Hou
- Department of Imaging, The Fourth People's Hospital of Taizhou in Jiangsu Province, Taizhou, 225300, China
| | - Rui Jiang
- School of Medical Informatics and Engineering, Xuzhou Medical University, Xuzhou, 221000, China
| | - Haowei Hu
- School of Medical Informatics and Engineering, Xuzhou Medical University, Xuzhou, 221000, China
| | - Peng Ren
- School of Medical Informatics and Engineering, Xuzhou Medical University, Xuzhou, 221000, China
| | - Ran Zhang
- School of Medical Informatics and Engineering, Xuzhou Medical University, Xuzhou, 221000, China
| | - Xiao Zhang
- School of Medical Informatics and Engineering, Xuzhou Medical University, Xuzhou, 221000, China.
- Yantai Longch Technologies Co., Ltd, Yantai, 264000, China.
| |
Collapse
|
6
|
Chierici A, Lareyre F, Salucki B, Iannelli A, Delingette H, Raffort J. Vascular liver segmentation: a narrative review on methods and new insights brought by artificial intelligence. J Int Med Res 2024; 52:3000605241263170. [PMID: 39291427 PMCID: PMC11418557 DOI: 10.1177/03000605241263170] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/21/2024] [Accepted: 05/28/2024] [Indexed: 09/19/2024] Open
Abstract
Liver vessel segmentation from routinely performed medical imaging is a useful tool for diagnosis, treatment planning and delivery, and prognosis evaluation for many diseases, particularly liver cancer. A precise representation of liver anatomy is crucial to define the extent of the disease and, when suitable, the consequent resective or ablative procedure, in order to guarantee a radical treatment without sacrificing an excessive volume of healthy liver. Once mainly performed manually, with notable cost in terms of time and human energies, vessel segmentation is currently realized through the application of artificial intelligence (AI), which has gained increased interest and development of the field. Many different AI-driven models adopted for this aim have been described and can be grouped into different categories: thresholding methods, edge- and region-based methods, model-based methods, and machine learning models. The latter includes neural network and deep learning models that now represent the principal algorithms exploited for vessel segmentation. The present narrative review describes how liver vessel segmentation can be realized through AI models, with a summary of model results in terms of accuracy, and an overview on the future progress of this topic.
Collapse
Affiliation(s)
- Andrea Chierici
- Department of Digestive Surgery, Hospital of Antibes Juan-les-Pins, Antibes, France
- Department of Digestive Surgery, University Hospital of Nice, Nice, France
- Université Côte d'Azur, Inria, Epione Team, Sophia Antipolis, France
| | - Fabien Lareyre
- Department of Vascular Surgery, Hospital of Antibes Juan-les-Pins, Antibes, France
- Université Côte d'Azur, Inserm U1065, C3M, Nice, France
| | - Benjamin Salucki
- Department of Digestive Surgery, Hospital of Antibes Juan-les-Pins, Antibes, France
| | - Antonio Iannelli
- Université Côte d'Azur, Inserm U1065, Team 8 “Hepatic complications of obesity and alcohol”, Nice, France
- ADIPOCIBLE Study Group, Université Côte d'Azur, Nice, France
| | - Hervé Delingette
- Université Côte d'Azur, Inria, Epione Team, Sophia Antipolis, France
| | - Juliette Raffort
- Université Côte d'Azur, Inserm U1065, C3M, Nice, France
- Clinical Chemistry Laboratory, University Hospital of Nice, Nice, France
- 3IA Institute, Université Côte d'Azur, Nice, France
| |
Collapse
|
7
|
Boneš E, Gergolet M, Bohak C, Lesar Ž, Marolt M. Automatic Segmentation and Alignment of Uterine Shapes from 3D Ultrasound Data. Comput Biol Med 2024; 178:108794. [PMID: 38941903 DOI: 10.1016/j.compbiomed.2024.108794] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/31/2023] [Revised: 06/18/2024] [Accepted: 06/19/2024] [Indexed: 06/30/2024]
Abstract
BACKGROUND The uterus is the most important organ in the female reproductive system. Its shape plays a critical role in fertility and pregnancy outcomes. Advances in medical imaging, such as 3D ultrasound, have significantly improved the exploration of the female genital tract, thereby enhancing gynecological healthcare. Despite well-documented data for organs like the liver and heart, large-scale studies on the uterus are lacking. Existing classifications, such as VCUAM and ESHRE/ESGE, provide different definitions for normal uterine shapes but are not based on real-world measurements. Moreover, the lack of comprehensive datasets significantly hinders research in this area. Our research, part of the larger NURSE study, aims to fill this gap by establishing the shape of a normal uterus using real-world 3D vaginal ultrasound scans. This will facilitate research into uterine shape abnormalities associated with infertility and recurrent miscarriages. METHODS We developed an automated system for the segmentation and alignment of uterine shapes from 3D ultrasound data, which consists of two steps: automatic segmentation of the uteri in 3D ultrasound scans using deep learning techniques, and alignment of the resulting shapes with standard geometrical approaches, enabling the extraction of the normal shape for future analysis. The system was trained and validated on a comprehensive dataset of 3D ultrasound images from multiple medical centers. Its performance was evaluated by comparing the automated results with manual annotations provided by expert clinicians. RESULTS The presented approach demonstrated high accuracy in segmenting and aligning uterine shapes from 3D ultrasound data. The segmentation achieved an average Dice similarity coefficient (DSC) of 0.90. Our method for aligning uterine shapes showed minimal translation and rotation errors compared to traditional methods, with the preliminary average shape exhibiting characteristics consistent with expert findings of a normal uterus. CONCLUSION We have presented an approach to automatically segment and align uterine shapes from 3D ultrasound data. We trained a deep learning nnU-Net model that achieved high accuracy and proposed an alignment method using a combination of standard geometrical techniques. Additionally, we have created a publicly available dataset of 3D transvaginal ultrasound volumes with manual annotations of uterine cavities to support further research and development in this field. The dataset and the trained models are available at https://github.com/UL-FRI-LGM/UterUS.
Collapse
Affiliation(s)
- Eva Boneš
- University of Ljubljana, Faculty of Computer and Information Science, Večna pot 113, Ljubljana, 1000, Slovenia.
| | - Marco Gergolet
- University of Ljubljana, Faculty of Medicine, Vrazov trg 2, Ljubljana, 1000, Slovenia.
| | - Ciril Bohak
- University of Ljubljana, Faculty of Computer and Information Science, Večna pot 113, Ljubljana, 1000, Slovenia; King Abdullah University of Science and Technology, Visual Computing Center, Thuwal, 23955-6900, Saudi Arabia.
| | - Žiga Lesar
- University of Ljubljana, Faculty of Computer and Information Science, Večna pot 113, Ljubljana, 1000, Slovenia.
| | - Matija Marolt
- University of Ljubljana, Faculty of Computer and Information Science, Večna pot 113, Ljubljana, 1000, Slovenia.
| |
Collapse
|
8
|
Yue Y, Li N, Zhang G, Xing W, Zhu Z, Liu X, Song S, Ta D. A transformer-guided cross-modality adaptive feature fusion framework for esophageal gross tumor volume segmentation. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 251:108216. [PMID: 38761412 DOI: 10.1016/j.cmpb.2024.108216] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 11/01/2023] [Revised: 04/17/2024] [Accepted: 05/06/2024] [Indexed: 05/20/2024]
Abstract
BACKGROUND AND OBJECTIVE Accurate segmentation of esophageal gross tumor volume (GTV) indirectly enhances the efficacy of radiotherapy for patients with esophagus cancer. In this domain, learning-based methods have been employed to fuse cross-modality positron emission tomography (PET) and computed tomography (CT) images, aiming to improve segmentation accuracy. This fusion is essential as it combines functional metabolic information from PET with anatomical information from CT, providing complementary information. While the existing three-dimensional (3D) segmentation method has achieved state-of-the-art (SOTA) performance, it typically relies on pure-convolution architectures, limiting its ability to capture long-range spatial dependencies due to convolution's confinement to a local receptive field. To address this limitation and further enhance esophageal GTV segmentation performance, this work proposes a transformer-guided cross-modality adaptive feature fusion network, referred to as TransAttPSNN, which is based on cross-modality PET/CT scans. METHODS Specifically, we establish an attention progressive semantically-nested network (AttPSNN) by incorporating the convolutional attention mechanism into the progressive semantically-nested network (PSNN). Subsequently, we devise a plug-and-play transformer-guided cross-modality adaptive feature fusion model, which is inserted between the multi-scale feature counterparts of a two-stream AttPSNN backbone (one for the PET modality flow and another for the CT modality flow), resulting in the proposed TransAttPSNN architecture. RESULTS Through extensive four-fold cross-validation experiments on the clinical PET/CT cohort. The proposed approach acquires a Dice similarity coefficient (DSC) of 0.76 ± 0.13, a Hausdorff distance (HD) of 9.38 ± 8.76 mm, and a Mean surface distance (MSD) of 1.13 ± 0.94 mm, outperforming the SOTA competing methods. The qualitative results show a satisfying consistency with the lesion areas. CONCLUSIONS The devised transformer-guided cross-modality adaptive feature fusion module integrates the strengths of PET and CT, effectively enhancing the segmentation performance of esophageal GTV. The proposed TransAttPSNN has further advanced the research of esophageal GTV segmentation.
Collapse
Affiliation(s)
- Yaoting Yue
- Department of Biomedical Engineering, School of Information Science and Technology, Fudan University, Shanghai 200438, PR China
| | - Nan Li
- Department of Nuclear Medicine, Fudan University Shanghai Cancer Center, Shanghai 201321, PR China
| | - Gaobo Zhang
- Department of Biomedical Engineering, School of Information Science and Technology, Fudan University, Shanghai 200438, PR China
| | - Wenyu Xing
- Academy for Engineering and Technology, Fudan University, Shanghai 200433, PR China
| | - Zhibin Zhu
- Department of Biomedical Engineering, School of Information Science and Technology, Fudan University, Shanghai 200438, PR China; School of Physics and Electromechanical Engineering, Hexi University, Zhangye 734000, Gansu, PR China
| | - Xin Liu
- Academy for Engineering and Technology, Fudan University, Shanghai 200433, PR China.
| | - Shaoli Song
- Department of Nuclear Medicine, Fudan University Shanghai Cancer Center, Shanghai 201321, PR China.
| | - Dean Ta
- Department of Biomedical Engineering, School of Information Science and Technology, Fudan University, Shanghai 200438, PR China; Academy for Engineering and Technology, Fudan University, Shanghai 200433, PR China.
| |
Collapse
|
9
|
Hu J, Cui Z, Zhang X, Zhang J, Ge Y, Zhang H, Lu Y, Shen D. Uncertainty-aware refinement framework for ovarian tumor segmentation in CECT volume. Med Phys 2024; 51:2678-2694. [PMID: 37862556 DOI: 10.1002/mp.16795] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/18/2023] [Revised: 09/05/2023] [Accepted: 09/26/2023] [Indexed: 10/22/2023] Open
Abstract
BACKGROUND Ovarian cancer is a highly lethal gynecological disease. Accurate and automated segmentation of ovarian tumors in contrast-enhanced computed tomography (CECT) images is crucial in the radiotherapy treatment of ovarian cancer, enabling radiologists to evaluate cancer progression and develop timely therapeutic plans. However, automatic ovarian tumor segmentation is challenging due to factors such as inhomogeneous background, ambiguous tumor boundaries, and imbalanced foreground-background, all of which contribute to high predictive uncertainty for a segmentation model. PURPOSE To tackle these challenges, we propose an uncertainty-aware refinement framework that aims to estimate and refine regions with high predictive uncertainty for accurate ovarian tumor segmentation in CECT images. METHODS To this end, we first employ an approximate Bayesian network to detect coarse regions of interest (ROIs) of both ovarian tumors and uncertain regions. These ROIs allow a subsequent segmentation network to narrow down the search area for tumors and prioritize uncertain regions, resulting in precise segmentation of ovarian tumors. Meanwhile, the framework integrates two guidance modules that learn two implicit functions capable of mapping query features sampled according to their uncertainty to organ or boundary manifolds, guiding the segmentation network to facilitate information encoding of uncertain regions. RESULTS Firstly, 367 CECT images are collected from the same hospital for experiments. Dice score, Jaccard, Recall, Positive predictive value (PPV), 95% Hausdorff distance (HD95) and Average symmetric surface distance (ASSD) for the testing group of 77 cases are 86.31%, 73.93%, 83.95%, 86.03%, 15.17 mm and 2.57 mm, all of which are significantly better than that of the other state-of-the-art models. And results of visual comparison shows that the compared methods have more mis-segmentation than our method. Furthermore, our method achieves a Dice score that is at least 20% higher than the Dice scores of other compared methods when tumor volumes are less than 20 cm3 $^3$ , indicating better recognition ability to small regions by our method. And then, 38 CECT images are collected from another hospital to form an external testing group. Our approach consistently outperform the compared methods significantly, with the external testing group exhibiting substantial improvements across key evaluation metrics: Dice score (83.74%), Jaccard (69.55%), Recall (82.12%), PPV (81.61%), HD95 (12.31 mm), and ASSD (2.32 mm), robustly establishing its superior performance. CONCLUSIONS Experimental results demonstrate that the framework significantly outperforms the compared state-of-the-art methods, with decreased under- or over-segmentation and better small tumor identification. It has the potential for clinical application.
Collapse
Affiliation(s)
- Jiaqi Hu
- Zhejiang Provincial Key Laboratory of Precision Diagnosis and Therapy for Major Gynecological Diseases, Department of Gynecologic Oncology, Women's Hospital and Institute of Translational Medicine, Zhejiang University School of Medicine, Hangzhou, Zhejiang, China
| | - Zhiming Cui
- School of Biomedical Engineering & State Key Laboratory of Advanced Medical Materials and Devices, ShanghaiTech University, Shanghai, China
| | - Xiao Zhang
- School of Biomedical Engineering & State Key Laboratory of Advanced Medical Materials and Devices, ShanghaiTech University, Shanghai, China
- School of Information Science and Technology, Northwest University, Xi'an, China
| | - Jiadong Zhang
- School of Biomedical Engineering & State Key Laboratory of Advanced Medical Materials and Devices, ShanghaiTech University, Shanghai, China
| | - Yuyan Ge
- School of Biomedical Engineering & State Key Laboratory of Advanced Medical Materials and Devices, ShanghaiTech University, Shanghai, China
- Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, China
| | - Honghe Zhang
- Department of Pathology, Research Unit of Intelligence Classification of Tumor Pathology and Precision Therapy, Chinese Academy of Medical Sciences, Zhejiang University School of Medicine, Hangzhou, Zhejiang, China
| | - Yan Lu
- Zhejiang Provincial Key Laboratory of Precision Diagnosis and Therapy for Major Gynecological Diseases, Department of Gynecologic Oncology, Women's Hospital and Institute of Translational Medicine, Zhejiang University School of Medicine, Hangzhou, Zhejiang, China
- Cancer center, Zhejiang University, Hangzhou, Zhejiang, China
| | - Dinggang Shen
- School of Biomedical Engineering & State Key Laboratory of Advanced Medical Materials and Devices, ShanghaiTech University, Shanghai, China
- Shanghai United Imaging Intelligence Co., Ltd. Shanghai, Shanghai, China
- Shanghai Clinical Research and Trial Center, Shanghai, China
| |
Collapse
|
10
|
Lee HS, Park JH, Lee SJ. Artificial intelligence-based speckle featurization and localization for ultrasound speckle tracking velocimetry. ULTRASONICS 2024; 138:107241. [PMID: 38232448 DOI: 10.1016/j.ultras.2024.107241] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 09/15/2023] [Revised: 12/25/2023] [Accepted: 01/02/2024] [Indexed: 01/19/2024]
Abstract
Deep learning-based super-resolution ultrasound (DL-SRU) framework has been successful in improving spatial resolution and measuring the velocity field information of a blood flows by localizing and tracking speckle signals of red blood cells (RBCs) without using any contrast agents. However, DL-SRU can localize only a small part of the speckle signals of blood flow owing to ambiguity problems encountered in the classification of blood flow signals from ultrasound B-mode images and the building up of suitable datasets required for training artificial neural networks, as well as the structural limitations of the neural network itself. An artificial intelligence-based speckle featurization and localization (AI-SFL) framework is proposed in this study. It includes a machine learning-based algorithm for classifying blood flow signals from ultrasound B-mode images, dimensionality reduction for featurizing speckle patterns of the classified blood flow signals by approximating them with quantitative values. A novel and robust neural network (ResSU-net) is trained using the online data generation (ODG) method and the extracted speckle features. The super-resolution performance of the proposed AI-SFL and ODG method is evaluated and compared with the results of previous U-net and conventional data augmentation methods under in silico conditions. The predicted locations of RBCs by the AI-SFL and DL-SRU for speckle patterns of blood flow are applied to a PTV algorithm to measure quantitative velocity fields of the flow. Finally, the feasibility of the proposed AI-SFL framework for measuring real blood flows is verified under in vivo conditions.
Collapse
Affiliation(s)
- Hyo Seung Lee
- Department of Mechanical Engineering, Pohang University of Science and Technology, 77 Cheongam-Ro, Nam-Gu, Pohang, Gyeongbuk, South Korea.
| | - Jun Hong Park
- Department of Radiology, Stanford University 450 Jane Stanford Way Stanford, CA 94305-2004, United States.
| | - Sang Joon Lee
- Department of Mechanical Engineering, Pohang University of Science and Technology, 77 Cheongam-Ro, Nam-Gu, Pohang, Gyeongbuk, South Korea.
| |
Collapse
|
11
|
Kumari V, Kumar N, Kumar K S, Kumar A, Skandha SS, Saxena S, Khanna NN, Laird JR, Singh N, Fouda MM, Saba L, Singh R, Suri JS. Deep Learning Paradigm and Its Bias for Coronary Artery Wall Segmentation in Intravascular Ultrasound Scans: A Closer Look. J Cardiovasc Dev Dis 2023; 10:485. [PMID: 38132653 PMCID: PMC10743870 DOI: 10.3390/jcdd10120485] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/27/2023] [Revised: 10/15/2023] [Accepted: 11/07/2023] [Indexed: 12/23/2023] Open
Abstract
BACKGROUND AND MOTIVATION Coronary artery disease (CAD) has the highest mortality rate; therefore, its diagnosis is vital. Intravascular ultrasound (IVUS) is a high-resolution imaging solution that can image coronary arteries, but the diagnosis software via wall segmentation and quantification has been evolving. In this study, a deep learning (DL) paradigm was explored along with its bias. METHODS Using a PRISMA model, 145 best UNet-based and non-UNet-based methods for wall segmentation were selected and analyzed for their characteristics and scientific and clinical validation. This study computed the coronary wall thickness by estimating the inner and outer borders of the coronary artery IVUS cross-sectional scans. Further, the review explored the bias in the DL system for the first time when it comes to wall segmentation in IVUS scans. Three bias methods, namely (i) ranking, (ii) radial, and (iii) regional area, were applied and compared using a Venn diagram. Finally, the study presented explainable AI (XAI) paradigms in the DL framework. FINDINGS AND CONCLUSIONS UNet provides a powerful paradigm for the segmentation of coronary walls in IVUS scans due to its ability to extract automated features at different scales in encoders, reconstruct the segmented image using decoders, and embed the variants in skip connections. Most of the research was hampered by a lack of motivation for XAI and pruned AI (PAI) models. None of the UNet models met the criteria for bias-free design. For clinical assessment and settings, it is necessary to move from a paper-to-practice approach.
Collapse
Affiliation(s)
- Vandana Kumari
- School of Computer Science and Engineering, Galgotias University, Greater Noida 201310, India; (V.K.); (S.K.K.)
| | - Naresh Kumar
- Department of Applied Computational Science and Engineering, G L Bajaj Institute of Technology and Management, Greater Noida 201310, India
| | - Sampath Kumar K
- School of Computer Science and Engineering, Galgotias University, Greater Noida 201310, India; (V.K.); (S.K.K.)
| | - Ashish Kumar
- School of CSET, Bennett University, Greater Noida 201310, India;
| | - Sanagala S. Skandha
- Department of CSE, CMR College of Engineering and Technology, Hyderabad 501401, India;
| | - Sanjay Saxena
- Department of Computer Science and Engineering, IIT Bhubaneswar, Bhubaneswar 751003, India;
| | - Narendra N. Khanna
- Department of Cardiology, Indraprastha APOLLO Hospitals, New Delhi 110076, India;
| | - John R. Laird
- Heart and Vascular Institute, Adventist Health St. Helena, St Helena, CA 94574, USA;
| | - Narpinder Singh
- Department of Food Science and Technology, Graphic Era, Deemed to be University, Dehradun 248002, India;
| | - Mostafa M. Fouda
- Department of Electrical and Computer Engineering, Idaho State University, Pocatello, ID 83209, USA;
| | - Luca Saba
- Department of Radiology, Azienda Ospedaliero Universitaria (A.O.U.), 09100 Cagliari, Italy;
| | - Rajesh Singh
- Department of Research and Innovation, Uttaranchal Institute of Technology, Uttaranchal University, Dehradun 248007, India;
| | - Jasjit S. Suri
- Stroke Diagnostics and Monitoring Division, AtheroPoint™, Roseville, CA 95661, USA
- Department of Computer Science & Engineering, Graphic Era, Deemed to be University, Dehradun 248002, India
- Monitoring and Diagnosis Division, AtheroPoint™, Roseville, CA 95661, USA
| |
Collapse
|
12
|
Torfeh T, Aouadi S, Yoganathan SA, Paloor S, Hammoud R, Al-Hammadi N. Deep Learning Approaches for Automatic Quality Assurance of Magnetic Resonance Images Using ACR Phantom. BMC Med Imaging 2023; 23:197. [PMID: 38031032 PMCID: PMC10685462 DOI: 10.1186/s12880-023-01157-5] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/04/2023] [Accepted: 11/17/2023] [Indexed: 12/01/2023] Open
Abstract
BACKGROUND In recent years, there has been a growing trend towards utilizing Artificial Intelligence (AI) and machine learning techniques in medical imaging, including for the purpose of automating quality assurance. In this research, we aimed to develop and evaluate various deep learning-based approaches for automatic quality assurance of Magnetic Resonance (MR) images using the American College of Radiology (ACR) standards. METHODS The study involved the development, optimization, and testing of custom convolutional neural network (CNN) models. Additionally, popular pre-trained models such as VGG16, VGG19, ResNet50, InceptionV3, EfficientNetB0, and EfficientNetB5 were trained and tested. The use of pre-trained models, particularly those trained on the ImageNet dataset, for transfer learning was also explored. Two-class classification models were employed for assessing spatial resolution and geometric distortion, while an approach classifying the image into 10 classes representing the number of visible spokes was used for the low contrast. RESULTS Our results showed that deep learning-based methods can be effectively used for MR image quality assurance and can improve the performance of these models. The low contrast test was one of the most challenging tests within the ACR phantom. CONCLUSIONS Overall, for geometric distortion and spatial resolution, all of the deep learning models tested produced prediction accuracy of 80% or higher. The study also revealed that training the models from scratch performed slightly better compared to transfer learning. For the low contrast, our investigation emphasized the adaptability and potential of deep learning models. The custom CNN models excelled in predicting the number of visible spokes, achieving commendable accuracy, recall, precision, and F1 scores.
Collapse
Affiliation(s)
- Tarraf Torfeh
- Department of Radiation Oncology, National Center for Cancer Care & Research (NCCCR), Hamad Medical Corporation, Doha, Qatar.
| | - Souha Aouadi
- Department of Radiation Oncology, National Center for Cancer Care & Research (NCCCR), Hamad Medical Corporation, Doha, Qatar
| | - S A Yoganathan
- Department of Radiation Oncology, National Center for Cancer Care & Research (NCCCR), Hamad Medical Corporation, Doha, Qatar
| | - Satheesh Paloor
- Department of Radiation Oncology, National Center for Cancer Care & Research (NCCCR), Hamad Medical Corporation, Doha, Qatar
| | - Rabih Hammoud
- Department of Radiation Oncology, National Center for Cancer Care & Research (NCCCR), Hamad Medical Corporation, Doha, Qatar
| | - Noora Al-Hammadi
- Department of Radiation Oncology, National Center for Cancer Care & Research (NCCCR), Hamad Medical Corporation, Doha, Qatar
| |
Collapse
|
13
|
Gao J, Lao Q, Liu P, Yi H, Kang Q, Jiang Z, Wu X, Li K, Chen Y, Zhang L. Anatomically Guided Cross-Domain Repair and Screening for Ultrasound Fetal Biometry. IEEE J Biomed Health Inform 2023; 27:4914-4925. [PMID: 37486830 DOI: 10.1109/jbhi.2023.3298096] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/26/2023]
Abstract
Ultrasound based estimation of fetal biometry is extensively used to diagnose prenatal abnormalities and to monitor fetal growth, for which accurate segmentation of the fetal anatomy is a crucial prerequisite. Although deep neural network-based models have achieved encouraging results on this task, inevitable distribution shifts in ultrasound images can still result in severe performance drop in real world deployment scenarios. In this article, we propose a complete ultrasound fetal examination system to deal with this troublesome problem by repairing and screening the anatomically implausible results. Our system consists of three main components: A routine segmentation network, a fetal anatomical key points guided repair network, and a shape-coding based selective screener. Guided by the anatomical key points, our repair network has stronger cross-domain repair capabilities, which can substantially improve the outputs of the segmentation network. By quantifying the distance between an arbitrary segmentation mask to its corresponding anatomical shape class, the proposed shape-coding based selective screener can then effectively reject the entire implausible results that cannot be fully repaired. Extensive experiments demonstrate that our proposed framework has strong anatomical guarantee and outperforms other methods in three different cross-domain scenarios.
Collapse
|
14
|
Song H, Liu C, Li S, Zhang P. TS-GCN: A novel tumor segmentation method integrating transformer and GCN. MATHEMATICAL BIOSCIENCES AND ENGINEERING : MBE 2023; 20:18173-18190. [PMID: 38052553 DOI: 10.3934/mbe.2023807] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/07/2023]
Abstract
As one of the critical branches of medical image processing, the task of segmentation of breast cancer tumors is of great importance for planning surgical interventions, radiotherapy and chemotherapy. Breast cancer tumor segmentation faces several challenges, including the inherent complexity and heterogeneity of breast tissue, the presence of various imaging artifacts and noise in medical images, low contrast between the tumor region and healthy tissue, and inconsistent size of the tumor region. Furthermore, the existing segmentation methods may not fully capture the rich spatial and contextual information in small-sized regions in breast images, leading to suboptimal performance. In this paper, we propose a novel breast tumor segmentation method, called the transformer and graph convolutional neural (TS-GCN) network, for medical imaging analysis. Specifically, we designed a feature aggregation network to fuse the features extracted from the transformer, GCN and convolutional neural network (CNN) networks. The CNN extract network is designed for the image's local deep feature, and the transformer and GCN networks can better capture the spatial and context dependencies among pixels in images. By leveraging the strengths of three feature extraction networks, our method achieved superior segmentation performance on the BUSI dataset and dataset B. The TS-GCN showed the best performance on several indexes, with Acc of 0.9373, Dice of 0.9058, IoU of 0.7634, F1 score of 0.9338, and AUC of 0.9692, which outperforms other state-of-the-art methods. The research of this segmentation method provides a promising future for medical image analysis and diagnosis of other diseases.
Collapse
Affiliation(s)
- Haiyan Song
- The Second Affiliated Hospital of Shandong University of Traditional Chinese Medicine, Jinan, China
| | - Cuihong Liu
- Affiliated Eye Hospital of Shandong University of Traditional Chinese Medicine, Jinan, China
- School of Nursing, Shandong University of Traditional Chinese Medicine, Jinan, China
| | - Shengnan Li
- The Second Affiliated Hospital of Shandong University of Traditional Chinese Medicine, Jinan, China
| | - Peixiao Zhang
- The Second Affiliated Hospital of Shandong University of Traditional Chinese Medicine, Jinan, China
| |
Collapse
|
15
|
Peng T, Wu Y, Gu Y, Xu D, Wang C, Li Q, Cai J. Intelligent contour extraction approach for accurate segmentation of medical ultrasound images. Front Physiol 2023; 14:1177351. [PMID: 37675280 PMCID: PMC10479019 DOI: 10.3389/fphys.2023.1177351] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/03/2023] [Accepted: 07/28/2023] [Indexed: 09/08/2023] Open
Abstract
Introduction: Accurate contour extraction in ultrasound images is of great interest for image-guided organ interventions and disease diagnosis. Nevertheless, it remains a problematic issue owing to the missing or ambiguous outline between organs (i.e., prostate and kidney) and surrounding tissues, the appearance of shadow artifacts, and the large variability in the shape of organs. Methods: To address these issues, we devised a method that includes four stages. In the first stage, the data sequence is acquired using an improved adaptive selection principal curve method, in which a limited number of radiologist defined data points are adopted as the prior. The second stage then uses an enhanced quantum evolution network to help acquire the optimal neural network. The third stage involves increasing the precision of the experimental outcomes after training the neural network, while using the data sequence as the input. In the final stage, the contour is smoothed using an explicable mathematical formula explained by the model parameters of the neural network. Results: Our experiments showed that our approach outperformed other current methods, including hybrid and Transformer-based deep-learning methods, achieving an average Dice similarity coefficient, Jaccard similarity coefficient, and accuracy of 95.7 ± 2.4%, 94.6 ± 2.6%, and 95.3 ± 2.6%, respectively. Discussion: This work develops an intelligent contour extraction approach on ultrasound images. Our approach obtained more satisfactory outcome compared with recent state-of-the-art approaches . The knowledge of precise boundaries of the organ is significant for the conservation of risk structures. Our developed approach has the potential to enhance disease diagnosis and therapeutic outcomes.
Collapse
Affiliation(s)
- Tao Peng
- School of Future Science and Engineering, Soochow University, Suzhou, China
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Kowloon, Hong Kong SAR, China
- Department of Radiation Oncology, UT Southwestern Medical Center, Dallas, TX, United States
| | - Yiyun Wu
- Department of Ultrasound, Jiangsu Province Hospital of Chinese Medicine, Nanjing, Jiangsu, China
| | - Yidong Gu
- Department of Medical Ultrasound, The Affiliated Suzhou Hospital of Nanjing Medical University, Suzhou Municipal Hospital, Suzhou, Jiangsu, China
| | - Daqiang Xu
- Department of Radiology, The Affiliated Suzhou Hospital of Nanjing Medical University, Suzhou Municipal Hospital, Suzhou, Jiangsu, China
| | - Caishan Wang
- Department of Ultrasound, The Second Affiliated Hospital of Soochow University, Suzhou, China
| | - Quan Li
- Center of Stomatology, The Second Affiliated Hospital of Soochow University, Suzhou, China
| | - Jing Cai
- Department of Health Technology and Informatics, The Hong Kong Polytechnic University, Kowloon, Hong Kong SAR, China
| |
Collapse
|
16
|
Xin C, Li B, Wang D, Chen W, Yue S, Meng D, Qiao X, Zhang Y. Deep learning for the rapid automatic segmentation of forearm muscle boundaries from ultrasound datasets. Front Physiol 2023; 14:1166061. [PMID: 37520832 PMCID: PMC10374344 DOI: 10.3389/fphys.2023.1166061] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/18/2023] [Accepted: 06/28/2023] [Indexed: 08/01/2023] Open
Abstract
Ultrasound (US) is widely used in the clinical diagnosis and treatment of musculoskeletal diseases. However, the low efficiency and non-uniformity of artificial recognition hinder the application and popularization of US for this purpose. Herein, we developed an automatic muscle boundary segmentation tool for US image recognition and tested its accuracy and clinical applicability. Our dataset was constructed from a total of 465 US images of the flexor digitorum superficialis (FDS) from 19 participants (10 men and 9 women, age 27.4 ± 6.3 years). We used the U-net model for US image segmentation. The U-net output often includes several disconnected regions. Anatomically, the target muscle usually only has one connected region. Based on this principle, we designed an algorithm written in C++ to eliminate redundantly connected regions of outputs. The muscle boundary images generated by the tool were compared with those obtained by professionals and junior physicians to analyze their accuracy and clinical applicability. The dataset was divided into five groups for experimentation, and the average Dice coefficient, recall, and accuracy, as well as the intersection over union (IoU) of the prediction set in each group were all about 90%. Furthermore, we propose a new standard to judge the segmentation results. Under this standard, 99% of the total 150 predicted images by U-net are excellent, which is very close to the segmentation result obtained by professional doctors. In this study, we developed an automatic muscle segmentation tool for US-guided muscle injections. The accuracy of the recognition of the muscle boundary was similar to that of manual labeling by a specialist sonographer, providing a reliable auxiliary tool for clinicians to shorten the US learning cycle, reduce the clinical workload, and improve injection safety.
Collapse
Affiliation(s)
- Chen Xin
- Rehabilitation Center, Qilu Hospital of Shandong University, Jinan, China
| | - Baoxu Li
- School of Mathematics, Shandong University, Jinan, China
| | - Dezheng Wang
- Rehabilitation Center, Qilu Hospital of Shandong University, Jinan, China
| | - Wei Chen
- Department of Biomedical Engineering, School of Control Science and Engineering, Shandong University, Jinan, Shandong, China
| | - Shouwei Yue
- Rehabilitation Center, Qilu Hospital of Shandong University, Jinan, China
| | - Dong Meng
- Rehabilitation Center, Qilu Hospital of Shandong University, Jinan, China
| | - Xu Qiao
- Department of Biomedical Engineering, School of Control Science and Engineering, Shandong University, Jinan, Shandong, China
| | - Yang Zhang
- Rehabilitation Center, Qilu Hospital of Shandong University, Jinan, China
| |
Collapse
|
17
|
Wan P, Xue H, Liu C, Chen F, Kong W, Zhang D. Dynamic Perfusion Representation and Aggregation Network for Nodule Segmentation Using Contrast-Enhanced US. IEEE J Biomed Health Inform 2023; 27:3431-3442. [PMID: 37097791 DOI: 10.1109/jbhi.2023.3270307] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/26/2023]
Abstract
Dynamic contrast-enhanced ultrasound (CEUS) imaging has been widely applied in lesion detection and characterization, due to its offered real-time observation of microvascular perfusion. Accurate lesion segmentation is of great importance to the quantitative and qualitative perfusion analysis. In this paper, we propose a novel dynamic perfusion representation and aggregation network (DpRAN) for the automatic segmentation of lesions using dynamic CEUS imaging. The core challenge of this work lies in enhancement dynamics modeling of various perfusion areas. Specifically, we divide enhancement features into the two scales: short-range enhancement patterns and long-range evolution tendency. To effectively represent real-time enhancement characteristics and aggregate them in a global view, we introduce the perfusion excitation (PE) gate and cross-attention temporal aggregation (CTA) module, respectively. Different from the common temporal fusion methods, we also introduce an uncertainty estimation strategy to assist the model to locate the critical enhancement point first, in which a relatively distinguished enhancement pattern is displayed. The segmentation performance of our DpRAN method is validated on our collected CEUS datasets of thyroid nodules. We obtain the mean dice coefficient (DSC) and intersection of union (IoU) of 0.794 and 0.676, respectively. Superior performance demonstrates its efficacy to capture distinguished enhancement characteristics for lesion recognition.
Collapse
|
18
|
Kinoshita T, Takahashi T, Murayama R, Nakagami G, Sanada H, Noguchi H. Creation of the Forearm 3D-Model with Veins from Transversal Ultrasonography Image Sequence. ANNUAL INTERNATIONAL CONFERENCE OF THE IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. IEEE ENGINEERING IN MEDICINE AND BIOLOGY SOCIETY. ANNUAL INTERNATIONAL CONFERENCE 2023; 2023:1-4. [PMID: 38083707 DOI: 10.1109/embc40787.2023.10340868] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/18/2023]
Abstract
This study developed an automatic detection algorithm of vessel and skin regions in a transversal ultrasonography image on the arm. We also developed an algorithm to generate a 3D model from detected areas to assist vein puncture. In the algorithm, the vessel's candidate regions in the ultrasonography image were detected using U-Net or Mask R-CNN, which are a kind of deep learning method for segmentation. Then vessel regions were selected among the candidates based on continuous properties in an image sequence. The skin regions were also detected. The 3D polygon data was created from paired pixels in sequential images. The experiments demonstrated that Mask R-CNN could correctly estimate the branch of vessel which were difficult to identify accurate region separately using U-Net, and achieved an overall IoU of 80%. The confirmation experiment of 3D model demonstrated that generated model have enough feasibility for assessment of appropriate veins and locations for puncture.Clinical relevance-The developed 3D model generation from ultrasonography images will be useful for support to identify the appropriate veins for puncture.
Collapse
|
19
|
Yang Y, Chen F, Liang H, Bai Y, Wang Z, Zhao L, Ma S, Niu Q, Li F, Xie T, Cai Y. CNN-based automatic segmentations and radiomics feature reliability on contrast-enhanced ultrasound images for renal tumors. Front Oncol 2023; 13:1166988. [PMID: 37333811 PMCID: PMC10272725 DOI: 10.3389/fonc.2023.1166988] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/15/2023] [Accepted: 05/22/2023] [Indexed: 06/20/2023] Open
Abstract
Objective To investigate the feasibility and efficiency of automatic segmentation of contrast-enhanced ultrasound (CEUS) images in renal tumors by convolutional neural network (CNN) based models and their further application in radiomic analysis. Materials and methods From 94 pathologically confirmed renal tumor cases, 3355 CEUS images were extracted and randomly divided into training set (3020 images) and test set (335 images). According to the histological subtypes of renal cell carcinoma, the test set was further split into clear cell renal cell carcinoma (ccRCC) set (225 images), renal angiomyolipoma (AML) set (77 images) and set of other subtypes (33 images). Manual segmentation was the gold standard and serves as ground truth. Seven CNN-based models including DeepLabV3+, UNet, UNet++, UNet3+, SegNet, MultilResUNet and Attention UNet were used for automatic segmentation. Python 3.7.0 and Pyradiomics package 3.0.1 were used for radiomic feature extraction. Performance of all approaches was evaluated by the metrics of mean intersection over union (mIOU), dice similarity coefficient (DSC), precision, and recall. Reliability and reproducibility of radiomics features were evaluated by the Pearson coefficient and the intraclass correlation coefficient (ICC). Results All seven CNN-based models achieved good performance with the mIOU, DSC, precision and recall ranging between 81.97%-93.04%, 78.67%-92.70%, 93.92%-97.56%, and 85.29%-95.17%, respectively. The average Pearson coefficients ranged from 0.81 to 0.95, and the average ICCs ranged from 0.77 to 0.92. The UNet++ model showed the best performance with the mIOU, DSC, precision and recall of 93.04%, 92.70%, 97.43% and 95.17%, respectively. For ccRCC, AML and other subtypes, the reliability and reproducibility of radiomic analysis derived from automatically segmented CEUS images were excellent, with the average Pearson coefficients of 0.95, 0.96 and 0.96, and the average ICCs for different subtypes were 0.91, 0.93 and 0.94, respectively. Conclusion This retrospective single-center study showed that the CNN-based models had good performance on automatic segmentation of CEUS images for renal tumors, especially the UNet++ model. The radiomics features extracted from automatically segmented CEUS images were feasible and reliable, and further validation by multi-center research is necessary.
Collapse
Affiliation(s)
- Yin Yang
- Department of Ultrasound, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Fei Chen
- Department of Pediatrics, Jiahui International Hospital, Shanghai, China
| | - Hongmei Liang
- Department of Ultrasound, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Yun Bai
- Department of Ultrasound, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Zhen Wang
- School of Computer Science and Technology, Taiyuan Normal University, Taiyuan, China
| | - Lei Zhao
- Department of Ultrasound, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Sai Ma
- Department of Ultrasound, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Qinghua Niu
- Department of Ultrasound, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Fan Li
- Department of Ultrasound, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| | - Tianwu Xie
- Institute of Radiation Medicine, Fudan University, Shanghai, China
| | - Yingyu Cai
- Department of Ultrasound, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
| |
Collapse
|
20
|
Shin J, Lee S, Yi J. Real-Time Deep Recognition of Standardized Liver Ultrasound Scan Locations. SENSORS (BASEL, SWITZERLAND) 2023; 23:4850. [PMID: 37430764 DOI: 10.3390/s23104850] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/08/2023] [Revised: 05/12/2023] [Accepted: 05/16/2023] [Indexed: 07/12/2023]
Abstract
Liver ultrasound (US) plays a critical role in diagnosing liver diseases. However, it is often difficult for examiners to accurately identify the liver segments captured in US images due to patient variability and the complexity of US images. Our study aim is automatic, real-time recognition of standardized US scans coordinated with reference liver segments to guide examiners. We propose a novel deep hierarchical architecture for classifying liver US images into 11 standardized US scans, which has yet to be properly established due to excessive variability and complexity. We address this problem based on a hierarchical classification of 11 US scans with different features applied to individual hierarchies as well as a novel feature space proximity analysis for handling ambiguous US images. Experiments were performed using US image datasets obtained from a hospital setting. To evaluate the performance under patient variability, we separated the training and testing datasets into distinct patient groups. The experimental results show that the proposed method achieved an F1-score of more than 93%, which is more than sufficient for a tool to guide examiners. The superior performance of the proposed hierarchical architecture was demonstrated by comparing its performance with that of non-hierarchical architecture.
Collapse
Affiliation(s)
- Jonghwan Shin
- Department of Electrical and Computer Engineering, Sungkyunkwan University, Suwon 16419, Republic of Korea
| | - Sukhan Lee
- Artificial Intelligence Department, Sungkyunkwan University, Suwon 16419, Republic of Korea
| | - Juneho Yi
- Department of Electrical and Computer Engineering, Sungkyunkwan University, Suwon 16419, Republic of Korea
| |
Collapse
|
21
|
Karri M, Annavarapu CSR, Acharya UR. Skin lesion segmentation using two-phase cross-domain transfer learning framework. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2023; 231:107408. [PMID: 36805279 DOI: 10.1016/j.cmpb.2023.107408] [Citation(s) in RCA: 5] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/25/2022] [Revised: 01/31/2023] [Accepted: 02/04/2023] [Indexed: 06/18/2023]
Abstract
BACKGROUND AND OBJECTIVE Deep learning (DL) models have been used for medical imaging for a long time but they did not achieve their full potential in the past because of insufficient computing power and scarcity of training data. In recent years, we have seen substantial growth in DL networks because of improved technology and an abundance of data. However, previous studies indicate that even a well-trained DL algorithm may struggle to generalize data from multiple sources because of domain shifts. Additionally, ineffectiveness of basic data fusion methods, complexity of segmentation target and low interpretability of current DL models limit their use in clinical decisions. To meet these challenges, we present a new two-phase cross-domain transfer learning system for effective skin lesion segmentation from dermoscopic images. METHODS Our system is based on two significant technical inventions. We examine a two- phase cross-domain transfer learning approach, including model-level and data-level transfer learning, by fine-tuning the system on two datasets, MoleMap and ImageNet. We then present nSknRSUNet, a high-performing DL network, for skin lesion segmentation using broad receptive fields and spatial edge attention feature fusion. We examine the trained model's generalization capabilities on skin lesion segmentation to quantify these two inventions. We cross-examine the model using two skin lesion image datasets, MoleMap and HAM10000, obtained from varied clinical contexts. RESULTS At data-level transfer learning for the HAM10000 dataset, the proposed model obtained 94.63% of DSC and 99.12% accuracy. In cross-examination at data-level transfer learning for the Molemap dataset, the proposed model obtained 93.63% of DSC and 97.01% of accuracy. CONCLUSION Numerous experiments reveal that our system produces excellent performance and improves upon state-of-the-art methods on both qualitative and quantitative measures.
Collapse
Affiliation(s)
- Meghana Karri
- Department of Computer Science and Engineering, Indian Institute of Technology (Indian School of Mines), Dhanbad, 826004, Jharkhand, India.
| | - Chandra Sekhara Rao Annavarapu
- Department of Computer Science and Engineering, Indian Institute of Technology (Indian School of Mines), Dhanbad, 826004, Jharkhand, India.
| | - U Rajendra Acharya
- Ngee Ann Polytechnic, Department of Electronics and Computer Engineering, 599489, Singapore; Department of Biomedical Engineering, School of science and Technology, SUSS university, Singapore; Department of Biomedical Informatics and Medical Engineering, Asia university, Taichung, Taiwan.
| |
Collapse
|
22
|
Monkam P, Jin S, Lu W. Annotation Cost Minimization for Ultrasound Image Segmentation Using Cross-Domain Transfer Learning. IEEE J Biomed Health Inform 2023; 27:2015-2025. [PMID: 37021897 DOI: 10.1109/jbhi.2023.3236989] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/18/2023]
Abstract
Deep learning techniques can help minimize inter-physician analysis variability and the medical expert workloads, thereby enabling more accurate diagnoses. However, their implementation requires large-scale annotated dataset whose acquisition incurs heavy time and human-expertise costs. Hence, to significantly minimize the annotation cost, this study presents a novel framework that enables the deployment of deep learning methods in ultrasound (US) image segmentation requiring only very limited manually annotated samples. We propose SegMix, a fast and efficient approach that exploits a segment-paste-blend concept to generate large number of annotated samples based on a few manually acquired labels. Besides, a series of US-specific augmentation strategies built upon image enhancement algorithms are introduced to make maximum use of the available limited number of manually delineated images. The feasibility of the proposed framework is validated on the left ventricle (LV) segmentation and fetal head (FH) segmentation tasks, respectively. Experimental results demonstrate that using only 10 manually annotated images, the proposed framework can achieve a Dice and JI of 82.61% and 83.92%, and 88.42% and 89.27% for LV segmentation and FH segmentation, respectively. Compared with training using the entire training set, there is over 98% of annotation cost reduction while achieving comparable segmentation performance. This indicates that the proposed framework enables satisfactory deep leaning performance when very limited number of annotated samples is available. Therefore, we believe that it can be a reliable solution for annotation cost reduction in medical image analysis.
Collapse
|
23
|
Zhu Y, Li C, Hu K, Luo H, Zhou M, Li X, Gao X. A new two-stream network based on feature separation and complementation for ultrasound image segmentation. Biomed Signal Process Control 2023. [DOI: 10.1016/j.bspc.2022.104567] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/31/2022]
|
24
|
Gong Y, Zhu H, Li J, Yang J, Cheng J, Chang Y, Bai X, Ji X. SCCNet: Self-correction boundary preservation with a dynamic class prior filter for high-variability ultrasound image segmentation. Comput Med Imaging Graph 2023; 104:102183. [PMID: 36623451 DOI: 10.1016/j.compmedimag.2023.102183] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/30/2022] [Revised: 12/06/2022] [Accepted: 12/29/2022] [Indexed: 01/06/2023]
Abstract
The highly ambiguous nature of boundaries and similar objects is difficult to address in some ultrasound image segmentation tasks, such as neck muscle segmentation, leading to unsatisfactory performance. Thus, this paper proposes a two-stage network called SCCNet (self-correction context network) using a self-correction boundary preservation module and class-context filter to alleviate these problems. The proposed self-correction boundary preservation module uses a dynamic key boundary point (KBP) map to increase the capability of iteratively discriminating ambiguous boundary points segments, and the predicted segmentation map from one stage is used to obtain a dynamic class prior filter to improve the segmentation performance at Stage 2. Finally, three datasets, Neck Muscle, CAMUS and Thyroid, are used to demonstrate that our proposed SCCNet outperforms other state-of-the art methods, such as BPBnet, DSNnet, and RAGCnet. Our proposed network shows at least a 1.2-3.7% improvement on the three datasets, Neck Muscle, Thyroid, and CAMUS. The source code is available at https://github.com/lijixing0425/SCCNet.
Collapse
Affiliation(s)
- Yuxin Gong
- School of Biological Science and Medical Engineering, Beihang University, Beijing, China; Department of Ultrasound Diagnosis, Xuanwu Hospital, Capital Medical University, China
| | - Haogang Zhu
- State Key Laboratory of Software Development Environment, School of Computer Science and Engineering, Beihang University, Beijing, China; Beijing Advanced Innovation Centre for Big Data-Based Precision Medicine, Beihang University, Beijing, China.
| | - Jixing Li
- Institute of Semiconductors, Chinese Academy of Sciences, Beijing 100083, China; University of Chinese Academy of Sciences, Beijing 100089, China
| | - Jingchun Yang
- Department of Ultrasound Diagnosis, Xuanwu Hospital, Capital Medical University, China.
| | - Jian Cheng
- State Key Laboratory of Software Development Environment, School of Computer Science and Engineering, Beihang University, Beijing, China
| | - Ying Chang
- Department of Ultrasound Diagnosis, Xuanwu Hospital, Capital Medical University, China
| | - Xiao Bai
- State Key Laboratory of Software Development Environment, School of Computer Science and Engineering, Beihang University, Beijing, China
| | - Xunming Ji
- Beijing Institute of Brain Disorders, Laboratory of Brain Disorders, Ministry of Science and Technology, Collaborative Innovation Center for Brain Disorders, Capital Medical University, Beijing, China
| |
Collapse
|
25
|
Prostate Ultrasound Image Segmentation Based on DSU-Net. Biomedicines 2023; 11:biomedicines11030646. [PMID: 36979625 PMCID: PMC10045621 DOI: 10.3390/biomedicines11030646] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/28/2022] [Revised: 02/14/2023] [Accepted: 02/16/2023] [Indexed: 02/23/2023] Open
Abstract
In recent years, the incidence of prostate cancer in the male population has been increasing year by year. Transrectal ultrasound (TRUS) is an important means of prostate cancer diagnosis. The accurate segmentation of the prostate in TRUS images can assist doctors in needle biopsy and surgery and is also the basis for the accurate identification of prostate cancer. Due to the asymmetric shape and blurred boundary line of the prostate in TRUS images, it is difficult to obtain accurate segmentation results with existing segmentation methods. Therefore, a prostate segmentation method called DSU-Net is proposed in this paper. This proposed method replaces the basic convolution in the U-Net model with the improved convolution combining shear transformation and deformable convolution, making the network more sensitive to border features and more suitable for prostate segmentation tasks. Experiments show that DSU-Net has higher accuracy than other existing traditional segmentation methods.
Collapse
|
26
|
Xia M, Yang H, Huang Y, Qu Y, Zhou G, Zhang F, Wang Y, Guo Y. 3D pyramidal densely connected network with cross-frame uncertainty guidance for intravascular ultrasound sequence segmentation. Phys Med Biol 2023; 68. [PMID: 36745930 DOI: 10.1088/1361-6560/acb988] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/25/2022] [Accepted: 02/06/2023] [Indexed: 02/08/2023]
Abstract
Objective. Automatic extraction of external elastic membrane border (EEM) and lumen-intima border (LIB) in intravascular ultrasound (IVUS) sequences aids atherosclerosis diagnosis. Existing IVUS segmentation networks ignored longitudinal relations among sequential images and neglected that IVUS images of different vascular conditions vary largely in intricacy and informativeness. As a result, they suffered from performance degradation in complicated parts in IVUS sequences.Approach. In this paper, we develop a 3D Pyramidal Densely-connected Network (PDN) with Adaptive learning and post-Correction guided by a novel cross-frame uncertainty (CFU). The proposed method is named PDN-AC. Specifically, the PDN enables the longitudinal information exploitation and the effective perception of size-varied vessel regions in IVUS samples, by pyramidally connecting multi-scale 3D dilated convolutions. Additionally, the CFU enhances the robustness of the method to complicated pathology from the frame-level (f-CFU) and pixel-level (p-CFU) via exploiting cross-frame knowledge in IVUS sequences. The f-CFU weighs the complexity of IVUS frames and steers an adaptive sampling during the PDN training. The p-CFU visualizes uncertain pixels probably misclassified by the PDN and guides an active contour-based post-correction.Main results. Human and animal experiments were conducted on IVUS datasets acquired from atherosclerosis patients and pigs. Results showed that the f-CFU weighted adaptive sampling reduced the Hausdorff distance (HD) by 10.53%/7.69% in EEM/LIB detection. Improvements achieved by the p-CFU guided post-correction were 2.94%/5.56%.Significance. The PDN-AC attained mean Jaccard values of 0.90/0.87 and HD values of 0.33/0.34 mm in EEM/LIB detection, preferable to state-of-the-art IVUS segmentation methods.
Collapse
Affiliation(s)
- Menghua Xia
- Department of Electronic Engineering, School of Information Science and Technology, Fudan University, Shanghai 200433, People's Republic of China
| | - Hongbo Yang
- Department of Cardiology, Zhongshan Hospital, Fudan University. Shanghai Institute of Cardiovascular Diseases, Shanghai 200032, People's Republic of China
| | - Yi Huang
- Department of Electronic Engineering, School of Information Science and Technology, Fudan University, Shanghai 200433, People's Republic of China
| | - Yanan Qu
- Department of Cardiology, Zhongshan Hospital, Fudan University. Shanghai Institute of Cardiovascular Diseases, Shanghai 200032, People's Republic of China
| | - Guohui Zhou
- Department of Electronic Engineering, School of Information Science and Technology, Fudan University, Shanghai 200433, People's Republic of China.,Key Laboratory of Medical Imaging Computing and Computer Assisted Intervention (MICCAI) of Shanghai, Fudan University, Shanghai 200032, People's Republic of China
| | - Feng Zhang
- Department of Cardiology, Zhongshan Hospital, Fudan University. Shanghai Institute of Cardiovascular Diseases, Shanghai 200032, People's Republic of China
| | - Yuanyuan Wang
- Department of Electronic Engineering, School of Information Science and Technology, Fudan University, Shanghai 200433, People's Republic of China.,Key Laboratory of Medical Imaging Computing and Computer Assisted Intervention (MICCAI) of Shanghai, Fudan University, Shanghai 200032, People's Republic of China
| | - Yi Guo
- Department of Electronic Engineering, School of Information Science and Technology, Fudan University, Shanghai 200433, People's Republic of China.,Key Laboratory of Medical Imaging Computing and Computer Assisted Intervention (MICCAI) of Shanghai, Fudan University, Shanghai 200032, People's Republic of China
| |
Collapse
|
27
|
Ansari MY, Yang Y, Meher PK, Dakua SP. Dense-PSP-UNet: A neural network for fast inference liver ultrasound segmentation. Comput Biol Med 2023; 153:106478. [PMID: 36603437 DOI: 10.1016/j.compbiomed.2022.106478] [Citation(s) in RCA: 52] [Impact Index Per Article: 26.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/05/2022] [Revised: 11/29/2022] [Accepted: 12/21/2022] [Indexed: 01/02/2023]
Abstract
Liver Ultrasound (US) or sonography is popularly used because of its real-time output, low-cost, ease-of-use, portability, and non-invasive nature. Segmentation of real-time liver US is essential for diagnosing and analyzing liver conditions (e.g., hepatocellular carcinoma (HCC)), assisting the surgeons/radiologists in therapeutic procedures. In this paper, we propose a method using a modified Pyramid Scene Parsing (PSP) module in tuned neural network backbones to achieve real-time segmentation without compromising the segmentation accuracy. Considering widespread noise in US data and its impact on outcomes, we study the impact of pre-processing and the influence of loss functions on segmentation performance. We have tested our method after annotating a publicly available US dataset containing 2400 images of 8 healthy volunteers (link to the annotated dataset is provided); the results show that the Dense-PSP-UNet model achieves a high Dice coefficient of 0.913±0.024 while delivering a real-time performance of 37 frames per second (FPS).
Collapse
Affiliation(s)
| | - Yin Yang
- Hamad Bin Khalifa Uinversity, Doha, Qatar
| | | | | |
Collapse
|
28
|
Khan A, Khan SH, Saif M, Batool A, Sohail A, Waleed Khan M. A Survey of Deep Learning Techniques for the Analysis of COVID-19 and their usability for Detecting Omicron. J EXP THEOR ARTIF IN 2023. [DOI: 10.1080/0952813x.2023.2165724] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/15/2023]
Affiliation(s)
- Asifullah Khan
- Pattern Recognition Lab, Department of Computer & Information Sciences, Pakistan Institute of Engineering & Applied Sciences, Nilore, Islamabad, Pakistan
- PIEAS Artificial Intelligence Center (PAIC), Pakistan Institute of Engineering & Applied Sciences, Islamabad, Pakistan
- Center for Mathematical Sciences, Pakistan Institute of Engineering & Applied Sciences, Islamabad, Pakistan
| | - Saddam Hussain Khan
- Pattern Recognition Lab, Department of Computer & Information Sciences, Pakistan Institute of Engineering & Applied Sciences, Nilore, Islamabad, Pakistan
- Department of Computer Systems Engineering, University of Engineering and Applied Sciences (UEAS), Swat, Pakistan
| | - Mahrukh Saif
- Pattern Recognition Lab, Department of Computer & Information Sciences, Pakistan Institute of Engineering & Applied Sciences, Nilore, Islamabad, Pakistan
| | - Asiya Batool
- Pattern Recognition Lab, Department of Computer & Information Sciences, Pakistan Institute of Engineering & Applied Sciences, Nilore, Islamabad, Pakistan
| | - Anabia Sohail
- Pattern Recognition Lab, Department of Computer & Information Sciences, Pakistan Institute of Engineering & Applied Sciences, Nilore, Islamabad, Pakistan
- Department of Computer Science, Faculty of Computing & Artificial Intelligence, Air University, Islamabad, Pakistan
| | - Muhammad Waleed Khan
- Pattern Recognition Lab, Department of Computer & Information Sciences, Pakistan Institute of Engineering & Applied Sciences, Nilore, Islamabad, Pakistan
- Department of Mechanical and Aerospace Engineering, Columbus, OH, USA
| |
Collapse
|
29
|
Ma L, Wang R, He Q, Huang L, Wei X, Lu X, Du Y, Luo J, Liao H. Artificial intelligence-based ultrasound imaging technologies for hepatic diseases. ILIVER 2022; 1:252-264. [DOI: 10.1016/j.iliver.2022.11.001] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/04/2025]
|
30
|
Xia X, Wang J, Liang S, Ye F, Tian MM, Hu W, Xu L. An attention base U-net for parotid tumor autosegmentation. Front Oncol 2022; 12:1028382. [PMID: 36505865 PMCID: PMC9730401 DOI: 10.3389/fonc.2022.1028382] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/25/2022] [Accepted: 10/26/2022] [Indexed: 11/25/2022] Open
Abstract
A parotid neoplasm is an uncommon condition that only accounts for less than 3% of all head and neck cancers, and they make up less than 0.3% of all new cancers diagnosed annually. Due to their nonspecific imaging features and heterogeneous nature, accurate preoperative diagnosis remains a challenge. Automatic parotid tumor segmentation may help physicians evaluate these tumors. Two hundred eighty-five patients diagnosed with benign or malignant parotid tumors were enrolled in this study. Parotid and tumor tissues were segmented by 3 radiologists on T1-weighted (T1w), T2-weighted (T2w) and T1-weighted contrast-enhanced (T1wC) MR images. These images were randomly divided into two datasets, including a training dataset (90%) and an validation dataset (10%). A 10-fold cross-validation was performed to assess the performance. An attention base U-net for parotid tumor autosegmentation was created on the MRI T1w, T2 and T1wC images. The results were evaluated in a separate dataset, and the mean Dice similarity coefficient (DICE) for both parotids was 0.88. The mean DICE for left and right tumors was 0.85 and 0.86, respectively. These results indicate that the performance of this model corresponds with the radiologist's manual segmentation. In conclusion, an attention base U-net for parotid tumor autosegmentation may assist physicians to evaluate parotid gland tumors.
Collapse
Affiliation(s)
- Xianwu Xia
- The Second Affiliated Hospital, School of Medicine, Zhejiang University, Hangzhou, Zhejiang, China
- Department of Oncology Intervention, The Affiliated Municipal Hospital of Taizhou University, Taizhou, China
- Department of Radiation Oncology, Fudan University Shanghai Cancer Center, Shanghai, China
- Department of Oncology, Shanghai Medical College, Fudan University, Shanghai, China
| | - Jiazhou Wang
- Department of Radiation Oncology, Fudan University Shanghai Cancer Center, Shanghai, China
- Department of Oncology, Shanghai Medical College, Fudan University, Shanghai, China
| | - Sheng Liang
- Department of Oncology Intervention, The Affiliated Municipal Hospital of Taizhou University, Taizhou, China
| | - Fangfang Ye
- Department of Oncology Intervention, The Affiliated Municipal Hospital of Taizhou University, Taizhou, China
| | - Min-Ming Tian
- Department of Oncology Intervention, Jiangxi University of Traditional Chinese Medicine, Nanchang, Jiangxi, China
| | - Weigang Hu
- Department of Radiation Oncology, Fudan University Shanghai Cancer Center, Shanghai, China
- Department of Oncology, Shanghai Medical College, Fudan University, Shanghai, China
| | - Leiming Xu
- The Second Affiliated Hospital, School of Medicine, Zhejiang University, Hangzhou, Zhejiang, China
| |
Collapse
|
31
|
Wang J, Chen G, Chen S, Joseph Raj AN, Zhuang Z, Xie L, Ma S. Ultrasonic breast tumor extraction based on adversarial mechanism and active contour. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 225:107052. [PMID: 35985149 DOI: 10.1016/j.cmpb.2022.107052] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Received: 04/24/2022] [Revised: 06/07/2022] [Accepted: 07/30/2022] [Indexed: 02/05/2023]
Abstract
BACKGROUND AND OBJECTIVE Breast cancer is a high incidence of gynecological diseases; breast ultrasound screening can effectively reduce the mortality rate of breast cancer. In breast ultrasound images, the localization and segmentation of tumor lesions are important steps for the extraction of lesions, which helps clinicians evaluate breast lesions quantitatively and makes better clinical diagnosis of the disease. However, the segmentation of breast lesions is difficult due to the blurred and uneven edges of some lesions. In this paper, we propose a segmentation framework combining active contour module and deep learning adversarial mechanism and apply it for the segmentation of breast tumor lesions. METHOD We use a conditional adversarial network as the main framework. The generator is a segmentation network consisting of a Deformed U-Net and an active contour module. Here, the Deformed U-Net performs pixel-level segmentation for breast ultrasound images. The active contour module refines the tumor lesion edges, and the refined result provides loss information for Deformed U-Net. Therefore, the Deformed U-Net can better classify the edge pixels. The discriminator is the Markov discriminator; this discriminator provides loss feedback for the segmentation network. We cross-train the discriminator and segmentation network to implement Adversarial Mechanism for getting a more optimized segmentation network. RESULTS The segmentation performance of the segmentation network for breast ultrasound images is improved by adding a Markov discriminator to provide discriminant loss training. The proposed method for segmenting the tumor lesions in breast ultrasound image obtains dice coefficient: 89.7%, accuracy: 98.1%, precision: 86.3%, mean-intersection-over-union: 82.2%, recall: 94.7%, specificity: 98.5% and F1score: 89.7%. CONCLUSION Comparing with traditional methods, the proposed method gives better performance. The experimental results show that the proposed method can effectively segment the lesions in breast ultrasound images, and then assist doctors to realize the diagnosis of breast lesions.
Collapse
Affiliation(s)
- Jinhong Wang
- Department of Ultrasound, The First Affiliated Hospital of Shantou University Medical College, 57 Changping Road, Longhu District, Shantou, Guangdong, China
| | - Guiqing Chen
- Department of Electronic Engineering, Shantou University, No.243, Daxue Road, Tuo Jiang Street, Jinping District, Shantou City, Guangdong, China
| | - Shiqiang Chen
- Department of Electronic Engineering, Shantou University, No.243, Daxue Road, Tuo Jiang Street, Jinping District, Shantou City, Guangdong, China
| | - Alex Noel Joseph Raj
- Department of Electronic Engineering, Shantou University, No.243, Daxue Road, Tuo Jiang Street, Jinping District, Shantou City, Guangdong, China
| | - Zhemin Zhuang
- Department of Electronic Engineering, Shantou University, No.243, Daxue Road, Tuo Jiang Street, Jinping District, Shantou City, Guangdong, China
| | - Lei Xie
- Department of Radiology, The First Affiliated Hospital of Shantou University Medical College, 57 Changping Road, Longhu District, Shantou, Guangdong, China
| | - Shuhua Ma
- Department of Radiology, The First Affiliated Hospital of Shantou University Medical College, 57 Changping Road, Longhu District, Shantou, Guangdong, China
| |
Collapse
|
32
|
Survarachakan S, Prasad PJR, Naseem R, Pérez de Frutos J, Kumar RP, Langø T, Alaya Cheikh F, Elle OJ, Lindseth F. Deep learning for image-based liver analysis — A comprehensive review focusing on malignant lesions. Artif Intell Med 2022; 130:102331. [DOI: 10.1016/j.artmed.2022.102331] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2021] [Revised: 05/23/2022] [Accepted: 05/30/2022] [Indexed: 11/26/2022]
|
33
|
Bi H, Sun J, Jiang Y, Ni X, Shu H. Structure boundary-preserving U-Net for prostate ultrasound image segmentation. Front Oncol 2022; 12:900340. [PMID: 35965563 PMCID: PMC9366193 DOI: 10.3389/fonc.2022.900340] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/20/2022] [Accepted: 06/30/2022] [Indexed: 11/19/2022] Open
Abstract
Prostate cancer diagnosis is performed under ultrasound-guided puncture for pathological cell extraction. However, determining accurate prostate location remains a challenge from two aspects: (1) prostate boundary in ultrasound images is always ambiguous; (2) the delineation of radiologists always occupies multiple pixels, leading to many disturbing points around the actual contour. We proposed a boundary structure-preserving U-Net (BSP U-Net) in this paper to achieve precise prostate contour. BSP U-Net incorporates prostate shape prior to traditional U-Net. The prior shape is built by the key point selection module, which is an active shape model-based method. Then, the module plugs into the traditional U-Net structure network to achieve prostate segmentation. The experiments were conducted on two datasets: PH2 + ISBI 2016 challenge and our private prostate ultrasound dataset. The results on PH2 + ISBI 2016 challenge achieved a Dice similarity coefficient (DSC) of 95.94% and a Jaccard coefficient (JC) of 88.58%. The results of prostate contour based on our method achieved a higher pixel accuracy of 97.05%, a mean intersection over union of 93.65%, a DSC of 92.54%, and a JC of 93.16%. The experimental results show that the proposed BSP U-Net has good performance on PH2 + ISBI 2016 challenge and prostate ultrasound image segmentation and outperforms other state-of-the-art methods.
Collapse
Affiliation(s)
- Hui Bi
- Department of Radiation Oncology, The Affiliated Changzhou No. 2 People’s Hospital of Nanjing Medical University, Changzhou, China
- School of Computer Science and Artificial Intelligence, Changzhou University, Changzhou, China
- Key Laboratory of Computer Network and Information Integration, Southeast University, Nanjing, China
| | - Jiawei Sun
- Jiangsu Province Engineering Research Center of Medical Physics, Changzhou, China
| | - Yibo Jiang
- School of Electrical and Information Engineering, Changzhou Institute of Technology, Changzhou, China
| | - Xinye Ni
- Department of Radiation Oncology, The Affiliated Changzhou No. 2 People’s Hospital of Nanjing Medical University, Changzhou, China
- Jiangsu Province Engineering Research Center of Medical Physics, Changzhou, China
- *Correspondence: Xinye Ni,
| | - Huazhong Shu
- Laboratory of Image Science and Technology, Southeast University, Nanjing, China
- Centre de Recherche en Information Biomédicale Sino-francais, Rennes, France
- Jiangsu Provincial Joint International Research Laboratory of Medical Information Processing, Southeast University, Nanjing, China
| |
Collapse
|
34
|
Jiang Z, Gao Y, Xie L, Navab N. Towards Autonomous Atlas-Based Ultrasound Acquisitions in Presence of Articulated Motion. IEEE Robot Autom Lett 2022. [DOI: 10.1109/lra.2022.3180440] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/08/2022]
Affiliation(s)
- Zhongliang Jiang
- Chair for Computer Aided Medical Procedures and Augmented Reality (CAMP), Technical University of Munich, Garching, Germany
| | - Yuan Gao
- Chair for Computer Aided Medical Procedures and Augmented Reality (CAMP), Technical University of Munich, Garching, Germany
| | - Le Xie
- Institute of Forming Technology and Equipment and Institute of Medical Robotics, Shanghai Jiao Tong University, Shanghai, China
| | - Nassir Navab
- Chair for Computer Aided Medical Procedures and Augmented Reality (CAMP), Technical University of Munich, Garching, Germany
| |
Collapse
|
35
|
Aux-MVNet: Auxiliary Classifier-Based Multi-View Convolutional Neural Network for Maxillary Sinusitis Diagnosis on Paranasal Sinuses View. Diagnostics (Basel) 2022; 12:diagnostics12030736. [PMID: 35328288 PMCID: PMC8947362 DOI: 10.3390/diagnostics12030736] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/21/2022] [Revised: 03/09/2022] [Accepted: 03/15/2022] [Indexed: 02/05/2023] Open
Abstract
Computed tomography (CT) is undoubtedly the most reliable and the only method for accurate diagnosis of sinusitis, while X-ray has long been used as the first imaging technique for early detection of sinusitis symptoms. More importantly, radiography plays a key role in determining whether or not a CT examination should be performed for further evaluation. In order to simplify the diagnostic process of paranasal sinus view and moreover to avoid the use of CT scans which have disadvantages such as high radiation dose, high cost, and high time consumption, this paper proposed a multi-view CNN able to faithfully estimate the severity of sinusitis. In this study, a multi-view convolutional neural network (CNN) is proposed which is able to accurately estimate the severity of sinusitis by analyzing only radiographs consisting of Waters’ view and Caldwell’s view without the aid of CT scans. The proposed network is designed as a cascaded architecture, and can simultaneously provide decisions for maxillary sinus localization and sinusitis classification. We obtained an average area under the curve (AUC) of 0.722 for maxillary sinusitis classification, and an AUC of 0.750 and 0.700 for the left and right maxillary sinusitis, respectively, using the proposed network.
Collapse
|
36
|
Luo Y, Ma Y, O’ Brien H, Jiang K, Kohli V, Maidelin S, Saeed M, Deng E, Pushparajah K, Rhode KS. Edge-enhancement densenet for X-ray fluoroscopy image denoising in cardiac electrophysiology procedures. Med Phys 2022; 49:1262-1275. [PMID: 34954836 PMCID: PMC9304258 DOI: 10.1002/mp.15426] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/15/2021] [Revised: 10/26/2021] [Accepted: 11/29/2021] [Indexed: 11/15/2022] Open
Abstract
PURPOSE Reducing X-ray dose increases safety in cardiac electrophysiology procedures but also increases image noise and artifacts which may affect the discernibility of devices and anatomical cues. Previous denoising methods based on convolutional neural networks (CNNs) have shown improvements in the quality of low-dose X-ray fluoroscopy images but may compromise clinically important details required by cardiologists. METHODS In order to obtain denoised X-ray fluoroscopy images whilst preserving details, we propose a novel deep-learning-based denoising framework, namely edge-enhancement densenet (EEDN), in which an attention-awareness edge-enhancement module is designed to increase edge sharpness. In this framework, a CNN-based denoiser is first used to generate an initial denoising result. Contours representing edge information are then extracted using an attention block and a group of interacted ultra-dense blocks for edge feature representation. Finally, the initial denoising result and enhanced edges are combined to generate the final X-ray image. The proposed denoising framework was tested on a total of 3262 clinical images taken from 100 low-dose X-ray sequences acquired from 20 patients. The performance was assessed by pairwise voting from five cardiologists as well as quantitative indicators. Furthermore, we evaluated our technique's effect on catheter detection using 416 images containing coronary sinus catheters in order to examine its influence as a pre-processing tool. RESULTS The average signal-to-noise ratio of X-ray images denoised with EEDN was 24.5, which was 2.2 times higher than that of the original images. The accuracy of catheter detection from EEDN denoised sequences showed no significant difference compared with their original counterparts. Moreover, EEDN received the highest average votes in our clinician assessment when compared to our existing technique and the original images. CONCLUSION The proposed deep learning-based framework shows promising capability for denoising interventional X-ray fluoroscopy images. The results from the catheter detection show that the network does not affect the results of such an algorithm when used as a pre-processing step. The extensive qualitative and quantitative evaluations suggest that the network may be of benefit to reduce radiation dose when applied in real time in the catheter laboratory.
Collapse
Affiliation(s)
- Yimin Luo
- School of Biomedical Engineering and Imaging SciencesKing's College LondonLondonUK
| | - Yingliang Ma
- School of ComputingElectronics and MathematicsCoventry UniversityCoventryUK
| | - Hugh O’ Brien
- School of Biomedical Engineering and Imaging SciencesKing's College LondonLondonUK
| | - Kui Jiang
- School of Computer ScienceWuhan UniversityWuhanChina
| | - Vikram Kohli
- School of Biomedical Engineering and Imaging SciencesKing's College LondonLondonUK
| | - Sesilia Maidelin
- School of Biomedical Engineering and Imaging SciencesKing's College LondonLondonUK
| | - Mahrukh Saeed
- School of Biomedical Engineering and Imaging SciencesKing's College LondonLondonUK
| | - Emily Deng
- School of Biomedical Engineering and Imaging SciencesKing's College LondonLondonUK
| | - Kuberan Pushparajah
- School of Biomedical Engineering and Imaging SciencesKing's College LondonLondonUK
| | - Kawal S. Rhode
- School of Biomedical Engineering and Imaging SciencesKing's College LondonLondonUK
| |
Collapse
|
37
|
Chen F, Ye H, Zhang D, Liao H. TypeSeg: A type-aware encoder-decoder network for multi-type ultrasound images co-segmentation. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2022; 214:106580. [PMID: 34953278 DOI: 10.1016/j.cmpb.2021.106580] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/25/2021] [Revised: 11/01/2021] [Accepted: 12/04/2021] [Indexed: 06/14/2023]
Abstract
PURPOSE As a portable and radiation-free imaging modality, ultrasound can be easily used to image various types of tissue structures. It is important to develop a method which supports the multi-type ultrasound images co-segmentation. However, state-of-the-art ultrasound segmentation methods commonly only focus on the single type images or ignore the type-aware information. METHODS To solve the above problem, this work proposes a novel type-aware encoder-decoder network (TypeSeg) for the multi-type ultrasound images co-segmentation. First, we develop a type-aware metric learning module to find an optimum latent feature space where the ultrasound images of the same types are close and that of the different types are separated by a certain margin. Second, depending on the extracted features, a decision module decides whether the input ultrasound images have the common tissue type or not, and the encoder-decoder network produces a segmentation mask accordingly. RESULTS We evaluate the performance of the proposed TypeSeg model on the ultrasound dataset that contains four types of tissues. The proposed TypeSeg model achieves the overall best results with the mean IOU score of 87.51% ± 3.93% for the multi-type ultrasound images. CONCLUSION The experimental results indicate that the proposed method outperforms all the compared state-of-the-art algorithms for the multi-type ultrasound images co-segmentation task.
Collapse
Affiliation(s)
- Fang Chen
- College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, MIIT Key Laboratory of Pattern Analysis and Machine Intelligence, China.
| | - Haoran Ye
- College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, MIIT Key Laboratory of Pattern Analysis and Machine Intelligence, China
| | - Daoqiang Zhang
- College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, MIIT Key Laboratory of Pattern Analysis and Machine Intelligence, China
| | - Hongen Liao
- Department of Biomedical Engineering, School of Medicine, Tsinghua University, China
| |
Collapse
|
38
|
Kim J, Kim HJ, Kim C, Lee JH, Kim KW, Park YM, Kim HW, Ki SY, Kim YM, Kim WH. Weakly-supervised deep learning for ultrasound diagnosis of breast cancer. Sci Rep 2021; 11:24382. [PMID: 34934144 PMCID: PMC8692405 DOI: 10.1038/s41598-021-03806-7] [Citation(s) in RCA: 24] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/01/2021] [Accepted: 11/30/2021] [Indexed: 11/21/2022] Open
Abstract
Conventional deep learning (DL) algorithm requires full supervision of annotating the region of interest (ROI) that is laborious and often biased. We aimed to develop a weakly-supervised DL algorithm that diagnosis breast cancer at ultrasound without image annotation. Weakly-supervised DL algorithms were implemented with three networks (VGG16, ResNet34, and GoogLeNet) and trained using 1000 unannotated US images (500 benign and 500 malignant masses). Two sets of 200 images (100 benign and 100 malignant masses) were used for internal and external validation sets. For comparison with fully-supervised algorithms, ROI annotation was performed manually and automatically. Diagnostic performances were calculated as the area under the receiver operating characteristic curve (AUC). Using the class activation map, we determined how accurately the weakly-supervised DL algorithms localized the breast masses. For internal validation sets, the weakly-supervised DL algorithms achieved excellent diagnostic performances, with AUC values of 0.92–0.96, which were not statistically different (all Ps > 0.05) from those of fully-supervised DL algorithms with either manual or automated ROI annotation (AUC, 0.92–0.96). For external validation sets, the weakly-supervised DL algorithms achieved AUC values of 0.86–0.90, which were not statistically different (Ps > 0.05) or higher (P = 0.04, VGG16 with automated ROI annotation) from those of fully-supervised DL algorithms (AUC, 0.84–0.92). In internal and external validation sets, weakly-supervised algorithms could localize 100% of malignant masses, except for ResNet34 (98%). The weakly-supervised DL algorithms developed in the present study were feasible for US diagnosis of breast cancer with well-performing localization and differential diagnosis.
Collapse
Affiliation(s)
- Jaeil Kim
- School of Computer Science and Engineering, Kyungpook National University, Daegu, Republic of Korea
| | - Hye Jung Kim
- Department of Radiology, School of Medicine, Kyungpook National University, Kyungpook National University Chilgok Hospital, Daegu, Republic of Korea
| | - Chanho Kim
- School of Computer Science and Engineering, Kyungpook National University, Daegu, Republic of Korea
| | - Jin Hwa Lee
- Department of Radiology, Dong-A University College of Medicine, Busan, Republic of Korea
| | - Keum Won Kim
- Departments of Radiology, School of Medicine, Konyang University, Konyang Univeristy Hospital, Daejeon, Republic of Korea
| | - Young Mi Park
- Department of Radiology, School of Medicine, Inje University, Busan Paik Hospital, Busan, Republic of Korea
| | - Hye Won Kim
- Department of Radiology, Wonkwang University Hospital, Wonkwang University School of Medicine, Iksan, Republic of Korea
| | - So Yeon Ki
- Department of Radiology, School of Medicine, Chonnam National University, Chonnam National University Hwasun Hospital, Hwasun, Republic of Korea
| | - You Me Kim
- Department of Radiology, School of Medicine, Dankook University, Dankook University Hospital, Cheonan, Republic of Korea
| | - Won Hwa Kim
- Department of Radiology, School of Medicine, Kyungpook National University, Kyungpook National University Chilgok Hospital, Daegu, Republic of Korea.
| |
Collapse
|
39
|
Hsu LM, Wang S, Walton L, Wang TWW, Lee SH, Shih YYI. 3D U-Net Improves Automatic Brain Extraction for Isotropic Rat Brain Magnetic Resonance Imaging Data. Front Neurosci 2021; 15:801008. [PMID: 34975392 PMCID: PMC8716693 DOI: 10.3389/fnins.2021.801008] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/24/2021] [Accepted: 11/15/2021] [Indexed: 12/24/2022] Open
Abstract
Brain extraction is a critical pre-processing step in brain magnetic resonance imaging (MRI) analytical pipelines. In rodents, this is often achieved by manually editing brain masks slice-by-slice, a time-consuming task where workloads increase with higher spatial resolution datasets. We recently demonstrated successful automatic brain extraction via a deep-learning-based framework, U-Net, using 2D convolutions. However, such an approach cannot make use of the rich 3D spatial-context information from volumetric MRI data. In this study, we advanced our previously proposed U-Net architecture by replacing all 2D operations with their 3D counterparts and created a 3D U-Net framework. We trained and validated our model using a recently released CAMRI rat brain database acquired at isotropic spatial resolution, including T2-weighted turbo-spin-echo structural MRI and T2*-weighted echo-planar-imaging functional MRI. The performance of our 3D U-Net model was compared with existing rodent brain extraction tools, including Rapid Automatic Tissue Segmentation, Pulse-Coupled Neural Network, SHape descriptor selected External Regions after Morphologically filtering, and our previously proposed 2D U-Net model. 3D U-Net demonstrated superior performance in Dice, Jaccard, center-of-mass distance, Hausdorff distance, and sensitivity. Additionally, we demonstrated the reliability of 3D U-Net under various noise levels, evaluated the optimal training sample sizes, and disseminated all source codes publicly, with a hope that this approach will benefit rodent MRI research community. Significant Methodological Contribution: We proposed a deep-learning-based framework to automatically identify the rodent brain boundaries in MRI. With a fully 3D convolutional network model, 3D U-Net, our proposed method demonstrated improved performance compared to current automatic brain extraction methods, as shown in several qualitative metrics (Dice, Jaccard, PPV, SEN, and Hausdorff). We trust that this tool will avoid human bias and streamline pre-processing steps during 3D high resolution rodent brain MRI data analysis. The software developed herein has been disseminated freely to the community.
Collapse
Affiliation(s)
- Li-Ming Hsu
- Center for Animal Magnetic Resonance Imaging, The University of North Carolina at Chapel Hill, Chapel Hill, NC, United States,Biomedical Research Imaging Center, The University of North Carolina at Chapel Hill, Chapel Hill, NC, United States,Department of Radiology, The University of North Carolina at Chapel Hill, Chapel Hill, NC, United States,*Correspondence: Li-Ming Hsu,
| | - Shuai Wang
- School of Mechanical, Electrical and Information Engineering, Shandong University, Weihai, China
| | - Lindsay Walton
- Center for Animal Magnetic Resonance Imaging, The University of North Carolina at Chapel Hill, Chapel Hill, NC, United States,Biomedical Research Imaging Center, The University of North Carolina at Chapel Hill, Chapel Hill, NC, United States,Department of Radiology, The University of North Carolina at Chapel Hill, Chapel Hill, NC, United States
| | - Tzu-Wen Winnie Wang
- Center for Animal Magnetic Resonance Imaging, The University of North Carolina at Chapel Hill, Chapel Hill, NC, United States,Biomedical Research Imaging Center, The University of North Carolina at Chapel Hill, Chapel Hill, NC, United States,Department of Radiology, The University of North Carolina at Chapel Hill, Chapel Hill, NC, United States
| | - Sung-Ho Lee
- Center for Animal Magnetic Resonance Imaging, The University of North Carolina at Chapel Hill, Chapel Hill, NC, United States,Biomedical Research Imaging Center, The University of North Carolina at Chapel Hill, Chapel Hill, NC, United States,Department of Radiology, The University of North Carolina at Chapel Hill, Chapel Hill, NC, United States
| | - Yen-Yu Ian Shih
- Center for Animal Magnetic Resonance Imaging, The University of North Carolina at Chapel Hill, Chapel Hill, NC, United States,Biomedical Research Imaging Center, The University of North Carolina at Chapel Hill, Chapel Hill, NC, United States,Department of Radiology, The University of North Carolina at Chapel Hill, Chapel Hill, NC, United States,Yen-Yu Ian Shih,
| |
Collapse
|
40
|
Dai X, Lei Y, Roper J, Chen Y, Bradley JD, Curran WJ, Liu T, Yang X. Deep learning-based motion tracking using ultrasound images. Med Phys 2021; 48:7747-7756. [PMID: 34724712 PMCID: PMC11742242 DOI: 10.1002/mp.15321] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/03/2021] [Revised: 10/13/2021] [Accepted: 10/22/2021] [Indexed: 12/25/2022] Open
Abstract
PURPOSE Ultrasound (US) imaging is an established imaging modality capable of offering video-rate volumetric images without ionizing radiation. It has the potential for intra-fraction motion tracking in radiation therapy. In this study, a deep learning-based method has been developed to tackle the challenges in motion tracking using US imaging. METHODS We present a Markov-like network, which is implemented via generative adversarial networks, to extract features from sequential US frames (one tracked frame followed by untracked frames) and thereby estimate a set of deformation vector fields (DVFs) through the registration of the tracked frame and the untracked frames. The positions of the landmarks in the untracked frames are finally determined by shifting landmarks in the tracked frame according to the estimated DVFs. The performance of the proposed method was evaluated on the testing dataset by calculating the tracking error (TE) between the predicted and ground truth landmarks on each frame. RESULTS The proposed method was evaluated using the MICCAI CLUST 2015 dataset which was collected using seven US scanners with eight types of transducers and the Cardiac Acquisitions for Multi-structure Ultrasound Segmentation (CAMUS) dataset which was acquired using GE Vivid E95 ultrasound scanners. The CLUST dataset contains 63 2D and 22 3D US image sequences respectively from 42 and 18 subjects, and the CAMUS dataset includes 2D US images from 450 patients. On CLUST dataset, our proposed method achieved a mean tracking error of 0.70 ± 0.38 mm for the 2D sequences and 1.71 ± 0.84 mm for the 3D sequences for those public available annotations. And on CAMUS dataset, a mean tracking error of 0.54 ± 1.24 mm for the landmarks in the left atrium was achieved. CONCLUSIONS A novel motion tracking algorithm using US images based on modern deep learning techniques has been demonstrated in this study. The proposed method can offer millimeter-level tumor motion prediction in real time, which has the potential to be adopted into routine tumor motion management in radiation therapy.
Collapse
Affiliation(s)
- Xianjin Dai
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
| | - Yang Lei
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
| | - Justin Roper
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
| | - Yue Chen
- The Wallace H. Coulter Department of Biomedical Engineering, Georgia Institute of Technology and Emory University School of Medicine, Atlanta, Georgia, USA
| | - Jeffrey D. Bradley
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
| | - Walter J. Curran
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
| | - Tian Liu
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
| | - Xiaofeng Yang
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
- The Wallace H. Coulter Department of Biomedical Engineering, Georgia Institute of Technology and Emory University School of Medicine, Atlanta, Georgia, USA
| |
Collapse
|
41
|
Chen Y, Liu J, Luo X, Luo J. ApodNet: Learning for High Frame Rate Synthetic Transmit Aperture Ultrasound Imaging. IEEE TRANSACTIONS ON MEDICAL IMAGING 2021; 40:3190-3204. [PMID: 34048340 DOI: 10.1109/tmi.2021.3084821] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Two-way dynamic focusing in synthetic transmit aperture (STA) beamforming can benefit high-quality ultrasound imaging with higher lateral spatial resolution and contrast resolution. However, STA requires the complete dataset for beamforming in a relatively low frame rate and transmit power. This paper proposes a deep-learning architecture to achieve high frame rate STA imaging with two-way dynamic focusing. The network consists of an encoder and a joint decoder. The encoder trains a set of binary weights as the apodizations of the high-frame-rate plane wave transmissions. In this respect, we term our network ApodNet. The decoder can recover the complete dataset from the acquired channel data to achieve dynamic transmit focusing. We evaluate the proposed method by simulations at different levels of noise and in-vivo experiments on the human biceps brachii and common carotid artery. The experimental results demonstrate that ApodNet provides a promising strategy for high frame rate STA imaging, obtaining comparable lateral resolution and contrast resolution with four-times higher frame rate than conventional STA imaging in the in-vivo experiments. Particularly, ApodNet improves contrast resolution of the hypoechoic targets with much shorter computational time when compared with other high-frame-rate methods in both simulations and in-vivo experiments.
Collapse
|
42
|
de Ruijter J, Muijsers JJM, van de Vosse FN, van Sambeek MRHM, Lopata RGP. A Generalized Approach for Automatic 3-D Geometry Assessment of Blood Vessels in Transverse Ultrasound Images Using Convolutional Neural Networks. IEEE TRANSACTIONS ON ULTRASONICS, FERROELECTRICS, AND FREQUENCY CONTROL 2021; 68:3326-3335. [PMID: 34143734 DOI: 10.1109/tuffc.2021.3090461] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/12/2023]
Abstract
Accurate 3-D geometries of arteries and veins are important clinical data for diagnosis of arterial disease and intervention planning. Automatic segmentation of vessels in the transverse view suffers from the low lateral resolution and contrast. Convolutional neural networks are a promising tool for automatic segmentation of medical images, outperforming the traditional segmentation methods with high robustness. In this study, we aim to create a general, robust, and accurate method to segment the lumen-wall boundary of healthy central and peripheral vessels in large field-of-view freehand ultrasound (US) datasets. Data were acquired using the freehand US, in combination with a probe tracker. A total of ±36 000 cross-sectional images, acquired in the common, internal, and external carotid artery ( N = 37 ), in the radial, ulnar artery, and cephalic vein ( N = 12 ), and in the femoral artery ( N = 5 ) were included. To create masks (of the lumen) for training data, a conventional automatic segmentation method was used. The neural networks were trained on: 1) data of all vessels and 2) the carotid artery only. The performance was compared and tested using an open-access dataset. The recall, precision, DICE, and intersection over union (IoU) were calculated. Overall, segmentation was successful in the carotid and peripheral arteries. The Multires U-net architecture performs best overall with DICE = 0.93 when trained on the total dataset. Future studies will focus on the inclusion of vascular pathologies.
Collapse
|
43
|
Automated Segmentation of Median Nerve in Dynamic Sonography Using Deep Learning: Evaluation of Model Performance. Diagnostics (Basel) 2021; 11:diagnostics11101893. [PMID: 34679591 PMCID: PMC8534332 DOI: 10.3390/diagnostics11101893] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/18/2021] [Revised: 10/01/2021] [Accepted: 10/10/2021] [Indexed: 11/21/2022] Open
Abstract
There is an emerging trend to employ dynamic sonography in the diagnosis of entrapment neuropathy, which exhibits aberrant spatiotemporal characteristics of the entrapped nerve when adjacent tissues move. However, the manual tracking of the entrapped nerve in consecutive images demands tons of human labors and impedes its popularity clinically. Here we evaluated the performance of automated median nerve segmentation in dynamic sonography using a variety of deep learning models pretrained with ImageNet, including DeepLabV3+, U-Net, FPN, and Mask-R-CNN. Dynamic ultrasound images of the median nerve at across wrist level were acquired from 52 subjects diagnosed as carpal tunnel syndrome when they moved their fingers. The videos of 16 subjects exhibiting diverse appearance and that of the remaining 36 subjects were used for model test and training, respectively. The centroid, circularity, perimeter, and cross section area of the median nerve in individual frame were automatically determined from the inferred nerve. The model performance was evaluated by the score of intersection over union (IoU) between the annotated and model-predicted data. We found that both DeepLabV3+ and Mask R-CNN predicted median nerve the best with averaged IOU scores close to 0.83, which indicates the feasibility of automated median nerve segmentation in dynamic sonography using deep learning.
Collapse
|
44
|
Lei Y, Wang T, Fu Y, Roper J, Jani AB, Liu T, Patel P, Yang X. Catheter position prediction using deep-learning-based multi-atlas registration for high-dose rate prostate brachytherapy. Med Phys 2021; 48:7261-7270. [PMID: 34480801 DOI: 10.1002/mp.15206] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/03/2021] [Revised: 07/26/2021] [Accepted: 08/28/2021] [Indexed: 12/19/2022] Open
Abstract
PURPOSE High-dose-rate (HDR) prostate brachytherapy involves treatment catheter placement, which is currently empirical and physician dependent. The lack of proper catheter placement guidance during the procedure has left the physicians to rely on a heuristic thinking-while-doing technique, which may cause large catheter placement variation and increased plan quality uncertainty. Therefore, the achievable dose distribution could not be quantified prior to the catheter placement. To overcome this challenge, we proposed a learning-based method to provide HDR catheter placement guidance for prostate cancer patients undergoing HDR brachytherapy. METHODS The proposed framework consists of deformable registration via registration network (Reg-Net), multi-atlas ranking, and catheter regression. To model the global spatial relationship among multiple organs, binary masks of the prostate and organs-at-risk are transformed into distance maps, which describe the distance of each local voxel to the organ surfaces. For a new patient, the generated distance map is used as fixed image. Reg-Net is utilized to deformably register the distance maps from multi-atlas set to match this patient's distance map and then bring catheter maps from multi-atlas to this patient via spatial transformation. Several criteria, namely prostate volume similarity, multi-organ semantic image similarity, and catheter position criteria (far from the urethra and within the partial prostate), are used for multi-atlas ranking. The top-ranked atlas' deformed catheter positions are selected as the predicted catheter positions for this patient. Finally, catheter regression is used to refine the final catheter positions. A retrospective study on 90 patients with a fivefold cross-validation scheme was used to evaluate the proposed method's feasibility. In order to investigate the impact of plan quality from the predicted catheter pattern, we optimized the source dwell position and time for both the clinical catheter pattern and predicted catheter pattern with the same optimization settings. Comparisons of clinically relevant dose volume histogram (DVH) metrics were completed. RESULTS For all patients, on average, both the clinical plan dose and predicted plan dose meet the common dose constraints when prostate dose coverage is kept at V100 = 95%. The plans from the predicted catheter pattern have slightly higher hotspot in terms of V150 by 5.0% and V200 by 2.9% on average. For bladder V75, rectum V75, and urethra V125, the average difference is close to zero, and the range of most patients is within ±1 cc. CONCLUSION We developed a new catheter placement prediction method for HDR prostate brachytherapy based on a deep-learning-based multi-atlas registration algorithm. It has great clinical potential since it can provide catheter location estimation prior to catheter placement, which could reduce the dependence on physicians' experience in catheter implantation and improve the quality of prostate HDR treatment plans. This approach merits further clinical evaluation and validation as a method of quality control for HDR prostate brachytherapy.
Collapse
Affiliation(s)
- Yang Lei
- Department of Radiation Oncology, Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
| | - Tonghe Wang
- Department of Radiation Oncology, Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
| | - Yabo Fu
- Department of Radiation Oncology, Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
| | - Justin Roper
- Department of Radiation Oncology, Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
| | - Ashesh B Jani
- Department of Radiation Oncology, Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
| | - Tian Liu
- Department of Radiation Oncology, Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
| | - Pretesh Patel
- Department of Radiation Oncology, Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
| | - Xiaofeng Yang
- Department of Radiation Oncology, Winship Cancer Institute, Emory University, Atlanta, Georgia, USA
| |
Collapse
|
45
|
|
46
|
Pham TT, Le MB, Le LH, Andersen J, Lou E. Assessment of hip displacement in children with cerebral palsy using machine learning approach. Med Biol Eng Comput 2021; 59:1877-1887. [PMID: 34357510 DOI: 10.1007/s11517-021-02416-9] [Citation(s) in RCA: 3] [Impact Index Per Article: 0.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/23/2020] [Accepted: 07/09/2021] [Indexed: 10/20/2022]
Abstract
Manual measurements of migration percentage (MP) on pelvis radiographs for assessing hip displacement are subjective and time consuming. A deep learning approach using convolution neural networks (CNNs) to automatically measure the MP was proposed. The pre-trained Inception ResNet v2 was fine tuned to detect locations of the eight reference landmarks used for MP measurements. A second network, fine-tuned MobileNetV2, was trained on the regions of interest to obtain more precise landmarks' coordinates. The MP was calculated from the final estimated landmarks' locations. A total of 122 radiographs were divided into 57 for training, 10 for validation, and 55 for testing. The mean absolute difference (MAD) and intra-class correlation coefficient (ICC [2,1]) of the comparison for the MP on 110 measurements (left and right hips) were 4.5 [Formula: see text] 4.3% (95% CI, 3.7-5.3%) and 0.91, respectively. Sensitivity and specificity were 87.8% and 93.4% for the classification of hip displacement (MP-threshold of 30%), and 63.2% and 94.5% for the classification of surgery-needed hips (MP-threshold of 40%). The prediction results were returned within 5 s. The developed fine-tuned CNNs detected the landmarks and provided automatic MP measurements with high accuracy and excellent reliability, which can assist clinicians to diagnose hip displacement in children with CP.
Collapse
Affiliation(s)
- Thanh-Tu Pham
- Department of Radiology and Diagnostic Imaging, University of Alberta, Edmonton, AB, Canada
| | - Minh-Binh Le
- Department of Radiology and Diagnostic Imaging, University of Alberta, Edmonton, AB, Canada.,Department of Computer Science, Ho Chi Minh City University of Science, Ho Chi Minh City, Vietnam
| | - Lawrence H Le
- Department of Radiology and Diagnostic Imaging, University of Alberta, Edmonton, AB, Canada
| | - John Andersen
- Department of Pediatrics, University of Alberta, Edmonton, AB, Canada
| | - Edmond Lou
- Department of Radiology and Diagnostic Imaging, University of Alberta, Edmonton, AB, Canada. .,Department of Electrical and Computer Engineering, 11-263 Donadeo Innovation Centre for Engineering, University of Alberta, 9211-116 Street, Edmonton, AB, T6G 1H9, Canada.
| |
Collapse
|
47
|
Szarski M, Chauhan S. Improved real-time segmentation of Intravascular Ultrasound images using coordinate-aware fully convolutional networks. Comput Med Imaging Graph 2021; 91:101955. [PMID: 34252744 DOI: 10.1016/j.compmedimag.2021.101955] [Citation(s) in RCA: 11] [Impact Index Per Article: 2.8] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/03/2020] [Revised: 07/01/2021] [Accepted: 07/02/2021] [Indexed: 11/26/2022]
Abstract
Segmentation of Intravascular Ultrasound (IVUS) images into Lumen and Media (interior and exterior) artery vessel walls is highly clinically relevant in the diagnosis and treatment of cardiovascular diseases such as atherosclerosis. When fused with position data, such segmentations also play a key role in reconstructing 3D representations of arteries. Automated segmentation in real-time is known to be a difficult image analysis problem, primarily due to artefacts commonly present in IVUS ultrasound images such as shadows, guide-wire effects, and side-branches. An additional challenge is the limited amount of expert labelled IVUS data, which limits the application of many well-performing deep learning models from other domains. To exploit the circular layered structure of the artery in B-Mode images, we propose a multi-class fully convolutional semantic segmentation network based on a minimal U-Net architecture augmented with learned translation dependence in the polar domain. The coordinate awareness in the multi-class segmentation allows the model to exploit relative spatial context about the interior and exterior vessel walls which are simply separable in polar coordinates. After training on 109 expert-labelled examples, our model significantly outperforms the state-of-the art in terms of mean Jaccard Measure (0.91 vs. 0.89) and Hausdorff distance (0.32 mm vs. 0.48 mm) on Media segmentation, and reaches equivalent performance on Lumen segmentation when evaluated on a standard publicly available dataset of 326 IVUS B-Mode images captured by 20 Mhz ultrasound probes. Using an order of magnitude fewer trainable parameters than the previous state-of-the-art, our model runs over 50 times faster and is able to execute in only 3 ms on a common GPU, achieving both leading accuracy and practical real-time performance.
Collapse
Affiliation(s)
- Martin Szarski
- Department of Mechanical and Aerospace Engineering, Monash University, Melbourne, Australia
| | - Sunita Chauhan
- Department of Mechanical and Aerospace Engineering, Monash University, Melbourne, Australia.
| |
Collapse
|
48
|
Bajaj R, Huang X, Kilic Y, Ramasamy A, Jain A, Ozkor M, Tufaro V, Safi H, Erdogan E, Serruys PW, Moon J, Pugliese F, Mathur A, Torii R, Baumbach A, Dijkstra J, Zhang Q, Bourantas CV. Advanced deep learning methodology for accurate, real-time segmentation of high-resolution intravascular ultrasound images. Int J Cardiol 2021; 339:185-191. [PMID: 34153412 DOI: 10.1016/j.ijcard.2021.06.030] [Citation(s) in RCA: 13] [Impact Index Per Article: 3.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 01/30/2021] [Revised: 04/10/2021] [Accepted: 06/15/2021] [Indexed: 11/29/2022]
Abstract
AIMS The aim of this study is to develop and validate a deep learning (DL) methodology capable of automated and accurate segmentation of intravascular ultrasound (IVUS) image sequences in real-time. METHODS AND RESULTS IVUS segmentation was performed by two experts who manually annotated the external elastic membrane (EEM) and lumen borders in the end-diastolic frames of 197 IVUS sequences portraying the native coronary arteries of 65 patients. The IVUS sequences of 177 randomly-selected vessels were used to train and optimise a novel DL model for the segmentation of IVUS images. Validation of the developed methodology was performed in 20 vessels using the estimations of two expert analysts as the reference standard. The mean difference for the EEM, lumen and plaque area between the DL-methodology and the analysts was ≤0.23mm2 (standard deviation ≤0.85mm2), while the Hausdorff and mean distance differences for the EEM and lumen borders was ≤0.19 mm (standard deviation≤0.17 mm). The agreement between DL and experts was similar to experts' agreement (Williams Index ranges: 0.754-1.061) with similar results in frames portraying calcific plaques or side branches. CONCLUSIONS The developed DL-methodology appears accurate and capable of segmenting high-resolution real-world IVUS datasets. These features are expected to facilitate its broad adoption and enhance the applications of IVUS in clinical practice and research.
Collapse
Affiliation(s)
- Retesh Bajaj
- Department of Cardiology, Barts Heart Centre, Barts Health NHS Trust, London, UK; Cardiovascular Devices Hub, Centre for Cardiovascular Medicine and Devices, William Harvey Research Institute, Queen Mary University of London, UK
| | - Xingru Huang
- School of Electronic Engineering and Computer Science, Queen Mary University of London, UK
| | - Yakup Kilic
- Department of Cardiology, Barts Heart Centre, Barts Health NHS Trust, London, UK
| | - Anantharaman Ramasamy
- Department of Cardiology, Barts Heart Centre, Barts Health NHS Trust, London, UK; Cardiovascular Devices Hub, Centre for Cardiovascular Medicine and Devices, William Harvey Research Institute, Queen Mary University of London, UK
| | - Ajay Jain
- Department of Cardiology, Barts Heart Centre, Barts Health NHS Trust, London, UK
| | - Mick Ozkor
- Department of Cardiology, Barts Heart Centre, Barts Health NHS Trust, London, UK
| | - Vincenzo Tufaro
- Department of Cardiology, Barts Heart Centre, Barts Health NHS Trust, London, UK
| | - Hannah Safi
- Department of Mechanical Engineering, University College London, London, UK
| | - Emrah Erdogan
- Department of Cardiology, Barts Heart Centre, Barts Health NHS Trust, London, UK
| | - Patrick W Serruys
- Faculty of Medicine, National Heart & Lung Institute, Imperial College London, UK
| | - James Moon
- Department of Cardiology, Barts Heart Centre, Barts Health NHS Trust, London, UK; Institute of Cardiovascular Sciences, University College London, London, UK
| | - Francesca Pugliese
- Department of Cardiology, Barts Heart Centre, Barts Health NHS Trust, London, UK; Cardiovascular Devices Hub, Centre for Cardiovascular Medicine and Devices, William Harvey Research Institute, Queen Mary University of London, UK
| | - Anthony Mathur
- Department of Cardiology, Barts Heart Centre, Barts Health NHS Trust, London, UK; Cardiovascular Devices Hub, Centre for Cardiovascular Medicine and Devices, William Harvey Research Institute, Queen Mary University of London, UK
| | - Ryo Torii
- Department of Mechanical Engineering, University College London, London, UK
| | - Andreas Baumbach
- Department of Cardiology, Barts Heart Centre, Barts Health NHS Trust, London, UK; Cardiovascular Devices Hub, Centre for Cardiovascular Medicine and Devices, William Harvey Research Institute, Queen Mary University of London, UK
| | - Jouke Dijkstra
- Department of Radiology, Division of Image Processing, Leiden University Medical Center, Leiden, the Netherlands
| | - Qianni Zhang
- School of Electronic Engineering and Computer Science, Queen Mary University of London, UK
| | - Christos V Bourantas
- Department of Cardiology, Barts Heart Centre, Barts Health NHS Trust, London, UK; Cardiovascular Devices Hub, Centre for Cardiovascular Medicine and Devices, William Harvey Research Institute, Queen Mary University of London, UK; Institute of Cardiovascular Sciences, University College London, London, UK.
| |
Collapse
|
49
|
Zhou Y, Chen H, Li Y, Cao X, Wang S, Shen D. Cross-Model Attention-Guided Tumor Segmentation for 3D Automated Breast Ultrasound (ABUS) Images. IEEE J Biomed Health Inform 2021; 26:301-311. [PMID: 34003755 DOI: 10.1109/jbhi.2021.3081111] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/06/2022]
Abstract
Tumor segmentation in 3D automated breast ultrasound (ABUS) plays an important role in breast disease diagnosis and surgical planning. However, automatic segmentation of tumors in 3D ABUS images is still challenging, due to the large tumor shape and size variations, and uncertain tumor locations among patients. In this paper, we develop a novel cross-model attention-guided tumor segmentation network with a hybrid loss for 3D ABUS images. Specifically, we incorporate the tumor location into a segmentation network by combining an improved 3D Mask R-CNN head into V-Net as an end-to-end architecture. Furthermore, we introduce a cross-model attention mechanism that is able to aggregate the segmentation probability map from the improved 3D Mask R-CNN to each feature extraction level in the V-Net. Then, we design a hybrid loss to balance the contribution of each part in the proposed cross-model segmentation network. We conduct extensive experiments on 170 3D ABUS from 107 patients. Experimental results show that our method outperforms other state-of-the-art methods, by achieving the Dice similarity coefficient (DSC) of 64.57%, Jaccard coefficient (JC) of 53.39%, recall (REC) of 64.43%, precision (PRE) of 74.51%, 95th Hausdorff distance (95HD) of 11.91mm, and average surface distance (ASD) of 4.63mm. Our code will be available online (https://github.com/zhouyuegithub/CMVNet).
Collapse
|
50
|
Lei Y, Wang T, Tian S, Fu Y, Patel P, Jani AB, Curran WJ, Liu T, Yang X. Male pelvic CT multi-organ segmentation using synthetic MRI-aided dual pyramid networks. Phys Med Biol 2021; 66:10.1088/1361-6560/abf2f9. [PMID: 33780918 PMCID: PMC11755409 DOI: 10.1088/1361-6560/abf2f9] [Citation(s) in RCA: 5] [Impact Index Per Article: 1.3] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/22/2020] [Accepted: 03/29/2021] [Indexed: 12/17/2022]
Abstract
The delineation of the prostate and organs-at-risk (OARs) is fundamental to prostate radiation treatment planning, but is currently labor-intensive and observer-dependent. We aimed to develop an automated computed tomography (CT)-based multi-organ (bladder, prostate, rectum, left and right femoral heads (RFHs)) segmentation method for prostate radiation therapy treatment planning. The proposed method uses synthetic MRIs (sMRIs) to offer superior soft-tissue information for male pelvic CT images. Cycle-consistent adversarial networks (CycleGAN) were used to generate CT-based sMRIs. Dual pyramid networks (DPNs) extracted features from both CTs and sMRIs. A deep attention strategy was integrated into the DPNs to select the most relevant features from both CTs and sMRIs to identify organ boundaries. The CT-based sMRI generated from our previously trained CycleGAN and its corresponding CT images were inputted to the proposed DPNs to provide complementary information for pelvic multi-organ segmentation. The proposed method was trained and evaluated using datasets from 140 patients with prostate cancer, and were then compared against state-of-art methods. The Dice similarity coefficients and mean surface distances between our results and ground truth were 0.95 ± 0.05, 1.16 ± 0.70 mm; 0.88 ± 0.08, 1.64 ± 1.26 mm; 0.90 ± 0.04, 1.27 ± 0.48 mm; 0.95 ± 0.04, 1.08 ± 1.29 mm; and 0.95 ± 0.04, 1.11 ± 1.49 mm for bladder, prostate, rectum, left and RFHs, respectively. Mean center of mass distances was within 3 mm for all organs. Our results performed significantly better than those of competing methods in most evaluation metrics. We demonstrated the feasibility of sMRI-aided DPNs for multi-organ segmentation on pelvic CT images, and its superiority over other networks. The proposed method could be used in routine prostate cancer radiotherapy treatment planning to rapidly segment the prostate and standard OARs.
Collapse
Affiliation(s)
| | | | - Sibo Tian
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, United States of America
| | - Yabo Fu
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, United States of America
| | - Pretesh Patel
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, United States of America
| | - Ashesh B Jani
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, United States of America
| | - Walter J Curran
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, United States of America
| | - Tian Liu
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, United States of America
| | - Xiaofeng Yang
- Department of Radiation Oncology and Winship Cancer Institute, Emory University, Atlanta, GA 30322, United States of America
| |
Collapse
|