1
|
Wu X, Li G, Wang X, Xu Z, Wang Y, Lei S, Xian J, Wang X, Zhang Y, Li G, Yuan K. Diagnosis assistant for liver cancer utilizing a large language model with three types of knowledge. Phys Med Biol 2025; 70:095009. [PMID: 40203862 DOI: 10.1088/1361-6560/adcb17] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/10/2024] [Accepted: 04/09/2025] [Indexed: 04/11/2025]
Abstract
Objective.Liver cancer has a high incidence rate, but experienced doctors are lacking in primary healthcare settings. The development of large models offers new possibilities for diagnosis. However, in liver cancer diagnosis, large models face certain limitations, such as insufficient understanding of specific medical images, inadequate consideration of liver vessel factors, and inaccuracies in reasoning logic. Therefore, this study proposes a diagnostic assistance tool specific to liver cancer to enhance the diagnostic capabilities of primary care doctors.Approach.A liver cancer diagnosis framework combining large and small models is proposed. A more accurate model for liver tumor segmentation and a more precise model for liver vessel segmentation are developed. The features extracted from the segmentation results of the small models are combined with the patient's medical records and then provided to the large model. The large model employs chain of thought prompts to simulate expert diagnostic reasoning and uses Retrieval-Augmented Generation to provide reliable answers based on trusted medical knowledge and cases.Main results.In the small model part, the proposed liver tumor and liver vessel segmentation methods achieve improved performance. In the large model part, this approach receives higher evaluation scores from doctors when analyzing patient imaging and medical records.Significance.First, a diagnostic framework combining small models and large models is proposed to optimize the liver cancer diagnosis process. Second, two segmentation models are introduced to compensate for the large model's shortcomings in extracting semantic information from images. Third, by simulating doctors' reasoning and integrating trusted knowledge, the framework enhances the reliability and interpretability of the large model's responses while reducing hallucination phenomena.
Collapse
Affiliation(s)
- Xuzhou Wu
- Institute of Biopharmaceutical and Health Engineering, SIGS, Tsinghua University, Beijing, People's Republic of China
| | - Guangxin Li
- Radiotherapy Department, Beijing Tsinghua Changgung Hospital, Beijing, People's Republic of China
| | - Xing Wang
- Radiotherapy Department, Beijing Tsinghua Changgung Hospital, Beijing, People's Republic of China
| | - Zeyu Xu
- Radiotherapy Department, Beijing Tsinghua Changgung Hospital, Beijing, People's Republic of China
| | - Yingni Wang
- School of Biomedical Engineering, Tsinghua University, Beijing, People's Republic of China
| | - Shuge Lei
- Department of Computer Science and Engineering, Molinaroli College of Engineering and Computing, University of South Carolina, Columbia, SC, United States of America
| | - Jianming Xian
- Institute of Biopharmaceutical and Health Engineering, SIGS, Tsinghua University, Beijing, People's Republic of China
| | - Xueyu Wang
- Institute of Biopharmaceutical and Health Engineering, SIGS, Tsinghua University, Beijing, People's Republic of China
| | - Yibao Zhang
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education/Beijing), Department of Radiation Oncology, Peking University Cancer Hospital & Institute, Beijing, People's Republic of China
| | - Gong Li
- Radiotherapy Department, Beijing Tsinghua Changgung Hospital, Beijing, People's Republic of China
| | - Kehong Yuan
- Institute of Biopharmaceutical and Health Engineering, SIGS, Tsinghua University, Beijing, People's Republic of China
| |
Collapse
|
2
|
Xu J, Dong A, Yang Y, Jin S, Zeng J, Xu Z, Jiang W, Zhang L, Dong J, Wang B. VSNet: Vessel Structure-aware Network for hepatic and portal vein segmentation. Med Image Anal 2025; 101:103458. [PMID: 39913966 DOI: 10.1016/j.media.2025.103458] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/14/2024] [Revised: 11/13/2024] [Accepted: 01/07/2025] [Indexed: 03/05/2025]
Abstract
Identifying and segmenting hepatic and portal veins (two predominant vascular systems in the liver, from CT scans) play a crucial role for clinicians in preoperative planning for treatment strategies. However, existing segmentation models often struggle to capture fine details of minor veins. In this article, we introduce Vessel Structure-aware Network (VSNet), a multi-task learning model with vessel-growing decoder, to address the challenge. VSNet excels at accurate segmentation by capturing the topological features of both minor veins while preserving correct connectivity from minor vessels to trucks. We also build and publish the largest dataset (303 cases) for hepatic and portal vessel segmentation. Through comprehensive experiments, we demonstrate that VSNet achieves the best Dice for hepatic vein of 0.824 and portal vein of 0.807 on our proposed dataset, significantly outperforming other popular segmentation models. The source code and dataset are publicly available at https://github.com/XXYZB/VSNet.
Collapse
Affiliation(s)
- Jichen Xu
- National Key Laboratory of Human-Machine Hybrid Augmented Intelligence, National Engineering Research Center for Visual Information and Applications, and Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, China
| | - Anqi Dong
- Division of Decision and Control Systems and Department of Mathematics, KTH Royal Institute of Technology, SE-100 44 Stockholm, Sweden
| | - Yang Yang
- School of Automation, Northwestern Polytechnical University, Xi'an, China
| | - Shuo Jin
- Beijing Tsinghua Changgung Hospital, School of Clinical Medicine, Tsinghua University, Beijing, China
| | - Jianping Zeng
- Beijing Tsinghua Changgung Hospital, School of Clinical Medicine, Tsinghua University, Beijing, China
| | - Zhengqing Xu
- Jingzhen Medical Technology Ltd., China; Matrix Medical Technology Ltd., China
| | - Wei Jiang
- Research Center of Artificial Intelligence of Shangluo, Shangluo University, Shangluo, China
| | - Liang Zhang
- School of Computer Science and Technology, Xidian University, Xi'an, China
| | - Jiahong Dong
- Beijing Tsinghua Changgung Hospital, School of Clinical Medicine, Tsinghua University, Beijing, China
| | - Bo Wang
- Jingzhen Medical Technology Ltd., China; Matrix Medical Technology Ltd., China; Institute of Medical Equipment Science and Engineering, Huazhong University of Science and Technology, Wuhan, China.
| |
Collapse
|
3
|
Wang H, Nakajima T, Shikano K, Nomura Y, Nakaguchi T. Diagnosis of Lung Cancer Using Endobronchial Ultrasonography Image Based on Multi-Scale Image and Multi-Feature Fusion Framework. Tomography 2025; 11:24. [PMID: 40137564 PMCID: PMC11945964 DOI: 10.3390/tomography11030024] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/16/2024] [Revised: 02/23/2025] [Accepted: 02/24/2025] [Indexed: 03/29/2025] Open
Abstract
Lung cancer is the leading cause of cancer-related deaths globally and ranks among the most common cancer types. Given its low overall five-year survival rate, early diagnosis and timely treatment are essential to improving patient outcomes. In recent years, advances in computer technology have enabled artificial intelligence to make groundbreaking progress in imaging-based lung cancer diagnosis. The primary aim of this study is to develop a computer-aided diagnosis (CAD) system for lung cancer using endobronchial ultrasonography (EBUS) images and deep learning algorithms to facilitate early detection and improve patient survival rates. We propose M3-Net, which is a multi-branch framework that integrates multiple features through an attention-based mechanism, enhancing diagnostic performance by providing more comprehensive information for lung cancer assessment. The framework was validated on a dataset of 95 patient cases, including 13 benign and 82 malignant cases. The dataset comprises 1140 EBUS images, with 540 images used for training, and 300 images each for the validation and test sets. The evaluation yielded the following results: accuracy of 0.76, F1-score of 0.75, AUC of 0.83, PPV of 0.80, NPV of 0.75, sensitivity of 0.72, and specificity of 0.80. These findings indicate that the proposed attention-based multi-feature fusion framework holds significant potential in assisting with lung cancer diagnosis.
Collapse
Affiliation(s)
- Huitao Wang
- Department of Medical Engineering, Graduate School of Science and Engineering, Chiba University, Chiba 263-8522, Japan;
| | - Takahiro Nakajima
- Department of General Thoracic Surgery, Dokkyo Medical University, Mibu 321-0293, Japan;
| | - Kohei Shikano
- Department of Respirology, Graduate School of Medicine, Chiba University, Chiba 260-8670, Japan;
| | - Yukihiro Nomura
- Center for Frontier Medical Engineering, Chiba University, Chiba 263-8522, Japan;
| | - Toshiya Nakaguchi
- Center for Frontier Medical Engineering, Chiba University, Chiba 263-8522, Japan;
| |
Collapse
|
4
|
Lim KY, Ko JE, Hwang YN, Lee SG, Kim SM. TransRAUNet: A Deep Neural Network with Reverse Attention Module Using HU Windowing Augmentation for Robust Liver Vessel Segmentation in Full Resolution of CT Images. Diagnostics (Basel) 2025; 15:118. [PMID: 39857002 PMCID: PMC11764155 DOI: 10.3390/diagnostics15020118] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/14/2024] [Revised: 12/30/2024] [Accepted: 01/03/2025] [Indexed: 01/27/2025] Open
Abstract
Background: Liver cancer has a high mortality rate worldwide, and clinicians segment liver vessels in CT images before surgical procedures. However, liver vessels have a complex structure, and the segmentation process is conducted manually, so it is time-consuming and labor-intensive. Consequently, it would be extremely useful to develop a deep learning-based automatic liver vessel segmentation method. Method: As a segmentation method, UNet is widely used as a baseline, and a multi-scale block or attention module has been introduced to extract context information. In recent machine learning efforts, not only has the global context extraction been improved by introducing Transformer, but a method to reinforce the edge area has been proposed. However, the data preprocessing step still commonly uses general augmentation methods, such as flip, rotation, and mirroring, so it does not perform robustly on images of varying brightness or contrast levels. We propose a method of applying image augmentation with different HU windowing values. In addition, to minimize the false negative area, we propose TransRAUNet, which introduces a reverse attention module (RAM) that can focus edge information to the baseline TransUNet. The proposed architecture solves context loss for small vessels by applying edge module (RAM) in the upsampling phase. It can also generate semantic feature maps that allows it to learn edge, global context, and detail location by combining high-level edge and low-level context features. Results: In the 3Dricadb dataset, the proposed model achieved a DSC of 0.948 and a sensitivity of 0.944 in liver vessel segmentation. This study demonstrated that the proposed augmentation method is effective and robust by comparisons with the model without augmentation and with the general augmentation method. Additionally, an ablation study showed that RAM has improved segmentation performance compared to TransUNet. Compared to prevailing state-of-the-art methods, the proposed model showed the best performance for liver vessel segmentation. Conclusions: TransRAUnet is expected to serve as a navigation aid for liver resection surgery through accurate liver vessel and tumor segmentation.
Collapse
Affiliation(s)
- Kyoung Yoon Lim
- Department of Medical Device and Healthcare, Dongguk University, Seoul 04620, Republic of Korea; (K.Y.L.); (Y.N.H.); (S.G.L.)
| | - Jae Eun Ko
- Department of Regulatory Science for Medical Device, Dongguk University, Seoul 04620, Republic of Korea;
| | - Yoo Na Hwang
- Department of Medical Device and Healthcare, Dongguk University, Seoul 04620, Republic of Korea; (K.Y.L.); (Y.N.H.); (S.G.L.)
| | - Sang Goo Lee
- Department of Medical Device and Healthcare, Dongguk University, Seoul 04620, Republic of Korea; (K.Y.L.); (Y.N.H.); (S.G.L.)
| | - Sung Min Kim
- Department of Medical Device and Healthcare, Dongguk University, Seoul 04620, Republic of Korea; (K.Y.L.); (Y.N.H.); (S.G.L.)
- Department of Regulatory Science for Medical Device, Dongguk University, Seoul 04620, Republic of Korea;
| |
Collapse
|
5
|
Zhang Q, Li J, Nan X, Zhang X. Detail-sensitive 3D-UNet for pulmonary airway segmentation from CT images. Med Biol Eng Comput 2024; 62:3749-3762. [PMID: 39017831 DOI: 10.1007/s11517-024-03169-x] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/03/2024] [Accepted: 07/06/2024] [Indexed: 07/18/2024]
Abstract
The segmentation of airway from computed tomography (CT) images plays a vital role in pulmonary disease diagnosis, evaluation, surgical planning, and treatment. Nevertheless, it is still challenging for current methods to handle distal thin and low-contrast airways, leading to mis-segmentation issues. This paper proposes a detail-sensitive 3D-UNet (DS-3D-UNet) that incorporates two new modules into 3D-UNet to segment airways accurately from CT images. The feature recalibration module is designed to give more attention to the foreground airway features through a new attention mechanism. The detail extractor module aims to restore multi-scale detailed features by fusion of features at different levels. Extensive experiments were conducted on the ATM'22 challenge dataset composed of 300 CT scans with airway annotations to evaluate its performance. Quantitative comparisons prove that the proposed model achieves the best performance in terms of Dice similarity coefficient (92.6%) and Intersection over Union (86.3%), outperforming other state-of-the-art methods. Qualitative comparisons further exhibit the superior performance of our method in segmenting thin and confused distal bronchi. The proposed model could provide important references for the diagnosis and treatment of pulmonary diseases, holding promising prospects in the field of digital medicine. Codes are available at https://github.com/nighlevil/DS-3D-UNet/tree/master .
Collapse
Affiliation(s)
- Qin Zhang
- School of Computer Science and Engineering, South China University of Technology, Guangzhou, Guangdong, 510000, China
| | - Jiajie Li
- School of Computer Science and Engineering, South China University of Technology, Guangzhou, Guangdong, 510000, China
| | - Xiangling Nan
- School of Computer Science and Engineering, South China University of Technology, Guangzhou, Guangdong, 510000, China
| | - Xiaodong Zhang
- Shenzhen Children's Hospital, Shenzhen, Guangdong, 518000, China.
- Shenzhen Institute of Advanced Technology, Chinese Academy of Sciences, Shenzhen, Guangdong, 518000, China.
| |
Collapse
|
6
|
Dai S, Liu X, Wei W, Yin X, Qiao L, Wang J, Zhang Y, Hou Y. A multi-scale, multi-task fusion UNet model for accurate breast tumor segmentation. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2024; 258:108484. [PMID: 39531807 DOI: 10.1016/j.cmpb.2024.108484] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/19/2023] [Revised: 10/12/2024] [Accepted: 10/29/2024] [Indexed: 11/16/2024]
Abstract
BACKGROUND AND OBJECTIVE Breast cancer is the most common cancer type among women worldwide and a leading cause of female death. Accurately interpreting these complex tumors, involving small size and morphology, requires a significant amount of expertise and time. Developing a breast tumor segmentation model to assist clinicians in treatment, therefore, holds great practical significance. METHODS We propose a multi-scale, multi-task model framework named MTF-UNet. Firstly, we differ from the common approach of using different convolution kernel sizes to extract multi-scale features, and instead use the same convolution kernel size with different numbers of convolutions to obtain multi-scale, multi-level features. Additionally, to better integrate features from different levels and sizes, we extract a new multi-branch feature fusion block (ADF). This block differs from using channel and spatial attention to fuse features, but considers fusion weights between various branches. Secondly, we propose to use the number of pixels predicted to be related to tumors and background to assist segmentation, which is different from the conventional approach of using classification tasks to assist segmentation. RESULTS We conducted extensive experiments on our proprietary DCE-MRI dataset, as well as two public datasets (BUSI and Kvasir-SEG). In the aforementioned datasets, our model achieved excellent MIoU scores of 90.4516%, 89.8408%, and 92.8431% on the respective test sets. Furthermore, our ablation study has demonstrated the efficacy of each component and the effective integration of our auxiliary prediction branch into other models. CONCLUSION Through comprehensive experiments and comparisons with other algorithms, the effectiveness, adaptability, and robustness of our proposed method have been demonstrated. We believe that MTF-UNet has great potential for further development in the field of medical image segmentation. The relevant code and data can be found at https://github.com/LCUDai/MTF-UNet.git.
Collapse
Affiliation(s)
- Shuo Dai
- School of Mathematics Science, Liaocheng University, Liaocheng, Shandong, 252000, China
| | - Xueyan Liu
- School of Mathematics Science, Liaocheng University, Liaocheng, Shandong, 252000, China.
| | - Wei Wei
- School of Electronics and information, Xi'An Polytechnic University, Xian, Shanxi, 710600, China
| | - Xiaoping Yin
- Department of Radiology, Affiliated Hospital of Hebei University, No. 212 of Yuhua Road, Lianchi District, Baoding, 071000, China
| | - Lishan Qiao
- School of Computer Science and Technology, Shandong Jianzhu University, Jinan, Shandong 250101, China
| | - Jianing Wang
- Department of Radiology, Affiliated Hospital of Hebei University, No. 212 of Yuhua Road, Lianchi District, Baoding, 071000, China
| | - Yu Zhang
- Department of Radiology, Affiliated Hospital of Hebei University, No. 212 of Yuhua Road, Lianchi District, Baoding, 071000, China
| | - Yan Hou
- Department of Radiology, Affiliated Hospital of Hebei University, No. 212 of Yuhua Road, Lianchi District, Baoding, 071000, China
| |
Collapse
|
7
|
Fu W, Hu H, Li X, Guo R, Chen T, Qian X. A Generalizable Causal-Invariance-Driven Segmentation Model for Peripancreatic Vessels. IEEE TRANSACTIONS ON MEDICAL IMAGING 2024; 43:3794-3806. [PMID: 38739508 DOI: 10.1109/tmi.2024.3400528] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 05/16/2024]
Abstract
Segmenting peripancreatic vessels in CT, including the superior mesenteric artery (SMA), the coeliac artery (CA), and the partial portal venous system (PPVS), is crucial for preoperative resectability analysis in pancreatic cancer. However, the clinical applicability of vessel segmentation methods is impeded by the low generalizability on multi-center data, mainly attributed to the wide variations in image appearance, namely the spurious correlation factor. Therefore, we propose a causal-invariance-driven generalizable segmentation model for peripancreatic vessels. It incorporates interventions at both image and feature levels to guide the model to capture causal information by enforcing consistency across datasets, thus enhancing the generalization performance. Specifically, firstly, a contrast-driven image intervention strategy is proposed to construct image-level interventions by generating images with various contrast-related appearances and seeking invariant causal features. Secondly, the feature intervention strategy is designed, where various patterns of feature bias across different centers are simulated to pursue invariant prediction. The proposed model achieved high DSC scores (79.69%, 82.62%, and 83.10%) for the three vessels on a cross-validation set containing 134 cases. Its generalizability was further confirmed on three independent test sets of 233 cases. Overall, the proposed method provides an accurate and generalizable segmentation model for peripancreatic vessels and offers a promising paradigm for increasing the generalizability of segmentation models from a causality perspective. Our source codes will be released at https://github.com/ SJTUBME-QianLab/PC_VesselSeg.
Collapse
|
8
|
Delmoral JC, R S Tavares JM. Semantic Segmentation of CT Liver Structures: A Systematic Review of Recent Trends and Bibliometric Analysis : Neural Network-based Methods for Liver Semantic Segmentation. J Med Syst 2024; 48:97. [PMID: 39400739 PMCID: PMC11473507 DOI: 10.1007/s10916-024-02115-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2024] [Accepted: 10/02/2024] [Indexed: 10/15/2024]
Abstract
The use of artificial intelligence (AI) in the segmentation of liver structures in medical images has become a popular research focus in the past half-decade. The performance of AI tools in screening for this task may vary widely and has been tested in the literature in various datasets. However, no scientometric report has provided a systematic overview of this scientific area. This article presents a systematic and bibliometric review of recent advances in neuronal network modeling approaches, mainly of deep learning, to outline the multiple research directions of the field in terms of algorithmic features. Therefore, a detailed systematic review of the most relevant publications addressing fully automatic semantic segmenting liver structures in Computed Tomography (CT) images in terms of algorithm modeling objective, performance benchmark, and model complexity is provided. The review suggests that fully automatic hybrid 2D and 3D networks are the top performers in the semantic segmentation of the liver. In the case of liver tumor and vasculature segmentation, fully automatic generative approaches perform best. However, the reported performance benchmark indicates that there is still much to be improved in segmenting such small structures in high-resolution abdominal CT scans.
Collapse
Affiliation(s)
- Jessica C Delmoral
- Instituto de Ciência e Inovação em Engenharia Mecânica e Engenharia Industrial, Faculdade de Engenharia, Universidade do Porto, Rua Dr. Roberto Frias, s/n, 4200-465, Porto, Portugal
| | - João Manuel R S Tavares
- Instituto de Ciência e Inovação em Engenharia Mecânica e Engenharia Industrial, Departamento de Engenharia Mecânica, Faculdade de Engenharia, Universidade do Porto, Rua Dr. Roberto Frias, s/n, 4200-465, Porto, Portugal.
| |
Collapse
|
9
|
Li H, Hussin N, He D, Geng Z, Li S. Design of image segmentation model based on residual connection and feature fusion. PLoS One 2024; 19:e0309434. [PMID: 39361568 PMCID: PMC11449362 DOI: 10.1371/journal.pone.0309434] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/29/2024] [Accepted: 08/12/2024] [Indexed: 10/05/2024] Open
Abstract
With the development of deep learning technology, convolutional neural networks have made great progress in the field of image segmentation. However, for complex scenes and multi-scale target images, the existing technologies are still unable to achieve effective image segmentation. In view of this, an image segmentation model based on residual connection and feature fusion is proposed. The model makes comprehensive use of the deep feature extraction ability of residual connections and the multi-scale feature integration ability of feature fusion. In order to solve the problem of background complexity and information loss in traditional image segmentation, experiments were carried out on two publicly available data sets. The results showed that in the ISPRS Vaihingen dataset and the Caltech UCSD Birds200 dataset, when the model completed the 56th and 84th iterations, respectively, the average accuracy of FRes-MFDNN was the highest, which was 97.89% and 98.24%, respectively. In the ISPRS Vaihingen dataset and the Caltech UCSD Birds200 dataset, when the system model ran to 0.20s and 0.26s, the F1 value of the FRes-MFDNN method was the largest, and the F1 value approached 100% infinitely. The FRes-MFDNN segmented four images in the ISPRS Vaihingen dataset, and the segmentation accuracy of images 1, 2, 3 and 4 were 91.44%, 92.12%, 94.02% and 91.41%, respectively. In practical applications, the MSRF-Net method, LBN-AA-SPN method, ARG-Otsu method, and FRes-MFDNN were used to segment unlabeled bird images. The results showed that the FRes-MFDNN was more complete in details, and the overall effect was significantly better than the other three models. Meanwhile, in ordinary scene images, although there was a certain degree of noise and occlusion, the model still accurately recognized and segmented the main bird images. The results show that compared with the traditional model, after FRes-MFDNN segmentation, the completeness, detail, and spatial continuity of pixels have been significantly improved, making it more suitable for complex scenes.
Collapse
Affiliation(s)
- Hong Li
- School of Information Engineering, Pingdingshan University, Pingdingshan, China
- Faculty of Engineering, Built Environment and Information Technology, SEGi University, Kota Damansara, Malaysia
| | - Norriza Hussin
- Faculty of Engineering, Built Environment and Information Technology, SEGi University, Kota Damansara, Malaysia
| | - Dandan He
- School of Information Engineering, Pingdingshan University, Pingdingshan, China
- Faculty of Engineering, Built Environment and Information Technology, SEGi University, Kota Damansara, Malaysia
| | - Zexun Geng
- School of Information Engineering, Pingdingshan University, Pingdingshan, China
| | - Shengpu Li
- School of Information Engineering, Pingdingshan University, Pingdingshan, China
| |
Collapse
|
10
|
Chierici A, Lareyre F, Salucki B, Iannelli A, Delingette H, Raffort J. Vascular liver segmentation: a narrative review on methods and new insights brought by artificial intelligence. J Int Med Res 2024; 52:3000605241263170. [PMID: 39291427 PMCID: PMC11418557 DOI: 10.1177/03000605241263170] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/21/2024] [Accepted: 05/28/2024] [Indexed: 09/19/2024] Open
Abstract
Liver vessel segmentation from routinely performed medical imaging is a useful tool for diagnosis, treatment planning and delivery, and prognosis evaluation for many diseases, particularly liver cancer. A precise representation of liver anatomy is crucial to define the extent of the disease and, when suitable, the consequent resective or ablative procedure, in order to guarantee a radical treatment without sacrificing an excessive volume of healthy liver. Once mainly performed manually, with notable cost in terms of time and human energies, vessel segmentation is currently realized through the application of artificial intelligence (AI), which has gained increased interest and development of the field. Many different AI-driven models adopted for this aim have been described and can be grouped into different categories: thresholding methods, edge- and region-based methods, model-based methods, and machine learning models. The latter includes neural network and deep learning models that now represent the principal algorithms exploited for vessel segmentation. The present narrative review describes how liver vessel segmentation can be realized through AI models, with a summary of model results in terms of accuracy, and an overview on the future progress of this topic.
Collapse
Affiliation(s)
- Andrea Chierici
- Department of Digestive Surgery, Hospital of Antibes Juan-les-Pins, Antibes, France
- Department of Digestive Surgery, University Hospital of Nice, Nice, France
- Université Côte d'Azur, Inria, Epione Team, Sophia Antipolis, France
| | - Fabien Lareyre
- Department of Vascular Surgery, Hospital of Antibes Juan-les-Pins, Antibes, France
- Université Côte d'Azur, Inserm U1065, C3M, Nice, France
| | - Benjamin Salucki
- Department of Digestive Surgery, Hospital of Antibes Juan-les-Pins, Antibes, France
| | - Antonio Iannelli
- Université Côte d'Azur, Inserm U1065, Team 8 “Hepatic complications of obesity and alcohol”, Nice, France
- ADIPOCIBLE Study Group, Université Côte d'Azur, Nice, France
| | - Hervé Delingette
- Université Côte d'Azur, Inria, Epione Team, Sophia Antipolis, France
| | - Juliette Raffort
- Université Côte d'Azur, Inserm U1065, C3M, Nice, France
- Clinical Chemistry Laboratory, University Hospital of Nice, Nice, France
- 3IA Institute, Université Côte d'Azur, Nice, France
| |
Collapse
|
11
|
Xu J, Jiang W, Wu J, Zhang W, Zhu Z, Xin J, Zheng N, Wang B. Hepatic and portal vein segmentation with dual-stream deep neural network. Med Phys 2024; 51:5441-5456. [PMID: 38648676 DOI: 10.1002/mp.17090] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/28/2023] [Revised: 02/16/2024] [Accepted: 03/01/2024] [Indexed: 04/25/2024] Open
Abstract
BACKGROUND Liver lesions mainly occur inside the liver parenchyma, which are difficult to locate and have complicated relationships with essential vessels. Thus, preoperative planning is crucial for the resection of liver lesions. Accurate segmentation of the hepatic and portal veins (PVs) on computed tomography (CT) images is of great importance for preoperative planning. However, manually labeling the mask of vessels is laborious and time-consuming, and the labeling results of different clinicians are prone to inconsistencies. Hence, developing an automatic segmentation algorithm for hepatic and PVs on CT images has attracted the attention of researchers. Unfortunately, existing deep learning based automatic segmentation methods are prone to misclassifying peripheral vessels into wrong categories. PURPOSE This study aims to provide a fully automatic and robust semantic segmentation algorithm for hepatic and PVs, guiding subsequent preoperative planning. In addition, to address the deficiency of the public dataset for hepatic and PV segmentation, we revise the annotations of the Medical Segmentation Decathlon (MSD) hepatic vessel segmentation dataset and add the masks of the hepatic veins (HVs) and PVs. METHODS We proposed a structure with a dual-stream encoder combining convolution and Transformer block, named Dual-stream Hepatic Portal Vein segmentation Network, to extract local features and long-distance spatial information, thereby extracting anatomical information of hepatic and portal vein, avoiding misdivisions of adjacent peripheral vessels. Besides, a multi-scale feature fusion block based on dilated convolution is proposed to extract multi-scale features on expanded perception fields for local features, and a multi-level fusing attention module is introduced for efficient context information extraction. Paired t-test is conducted to evaluate the significant difference in dice between the proposed methods and the comparing methods. RESULTS Two datasets are constructed from the original MSD dataset. For each dataset, 50 cases are randomly selected for model evaluation in the scheme of 5-fold cross-validation. The results show that our method outperforms the state-of-the-art Convolutional Neural Network-based and transformer-based methods. Specifically, for the first dataset, our model reaches 0.815, 0.830, and 0.807 at overall dice, precision, and sensitivity. The dice of the hepatic and PVs are 0.835 and 0.796, which also exceed the numeric result of the comparing methods. Almost all the p-values of paired t-tests on the proposed approach and comparing approaches are smaller than 0.05. On the second dataset, the proposed algorithm achieves 0.749, 0.762, 0.726, 0.835, and 0.796 for overall dice, precision, sensitivity, dice for HV, and dice for PV, among which the first four numeric results exceed comparing methods. CONCLUSIONS The proposed method is effective in solving the problem of misclassifying interlaced peripheral veins for the HV and PV segmentation task and outperforming the comparing methods on the relabeled dataset.
Collapse
Affiliation(s)
- Jichen Xu
- National Key Laboratory of Human-Machine Hybrid Augmented Intelligence, National Engineering Research Center for Visual Information and Applications, and Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, China
| | - Wei Jiang
- Research Center of Artificial Intelligence of Shangluo, Shangluo University, Shangluo, China
| | - Jiayi Wu
- National Key Laboratory of Human-Machine Hybrid Augmented Intelligence, National Engineering Research Center for Visual Information and Applications, and Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, China
| | - Wei Zhang
- Beijing Jingzhen Medical Technology Ltd., Beijing, China
- Xi'an Zhizhenzhineng Technology Ltd., Xi'an, China
- School of Telecommunications Engineering, Xidian University, Xi'an, China
| | - Zhenyu Zhu
- Hepatobiliary Surgery Center, The Fifth Medical Center of PLA General Hospital, Beijing, China
| | - Jingmin Xin
- National Key Laboratory of Human-Machine Hybrid Augmented Intelligence, National Engineering Research Center for Visual Information and Applications, and Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, China
| | - Nanning Zheng
- National Key Laboratory of Human-Machine Hybrid Augmented Intelligence, National Engineering Research Center for Visual Information and Applications, and Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, China
| | - Bo Wang
- Beijing Jingzhen Medical Technology Ltd., Beijing, China
- Xi'an Zhizhenzhineng Technology Ltd., Xi'an, China
- Huazhong University of Science and Technology, the Institute of Medical Equipment Science and Engineering, Wuhan, China
| |
Collapse
|
12
|
Chen W, Zhao L, Bian R, Li Q, Zhao X, Zhang M. Compensation of small data with large filters for accurate liver vessel segmentation from contrast-enhanced CT images. BMC Med Imaging 2024; 24:129. [PMID: 38822274 PMCID: PMC11143594 DOI: 10.1186/s12880-024-01309-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/14/2023] [Accepted: 05/27/2024] [Indexed: 06/02/2024] Open
Abstract
BACKGROUND Segmenting liver vessels from contrast-enhanced computed tomography images is essential for diagnosing liver diseases, planning surgeries and delivering radiotherapy. Nevertheless, identifying vessels is a challenging task due to the tiny cross-sectional areas occupied by vessels, which has posed great challenges for vessel segmentation, such as limited features to be learned and difficult to construct high-quality as well as large-volume data. METHODS We present an approach that only requires a few labeled vessels but delivers significantly improved results. Our model starts with vessel enhancement by fading out liver intensity and generates candidate vessels by a classifier fed with a large number of image filters. Afterwards, the initial segmentation is refined using Markov random fields. RESULTS In experiments on the well-known dataset 3D-IRCADb, the averaged Dice coefficient is lifted to 0.63, and the mean sensitivity is increased to 0.71. These results are significantly better than those obtained from existing machine-learning approaches and comparable to those generated from deep-learning models. CONCLUSION Sophisticated integration of a large number of filters is able to pinpoint effective features from liver images that are sufficient to distinguish vessels from other liver tissues under a scarcity of large-volume labeled data. The study can shed light on medical image segmentation, especially for those without sufficient data.
Collapse
Affiliation(s)
- Wen Chen
- Department of Medical Imaging, The First Affiliated Hospital of Xi'an Jiaotong University, Xi'an, China
- Taihe Hospital, Hubei University of Medicine, Shiyan, China
| | - Liang Zhao
- Taihe Hospital, Hubei University of Medicine, Shiyan, China.
| | - Rongrong Bian
- Taihe Hospital, Hubei University of Medicine, Shiyan, China
| | - Qingzhou Li
- Taihe Hospital, Hubei University of Medicine, Shiyan, China
| | - Xueting Zhao
- Taihe Hospital, Hubei University of Medicine, Shiyan, China
| | - Ming Zhang
- Department of Medical Imaging, The First Affiliated Hospital of Xi'an Jiaotong University, Xi'an, China.
| |
Collapse
|
13
|
Zhou Y, Zheng Y, Tian Y, Bai Y, Cai N, Wang P. SCAN: sequence-based context-aware association network for hepatic vessel segmentation. Med Biol Eng Comput 2024; 62:817-827. [PMID: 38032458 DOI: 10.1007/s11517-023-02975-z] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/29/2023] [Accepted: 11/22/2023] [Indexed: 12/01/2023]
Abstract
Accurate segmentation of hepatic vessel is significant for the surgeons to design the preoperative planning of liver surgery. In this paper, a sequence-based context-aware association network (SCAN) is designed for hepatic vessel segmentation, in which three schemes are incorporated to simultaneously extract the 2D features of hepatic vessels and capture the correlations between adjacent CT slices. The two schemes of slice-level attention module and graph association module are designed to bridge feature gaps between the encoder and the decoder in the low- and high-dimensional spaces. The region-edge constrained loss is designed to well optimize the proposed SCAN, which integrates cross-entropy loss, dice loss, and edge-constrained loss. Experimental results indicate that the proposed SCAN is superior to several existing deep learning frameworks, in terms of 0.845 DSC, 0.856 precision, 0.866 sensitivity, and 0.861 F1-score.
Collapse
Affiliation(s)
- Yinghong Zhou
- School of Information Engineering, Guangdong University of Technology, Guangzhou, China
| | - Yu Zheng
- School of Information Engineering, Guangdong University of Technology, Guangzhou, China
| | - Yinfeng Tian
- School of Information Engineering, Guangdong University of Technology, Guangzhou, China
| | - Youfang Bai
- School of Information Engineering, Guangdong University of Technology, Guangzhou, China
| | - Nian Cai
- School of Information Engineering, Guangdong University of Technology, Guangzhou, China.
| | - Ping Wang
- Department of Hepatobiliary Surgery in the First Affiliated Hospital of Guangzhou Medical University, Guangzhou, China.
| |
Collapse
|
14
|
Wang J, Liang J, Xiao Y, Zhou JT, Fang Z, Yang F. TaiChiNet: Negative-Positive Cross-Attention Network for Breast Lesion Segmentation in Ultrasound Images. IEEE J Biomed Health Inform 2024; 28:1516-1527. [PMID: 38206781 DOI: 10.1109/jbhi.2024.3352984] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/13/2024]
Abstract
Breast lesion segmentation in ultrasound images is essential for computer-aided breast-cancer diagnosis. To improve the segmentation performance, most approaches design sophisticated deep-learning models by mining the patterns of foreground lesions and normal backgrounds simultaneously or by unilaterally enhancing foreground lesions via various focal losses. However, the potential of normal backgrounds is underutilized, which could reduce false positives by compacting the feature representation of all normal backgrounds. From a novel viewpoint of bilateral enhancement, we propose a negative-positive cross-attention network to concentrate on normal backgrounds and foreground lesions, respectively. Derived from the complementing opposites of bipolarity in TaiChi, the network is denoted as TaiChiNet, which consists of the negative normal-background and positive foreground-lesion paths. To transmit the information across the two paths, a cross-attention module, a complementary MLP-head, and a complementary loss are built for deep-layer features, shallow-layer features, and mutual-learning supervision, separately. To the best of our knowledge, this is the first work to formulate breast lesion segmentation as a mutual supervision task from the foreground-lesion and normal-background views. Experimental results have demonstrated the effectiveness of TaiChiNet on two breast lesion segmentation datasets with a lightweight architecture. Furthermore, extensive experiments on the thyroid nodule segmentation and retinal optic cup/disc segmentation datasets indicate the application potential of TaiChiNet.
Collapse
|
15
|
Chen S, Fan J, Ding Y, Geng H, Ai D, Xiao D, Song H, Wang Y, Yang J. PEA-Net: A progressive edge information aggregation network for vessel segmentation. Comput Biol Med 2024; 169:107766. [PMID: 38150885 DOI: 10.1016/j.compbiomed.2023.107766] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/11/2023] [Revised: 10/18/2023] [Accepted: 11/21/2023] [Indexed: 12/29/2023]
Abstract
Automatic vessel segmentation is a critical area of research in medical image analysis, as it can greatly assist doctors in accurately and efficiently diagnosing vascular diseases. However, accurately extracting the complete vessel structure from images remains a challenge due to issues such as uneven contrast and background noise. Existing methods primarily focus on segmenting individual pixels and often fail to consider vessel features and morphology. As a result, these methods often produce fragmented results and misidentify vessel-like background noise, leading to missing and outlier points in the overall segmentation. To address these issues, this paper proposes a novel approach called the progressive edge information aggregation network for vessel segmentation (PEA-Net). The proposed method consists of several key components. First, a dual-stream receptive field encoder (DRE) is introduced to preserve fine structural features and mitigate false positive predictions caused by background noise. This is achieved by combining vessel morphological features obtained from different receptive field sizes. Second, a progressive complementary fusion (PCF) module is designed to enhance fine vessel detection and improve connectivity. This module complements the decoding path by combining features from previous iterations and the DRE, incorporating nonsalient information. Additionally, segmentation-edge decoupling enhancement (SDE) modules are employed as decoders to integrate upsampling features with nonsalient information provided by the PCF. This integration enhances both edge and segmentation information. The features in the skip connection and decoding path are iteratively updated to progressively aggregate fine structure information, thereby optimizing segmentation results and reducing topological disconnections. Experimental results on multiple datasets demonstrate that the proposed PEA-Net model and strategy achieve optimal performance in both pixel-level and topology-level metrics.
Collapse
Affiliation(s)
- Sigeng Chen
- Laboratory of Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Photonics, Beijing Institute of Technology, Beijing, 100081, China
| | - Jingfan Fan
- Laboratory of Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Photonics, Beijing Institute of Technology, Beijing, 100081, China.
| | - Yang Ding
- Laboratory of Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Photonics, Beijing Institute of Technology, Beijing, 100081, China
| | - Haixiao Geng
- Laboratory of Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Photonics, Beijing Institute of Technology, Beijing, 100081, China
| | - Danni Ai
- Laboratory of Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Photonics, Beijing Institute of Technology, Beijing, 100081, China
| | - Deqiang Xiao
- Laboratory of Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Photonics, Beijing Institute of Technology, Beijing, 100081, China
| | - Hong Song
- School of Computer Science and Technology, Beijing Institute of Technology, Beijing, 100081, China
| | - Yining Wang
- Department of Radiology, Peking Union Medical College Hospital, Chinese Academy of Medical Sciences and Peking Union Medical College, Beijing, 100730, China.
| | - Jian Yang
- Laboratory of Beijing Engineering Research Center of Mixed Reality and Advanced Display, School of Optics and Photonics, Beijing Institute of Technology, Beijing, 100081, China.
| |
Collapse
|
16
|
Rong Z. Optimization of table tennis target detection algorithm guided by multi-scale feature fusion of deep learning. Sci Rep 2024; 14:1401. [PMID: 38228726 PMCID: PMC10792085 DOI: 10.1038/s41598-024-51865-3] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/27/2023] [Accepted: 01/10/2024] [Indexed: 01/18/2024] Open
Abstract
This paper aims to propose a table tennis target detection (TD) method based on deep learning (DL) and multi-scale feature fusion (MFF) to improve the detection accuracy of the ball in table tennis competition, optimize the training process of athletes, and improve the technical level. In this paper, DL technology is used to improve the accuracy of table tennis TD through MFF guidance. Initially, based on the FAST Region-based Convolutional Neural Network (FAST R-CNN), the TD is carried out in the table tennis match. Then, through the method of MFF guidance, different levels of feature information are fused, which improves the accuracy of TD. Through the experimental verification on the test set, it is found that the mean Average Precision (mAP) value of the target detection algorithm (TDA) proposed here reaches 87.3%, which is obviously superior to other TDAs and has higher robustness. The DL TDA combined with the proposed MFF can be applied to various detection fields and can help the application of TD in real life.
Collapse
Affiliation(s)
- Zhang Rong
- Shaanxi Energy Institute, Xi'an, 71000, Shaanxi, China.
| |
Collapse
|
17
|
Xu B, Yang J, Hong P, Fan X, Sun Y, Zhang L, Yang B, Xu L, Avolio A. Coronary artery segmentation in CCTA images based on multi-scale feature learning. JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY 2024; 32:973-991. [PMID: 38943423 DOI: 10.3233/xst-240093] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 07/01/2024]
Abstract
BACKGROUND Coronary artery segmentation is a prerequisite in computer-aided diagnosis of Coronary Artery Disease (CAD). However, segmentation of coronary arteries in Coronary Computed Tomography Angiography (CCTA) images faces several challenges. The current segmentation approaches are unable to effectively address these challenges and existing problems such as the need for manual interaction or low segmentation accuracy. OBJECTIVE A Multi-scale Feature Learning and Rectification (MFLR) network is proposed to tackle the challenges and achieve automatic and accurate segmentation of coronary arteries. METHODS The MFLR network introduces a multi-scale feature extraction module in the encoder to effectively capture contextual information under different receptive fields. In the decoder, a feature correction and fusion module is proposed, which employs high-level features containing multi-scale information to correct and guide low-level features, achieving fusion between the two-level features to further improve segmentation performance. RESULTS The MFLR network achieved the best performance on the dice similarity coefficient, Jaccard index, Recall, F1-score, and 95% Hausdorff distance, for both in-house and public datasets. CONCLUSION Experimental results demonstrate the superiority and good generalization ability of the MFLR approach. This study contributes to the accurate diagnosis and treatment of CAD, and it also informs other segmentation applications in medicine.
Collapse
Affiliation(s)
- Bu Xu
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
| | - Jinzhong Yang
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
| | - Peng Hong
- Software College, Northeastern University, Shenyang, China
| | - Xiaoxue Fan
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
| | - Yu Sun
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
- Department of Radiology, General Hospital of North Theater Command, Shenyang, China
| | - Libo Zhang
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
- Department of Radiology, General Hospital of North Theater Command, Shenyang, China
| | - Benqiang Yang
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
- Department of Radiology, General Hospital of North Theater Command, Shenyang, China
| | - Lisheng Xu
- College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
- Key Laboratory of Medical Image Computing, Ministry of Education, Shenyang, China
- Engineering Research Center of Medical Imaging and Intelligent Analysis, Ministry of Education, Shenyang, China
| | - Alberto Avolio
- Macquarie Medical School, Faculty of Medicine, Health and Human Sciences, Macquarie University, Sydney, Australia
| |
Collapse
|
18
|
Zhang T, Yang F, Zhang P. Progress and clinical translation in hepatocellular carcinoma of deep learning in hepatic vascular segmentation. Digit Health 2024; 10:20552076241293498. [PMID: 39502486 PMCID: PMC11536605 DOI: 10.1177/20552076241293498] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/06/2024] [Accepted: 10/03/2024] [Indexed: 11/08/2024] Open
Abstract
This paper reviews the advancements in deep learning for hepatic vascular segmentation and its clinical implications in the holistic management of hepatocellular carcinoma (HCC). The key to the diagnosis and treatment of HCC lies in imaging examinations, with the challenge in liver surgery being the precise assessment of Hepatic vasculature. In this regard, deep learning methods, including convolutional neural networksamong various other approaches, have significantly improved accuracy and speed. The review synthesizes findings from 30 studies, covering aspects such as network architectures, applications, supervision techniques, evaluation metrics, and motivations. Furthermore, we also examine the challenges and future prospects of deep learning technologies in enhancing the comprehensive diagnosis and treatment of HCC, discussing anticipated breakthroughs that could transform patient management. By combining clinical needs with technological advancements, deep learning is expected to make greater breakthroughs in the field of hepatic vascular segmentation, thereby providing stronger support for the diagnosis and treatment of HCC.
Collapse
Affiliation(s)
- Tianyang Zhang
- The First Hospital of Jilin University, Changchun, Jilin, China
| | - Feiyang Yang
- College of Computer Science and Technology, Jilin University, Changchun, China
| | - Ping Zhang
- The First Hospital of Jilin University, Changchun, Jilin, China
| |
Collapse
|
19
|
Liang J, Feng J, Lin Z, Wei J, Luo X, Wang QM, He B, Chen H, Ye Y. Research on prognostic risk assessment model for acute ischemic stroke based on imaging and multidimensional data. Front Neurol 2023; 14:1294723. [PMID: 38192576 PMCID: PMC10773779 DOI: 10.3389/fneur.2023.1294723] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/18/2023] [Accepted: 11/30/2023] [Indexed: 01/10/2024] Open
Abstract
Accurately assessing the prognostic outcomes of patients with acute ischemic stroke and adjusting treatment plans in a timely manner for those with poor prognosis is crucial for intervening in modifiable risk factors. However, there is still controversy regarding the correlation between imaging-based predictions of complications in acute ischemic stroke. To address this, we developed a cross-modal attention module for integrating multidimensional data, including clinical information, imaging features, treatment plans, prognosis, and complications, to achieve complementary advantages. The fused features preserve magnetic resonance imaging (MRI) characteristics while supplementing clinical relevant information, providing a more comprehensive and informative basis for clinical diagnosis and treatment. The proposed framework based on multidimensional data for activity of daily living (ADL) scoring in patients with acute ischemic stroke demonstrates higher accuracy compared to other state-of-the-art network models, and ablation experiments confirm the effectiveness of each module in the framework.
Collapse
Affiliation(s)
- Jiabin Liang
- Postgraduate Cultivation Base of Guangzhou University of Chinese Medicine, Panyu Central Hospital, Guangzhou, China
- Graduate School, Guangzhou University of Chinese Medicine, Guangzhou, China
- Medical Imaging Institute of Panyu, Guangzhou, China
| | - Jie Feng
- Radiology Department of Sun Yat-sen Memorial Hospital, Sun Yat-sen University, Guangzhou, China
| | - Zhijie Lin
- Laboratory for Intelligent Information Processing, Guangdong University of Technology, Guangzhou, China
| | - Jinbo Wei
- Postgraduate Cultivation Base of Guangzhou University of Chinese Medicine, Panyu Central Hospital, Guangzhou, China
| | - Xun Luo
- Kerry Rehabilitation Medicine Research Institute, Shenzhen, China
| | - Qing Mei Wang
- Stroke Biological Recovery Laboratory, Spaulding Rehabilitation Hospital, Teaching Affiliate of Harvard Medical School, Charlestown, MA, United States
| | - Bingjie He
- Panyu Health Management Center, Guangzhou, China
| | - Hanwei Chen
- Postgraduate Cultivation Base of Guangzhou University of Chinese Medicine, Panyu Central Hospital, Guangzhou, China
- Medical Imaging Institute of Panyu, Guangzhou, China
- Panyu Health Management Center, Guangzhou, China
| | - Yufeng Ye
- Postgraduate Cultivation Base of Guangzhou University of Chinese Medicine, Panyu Central Hospital, Guangzhou, China
- Medical Imaging Institute of Panyu, Guangzhou, China
| |
Collapse
|
20
|
Zhang C, Liu Y, Wang K, Tian J. Specular highlight removal for endoscopic images using partial attention network. Phys Med Biol 2023; 68:225009. [PMID: 37827170 DOI: 10.1088/1361-6560/ad02d9] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/03/2023] [Accepted: 10/12/2023] [Indexed: 10/14/2023]
Abstract
Objective.Endoscopic imaging is a visualization method widely used in minimally invasive surgery. However, owing to the strong reflection of the mucus layer on the organs, specular highlights often appear to degrade the imaging performance. Thus, it is necessary to develop an effective highlight removal method for endoscopic imaging.Approach.A specular highlight removal method using a partial attention network (PatNet) for endoscopic imaging is proposed to reduce the interference of bright light in endoscopic surgery. The method is designed as two procedures: highlight segmentation and endoscopic image inpainting. Image segmentation uses brightness threshold based on illumination compensation to divide the endoscopic image into the highlighted mask and the non-highlighted area. The image inpainting algorithm uses a partial convolution network that integrates an attention mechanism. A mask dataset with random hopping points is designed to simulate specular highlight in endoscopic imaging for network training. Through the filtering of masks, the method can focus on recovering defective pixels and preserving valid pixels as much as possible.Main results.The PatNet is compared with 3 highlight segmentation methods, 3 imaging inpainting methods and 5 highlight removal methods for effective analysis. Experimental results show that the proposed method provides better performance in terms of both perception and quantification. In addition, surgeons are invited to score the processing results for different highlight removal methods under realistic reflection conditions. The PatNet received the highest score of 4.18. Correspondingly, the kendall's W is 0.757 and the asymptotic significancep= 0.000 < 0.01, revealing that the subjective scores have good consistency and confidence.Significance.Generally, the method can realize irregular shape highlight reflection removal and image restoration close to the ground truth of endoscopic images. This method can improve the quality of endoscopic imaging for accurate image analysis.
Collapse
Affiliation(s)
- Chong Zhang
- Department of Big Data Management and Application, School of International Economics and Management, Beijing Technology and Business University, Beijing, People's Republic of China
- CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing, People's Republic of China
| | - Yueliang Liu
- Department of Big Data Management and Application, School of International Economics and Management, Beijing Technology and Business University, Beijing, People's Republic of China
| | - Kun Wang
- CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing, People's Republic of China
| | - Jie Tian
- CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing, People's Republic of China
- Beijing Advanced Innovation Center for Big Data-Based Precision Medicine, Beihang University, Beijing, People's Republic of China
| |
Collapse
|
21
|
Sun L, Zhang Y, Liu T, Ge H, Tian J, Qi X, Sun J, Zhao Y. A collaborative multi-task learning method for BI-RADS category 4 breast lesion segmentation and classification of MRI images. COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE 2023; 240:107705. [PMID: 37454498 DOI: 10.1016/j.cmpb.2023.107705] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Received: 12/03/2022] [Revised: 06/15/2023] [Accepted: 07/01/2023] [Indexed: 07/18/2023]
Abstract
BACKGROUND AND OBJECTIVE The diagnosis of BI-RADS category 4 breast lesion is difficult because its probability of malignancy ranges from 2% to 95%. For BI-RADS category 4 breast lesions, MRI is one of the prominent noninvasive imaging techniques. In this paper, we research computer algorithms to segment lesions and classify the benign or malignant lesions in MRI images. However, this task is challenging because the BI-RADS category 4 lesions are characterized by irregular shape, imbalanced class, and low contrast. METHODS We fully utilize the intrinsic correlation between segmentation and classification tasks, where accurate segmentation will yield accurate classification results, and classification results will promote better segmentation. Therefore, we propose a collaborative multi-task algorithm (CMTL-SC). Specifically, a preliminary segmentation subnet is designed to identify the boundaries, locations and segmentation masks of lesions; a classification subnet, which combines the information provided by the preliminary segmentation, is designed to achieve benign or malignant classification; a repartition segmentation subnet which aggregates the benign or malignant results, is designed to refine the lesion segment. The three subnets work cooperatively so that the CMTL-SC can identify the lesions better which solves the three challenges. RESULTS AND CONCLUSION We collect MRI data from 248 patients in the Second Hospital of Dalian Medical University. The results show that the lesion boundaries delineated by the CMTL-SC are close to the boundaries delineated by the physicians. Moreover, the CMTL-SC yields better results than the single-task and multi-task state-of-the-art algorithms. Therefore, CMTL-SC can help doctors make precise diagnoses and refine treatments for patients.
Collapse
Affiliation(s)
- Liang Sun
- College of Computer Science and Technology, Dalian University of Technology, Dalian, China
| | - Yunling Zhang
- College of Computer Science and Technology, Dalian University of Technology, Dalian, China
| | - Tang Liu
- Department of Radiology, The Second Affiliated Hospital of Dalian Medical University, Dalian, China
| | - Hongwei Ge
- College of Computer Science and Technology, Dalian University of Technology, Dalian, China
| | - Juan Tian
- Department of Radiology, The Second Affiliated Hospital of Dalian Medical University, Dalian, China
| | - Xin Qi
- Department of Radiology, The Second Affiliated Hospital of Dalian Medical University, Dalian, China
| | - Jian Sun
- Health Management Center, The Second Affiliated Hospital of Dalian Medical University, Dalian, China
| | - Yiping Zhao
- Department of Radiology, The Second Affiliated Hospital of Dalian Medical University, Dalian, China.
| |
Collapse
|
22
|
Li H, Zeng P, Bai C, Wang W, Yu Y, Yu P. PMJAF-Net: Pyramidal multi-scale joint attention and adaptive fusion network for explainable skin lesion segmentation. Comput Biol Med 2023; 165:107454. [PMID: 37716246 DOI: 10.1016/j.compbiomed.2023.107454] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/14/2023] [Revised: 08/18/2023] [Accepted: 09/04/2023] [Indexed: 09/18/2023]
Abstract
Traditional convolutional neural networks have achieved remarkable success in skin lesion segmentation. However, the successive pooling operations and convolutional spans reduce the feature resolution and hinder the dense prediction for spatial information, resulting in blurred boundaries, low accuracy and poor interpretability for irregular lesion segmentation under low contrast. To solve the above issues, a pyramidal multi-scale joint attention and adaptive fusion network for explainable (PMJAF-Net) skin lesion segmentation is proposed. Firstly, an adaptive spatial attention module is designed to establish the long-term correlation between pixels, enrich the global and local contextual information, and refine the detailed features. Subsequently, an efficient pyramidal multi-scale channel attention module is proposed to capture the multi-scale information and edge features by using the pyramidal module. Meanwhile, a channel attention module is devised to establish the long-term correlation between channels and highlight the most related feature channels to capture the multi-scale key information on each channel. Thereafter, a multi-scale adaptive fusion attention module is put forward to efficiently fuse the scale features at different decoding stages. Finally, a novel hybrid loss function based on region salient features and boundary quality is presented to guide the network to learn from map-level, patch-level and pixel-level and to accurately predict the lesion regions with clear boundaries. In addition, visualizing attention weight maps are utilized to visually enhance the interpretability of our proposed model. Comprehensive experiments are conducted on four public skin lesion datasets, and the results demonstrate that the proposed network outperforms the state-of-the-art methods, with the segmentation assessment evaluation metrics Dice, JI, and ACC improved to 92.65%, 87.86% and 96.26%, respectively.
Collapse
Affiliation(s)
- Haiyan Li
- School of Information, Yunnan University, Kunming, 650504, China
| | - Peng Zeng
- School of Information, Yunnan University, Kunming, 650504, China
| | - Chongbin Bai
- Otolaryngology Department, Honghe Prefecture Second People's Hospital, Jianshui, 654300, China
| | - Wei Wang
- School of Software, Yunnan University, Kunming, 650504, China.
| | - Ying Yu
- School of Information, Yunnan University, Kunming, 650504, China
| | - Pengfei Yu
- School of Information, Yunnan University, Kunming, 650504, China
| |
Collapse
|
23
|
Li Y, Zhang Y, Liu JY, Wang K, Zhang K, Zhang GS, Liao XF, Yang G. Global Transformer and Dual Local Attention Network via Deep-Shallow Hierarchical Feature Fusion for Retinal Vessel Segmentation. IEEE TRANSACTIONS ON CYBERNETICS 2023; 53:5826-5839. [PMID: 35984806 DOI: 10.1109/tcyb.2022.3194099] [Citation(s) in RCA: 12] [Impact Index Per Article: 6.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/15/2023]
Abstract
Clinically, retinal vessel segmentation is a significant step in the diagnosis of fundus diseases. However, recent methods generally neglect the difference of semantic information between deep and shallow features, which fail to capture the global and local characterizations in fundus images simultaneously, resulting in the limited segmentation performance for fine vessels. In this article, a global transformer (GT) and dual local attention (DLA) network via deep-shallow hierarchical feature fusion (GT-DLA-dsHFF) are investigated to solve the above limitations. First, the GT is developed to integrate the global information in the retinal image, which effectively captures the long-distance dependence between pixels, alleviating the discontinuity of blood vessels in the segmentation results. Second, DLA, which is constructed using dilated convolutions with varied dilation rates, unsupervised edge detection, and squeeze-excitation block, is proposed to extract local vessel information, consolidating the edge details in the segmentation result. Finally, a novel deep-shallow hierarchical feature fusion (dsHFF) algorithm is studied to fuse the features in different scales in the deep learning framework, respectively, which can mitigate the attenuation of valid information in the process of feature fusion. We verified the GT-DLA-dsHFF on four typical fundus image datasets. The experimental results demonstrate our GT-DLA-dsHFF achieves superior performance against the current methods and detailed discussions verify the efficacy of the proposed three modules. Segmentation results of diseased images show the robustness of our proposed GT-DLA-dsHFF. Implementation codes will be available on https://github.com/YangLibuaa/GT-DLA-dsHFF.
Collapse
|
24
|
Patro KK, Allam JP, Neelapu BC, Tadeusiewicz R, Acharya UR, Hammad M, Yildirim O, Pławiak P. Application of Kronecker convolutions in deep learning technique for automated detection of kidney stones with coronal CT images. Inf Sci (N Y) 2023; 640:119005. [DOI: 10.1016/j.ins.2023.119005] [Citation(s) in RCA: 2] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 07/23/2024]
|
25
|
Yu Z, Yu L, Zheng W, Wang S. EIU-Net: Enhanced feature extraction and improved skip connections in U-Net for skin lesion segmentation. Comput Biol Med 2023; 162:107081. [PMID: 37301097 DOI: 10.1016/j.compbiomed.2023.107081] [Citation(s) in RCA: 13] [Impact Index Per Article: 6.5] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/20/2022] [Revised: 04/25/2023] [Accepted: 05/27/2023] [Indexed: 06/12/2023]
Abstract
Skin lesion segmentation is a computer-aided diagnosis method for quantitative analysis of melanoma that can improve efficiency and accuracy. Although many methods based on U-Net have achieved tremendous success, they still cannot handle challenging tasks well due to weak feature extraction. In response to skin lesion segmentation, a novel method called EIU-Net is proposed to tackle the challenging task. To capture the local and global contextual information, we employ inverted residual blocks and an efficient pyramid squeeze attention (EPSA) block as the main encoders at different stages, while atrous spatial pyramid pooling (ASPP) is utilized after the last encoder and the soft-pool method is introduced for downsampling. Also, we propose a novel method named multi-layer fusion (MLF) module to effectively fuse the feature distributions and capture significant boundary information of skin lesions in different encoders to improve the performance of the network. Furthermore, a reshaped decoders fusion module is used to obtain multi-scale information by fusing feature maps of different decoders to improve the final results of skin lesion segmentation. To validate the performance of our proposed network, we compare it with other methods on four public datasets, including the ISIC 2016, ISIC 2017, ISIC 2018, and PH2 datasets. And the main metric Dice scores achieved by our proposed EIU-Net are 0.919, 0.855, 0.902, and 0.916 on the four datasets, respectively, outperforming other methods. Ablation experiments also demonstrate the effectiveness of the main modules in our proposed network. Our code is available at https://github.com/AwebNoob/EIU-Net.
Collapse
Affiliation(s)
- Zimin Yu
- Department of Computer Science and Engineering, School of Information Science and Engineering, Yunnan University, Kunming, 650504, Yunnan, China
| | - Li Yu
- Department of Computer Science and Engineering, School of Information Science and Engineering, Yunnan University, Kunming, 650504, Yunnan, China
| | - Weihua Zheng
- Department of Computer Science and Engineering, School of Information Science and Engineering, Yunnan University, Kunming, 650504, Yunnan, China
| | - Shunfang Wang
- Department of Computer Science and Engineering, School of Information Science and Engineering, Yunnan University, Kunming, 650504, Yunnan, China; The Key Lab of Intelligent Systems and Computing of Yunnan Province, Yunnan University, Kunming, Yunnan, China.
| |
Collapse
|
26
|
Shi T, Ding X, Zhou W, Pan F, Yan Z, Bai X, Yang X. Affinity Feature Strengthening for Accurate, Complete and Robust Vessel Segmentation. IEEE J Biomed Health Inform 2023; 27:4006-4017. [PMID: 37163397 DOI: 10.1109/jbhi.2023.3274789] [Citation(s) in RCA: 4] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 05/12/2023]
Abstract
Vessel segmentation is crucial in many medical image applications, such as detecting coronary stenoses, retinal vessel diseases and brain aneurysms. However, achieving high pixel-wise accuracy, complete topology structure and robustness to various contrast variations are critical and challenging, and most existing methods focus only on achieving one or two of these aspects. In this paper, we present a novel approach, the affinity feature strengthening network (AFN), which jointly models geometry and refines pixel-wise segmentation features using a contrast-insensitive, multiscale affinity approach. Specifically, we compute a multiscale affinity field for each pixel, capturing its semantic relationships with neighboring pixels in the predicted mask image. This field represents the local geometry of vessel segments of different sizes, allowing us to learn spatial- and scale-aware adaptive weights to strengthen vessel features. We evaluate our AFN on four different types of vascular datasets: X-ray angiography coronary vessel dataset (XCAD), portal vein dataset (PV), digital subtraction angiography cerebrovascular vessel dataset (DSA) and retinal vessel dataset (DRIVE). Extensive experimental results demonstrate that our AFN outperforms the state-of-the-art methods in terms of both higher accuracy and topological metrics, while also being more robust to various contrast changes.
Collapse
|
27
|
Wu M, Qian Y, Liao X, Wang Q, Heng PA. Hepatic vessel segmentation based on 3D swin-transformer with inductive biased multi-head self-attention. BMC Med Imaging 2023; 23:91. [PMID: 37422639 PMCID: PMC10329304 DOI: 10.1186/s12880-023-01045-y] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/27/2022] [Accepted: 06/05/2023] [Indexed: 07/10/2023] Open
Abstract
PURPOSE Segmentation of liver vessels from CT images is indispensable prior to surgical planning and aroused a broad range of interest in the medical image analysis community. Due to the complex structure and low-contrast background, automatic liver vessel segmentation remains particularly challenging. Most of the related researches adopt FCN, U-net, and V-net variants as a backbone. However, these methods mainly focus on capturing multi-scale local features which may produce misclassified voxels due to the convolutional operator's limited locality reception field. METHODS We propose a robust end-to-end vessel segmentation network called Inductive BIased Multi-Head Attention Vessel Net(IBIMHAV-Net) by expanding swin transformer to 3D and employing an effective combination of convolution and self-attention. In practice, we introduce voxel-wise embedding rather than patch-wise embedding to locate precise liver vessel voxels and adopt multi-scale convolutional operators to gain local spatial information. On the other hand, we propose the inductive biased multi-head self-attention which learns inductively biased relative positional embedding from initialized absolute position embedding. Based on this, we can gain more reliable queries and key matrices. RESULTS We conducted experiments on the 3DIRCADb dataset. The average dice and sensitivity of the four tested cases were 74.8[Formula: see text] and 77.5[Formula: see text], which exceed the results of existing deep learning methods and improved graph cuts method. The Branches Detected(BD)/Tree-length Detected(TD) indexes also proved the global/local feature capture ability better than other methods. CONCLUSION The proposed model IBIMHAV-Net provides an automatic, accurate 3D liver vessel segmentation with an interleaved architecture that better utilizes both global and local spatial features in CT volumes. It can be further extended for other clinical data.
Collapse
Affiliation(s)
- Mian Wu
- Guangdong Provincial Key Laboratory of Computer Vision and Virtual Reality Technology, Shenzhen Institute of Advanced Technology, Chinese Academy of Science, Shenzhen, China
| | - Yinling Qian
- Guangdong Provincial Key Laboratory of Computer Vision and Virtual Reality Technology, Shenzhen Institute of Advanced Technology, Chinese Academy of Science, Shenzhen, China
| | - Xiangyun Liao
- Guangdong Provincial Key Laboratory of Computer Vision and Virtual Reality Technology, Shenzhen Institute of Advanced Technology, Chinese Academy of Science, Shenzhen, China.
| | - Qiong Wang
- Guangdong Provincial Key Laboratory of Computer Vision and Virtual Reality Technology, Shenzhen Institute of Advanced Technology, Chinese Academy of Science, Shenzhen, China
| | - Pheng-Ann Heng
- Guangdong Provincial Key Laboratory of Computer Vision and Virtual Reality Technology, Shenzhen Institute of Advanced Technology, Chinese Academy of Science, Shenzhen, China
- The Chinese University of Hong Kong, Hong Kong SAR, China
| |
Collapse
|
28
|
Li R, Xu L, Xie K, Song J, Ma X, Chang L, Yan Q. DHT-Net: Dynamic Hierarchical Transformer Network for Liver and Tumor Segmentation. IEEE J Biomed Health Inform 2023; 27:3443-3454. [PMID: 37079414 DOI: 10.1109/jbhi.2023.3268218] [Citation(s) in RCA: 9] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 04/21/2023]
Abstract
Automatic segmentation of liver tumors is crucial to assist radiologists in clinical diagnosis. While various deep learningbased algorithms have been proposed, such as U-Net and its variants, the inability to explicitly model long-range dependencies in CNN limits the extraction of complex tumor features. Some researchers have applied Transformer-based 3D networks to analyze medical images. However, the previous methods focus on modeling the local information (eg. edge) or global information (eg. morphology) with fixed network weights. To learn and extract complex tumor features of varied tumor size, location, and morphology for more accurate segmentation, we propose a Dynamic Hierarchical Transformer Network, named DHT-Net. The DHT-Net mainly contains a Dynamic Hierarchical Transformer (DHTrans) structure and an Edge Aggregation Block (EAB). The DHTrans first automatically senses the tumor location by Dynamic Adaptive Convolution, which employs hierarchical operations with the different receptive field sizes to learn the features of various tumors, thus enhancing the semantic representation ability of tumor features. Then, to adequately capture the irregular morphological features in the tumor region, DHTrans aggregates global and local texture information in a complementary manner. In addition, we introduce the EAB to extract detailed edge features in the shallow fine-grained details of the network, which provides sharp boundaries of liver and tumor regions. We evaluate DHT-Net on two challenging public datasets, LiTS and 3DIRCADb. The proposed method has shown superior liver and tumor segmentation performance compared to several state-of-the-art 2D, 3D, and 2.5D hybrid models.
Collapse
|
29
|
Chen C, Zhou K, Wang Z, Zhang Q, Xiao R. All answers are in the images: A review of deep learning for cerebrovascular segmentation. Comput Med Imaging Graph 2023; 107:102229. [PMID: 37043879 DOI: 10.1016/j.compmedimag.2023.102229] [Citation(s) in RCA: 11] [Impact Index Per Article: 5.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 10/24/2022] [Revised: 03/03/2023] [Accepted: 04/03/2023] [Indexed: 04/14/2023]
Abstract
Cerebrovascular imaging is a common examination. Its accurate cerebrovascular segmentation become an important auxiliary method for the diagnosis and treatment of cerebrovascular diseases, which has received extensive attention from researchers. Deep learning is a heuristic method that encourages researchers to derive answers from the images by driving datasets. With the continuous development of datasets and deep learning theory, it has achieved important success for cerebrovascular segmentation. Detailed survey is an important reference for researchers. To comprehensively analyze the newest cerebrovascular segmentation, we have organized and discussed researches centered on deep learning. This survey comprehensively reviews deep learning for cerebrovascular segmentation since 2015, it mainly includes sliding window based models, U-Net based models, other CNNs based models, small-sample based models, semi-supervised or unsupervised models, fusion based models, Transformer based models, and graphics based models. We organize the structures, improvement, and important parameters of these models, as well as analyze development trends and quantitative assessment. Finally, we have discussed the challenges and opportunities of possible research directions, hoping that our survey can provide researchers with convenient reference.
Collapse
Affiliation(s)
- Cheng Chen
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
| | - Kangneng Zhou
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
| | - Zhiliang Wang
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China
| | - Qian Zhang
- Department of Neurosurgery, Beijing Tiantan Hospital, Capital Medical University, Beijing 100070, China; China National Clinical Research Center for Neurological Diseases, Beijing 100070, China
| | - Ruoxiu Xiao
- School of Computer and Communication Engineering, University of Science and Technology Beijing, Beijing 100083, China; Shunde Innovation School, University of Science and Technology Beijing, Foshan 100024, China.
| |
Collapse
|
30
|
A novel multi-attention, multi-scale 3D deep network for coronary artery segmentation. Med Image Anal 2023; 85:102745. [PMID: 36630869 DOI: 10.1016/j.media.2023.102745] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/03/2022] [Revised: 12/13/2022] [Accepted: 01/05/2023] [Indexed: 01/11/2023]
Abstract
Automatic segmentation of coronary arteries provides vital assistance to enable accurate and efficient diagnosis and evaluation of coronary artery disease (CAD). However, the task of coronary artery segmentation (CAS) remains highly challenging due to the large-scale variations exhibited by coronary arteries, their complicated anatomical structures and morphologies, as well as the low contrast between vessels and their background. To comprehensively tackle these challenges, we propose a novel multi-attention, multi-scale 3D deep network for CAS, which we call CAS-Net. Specifically, we first propose an attention-guided feature fusion (AGFF) module to efficiently fuse adjacent hierarchical features in the encoding and decoding stages to capture more effectively latent semantic information. Then, we propose a scale-aware feature enhancement (SAFE) module, aiming to dynamically adjust the receptive fields to extract more expressive features effectively, thereby enhancing the feature representation capability of the network. Furthermore, we employ the multi-scale feature aggregation (MSFA) module to learn a more distinctive semantic representation for refining the vessel maps. In addition, considering that the limited training data annotated with a quality golden standard are also a significant factor restricting the development of CAS, we construct a new dataset containing 119 cases consisting of coronary computed tomographic angiography (CCTA) volumes and annotated coronary arteries. Extensive experiments on our self-collected dataset and three publicly available datasets demonstrate that the proposed method has good segmentation performance and generalization ability, outperforming multiple state-of-the-art algorithms on various metrics. Compared with U-Net3D, the proposed method significantly improves the Dice similarity coefficient (DSC) by at least 4% on each dataset, due to the synergistic effect among the three core modules, AGFF, SAFE, and MSFA. Our implementation is released at https://github.com/Cassie-CV/CAS-Net.
Collapse
|
31
|
Qu Z, Zhuo L, Cao J, Li X, Yin H, Wang Z. TP-Net: Two-Path Network for Retinal Vessel Segmentation. IEEE J Biomed Health Inform 2023; 27:1979-1990. [PMID: 37021912 DOI: 10.1109/jbhi.2023.3237704] [Citation(s) in RCA: 9] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/19/2023]
Abstract
Refined and automatic retinal vessel segmentation is crucial for computer-aided early diagnosis of retinopathy. However, existing methods often suffer from mis-segmentation when dealing with thin and low-contrast vessels. In this paper, a two-path retinal vessel segmentation network is proposed, namely TP-Net, which consists of three core parts, i.e., main-path, sub-path, and multi-scale feature aggregation module (MFAM). Main-path is to detect the trunk area of the retinal vessels, and the sub-path to effectively capture edge information of the retinal vessels. The prediction results of the two paths are combined by MFAM, obtaining refined segmentation of retinal vessels. In the main-path, a three-layer lightweight backbone network is elaborately designed according to the characteristics of retinal vessels, and then a global feature selection mechanism (GFSM) is proposed, which can autonomously select features that are more important for the segmentation task from the features at different layers of the network, thereby, enhancing the segmentation capability for low-contrast vessels. In the sub-path, an edge feature extraction method and an edge loss function are proposed, which can enhance the ability of the network to capture edge information and reduce the mis-segmentation of thin vessels. Finally, MFAM is proposed to fuse the prediction results of main-path and sub-path, which can remove background noises while preserving edge details, and thus, obtaining refined segmentation of retinal vessels. The proposed TP-Net has been evaluated on three public retinal vessel datasets, namely DRIVE, STARE, and CHASE DB1. The experimental results show that the TP-Net achieved a superior performance and generalization ability with fewer model parameters compared with the state-of-the-art methods.
Collapse
|
32
|
Pan L, Li Z, Shen Z, Liu Z, Huang L, Yang M, Zheng B, Zeng T, Zheng S. Learning multi-view and centerline topology connectivity information for pulmonary artery-vein separation. Comput Biol Med 2023; 155:106669. [PMID: 36803793 DOI: 10.1016/j.compbiomed.2023.106669] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/23/2022] [Revised: 02/06/2023] [Accepted: 02/10/2023] [Indexed: 02/16/2023]
Abstract
BACKGROUND Automatic pulmonary artery-vein separation has considerable importance in the diagnosis and treatment of lung diseases. However, insufficient connectivity and spatial inconsistency have always been the problems of artery-vein separation. METHODS A novel automatic method for artery-vein separation in CT images is presented in this work. Specifically, a multi-scale information aggregated network (MSIA-Net) including multi-scale fusion blocks and deep supervision, is proposed to learn the features of artery-vein and aggregate additional semantic information, respectively. The proposed method integrates nine MSIA-Net models for artery-vein separation, vessel segmentation, and centerline separation tasks along with axial, coronal, and sagittal multi-view slices. First, the preliminary artery-vein separation results are obtained by the proposed multi-view fusion strategy (MVFS). Then, centerline correction algorithm (CCA) is used to correct the preliminary results of artery-vein separation by the centerline separation results. Finally, the vessel segmentation results are utilized to reconstruct the artery-vein morphology. In addition, weighted cross-entropy and dice loss are employed to solve the class imbalance problem. RESULTS We constructed 50 manually labeled contrast-enhanced computed CT scans for five-fold cross-validation, and experimental results demonstrated that our method achieves superior segmentation performance of 97.7%, 85.1%, and 84.9% on ACC, Pre, and DSC, respectively. Additionally, a series of ablation studies demonstrate the effectiveness of the proposed components. CONCLUSION The proposed method can effectively solve the problem of insufficient vascular connectivity and correct the spatial inconsistency of artery-vein.
Collapse
Affiliation(s)
- Lin Pan
- College of Physics and Information Engineering, Fuzhou University, Fuzhou, China
| | - Zhaopei Li
- College of Physics and Information Engineering, Fuzhou University, Fuzhou, China
| | - Zhiqiang Shen
- College of Physics and Information Engineering, Fuzhou University, Fuzhou, China
| | - Zheng Liu
- Faculty of Applied Science, School of Engineering, University of British Columbia, Kelowna, BC, Canada
| | - Liqin Huang
- College of Physics and Information Engineering, Fuzhou University, Fuzhou, China
| | - Mingjing Yang
- College of Physics and Information Engineering, Fuzhou University, Fuzhou, China
| | - Bin Zheng
- Key Laboratory of Cardio-Thoracic Surgery, Fujian Medical University, Fuzhou, China
| | - Taidui Zeng
- Key Laboratory of Cardio-Thoracic Surgery, Fujian Medical University, Fuzhou, China
| | - Shaohua Zheng
- College of Physics and Information Engineering, Fuzhou University, Fuzhou, China.
| |
Collapse
|
33
|
He T, Guo C, Liu H, Jiang L. A venipuncture robot with decoupled position and attitude guided by near-infrared vision and force feedback. Int J Med Robot 2023:e2512. [PMID: 36809654 DOI: 10.1002/rcs.2512] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 11/02/2022] [Revised: 12/27/2022] [Accepted: 02/16/2023] [Indexed: 02/23/2023]
Abstract
BACKGROUND This study aims to develop a venipuncture robot to replace manual venipuncture to ease the heavy workload, lower the risk of 2019-nCoV infection, and boost venipuncture success rates. METHOD The robot is designed with decoupled position and attitude. It consists of a 3-degree-of-freedom positioning manipulator to locate the needle and a 3-degree-of-freedom end-effector that is always vertical to adjust the yaw and pitch angles of the needle. The near-infrared vision and laser sensors obtain the three-dimensional information of puncture positions, while the change in force detects the state feedback of punctures. RESULTS The experimental results demonstrate that the venipuncture robot has a compact design, flexible motion, high positioning accuracy and repeatability (0.11 and 0.04 mm), and a high success rate when puncturing the phantom. CONCLUSION This paper presents a decoupled position and attitude venipuncture robot guided by near-infrared vision and force feedback to replace manual venipuncture. The robot is compact, dexterous, and accurate, which helps to improve the success rate of venipuncture, and it is expected to achieve fully automatic venipuncture in the future.
Collapse
Affiliation(s)
- Tianbao He
- State Key Laboratory of Robotics and System, Harbin Institute of Technology, Harbin, China
| | - Chuangqiang Guo
- State Key Laboratory of Robotics and System, Harbin Institute of Technology, Harbin, China
| | - Hansong Liu
- State Key Laboratory of Robotics and System, Harbin Institute of Technology, Harbin, China
| | - Li Jiang
- State Key Laboratory of Robotics and System, Harbin Institute of Technology, Harbin, China
| |
Collapse
|
34
|
Chen Y, Zheng C, Zhou T, Feng L, Liu L, Zeng Q, Wang G. A deep residual attention-based U-Net with a biplane joint method for liver segmentation from CT scans. Comput Biol Med 2023; 152:106421. [PMID: 36527780 DOI: 10.1016/j.compbiomed.2022.106421] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/29/2022] [Revised: 11/17/2022] [Accepted: 12/11/2022] [Indexed: 12/15/2022]
Abstract
Liver tumours are diseases with high morbidity and high deterioration probabilities, and accurate liver area segmentation from computed tomography (CT) scans is a prerequisite for quick tumour diagnosis. While 2D network segmentation methods can perform segmentation with lower device performance requirements, they often discard the rich 3D spatial information contained in CT scans, limiting their segmentation accuracy. Hence, a deep residual attention-based U-shaped network (DRAUNet) with a biplane joint method for liver segmentation is proposed in this paper, where the biplane joint method introduces coronal CT slices to assist the transverse slices with segmentation, incorporating more 3D spatial information into the segmentation results to improve the segmentation performance of the network. Additionally, a novel deep residual block (DR block) and dual-effect attention module (DAM) are introduced in DRAUNet, where the DR block has deeper layers and two shortcut paths. The DAM efficiently combines the correlations of feature channels and the spatial locations of feature maps. The DRAUNet with the biplane joint method is tested on three datasets, Liver Tumour Segmentation (LiTS), 3D Image Reconstruction for Comparison of Algorithms Database (3DIRCADb), and Segmentation of the Liver Competition 2007 (Sliver07), and it achieves 97.3%, 97.4%, and 96.9% Dice similarity coefficients (DSCs) for liver segmentation, respectively, outperforming most state-of-the-art networks; this strongly demonstrates the segmentation performance of DRAUNet and the ability of the biplane joint method to obtain 3D spatial information from 3D images.
Collapse
Affiliation(s)
- Ying Chen
- School of Software, Nanchang Hangkong University, Nanchang, 330063, PR China
| | - Cheng Zheng
- School of Software, Nanchang Hangkong University, Nanchang, 330063, PR China.
| | - Taohui Zhou
- School of Software, Nanchang Hangkong University, Nanchang, 330063, PR China
| | - Longfeng Feng
- School of Software, Nanchang Hangkong University, Nanchang, 330063, PR China
| | - Lan Liu
- Department of Radiology, Jiangxi Cancer Hospital, Nanchang, 330029, PR China.
| | - Qiao Zeng
- Department of Radiology, Jiangxi Cancer Hospital, Nanchang, 330029, PR China
| | - Guoqing Wang
- Zhejiang Suosi Technology Co. Ltd, Wenzhou, 325000, PR China.
| |
Collapse
|
35
|
Niu K, Guo Z, Peng X, Pei S. P-ResUnet: Segmentation of brain tissue with Purified Residual Unet. Comput Biol Med 2022; 151:106294. [PMID: 36435055 DOI: 10.1016/j.compbiomed.2022.106294] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/20/2022] [Revised: 10/14/2022] [Accepted: 11/06/2022] [Indexed: 11/13/2022]
Abstract
Brain tissue of Magnetic Resonance Imaging is precisely segmented and quantified, which aids in the diagnosis of neurological diseases such as epilepsy, Alzheimer's, and multiple sclerosis. Recently, UNet-like architectures are widely used for medical image segmentation, which achieved promising performance by using the skip connection to fuse the low-level and high-level information. However, In the process of integrating the low-level and high-level information, the non-object information (noise) will be added, which reduces the accuracy of medical image segmentation. Likewise, the same problem also exists in the residual unit. Since the output and input of the residual unit are fused, the non-object information (noise) of the input of the residual unit will be in the integration. To address this challenging problem, in this paper we propose a Purified Residual U-net for the segmentation of brain tissue. This model encodes the image to obtain deep semantic information and purifies the information of low-level features and the residual unit from the image, and acquires the result through a decoder at last. We use the Dilated Pyramid Separate Block (DPSB) as the first block to purify the features for each layer in the encoder without the first layer, which expands the receptive field of the convolution kernel with only a few parameters added. In the first layer, we have explored the best performance achieved with DPB. We find the most non-object information (noise) in the initial image, so it is good for the accuracy to exchange the information to the max degree. We have conducted experiments with the widely used IBSR-18 dataset composed of T-1 weighted MRI volumes from 18 subjects. The results show that compared with some of the cutting-edge methods, our method enhances segmentation performance with the mean dice score reaching 91.093% and the mean Hausdorff distance decreasing to 3.2606.
Collapse
Affiliation(s)
- Ke Niu
- Beijing Information Science and Technology University, Beijing, China.
| | - Zhongmin Guo
- Beijing Information Science and Technology University, Beijing, China.
| | - Xueping Peng
- Australian Artificial Intelligence Institute, Faculty of Engineering and Information Technology, University of Technology Sydney, Australia.
| | - Su Pei
- Beijing Information Science and Technology University, Beijing, China.
| |
Collapse
|
36
|
Improving vessel connectivity in retinal vessel segmentation via adversarial learning. Knowl Based Syst 2022. [DOI: 10.1016/j.knosys.2022.110243] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/29/2022]
|
37
|
Kuang H, Yang Z, Zhang X, Tan J, Wang X, Zhang L. Hepatic Vein and Arterial Vessel Segmentation in Liver Tumor Patients. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2022; 2022:2303733. [PMID: 36188682 PMCID: PMC9525193 DOI: 10.1155/2022/2303733] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 05/27/2022] [Revised: 08/05/2022] [Accepted: 08/29/2022] [Indexed: 11/24/2022]
Abstract
Preoperative observation of liver status in patients with liver tumors by abdominal Computed Tomography (CT) imaging is one of the essential references for formulating surgical plans. Preoperative vessel segmentation in the patient's liver region has become an increasingly important and challenging problem. Almost all existing methods first segment arterial and venous vessels on CT in the arterial and venous phases, respectively. Then, the two are directly registered to complete the reconstruction of the vascular system, ignoring the displacement and deformation of blood vessels caused by changes in body position and respiration in the two phases. We propose an unsupervised domain-adaptive two-stage vessel segmentation framework for simultaneous fine segmentation of arterial and venous vessels on venous phase CT. Specifically, we first achieve domain adaptation for arterial and venous phase CT using a modified cycle-consistent adversarial network. The newly added discriminator can improve the ability to generate and discriminate tiny blood vessels, making the domain-adaptive network more robust. The second-stage supervised training of arterial vessels was then performed on the translated arterial phase CT. In this process, we propose an orthogonal depth projection loss function to enhance the representation ability of the 3D U-shape segmentation network for the geometric information of the vessel model. The segmented venous vessels were also performed on venous phase CT in the second stage. Finally, we invited professional doctors to annotate arterial and venous vessels on the venous phase CT of the test set. The experimental results show that the segmentation accuracy of arterial and venous vessels on venous phase CT is 0.8454 and 0.8087, respectively. Our proposed framework can simultaneously achieve supervised segmentation of venous vessels and unsupervised segmentation of arterial vessels on venous phase CT. Our approach can be extended to other fields of medical segmentation, such as unsupervised domain adaptive segmentation of liver tumors at different CT phases, to facilitate the development of the community.
Collapse
Affiliation(s)
- Haopeng Kuang
- Academy for Engineering & Technology, Fudan University, Shanghai 200000, China
| | - Zhongwei Yang
- Academy for Engineering & Technology, Fudan University, Shanghai 200000, China
| | - Xukun Zhang
- Academy for Engineering & Technology, Fudan University, Shanghai 200000, China
| | - Jinpeng Tan
- Liver Surgery Department, Zhongshan Hospital, Fudan University, Shanghai 200000, China
| | - Xiaoying Wang
- Liver Surgery Department, Zhongshan Hospital, Fudan University, Shanghai 200000, China
| | - Lihua Zhang
- Academy for Engineering & Technology, Fudan University, Shanghai 200000, China
| |
Collapse
|
38
|
Li Y, Zhang Y, Cui W, Lei B, Kuang X, Zhang T. Dual Encoder-Based Dynamic-Channel Graph Convolutional Network With Edge Enhancement for Retinal Vessel Segmentation. IEEE TRANSACTIONS ON MEDICAL IMAGING 2022; 41:1975-1989. [PMID: 35167444 DOI: 10.1109/tmi.2022.3151666] [Citation(s) in RCA: 47] [Impact Index Per Article: 15.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
Retinal vessel segmentation with deep learning technology is a crucial auxiliary method for clinicians to diagnose fundus diseases. However, the deep learning approaches inevitably lose the edge information, which contains spatial features of vessels while performing down-sampling, leading to the limited segmentation performance of fine blood vessels. Furthermore, the existing methods ignore the dynamic topological correlations among feature maps in the deep learning framework, resulting in the inefficient capture of the channel characterization. To address these limitations, we propose a novel dual encoder-based dynamic-channel graph convolutional network with edge enhancement (DE-DCGCN-EE) for retinal vessel segmentation. Specifically, we first design an edge detection-based dual encoder to preserve the edge of vessels in down-sampling. Secondly, we investigate a dynamic-channel graph convolutional network to map the image channels to the topological space and synthesize the features of each channel on the topological map, which solves the limitation of insufficient channel information utilization. Finally, we study an edge enhancement block, aiming to fuse the edge and spatial features in the dual encoder, which is beneficial to improve the accuracy of fine blood vessel segmentation. Competitive experimental results on five retinal image datasets validate the efficacy of the proposed DE-DCGCN-EE, which achieves more remarkable segmentation results against the other state-of-the-art methods, indicating its potential clinical application.
Collapse
|
39
|
Survarachakan S, Prasad PJR, Naseem R, Pérez de Frutos J, Kumar RP, Langø T, Alaya Cheikh F, Elle OJ, Lindseth F. Deep learning for image-based liver analysis — A comprehensive review focusing on malignant lesions. Artif Intell Med 2022; 130:102331. [DOI: 10.1016/j.artmed.2022.102331] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 07/23/2021] [Revised: 05/23/2022] [Accepted: 05/30/2022] [Indexed: 11/26/2022]
|
40
|
Qian X, Hu W, Gao L, Xu J, Wang B, Song J, Yang S, Lu Q, Zhang L, Yan J, Dong J. Trans-arterial positive ICG staining-guided laparoscopic liver watershed resection for hepatocellular carcinoma. Front Oncol 2022; 12:966626. [PMID: 35936704 PMCID: PMC9354495 DOI: 10.3389/fonc.2022.966626] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/11/2022] [Accepted: 06/27/2022] [Indexed: 11/29/2022] Open
Abstract
Introduction Anatomical liver resection is the optimal treatment for patients with resectable hepatocellular carcinoma (HCC). Laparoscopic Couinaud liver segment resection could be performed easily as liver segments could be stained by ultrasound-guided indocyanine green (ICG) injection into the corresponding segment portal vein. Several smaller liver anatomical units (liver watersheds) have been identified (such as S8v, S8d, S4a, and S4b). However, since portal veins of liver watersheds are too thin to be identified under ultrasound, the boundaries of these liver watersheds could not be stained intraoperatively, making laparoscopic resection of these liver watersheds demanding. Digital subtraction angiography (DSA) could identify arteries of liver watersheds with a diameter of less than 2 mm. Yet, its usage for liver watershed staining has not been explored so far. Purpose The aim of this study is to explore the possibility of positive liver watershed staining via trans-arterial ICG injection under DSA examination for navigating laparoscopic watershed-oriented hepatic resection. Methods We describe, in a step-by-step approach, the application of trans-arterial ICG injection to stain aimed liver watershed during laparoscopic anatomical hepatectomy. The efficiency and safety of the technique are illustrated and discussed in comparison with the laparoscopic anatomical liver resection via ultrasound-guided liver segment staining. Results Eight of 10 HCC patients received successful trans-arterial liver watershed staining. The success rate of the trans-artery staining approach was 80%, higher than that of the ultrasound-guided portal vein staining approach (60%). Longer surgical duration was found in patients who underwent the trans-artery staining approach (305.3 ± 23.2 min vs. 268.4 ± 34.7 min in patients who underwent the ultrasound-guided portal vein staining approach, p = 0.004). No significant difference was found in major morbidity, reoperation rate, hospital stay duration, and 30-day and 90-day mortality between the 2 groups. Conclusions Trans-arterial ICG staining is safe and feasible for staining the aimed liver watershed, navigating watershed-oriented hepatic resection under fluorescence laparoscopy for surgeons.
Collapse
Affiliation(s)
- Xinye Qian
- Center of Hepatobiliary Pancreatic Disease, Beijing Tsinghua Changgung Hospital, School of Clinical Medicine, Tsinghua University, Beijing, China
- School of Clinical Medicine, Tsinghua University, Beijing, China
| | - Wang Hu
- Center of Hepatobiliary Pancreatic Disease, Beijing Tsinghua Changgung Hospital, School of Clinical Medicine, Tsinghua University, Beijing, China
- School of Clinical Medicine, Tsinghua University, Beijing, China
| | - Lu Gao
- Center of Hepatobiliary Pancreatic Disease, Beijing Tsinghua Changgung Hospital, School of Clinical Medicine, Tsinghua University, Beijing, China
- School of Clinical Medicine, Tsinghua University, Beijing, China
| | - Jingyi Xu
- Center of Hepatobiliary Pancreatic Disease, Beijing Tsinghua Changgung Hospital, School of Clinical Medicine, Tsinghua University, Beijing, China
- School of Clinical Medicine, Tsinghua University, Beijing, China
| | - Bo Wang
- The State Key Laboratory of Precision Measurement Technology and Instruments, Department of Precision Instrument, Tsinghua University, Beijing, China
- Beijing Jingzhen Medical Technology Ltd., Beijing, China
| | - Jiyong Song
- Center of Hepatobiliary Pancreatic Disease, Beijing Tsinghua Changgung Hospital, School of Clinical Medicine, Tsinghua University, Beijing, China
| | - Shizhong Yang
- Center of Hepatobiliary Pancreatic Disease, Beijing Tsinghua Changgung Hospital, School of Clinical Medicine, Tsinghua University, Beijing, China
| | - Qian Lu
- Center of Hepatobiliary Pancreatic Disease, Beijing Tsinghua Changgung Hospital, School of Clinical Medicine, Tsinghua University, Beijing, China
- School of Clinical Medicine, Tsinghua University, Beijing, China
| | - Lin Zhang
- Center of Hepatobiliary Pancreatic Disease, Beijing Tsinghua Changgung Hospital, School of Clinical Medicine, Tsinghua University, Beijing, China
| | - Jun Yan
- Center of Hepatobiliary Pancreatic Disease, Beijing Tsinghua Changgung Hospital, School of Clinical Medicine, Tsinghua University, Beijing, China
- School of Clinical Medicine, Tsinghua University, Beijing, China
- Department of Hepatobiliary surgery, Xuzhou Central Hospital, Xuzhou, China
- Department of Hepatobiliary Surgery, The No.2 Hospital of Baoding, Baoding, China
| | - Jiahong Dong
- Center of Hepatobiliary Pancreatic Disease, Beijing Tsinghua Changgung Hospital, School of Clinical Medicine, Tsinghua University, Beijing, China
- School of Clinical Medicine, Tsinghua University, Beijing, China
- The State Key Laboratory of Precision Measurement Technology and Instruments, Department of Precision Instrument, Tsinghua University, Beijing, China
| |
Collapse
|
41
|
Li Y, Ren T, Li J, Li X, Li A. Multi-perspective label based deep learning framework for cerebral vasculature segmentation in whole-brain fluorescence images. BIOMEDICAL OPTICS EXPRESS 2022; 13:3657-3671. [PMID: 35781963 PMCID: PMC9208593 DOI: 10.1364/boe.458111] [Citation(s) in RCA: 3] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 03/09/2022] [Revised: 04/23/2022] [Accepted: 05/22/2022] [Indexed: 06/15/2023]
Abstract
The popularity of fluorescent labelling and mesoscopic optical imaging techniques enable the acquisition of whole mammalian brain vasculature images at capillary resolution. Segmentation of the cerebrovascular network is essential for analyzing the cerebrovascular structure and revealing the pathogenesis of brain diseases. Existing deep learning methods use a single type of annotated labels with the same pixel weight to train the neural network and segment vessels. Due to the variation in the shape, density and brightness of vessels in whole-brain fluorescence images, it is difficult for a neural network trained with a single type of label to segment all vessels accurately. To address this problem, we proposed a deep learning cerebral vasculature segmentation framework based on multi-perspective labels. First, the pixels in the central region of thick vessels and the skeleton region of vessels were extracted separately using morphological operations based on the binary annotated labels to generate two different labels. Then, we designed a three-stage 3D convolutional neural network containing three sub-networks, namely thick-vessel enhancement network, vessel skeleton enhancement network and multi-channel fusion segmentation network. The first two sub-networks were trained by the two labels generated in the previous step, respectively, and pre-segmented the vessels. The third sub-network was responsible for fusing the pre-segmented results to precisely segment the vessels. We validated our method on two mouse cerebral vascular datasets generated by different fluorescence imaging modalities. The results showed that our method outperforms the state-of-the-art methods, and the proposed method can be applied to segment the vasculature on large-scale volumes.
Collapse
Affiliation(s)
- Yuxin Li
- Shaanxi Key Laboratory of Network Computing and Security Technology, School of Computer Science and Engineering, Xi’an University of Technology, Xi’an, 710048, China
| | - Tong Ren
- Shaanxi Key Laboratory of Network Computing and Security Technology, School of Computer Science and Engineering, Xi’an University of Technology, Xi’an, 710048, China
| | - Junhuai Li
- Shaanxi Key Laboratory of Network Computing and Security Technology, School of Computer Science and Engineering, Xi’an University of Technology, Xi’an, 710048, China
| | - Xiangning Li
- Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, MoE Key Laboratory for Biomedical Photonics, School of Engineering Sciences, Huazhong University of Science and Technology, Wuhan, 430074, China
- HUST-Suzhou Institute for Brainsmatics, Suzhou, 215123, China
| | - Anan Li
- Britton Chance Center for Biomedical Photonics, Wuhan National Laboratory for Optoelectronics, MoE Key Laboratory for Biomedical Photonics, School of Engineering Sciences, Huazhong University of Science and Technology, Wuhan, 430074, China
- HUST-Suzhou Institute for Brainsmatics, Suzhou, 215123, China
| |
Collapse
|
42
|
Li S, Lu C, Kong X, Zhu J, He X, Zhang N. MSFF-Net: Multi-Scale Feature Fusion Network for Gastrointestinal Vessel Segmentation. J Med Biol Eng 2022. [DOI: 10.1007/s40846-022-00704-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/18/2022]
|
43
|
Sun M, Wang Y, Fu Z, Li L, Liu Y, Zhao X. A Machine Learning Method for Automated In Vivo Transparent Vessel Segmentation and Identification Based on Blood Flow Characteristics. MICROSCOPY AND MICROANALYSIS : THE OFFICIAL JOURNAL OF MICROSCOPY SOCIETY OF AMERICA, MICROBEAM ANALYSIS SOCIETY, MICROSCOPICAL SOCIETY OF CANADA 2022; 28:1-14. [PMID: 35387704 DOI: 10.1017/s1431927622000514] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 06/14/2023]
Abstract
In vivo transparent vessel segmentation is important to life science research. However, this task remains very challenging because of the fuzzy edges and the barely noticeable tubular characteristics of vessels under a light microscope. In this paper, we present a new machine learning method based on blood flow characteristics to segment the global vascular structure in vivo. Specifically, the videos of blood flow in transparent vessels are used as input. We use the machine learning classifier to classify the vessel pixels through the motion features extracted from moving red blood cells and achieve vessel segmentation based on a region-growing algorithm. Moreover, we utilize the moving characteristics of blood flow to distinguish between the types of vessels, including arteries, veins, and capillaries. In the experiments, we evaluate the performance of our method on videos of zebrafish embryos. The experimental results indicate the high accuracy of vessel segmentation, with an average accuracy of 97.98%, which is much more superior than other segmentation or motion-detection algorithms. Our method has good robustness when applied to input videos with various time resolutions, with a minimum of 3.125 fps.
Collapse
Affiliation(s)
- Mingzhu Sun
- Institute of Robotics and Automatic Information System (IRAIS) and the Tianjin Key Laboratory of Intelligent Robotic (tjKLIR), Nankai University, Tianjin300350, China
| | - Yiwen Wang
- Institute of Robotics and Automatic Information System (IRAIS) and the Tianjin Key Laboratory of Intelligent Robotic (tjKLIR), Nankai University, Tianjin300350, China
| | - Zhenhua Fu
- Institute of Robotics and Automatic Information System (IRAIS) and the Tianjin Key Laboratory of Intelligent Robotic (tjKLIR), Nankai University, Tianjin300350, China
| | - Lu Li
- Institute of Robotics and Automatic Information System (IRAIS) and the Tianjin Key Laboratory of Intelligent Robotic (tjKLIR), Nankai University, Tianjin300350, China
| | - Yaowei Liu
- Institute of Robotics and Automatic Information System (IRAIS) and the Tianjin Key Laboratory of Intelligent Robotic (tjKLIR), Nankai University, Tianjin300350, China
| | - Xin Zhao
- Institute of Robotics and Automatic Information System (IRAIS) and the Tianjin Key Laboratory of Intelligent Robotic (tjKLIR), Nankai University, Tianjin300350, China
| |
Collapse
|
44
|
Zhang J, Wu F, Chang W, Kong D. Techniques and Algorithms for Hepatic Vessel Skeletonization in Medical Images: A Survey. ENTROPY (BASEL, SWITZERLAND) 2022; 24:465. [PMID: 35455128 PMCID: PMC9031516 DOI: 10.3390/e24040465] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 02/28/2022] [Revised: 03/21/2022] [Accepted: 03/23/2022] [Indexed: 02/01/2023]
Abstract
Hepatic vessel skeletonization serves as an important means of hepatic vascular analysis and vessel segmentation. This paper presents a survey of techniques and algorithms for hepatic vessel skeletonization in medical images. We summarized the latest developments and classical approaches in this field. These methods are classified into five categories according to their methodological characteristics. The overview and brief assessment of each category are provided in the corresponding chapters, respectively. We provide a comprehensive summary among the cited publications, image modalities and datasets from various aspects, which hope to reveal the pros and cons of every method, summarize its achievements and discuss the challenges and future trends.
Collapse
Affiliation(s)
- Jianfeng Zhang
- School of Mathematical Sciences, Zhejiang University, Hangzhou 310027, China; (J.Z.); (W.C.)
- College of Mathematical Medicine, Zhejiang Normal University, Jinhua 321004, China
| | - Fa Wu
- Zhejiang Demetics Medical Technology Co., Ltd., Hangzhou 310012, China;
| | - Wanru Chang
- School of Mathematical Sciences, Zhejiang University, Hangzhou 310027, China; (J.Z.); (W.C.)
| | - Dexing Kong
- School of Mathematical Sciences, Zhejiang University, Hangzhou 310027, China; (J.Z.); (W.C.)
- College of Mathematical Medicine, Zhejiang Normal University, Jinhua 321004, China
| |
Collapse
|
45
|
Zhang C, Wang K, Tian J. Adaptive brightness fusion method for intraoperative near-infrared fluorescence and visible images. BIOMEDICAL OPTICS EXPRESS 2022; 13:1243-1260. [PMID: 35414996 PMCID: PMC8973195 DOI: 10.1364/boe.446176] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Received: 10/18/2021] [Revised: 01/27/2022] [Accepted: 01/27/2022] [Indexed: 06/14/2023]
Abstract
An adaptive brightness fusion method (ABFM) for near-infrared fluorescence imaging is proposed to adapt to different lighting conditions and make the equipment operation more convenient in clinical applications. The ABFM is designed based on the network structure of Attention Unet, which is an image segmentation technique. Experimental results show that ABFM has the function of adaptive brightness adjustment and has better fusion performance in terms of both perception and quantification. Generally, the proposed method can realize an adaptive brightness fusion of fluorescence and visible images to enhance the usability of fluorescence imaging technology during surgery.
Collapse
Affiliation(s)
- Chong Zhang
- Department of Big Data Management and Application, School of International Economics and Management, Beijing Technology and Business University, Beijing 100048, China
- CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
| | - Kun Wang
- CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100049, China
| | - Jie Tian
- CAS Key Laboratory of Molecular Imaging, Institute of Automation, Chinese Academy of Sciences, Beijing 100190, China
- School of Artificial Intelligence, University of Chinese Academy of Sciences, Beijing 100049, China
- BUAA-CCMU Advanced Innovation Center for Big Data-Based Precision Medicine, Beijing 100083, China
| |
Collapse
|
46
|
Lv W, Jian J, Liu J, Zhao X, Xin X, Hu C. Use of the volume-averaged Murray's deviation method for the characterization of branching geometry in liver fibrosis: a preliminary study on vascular circulation. Quant Imaging Med Surg 2022; 12:979-991. [PMID: 35111599 DOI: 10.21037/qims-21-47] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/12/2021] [Accepted: 09/24/2021] [Indexed: 11/06/2022]
Abstract
Background Vascular changes in liver fibrosis can result in increased intrahepatic vascular resistance and impaired blood circulation. This can hinder the recovery from fibrosis and may eventually lead to portal hypertension, a major cirrhosis complication. This report proposed a volume-averaged Murray's deviation method to characterize intrahepatic circulation in the liver during fibrosis and its subsequent regression via X-ray phase-contrast computed tomography (PCCT). Methods Liver fibrosis was induced in 24 Sprague-Dawley rats by exposure to carbon tetrachloride (CCl4) for up to 10 weeks, after which, spontaneous regression commenced and continued until week 30. High-resolution three-dimensional (3D) imaging of the livers was performed with PCCT. The values of Murray's deviation based on the volume-averaged and the conventional diameter-based methods were compared. After that, the intrahepatic circulation at different stages of fibrosis was evaluated using the volume-averaged method. The increase in collagen during liver fibrosis was assessed by pathological analyses. Results A comparison of the 2 methods showed that with an increase in the number of diameter measurements, the value of Murrary's deviation obtained using the diameter-based method gradually approaches those of the volume-averaged method, with minimal variations. The value of Murray's deviation increased with the development of fibrosis. After reversal, the value rapidly decreased and approached that of the normal state in both the main branches (1.05±0.17, 1.17±0.21, 1.34±0.18, and 1.17±0.19 in the normal, moderate, severe, and regressive groups, respectively; P<0.05 between the severe group and other groups) and the small branches (1.05±0.09, 1.42±0.48, 1.79±0.57, and 1.18±0.28 in the normal, moderate, severe, and regressive group, respectively; P<0.05 between adjacent groups). An analysis of Murray's deviation and the pathological results showed that the vascular circulation in this disease model was consistent with the progression and recovery from fibrosis. Conclusions This study showed the validity of the volume-averaged method for calculating Murray's deviation and demonstrated that it could accurately evaluate the blood circulation state of the liver during fibrosis and its subsequent regression. Thus, the volume-averaged method of calculating Murray's deviation may be an objective and valuable staging criterion to evaluate intrahepatic circulation during liver fibrosis.
Collapse
Affiliation(s)
- Wenjuan Lv
- School of Biomedical Engineering and Technology, Tianjin Medical University, Tianjin, China
| | - Jianbo Jian
- Department of Radiation Oncology, Tianjin Medical University General Hospital, Tianjin, China
| | - Jingyi Liu
- School of Biomedical Engineering and Technology, Tianjin Medical University, Tianjin, China
| | - Xinyan Zhao
- Liver Research Center, Beijing Friendship Hospital, Capital Medical University, Beijing, China.,Beijing Key Laboratory of Translational Medicine in Liver Cirrhosis and National Clinical Research Center of Digestive Diseases, Beijing, China
| | - Xiaohong Xin
- School of Biomedical Engineering and Technology, Tianjin Medical University, Tianjin, China
| | - Chunhong Hu
- School of Biomedical Engineering and Technology, Tianjin Medical University, Tianjin, China
| |
Collapse
|
47
|
Zhang X, Lee VCS, Rong J, Liu F, Kong H. Multi-channel convolutional neural network architectures for thyroid cancer detection. PLoS One 2022; 17:e0262128. [PMID: 35061759 PMCID: PMC8782508 DOI: 10.1371/journal.pone.0262128] [Citation(s) in RCA: 2] [Impact Index Per Article: 0.7] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/03/2021] [Accepted: 12/17/2021] [Indexed: 02/05/2023] Open
Abstract
Early detection of malignant thyroid nodules leading to patient-specific treatments can reduce morbidity and mortality rates. Currently, thyroid specialists use medical images to diagnose then follow the treatment protocols, which have limitations due to unreliable human false-positive diagnostic rates. With the emergence of deep learning, advances in computer-aided diagnosis techniques have yielded promising earlier detection and prediction accuracy; however, clinicians' adoption is far lacking. The present study adopts Xception neural network as the base structure and designs a practical framework, which comprises three adaptable multi-channel architectures that were positively evaluated using real-world data sets. The proposed architectures outperform existing statistical and machine learning techniques and reached a diagnostic accuracy rate of 0.989 with ultrasound images and 0.975 with computed tomography scans through the single input dual-channel architecture. Moreover, the patient-specific design was implemented for thyroid cancer detection and has obtained an accuracy of 0.95 for double inputs dual-channel architecture and 0.94 for four-channel architecture. Our evaluation suggests that ultrasound images and computed tomography (CT) scans yield comparable diagnostic results through computer-aided diagnosis applications. With ultrasound images obtained slightly higher results, CT, on the other hand, can achieve the patient-specific diagnostic design. Besides, with the proposed framework, clinicians can select the best fitting architecture when making decisions regarding a thyroid cancer diagnosis. The proposed framework also incorporates interpretable results as evidence, which potentially improves clinicians' trust and hence their adoption of the computer-aided diagnosis techniques proposed with increased efficiency and accuracy.
Collapse
Affiliation(s)
- Xinyu Zhang
- Department of Data Science and AI/Faculty of IT, Monash University, Melbourne, Victoria, Australia
| | - Vincent C. S. Lee
- Department of Data Science and AI/Faculty of IT, Monash University, Melbourne, Victoria, Australia
| | - Jia Rong
- Department of Data Science and AI/Faculty of IT, Monash University, Melbourne, Victoria, Australia
| | - Feng Liu
- West China Hospital of Sichuan University, Chengdu City, Sichuan Province, China
| | - Haoyu Kong
- Department of Human-Centred Computing/Faculty of IT, Monash University, Melbourne, Victoria, Australia
| |
Collapse
|
48
|
Li R, Huang YJ, Chen H, Liu X, Yu Y, Qian D, Wang L. 3D Graph-Connectivity Constrained Network for Hepatic Vessel Segmentation. IEEE J Biomed Health Inform 2021; 26:1251-1262. [PMID: 34613925 DOI: 10.1109/jbhi.2021.3118104] [Citation(s) in RCA: 10] [Impact Index Per Article: 2.5] [Reference Citation Analysis] [Abstract] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/09/2022]
Abstract
Segmentation of hepatic vessels from 3D CT images is necessary for accurate diagnosis and preoper-ative planning for liver cancer. However, due to the low contrast and high noises of CT images, automatic hepatic vessel segmentation is a challenging task. Hepatic vessels are connected branches containing thick and thin blood vessels, showing an important structural characteristic or a prior: the connectivity of blood vessels. However, this is rarely applied in existing methods. In this paper, we segment hepatic vessels from 3D CT images by utilizing the connectivity prior. To this end, a graph neural network (GNN) used to describe the connectivity prior of hepatic vessels is integrated into a general convolutional neu-ral network (CNN). Specifically, a graph attention network (GAT) is first used to model the graphical connectivity information of hepatic vessels, which can be trained with the vascular connectivity graph constructed directly from the ground truths. Second, the GAT is integrated with a lightweight 3D U-Net by an efficient mechanism called the plug-in mode, in which the GAT is incorporated into the U-Net as a multi-task branch and is only used to supervise the training procedure of the U-Net with the connectivity prior. The GAT will not be used in the inference stage, and thus will not increase the hardware and time costs of the inference stage compared with the U-Net. Therefore, hepatic vessel segmentation can be well improved in an efficient mode. Extensive experiments on two public datasets show that the proposed method is superior to related works in accuracy and connectivity of hepatic vessel segmentation.
Collapse
|
49
|
Owais M, Lee YW, Mahmood T, Haider A, Sultan H, Park KR. Multilevel Deep-Aggregated Boosted Network to Recognize COVID-19 Infection from Large-Scale Heterogeneous Radiographic Data. IEEE J Biomed Health Inform 2021; 25:1881-1891. [PMID: 33835928 PMCID: PMC8545161 DOI: 10.1109/jbhi.2021.3072076] [Citation(s) in RCA: 8] [Impact Index Per Article: 2.0] [Reference Citation Analysis] [Abstract] [MESH Headings] [Grants] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 11/21/2022]
Abstract
In the present epidemic of the coronavirus disease 2019 (COVID-19), radiological imaging modalities, such as X-ray and computed tomography (CT), have been identified as effective diagnostic tools. However, the subjective assessment of radiographic examination is a time-consuming task and demands expert radiologists. Recent advancements in artificial intelligence have enhanced the diagnostic power of computer-aided diagnosis (CAD) tools and assisted medical specialists in making efficient diagnostic decisions. In this work, we propose an optimal multilevel deep-aggregated boosted network to recognize COVID-19 infection from heterogeneous radiographic data, including X-ray and CT images. Our method leverages multilevel deep-aggregated features and multistage training via a mutually beneficial approach to maximize the overall CAD performance. To improve the interpretation of CAD predictions, these multilevel deep features are visualized as additional outputs that can assist radiologists in validating the CAD results. A total of six publicly available datasets were fused to build a single large-scale heterogeneous radiographic collection that was used to analyze the performance of the proposed technique and other baseline methods. To preserve generality of our method, we selected different patient data for training, validation, and testing, and consequently, the data of same patient were not included in training, validation, and testing subsets. In addition, fivefold cross-validation was performed in all the experiments for a fair evaluation. Our method exhibits promising performance values of 95.38%, 95.57%, 92.53%, 98.14%, 93.16%, and 98.55% in terms of average accuracy, F-measure, specificity, sensitivity, precision, and area under the curve, respectively and outperforms various state-of-the-art methods.
Collapse
|
50
|
Wang B, Yang J, Ai J, Luo N, An L, Feng H, Yang B, You Z. Accurate Tumor Segmentation via Octave Convolution Neural Network. Front Med (Lausanne) 2021; 8:653913. [PMID: 34095168 PMCID: PMC8169966 DOI: 10.3389/fmed.2021.653913] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/15/2021] [Accepted: 03/24/2021] [Indexed: 11/13/2022] Open
Abstract
Three-dimensional (3D) liver tumor segmentation from Computed Tomography (CT) images is a prerequisite for computer-aided diagnosis, treatment planning, and monitoring of liver cancer. Despite many years of research, 3D liver tumor segmentation remains a challenging task. In this paper, we propose an effective and efficient method for tumor segmentation in liver CT images using encoder-decoder based octave convolution networks. Compared with other convolution networks utilizing standard convolution for feature extraction, the proposed method utilizes octave convolutions for learning multiple-spatial-frequency features, thus can better capture tumors with varying sizes and shapes. The proposed network takes advantage of a fully convolutional architecture which performs efficient end-to-end learning and inference. More importantly, we introduce a deep supervision mechanism during the learning process to combat potential optimization difficulties, and thus the model can acquire a much faster convergence rate and more powerful discrimination capability. Finally, we integrate octave convolutions into the encoder-decoder architecture of UNet, which can generate high resolution tumor segmentation in one single forward feeding without post-processing steps. Both architectures are trained on a subset of the LiTS (Liver Tumor Segmentation) Challenge. The proposed approach is shown to significantly outperform other networks in terms of various accuracy measures and processing speed.
Collapse
Affiliation(s)
- Bo Wang
- The State Key Laboratory of Precision Measurement Technology and Instruments, Department of Precision Instrument, Tsinghua University, Beijing, China.,Innovation Center for Future Chips, Tsinghua University, Beijing, China.,Beijing Jingzhen Medical Technology Ltd., Beijing, China
| | - Jingyi Yang
- School of Artificial Intelligence, Xidian University, Xi'an, China
| | - Jingyang Ai
- Beijing Jingzhen Medical Technology Ltd., Beijing, China
| | - Nana Luo
- Affiliated Hospital of Jining Medical University, Jining, China
| | - Lihua An
- Affiliated Hospital of Jining Medical University, Jining, China
| | - Haixia Feng
- Affiliated Hospital of Jining Medical University, Jining, China
| | - Bo Yang
- China Institute of Marine Technology & Economy, Beijing, China
| | - Zheng You
- The State Key Laboratory of Precision Measurement Technology and Instruments, Department of Precision Instrument, Tsinghua University, Beijing, China.,Innovation Center for Future Chips, Tsinghua University, Beijing, China
| |
Collapse
|