1
|
Ghobadi V, Ismail LI, Wan Hasan WZ, Ahmad H, Ramli HR, Norsahperi NMH, Tharek A, Hanapiah FA. Challenges and solutions of deep learning-based automated liver segmentation: A systematic review. Comput Biol Med 2025; 185:109459. [PMID: 39642700 DOI: 10.1016/j.compbiomed.2024.109459] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/05/2024] [Revised: 11/12/2024] [Accepted: 11/19/2024] [Indexed: 12/09/2024]
Abstract
The liver is one of the vital organs in the body. Precise liver segmentation in medical images is essential for liver disease treatment. The deep learning-based liver segmentation process faces several challenges. This research aims to analyze the challenges of liver segmentation in prior studies and identify the modifications made to network models and other enhancements implemented by researchers to tackle each challenge. In total, 88 articles from Scopus and ScienceDirect databases published between January 2016 and January 2022 have been studied. The liver segmentation challenges are classified into five main categories, each containing some subcategories. For each challenge, the proposed technique to overcome the challenge is investigated. The provided report details the authors, publication years, dataset types, imaging technologies, and evaluation metrics of all references for comparison. Additionally, a summary table outlines the challenges and solutions.
Collapse
Affiliation(s)
- Vahideh Ghobadi
- Faculty of Engineering, Universiti Putra Malaysia, Serdang, 43400, Selangor, Malaysia.
| | - Luthffi Idzhar Ismail
- Faculty of Engineering, Universiti Putra Malaysia, Serdang, 43400, Selangor, Malaysia.
| | - Wan Zuha Wan Hasan
- Faculty of Engineering, Universiti Putra Malaysia, Serdang, 43400, Selangor, Malaysia.
| | - Haron Ahmad
- KPJ Specialist Hospital, Damansara Utama, Petaling Jaya, 47400, Selangor, Malaysia.
| | - Hafiz Rashidi Ramli
- Faculty of Engineering, Universiti Putra Malaysia, Serdang, 43400, Selangor, Malaysia.
| | | | - Anas Tharek
- Hospital Sultan Abdul Aziz Shah, University Putra Malaysia, Serdang, 43400, Selangor, Malaysia.
| | - Fazah Akhtar Hanapiah
- Faculty of Medicine, Universiti Teknologi MARA, Damansara Utama, Sungai Buloh, 47000, Selangor, Malaysia.
| |
Collapse
|
2
|
Vadlamudi S, Kumar V, Ghosh D, Abraham A. Artificial intelligence-powered precision: Unveiling the landscape of liver disease diagnosis—A comprehensive review. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE 2024; 138:109452. [DOI: 10.1016/j.engappai.2024.109452] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 01/07/2025]
|
3
|
Sun B, Gan C, Tang Y, Xu Q, Wang K, Zhu F. Identification and validation of a prognostic model based on three TLS-Related genes in oral squamous cell carcinoma. Cancer Cell Int 2024; 24:350. [PMID: 39462422 PMCID: PMC11515094 DOI: 10.1186/s12935-024-03543-7] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/02/2024] [Accepted: 10/21/2024] [Indexed: 10/29/2024] Open
Abstract
BACKGROUND The tertiary lymphoid structures (TLSs) have an immunomodulatory function and have a positive impact on the survival outcomes of patients with oral squamous cell carcinoma (OSCC). However, there is a lack of standard approaches for quantifying TLSs and prognostic models using TLS-related genes (TLSRGs). These limitations limit the widespread use of TLSs in clinical practice. METHODS A convolutional neural network was used to automatically detect and quantify TLSs in HE-stained whole slide images. By employing bioinformatics and diverse statistical methods, this research created a prognostic model using TCGA cohorts and explored the connection between this model and immune infiltration. The expression levels of three TLSRGs in clinical specimens were detected by immunohistochemistry. To facilitate the assessment of individual prognostic outcomes, we further constructed a nomogram based on the risk score and other clinical factors. RESULTS TLSs were found to be an independent predictor of both overall survival (OS) and disease-free survival in OSCC patients. A larger proportion of the TLS area represented a better prognosis. After analysis, we identified 69 differentially expressed TLSRGs and selected three pivotal TLSRGs to construct the risk score model. This model emerged as a standalone predictor for OS and exhibited close associations with CD4 + T cells, CD8 + T cells, and macrophages. Immunohistochemistry revealed high expression levels of CCR7 and CXCR5 in TLS + OSCC samples, while CD86 was highly expressed in TLS- OSCC samples. The nomogram demonstrates excellent predictive ability for overall survival in OSCC patients. CONCLUSIONS This is the first prognostic nomogram based on TLSRGs, that can effectively predict survival outcomes and contribute to individual treatment strategies for OSCC patients.
Collapse
Affiliation(s)
- Bincan Sun
- Department of Oral and Maxillofacial Surgery, Center of Stomatology, Xiangya Hospital, Central South University, Changsha, Hunan, P. R. China
- Xiangya Stomatological Hospital & Xiangya School of Stomatology, Central South University, Changsha, Hunan, P. R. China
| | - Chengwen Gan
- Department of Oral Maxillofacial Surgery, Hainan General Hospital, Haikou, Hainan, P. R. China
| | - Yan Tang
- Department of Nursing, The Second Affiliated Hospital of Hunan University of Chinese Medicine, Changsha, Hunan, P. R. China
| | - Qian Xu
- Department of Oral and Maxillofacial Surgery, Center of Stomatology, Xiangya Hospital, Central South University, Changsha, Hunan, P. R. China
- Research Center of Oral and Maxillofacial Tumor, Xiangya Hospital, Central South University, Changsha, Hunan, P. R. China
- Institute of Oral Cancer and Precancerous Lesions, Central South University, Changsha, Hunan, P. R. China
| | - Kai Wang
- Department of Oral and Maxillofacial Surgery, The Second Xiangya Hospital of Central South University, Changsha, Hunan, P. R. China
| | - Feiya Zhu
- Department of Oral and Maxillofacial Surgery, Center of Stomatology, Xiangya Hospital, Central South University, Changsha, Hunan, P. R. China.
- Research Center of Oral and Maxillofacial Tumor, Xiangya Hospital, Central South University, Changsha, Hunan, P. R. China.
- Institute of Oral Cancer and Precancerous Lesions, Central South University, Changsha, Hunan, P. R. China.
- National Engineering Research Center of Oral Biomaterials and Digital Medical Devices, Peking University School and Hospital of Stomatology, Beijing, P. R. China.
| |
Collapse
|
4
|
Delmoral JC, R S Tavares JM. Semantic Segmentation of CT Liver Structures: A Systematic Review of Recent Trends and Bibliometric Analysis : Neural Network-based Methods for Liver Semantic Segmentation. J Med Syst 2024; 48:97. [PMID: 39400739 PMCID: PMC11473507 DOI: 10.1007/s10916-024-02115-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 05/22/2024] [Accepted: 10/02/2024] [Indexed: 10/15/2024]
Abstract
The use of artificial intelligence (AI) in the segmentation of liver structures in medical images has become a popular research focus in the past half-decade. The performance of AI tools in screening for this task may vary widely and has been tested in the literature in various datasets. However, no scientometric report has provided a systematic overview of this scientific area. This article presents a systematic and bibliometric review of recent advances in neuronal network modeling approaches, mainly of deep learning, to outline the multiple research directions of the field in terms of algorithmic features. Therefore, a detailed systematic review of the most relevant publications addressing fully automatic semantic segmenting liver structures in Computed Tomography (CT) images in terms of algorithm modeling objective, performance benchmark, and model complexity is provided. The review suggests that fully automatic hybrid 2D and 3D networks are the top performers in the semantic segmentation of the liver. In the case of liver tumor and vasculature segmentation, fully automatic generative approaches perform best. However, the reported performance benchmark indicates that there is still much to be improved in segmenting such small structures in high-resolution abdominal CT scans.
Collapse
Affiliation(s)
- Jessica C Delmoral
- Instituto de Ciência e Inovação em Engenharia Mecânica e Engenharia Industrial, Faculdade de Engenharia, Universidade do Porto, Rua Dr. Roberto Frias, s/n, 4200-465, Porto, Portugal
| | - João Manuel R S Tavares
- Instituto de Ciência e Inovação em Engenharia Mecânica e Engenharia Industrial, Departamento de Engenharia Mecânica, Faculdade de Engenharia, Universidade do Porto, Rua Dr. Roberto Frias, s/n, 4200-465, Porto, Portugal.
| |
Collapse
|
5
|
Khan R, Su L, Zaman A, Hassan H, Kang Y, Huang B. Customized m-RCNN and hybrid deep classifier for liver cancer segmentation and classification. Heliyon 2024; 10:e30528. [PMID: 38765046 PMCID: PMC11096931 DOI: 10.1016/j.heliyon.2024.e30528] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 12/30/2023] [Revised: 04/28/2024] [Accepted: 04/29/2024] [Indexed: 05/21/2024] Open
Abstract
Diagnosing liver disease presents a significant medical challenge in impoverished countries, with over 30 billion individuals succumbing to it each year. Existing models for detecting liver abnormalities suffer from lower accuracy and higher constraint metrics. As a result, there is a pressing need for improved, efficient, and effective liver disease detection methods. To address the limitations of current models, this method introduces a deep liver segmentation and classification system based on a Customized Mask-Region Convolutional Neural Network (cm-RCNN). The process begins with preprocessing the input liver image, which includes Adaptive Histogram Equalization (AHE). AHE helps dehaze the input image, remove color distortion, and apply linear transformations to obtain the preprocessed image. Next, a precise region of interest is segmented from the preprocessed image using a novel deep strategy called cm-RCNN. To enhance segmentation accuracy, the architecture incorporates the ReLU activation function and the modified sigmoid activation function. Subsequently, a variety of features are extracted from the segmented image, including ResNet features, shape features (area, perimeter, approximation, and convex hull), and enhanced median binary pattern. These extracted features are then used to train a hybrid classification model, which incorporates classifiers like SqueezeNet and DeepMaxout models. The final classification outcome is determined by averaging the scores obtained from both classifiers.
Collapse
Affiliation(s)
- Rashid Khan
- College of Applied Sciences, Shenzhen University, Shenzhen, 518060, China
- Guangdong Key Laboratory for Biomedical Measurements and Ultrasound Imaging, National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, School of Biomedical Engineering, Shenzhen University Medical School, Shenzhen, 518060, China
- College of Big Data and Internet, Shenzhen Technology University, Shenzhen, 518188, China
| | - Liyilei Su
- College of Applied Sciences, Shenzhen University, Shenzhen, 518060, China
- Guangdong Key Laboratory for Biomedical Measurements and Ultrasound Imaging, National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, School of Biomedical Engineering, Shenzhen University Medical School, Shenzhen, 518060, China
- College of Big Data and Internet, Shenzhen Technology University, Shenzhen, 518188, China
| | - Asim Zaman
- College of Health Science and Environmental Engineering, Shenzhen Technology University, Shenzhen, 518188, China
| | - Haseeb Hassan
- College of Health Science and Environmental Engineering, Shenzhen Technology University, Shenzhen, 518188, China
| | - Yan Kang
- Guangdong Key Laboratory for Biomedical Measurements and Ultrasound Imaging, National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, School of Biomedical Engineering, Shenzhen University Medical School, Shenzhen, 518060, China
- College of Health Science and Environmental Engineering, Shenzhen Technology University, Shenzhen, 518188, China
| | - Bingding Huang
- College of Applied Sciences, Shenzhen University, Shenzhen, 518060, China
- College of Big Data and Internet, Shenzhen Technology University, Shenzhen, 518188, China
| |
Collapse
|
6
|
Li Y, Yan B, Hou J, Bai B, Huang X, Xu C, Fang L. UNet based on dynamic convolution decomposition and triplet attention. Sci Rep 2024; 14:271. [PMID: 38168684 PMCID: PMC10761743 DOI: 10.1038/s41598-023-50989-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 09/17/2023] [Accepted: 12/28/2023] [Indexed: 01/05/2024] Open
Abstract
The robustness and generalization of medical image segmentation models are being challenged by the differences between different disease types, different image types, and different cases.Deep learning based semantic segmentation methods have been providing state-of-the-art performance in the last few years. One deep learning technique, U-Net, has become the most popular architecture in the medical imaging segmentation. Despite outstanding overall performance in segmenting medical images, it still has the problems of limited feature expression ability and inaccurate segmentation. To this end, we propose a DTA-UNet based on Dynamic Convolution Decomposition (DCD) and Triple Attention (TA). Firstly, the model with Attention U-Net as the baseline network uses DCD to replace all the conventional convolution in the encoding-decoding process to enhance its feature extraction capability. Secondly, we combine TA with Attention Gate (AG) to be used for skip connection in order to highlight lesion regions by removing redundant information in both spatial and channel dimensions. The proposed model are tested on the two public datasets and actual clinical dataset such as the public COVID-SemiSeg dataset, the ISIC 2018 dataset, and the cooperative hospital stroke segmentation dataset. Ablation experiments on the clinical stroke segmentation dataset show the effectiveness of DCD and TA with only a 0.7628 M increase in the number of parameters compared to the baseline model. The proposed DTA-UNet is further evaluated on the three datasets of different types of images to verify its universality. Extensive experimental results show superior performance on different segmentation metrics compared to eight state-of-art methods.The GitHub URL of our code is https://github.com/shuaihou1234/DTA-UNet .
Collapse
Affiliation(s)
- Yang Li
- Academy for Advanced Interdisciplinary Studies, Northeast Normal University, Changchun, 130024, Jilin, China
- Shanghai Zhangjiang Institute of Mathematics, Shanghai, 201203, China
| | - Bobo Yan
- School of Computer Science and Engineering, Changchun University of Technology, Changchun, 130012, Jilin, China
- Pazhou Lab, Guangzhou, China
| | - Jianxin Hou
- School of Computer Science and Engineering, Changchun University of Technology, Changchun, 130012, Jilin, China
| | - Bingyang Bai
- School of Computer Science and Engineering, Changchun University of Technology, Changchun, 130012, Jilin, China
| | - Xiaoyu Huang
- School of Computer Science and Engineering, Changchun University of Technology, Changchun, 130012, Jilin, China
| | - Canfei Xu
- The Third Hospital of Jilin University, Changchun, 130117, Jilin, China
| | - Limei Fang
- Encephalopathy Center, The Third Affiliated Hospital of Changchun University of Chinese Medicine, Changchun, 130117, Jilin, China.
| |
Collapse
|
7
|
Saumiya S, Franklin SW. Residual Deformable Split Channel and Spatial U-Net for Automated Liver and Liver Tumour Segmentation. J Digit Imaging 2023; 36:2164-2178. [PMID: 37464213 PMCID: PMC10501969 DOI: 10.1007/s10278-023-00874-1] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/10/2023] [Revised: 06/20/2023] [Accepted: 06/21/2023] [Indexed: 07/20/2023] Open
Abstract
Accurate segmentation of the liver and liver tumour (LT) is challenging due to its hazy boundaries and large shape variability. Although using U-Net for liver and LT segmentation achieves better results than manual segmentation, it loses spatial and channel features during segmentation, leading to inaccurate liver and LT segmentation. A residual deformable split depth-wise separable U-Net (RDSDSU-Net) is proposed to increase the accuracy of liver and LT segmentation. The residual deformable convolution layer (DCL) with deformable pooling (DP) is used in the encoder as an attention mechanism to adaptively extract liver and LT shape and position characteristics. Afterward, a convolutional spatial and channel features split graph network (CSCFSG-Net) is introduced in the middle processing layer to improve the expression capability of the liver and LT features by capturing spatial and channel features separately and to extract global contextual liver and LT information from spatial and channel features. Sub-pixel convolutions (SPC) are used in the decoder section to prevent the segmentation results from having a chequerboard artefact effect. Also, the residual deformable encoder features are combined with the decoder through summation to avoid increasing the number of feature maps (FM). Finally, the efficiency of the RDSDSU-Net is evaluated on the 3DIRCADb and LiTS datasets. The DICE score of the proposed RDSDSU-Net achieved 98.21% for liver segmentation and 93.25% for LT segmentation on 3DIRCADb. The experimental outcomes illustrate that the proposed RDSDSU-Net model achieved better segmentation results than the existing techniques.
Collapse
Affiliation(s)
- S Saumiya
- Department of ECE, Bethlahem Institute of Engineering, Karungal, Tamil Nadu India
| | - S Wilfred Franklin
- Department of ECE, CSI Institute of Technology, Thovalai, Tamil Nadu India
| |
Collapse
|
8
|
Ananda S, Jain RK, Li Y, Iwamoto Y, Han XH, Kanasaki S, Hu H, Chen YW. A Boundary-Enhanced Liver Segmentation Network for Multi-Phase CT Images with Unsupervised Domain Adaptation. Bioengineering (Basel) 2023; 10:899. [PMID: 37627784 PMCID: PMC10451706 DOI: 10.3390/bioengineering10080899] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/07/2023] [Revised: 07/23/2023] [Accepted: 07/24/2023] [Indexed: 08/27/2023] Open
Abstract
Multi-phase computed tomography (CT) images have gained significant popularity in the diagnosis of hepatic disease. There are several challenges in the liver segmentation of multi-phase CT images. (1) Annotation: due to the distinct contrast enhancements observed in different phases (i.e., each phase is considered a different domain), annotating all phase images in multi-phase CT images for liver or tumor segmentation is a task that consumes substantial time and labor resources. (2) Poor contrast: some phase images may have poor contrast, making it difficult to distinguish the liver boundary. In this paper, we propose a boundary-enhanced liver segmentation network for multi-phase CT images with unsupervised domain adaptation. The first contribution is that we propose DD-UDA, a dual discriminator-based unsupervised domain adaptation, for liver segmentation on multi-phase images without multi-phase annotations, effectively tackling the annotation problem. To improve accuracy by reducing distribution differences between the source and target domains, we perform domain adaptation at two levels by employing two discriminators, one at the feature level and the other at the output level. The second contribution is that we introduce an additional boundary-enhanced decoder to the encoder-decoder backbone segmentation network to effectively recognize the boundary region, thereby addressing the problem of poor contrast. In our study, we employ the public LiTS dataset as the source domain and our private MPCT-FLLs dataset as the target domain. The experimental findings validate the efficacy of our proposed methods, producing substantially improved results when tested on each phase of the multi-phase CT image even without the multi-phase annotations. As evaluated on the MPCT-FLLs dataset, the existing baseline (UDA) method achieved IoU scores of 0.785, 0.796, and 0.772 for the PV, ART, and NC phases, respectively, while our proposed approach exhibited superior performance, surpassing both the baseline and other state-of-the-art methods. Notably, our method achieved remarkable IoU scores of 0.823, 0.811, and 0.800 for the PV, ART, and NC phases, respectively, emphasizing its effectiveness in achieving accurate image segmentation.
Collapse
Affiliation(s)
- Swathi Ananda
- Graduate School of Information Science and Engineering, Ritsumeikan University, Kusatsu-shi 525-0058, Japan; (S.A.); (R.K.J.); (Y.L.)
| | - Rahul Kumar Jain
- Graduate School of Information Science and Engineering, Ritsumeikan University, Kusatsu-shi 525-0058, Japan; (S.A.); (R.K.J.); (Y.L.)
| | - Yinhao Li
- Graduate School of Information Science and Engineering, Ritsumeikan University, Kusatsu-shi 525-0058, Japan; (S.A.); (R.K.J.); (Y.L.)
| | - Yutaro Iwamoto
- Faculty of Information and Communication Engineering, Osaka Electro-Communication University, Neyagawa-shi 572-0833, Japan;
| | - Xian-Hua Han
- Artificial Intelligence Research Center, Yamaguchi University, Yamaguchi-shi 753-8511, Japan;
| | | | - Hongjie Hu
- Department of Radiology Sir Run Run Shaw, Zhejiang University, Hangzhou 310016, China;
| | - Yen-Wei Chen
- Graduate School of Information Science and Engineering, Ritsumeikan University, Kusatsu-shi 525-0058, Japan; (S.A.); (R.K.J.); (Y.L.)
| |
Collapse
|
9
|
Wu S, Yu H, Li C, Zheng R, Xia X, Wang C, Wang H. A Coarse-to-Fine Fusion Network for Small Liver Tumor Detection and Segmentation: A Real-World Study. Diagnostics (Basel) 2023; 13:2504. [PMID: 37568868 PMCID: PMC10417427 DOI: 10.3390/diagnostics13152504] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 06/19/2023] [Revised: 07/14/2023] [Accepted: 07/18/2023] [Indexed: 08/13/2023] Open
Abstract
Liver tumor semantic segmentation is a crucial task in medical image analysis that requires multiple MRI modalities. This paper proposes a novel coarse-to-fine fusion segmentation approach to detect and segment small liver tumors of various sizes. To enhance the segmentation accuracy of small liver tumors, the method incorporates a detection module and a CSR (convolution-SE-residual) module, which includes a convolution block, an SE (squeeze and excitation) module, and a residual module for fine segmentation. The proposed method demonstrates superior performance compared to conventional single-stage end-to-end networks. A private liver MRI dataset comprising 218 patients with a total of 3605 tumors, including 3273 tumors smaller than 3.0 cm, were collected for the proposed method. There are five types of liver tumors identified in this dataset: hepatocellular carcinoma (HCC); metastases of the liver; cholangiocarcinoma (ICC); hepatic cyst; and liver hemangioma. The results indicate that the proposed method outperforms the single segmentation networks 3D UNet and nnU-Net as well as the fusion networks of 3D UNet and nnU-Net with nnDetection. The proposed architecture was evaluated on a test set of 44 images, with an average Dice similarity coefficient (DSC) and recall of 86.9% and 86.7%, respectively, which is a 1% improvement compared to the comparison method. More importantly, compared to existing methods, our proposed approach demonstrates state-of-the-art performance in segmenting small objects with sizes smaller than 10 mm, achieving a Dice score of 85.3% and a malignancy detection rate of 87.5%.
Collapse
Affiliation(s)
- Shu Wu
- Zhiyu Software Information Co., Ltd., Shanghai 200030, China
| | - Hang Yu
- Institute of Science and Technology for Brain-Inspired Intelligence, Fudan University, Shanghai 200433, China
| | - Cuiping Li
- Zhiyu Software Information Co., Ltd., Shanghai 200030, China
| | - Rencheng Zheng
- Institute of Science and Technology for Brain-Inspired Intelligence, Fudan University, Shanghai 200433, China
| | - Xueqin Xia
- Institute of Science and Technology for Brain-Inspired Intelligence, Fudan University, Shanghai 200433, China
| | - Chengyan Wang
- Human Phenome Institute, Fudan University, Shanghai 200433, China
| | - He Wang
- Institute of Science and Technology for Brain-Inspired Intelligence, Fudan University, Shanghai 200433, China
- Human Phenome Institute, Fudan University, Shanghai 200433, China
- Department of Neurology, Zhongshan Hospital, Fudan University, Shanghai 200032, China
- Key Laboratory of Computational Neuroscience and Brain-Inspired Intelligence (Fudan University), Ministry of Education, Shanghai 200433, China
| |
Collapse
|
10
|
Aldughayfiq B, Ashfaq F, Jhanjhi NZ, Humayun M. YOLO-Based Deep Learning Model for Pressure Ulcer Detection and Classification. Healthcare (Basel) 2023; 11:healthcare11091222. [PMID: 37174764 PMCID: PMC10178524 DOI: 10.3390/healthcare11091222] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [Grants] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/12/2023] [Revised: 04/15/2023] [Accepted: 04/22/2023] [Indexed: 05/15/2023] Open
Abstract
Pressure ulcers are significant healthcare concerns affecting millions of people worldwide, particularly those with limited mobility. Early detection and classification of pressure ulcers are crucial in preventing their progression and reducing associated morbidity and mortality. In this work, we present a novel approach that uses YOLOv5, an advanced and robust object detection model, to detect and classify pressure ulcers into four stages and non-pressure ulcers. We also utilize data augmentation techniques to expand our dataset and strengthen the resilience of our model. Our approach shows promising results, achieving an overall mean average precision of 76.9% and class-specific mAP50 values ranging from 66% to 99.5%. Compared to previous studies that primarily utilize CNN-based algorithms, our approach provides a more efficient and accurate solution for the detection and classification of pressure ulcers. The successful implementation of our approach has the potential to improve the early detection and treatment of pressure ulcers, resulting in better patient outcomes and reduced healthcare costs.
Collapse
Affiliation(s)
- Bader Aldughayfiq
- Department of Information Systems, College of Computer and Information Sciences, Jouf University, Sakaka 72388, Saudi Arabia
| | - Farzeen Ashfaq
- School of Computer Science, SCS, Taylor's University, Subang Jaya 47500, Malaysia
| | - N Z Jhanjhi
- School of Computer Science, SCS, Taylor's University, Subang Jaya 47500, Malaysia
| | - Mamoona Humayun
- Department of Information Systems, College of Computer and Information Sciences, Jouf University, Sakaka 72388, Saudi Arabia
| |
Collapse
|
11
|
Zhang F, Wang Q, Fan E, Lu N, Chen D, Jiang H, Wang Y. Automatic segmentation of the tumor in nonsmall-cell lung cancer by combining coarse and fine segmentation. Med Phys 2022. [PMID: 36514264 DOI: 10.1002/mp.16158] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 03/24/2022] [Revised: 08/05/2022] [Accepted: 11/26/2022] [Indexed: 12/15/2022] Open
Abstract
OBJECTIVES Radiotherapy plays an important role in the treatment of nonsmall-cell lung cancer (NSCLC). Accurate delineation of tumor is the key to successful radiotherapy. Compared with the commonly used manual delineation ways, which are time-consuming and laborious, the automatic segmentation methods based on deep learning can greatly improve the treatment efficiency. METHODS In this paper, we introduce an automatic segmentation method by combining coarse and fine segmentations for NSCLC. Coarse segmentation network is the first level, identifing the rough region of the tumor. In this network, according to the tissue structure distribution of the thoracic cavity where tumor is located, we designed a competition method between tumors and organs at risk (OARs), which can increase the proportion of the identified tumor covering the ground truth and reduce false identification. Fine segmentation network is the second level, carrying out precise segmentation on the results of the coarse level. These two networks are independent of each other during training. When they are used, morphological processing of small scale corrosion and large scale expansion is used for the coarse segmentation results, and the outcomes are sent to the fine segmentation part as input, so as to achieve the complementary advantages of the two networks. RESULTS In the experiment, CT images of 200 patients with NSCLC are used to train the network, and CT images of 60 patients are used to test. Finally, our method produced the Dice similarity coefficient of 0.78 ± 0.10. CONCLUSIONS The experimental results show that the proposed method can accurately segment the tumor with NSCLC, and can also provide support for clinical diagnosis and treatment.
Collapse
Affiliation(s)
- Fuli Zhang
- Radiation Oncology Department, The Seventh Medical Center of Chinese PLA General Hospital, Beijing, China
| | - Qiusheng Wang
- School of Automation Science and Electrical Engineering, Beihang University, Beijing, China
| | - Enyu Fan
- School of Automation Science and Electrical Engineering, Beihang University, Beijing, China
| | - Na Lu
- Senior Department of Oncology, The Fifth Medical Center of PLA General Hospital, Beijing, China
| | - Diandian Chen
- Senior Department of Oncology, The Fifth Medical Center of PLA General Hospital, Beijing, China
| | - Huayong Jiang
- Senior Department of Oncology, The Fifth Medical Center of PLA General Hospital, Beijing, China
| | - Yadi Wang
- Senior Department of Oncology, The Fifth Medical Center of PLA General Hospital, Beijing, China
| |
Collapse
|
12
|
Gojić G, Petrović VB, Dragan D, Gajić DB, Mišković D, Džinić V, Grgić Z, Pantelić J, Oros A. Comparing the Clinical Viability of Automated Fundus Image Segmentation Methods. SENSORS (BASEL, SWITZERLAND) 2022; 22:9101. [PMID: 36501801 PMCID: PMC9735987 DOI: 10.3390/s22239101] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Figures] [Subscribe] [Scholar Register] [Received: 11/02/2022] [Revised: 11/16/2022] [Accepted: 11/17/2022] [Indexed: 06/17/2023]
Abstract
Recent methods for automatic blood vessel segmentation from fundus images have been commonly implemented as convolutional neural networks. While these networks report high values for objective metrics, the clinical viability of recovered segmentation masks remains unexplored. In this paper, we perform a pilot study to assess the clinical viability of automatically generated segmentation masks in the diagnosis of diseases affecting retinal vascularization. Five ophthalmologists with clinical experience were asked to participate in the study. The results demonstrate low classification accuracy, inferring that generated segmentation masks cannot be used as a standalone resource in general clinical practice. The results also hint at possible clinical infeasibility in experimental design. In the follow-up experiment, we evaluate the clinical quality of masks by having ophthalmologists rank generation methods. The ranking is established with high intra-observer consistency, indicating better subjective performance for a subset of tested networks. The study also demonstrates that objective metrics are not correlated with subjective metrics in retinal segmentation tasks for the methods involved, suggesting that objective metrics commonly used in scientific papers to measure the method's performance are not plausible criteria for choosing clinically robust solutions.
Collapse
Affiliation(s)
- Gorana Gojić
- The Institute for Artificial Intelligence Research and Development of Serbia, 21102 Novi Sad, Serbia
- Faculty of Technical Sciences, University of Novi Sad, 21102 Novi Sad, Serbia
| | - Veljko B. Petrović
- Faculty of Technical Sciences, University of Novi Sad, 21102 Novi Sad, Serbia
| | - Dinu Dragan
- Faculty of Technical Sciences, University of Novi Sad, 21102 Novi Sad, Serbia
| | - Dušan B. Gajić
- Faculty of Technical Sciences, University of Novi Sad, 21102 Novi Sad, Serbia
| | - Dragiša Mišković
- The Institute for Artificial Intelligence Research and Development of Serbia, 21102 Novi Sad, Serbia
| | | | | | - Jelica Pantelić
- Institute of Eye Diseases, University Clinical Center of Serbia, 11000 Belgrade, Serbia
| | - Ana Oros
- Eye Clinic Džinić, 21107 Novi Sad, Serbia
- Institute of Neonatology, 11000 Belgrade, Serbia
| |
Collapse
|
13
|
Yin TK, Huang KL, Chiu SR, Yang YQ, Chang BR. Endoscopy Artefact Detection by Deep Transfer Learning of Baseline Models. J Digit Imaging 2022; 35:1101-1110. [PMID: 35478060 PMCID: PMC9582060 DOI: 10.1007/s10278-022-00627-6] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 02/28/2021] [Revised: 03/28/2022] [Accepted: 03/30/2022] [Indexed: 10/18/2022] Open
Abstract
To visualise the tumours inside the body on a screen, a long and thin tube is inserted with a light source and a camera at the tip to obtain video frames inside organs in endoscopy. However, multiple artefacts exist in these video frames that cause difficulty during the diagnosis of cancers. In this research, deep learning was applied to detect eight kinds of artefacts: specularity, bubbles, saturation, contrast, blood, instrument, blur, and imaging artefacts. Based on transfer learning with pre-trained parameters and fine-tuning, two state-of-the-art methods were applied for detection: faster region-based convolutional neural networks (Faster R-CNN) and EfficientDet. Experiments were implemented on the grand challenge dataset, Endoscopy Artefact Detection and Segmentation (EAD2020). To validate our approach in this study, we used phase I of 2,200 frames and phase II of 331 frames in the original training dataset with ground-truth annotations as training and testing dataset, respectively. Among the tested methods, EfficientDet-D2 achieves a score of 0.2008 (mAPd[Formula: see text]0.6+mIoUd[Formula: see text]0.4) on the dataset that is better than three other baselines: Faster-RCNN, YOLOv3, and RetinaNet, and competitive to the best non-baseline result scored 0.25123 on the leaderboard although our testing was on phase II of 331 frames instead of the original 200 testing frames. Without extra improvement techniques beyond basic neural networks such as test-time augmentation, we showed that a simple baseline could achieve state-of-the-art performance in detecting artefacts in endoscopy. In conclusion, we proposed the combination of EfficientDet-D2 with suitable data augmentation and pre-trained parameters during fine-tuning training to detect the artefacts in endoscopy.
Collapse
Affiliation(s)
- Tang-Kai Yin
- Department of Computer Science and Information Engineering, National University of Kaohsiung, No. 700, Kaohsiung University Rd., Nan-Tzu Dist., 811, Kaohsiung, Taiwan.
| | - Kai-Lun Huang
- Department of Computer Science and Information Engineering, National University of Kaohsiung, No. 700, Kaohsiung University Rd., Nan-Tzu Dist., 811, Kaohsiung, Taiwan
| | - Si-Rong Chiu
- Department of Computer Science and Information Engineering, National University of Kaohsiung, No. 700, Kaohsiung University Rd., Nan-Tzu Dist., 811, Kaohsiung, Taiwan
| | - Yu-Qi Yang
- Department of Computer Science and Information Engineering, National University of Kaohsiung, No. 700, Kaohsiung University Rd., Nan-Tzu Dist., 811, Kaohsiung, Taiwan
| | - Bao-Rong Chang
- Department of Computer Science and Information Engineering, National University of Kaohsiung, No. 700, Kaohsiung University Rd., Nan-Tzu Dist., 811, Kaohsiung, Taiwan
| |
Collapse
|
14
|
A teacher-student framework for liver and tumor segmentation under mixed supervision from abdominal CT scans. Neural Comput Appl 2022. [DOI: 10.1007/s00521-022-07240-2] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/17/2022]
|
15
|
Image Segmentation via Multiscale Perceptual Grouping. Symmetry (Basel) 2022. [DOI: 10.3390/sym14061076] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 12/10/2022] Open
Abstract
The human eyes observe an image through perceptual units surrounded by symmetrical or asymmetrical object contours at a proper scale, which enables them to quickly extract the foreground of the image. Inspired by this characteristic, a model combined with multiscale perceptual grouping and unit-based segmentation is proposed in this paper. In the multiscale perceptual grouping part, a novel total variation regularization is proposed to smooth the image into different scales, which removes the inhomogeneity and preserves the edges. To simulate perceptual units surrounded by contours, the watershed method is utilized to cluster pixels into groups. The scale of smoothness is determined by the number of perceptual units. In the segmentation part, perceptual units are regarded as the basic element instead of discrete pixels in the graph cut. The appearance models of the foreground and background are constructed by combining the perceptual units. According to the relationship between perceptual units and the appearance model, the foreground can be segmented through a minimum-cut/maximum-flow algorithm. The experiment conducted on the CMU-Cornell iCoseg database shows that the proposed model has a promising performance.
Collapse
|
16
|
Artificial intelligence in gastrointestinal and hepatic imaging: past, present and future scopes. Clin Imaging 2022; 87:43-53. [DOI: 10.1016/j.clinimag.2022.04.007] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/17/2021] [Revised: 03/09/2022] [Accepted: 04/11/2022] [Indexed: 11/19/2022]
|
17
|
Chain Graph Explanation of Neural Network Based on Feature-Level Class Confusion. APPLIED SCIENCES-BASEL 2022. [DOI: 10.3390/app12031523] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 02/05/2023]
Abstract
Despite increasing interest in developing interpretable machine learning methods, most recent studies have provided explanations only for single instances, require additional datasets, and are sensitive to hyperparameters. This paper proposes a confusion graph that reveals model weaknesses by constructing a confusion dictionary. Unlike other methods, which focus on the performance variation caused by single-neuron suppression, it defines the role of each neuron in two different perspectives: ‘correction’ and ‘violation’. Furthermore, our method can identify the class relationships in similar positions at the feature level, which can suggest improvements to the model. Finally, the proposed graph construction is model-agnostic and does not require additional data or tedious hyperparameter tuning. Experimental results show that the information loss from omitting the channels guided by the proposed graph can result in huge performance degradation, from 91% to 33%, while the proposed graph only retains 1% of total neurons.
Collapse
|
18
|
Detection of Aerobics Action Based on Convolutional Neural Network. COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE 2022; 2022:1857406. [PMID: 35035453 PMCID: PMC8754619 DOI: 10.1155/2022/1857406] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Abstract] [Track Full Text] [Download PDF] [Figures] [Subscribe] [Scholar Register] [Received: 11/04/2021] [Revised: 12/13/2021] [Accepted: 12/16/2021] [Indexed: 11/19/2022]
Abstract
To further improve the accuracy of aerobics action detection, a method of aerobics action detection based on improving multiscale characteristics is proposed. In this method, based on faster R-CNN and aiming at the problems existing in faster R-CNN, the feature pyramid network (FPN) is used to extract aerobics action image features. So, the low-level semantic information in the images can be extracted, and it can be converted into high-resolution deep-level semantic information. Finally, the target detector is constructed by the above-extracted anchor points so as to realize the detection of aerobics action. The results show that the loss function of the neural network is reduced to 0.2 by using the proposed method, and the accuracy of the proposed method can reach 96.5% compared with other methods, which proves the feasibility of this study.
Collapse
|
19
|
Xu P, Kim K, Koh J, Wu D, Rim Lee Y, Young Park S, Young Tak W, Liu H, Li Q. Efficient knowledge distillation for liver CT segmentation using growing assistant network. Phys Med Biol 2021; 66. [PMID: 34768246 DOI: 10.1088/1361-6560/ac3935] [Citation(s) in RCA: 4] [Impact Index Per Article: 1.0] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 08/04/2021] [Accepted: 11/12/2021] [Indexed: 12/21/2022]
Abstract
Segmentation has been widely used in diagnosis, lesion detection, and surgery planning. Although deep learning (DL)-based segmentation methods currently outperform traditional methods, most DL-based segmentation models are computationally expensive and memory inefficient, which are not suitable for the intervention of liver surgery. To address this issue, a simple solution is to make a segmentation model very small for the fast inference time, however, there is a trade-off between the model size and performance. In this paper, we propose a DL-based real-time 3-D liver CT segmentation method, where knowledge distillation (KD) method, known as knowledge transfer from teacher to student models, is incorporated to compress the model while preserving the performance. Because it is well known that the knowledge transfer is inefficient when the disparity of teacher and student model sizes is large, we propose a growing teacher assistant network (GTAN) to gradually learn the knowledge without extra computational cost, which can efficiently transfer knowledge even with the large gap of teacher and student model sizes. In our results, dice similarity coefficient of the student model with KD improved 1.2% (85.9% to 87.1%) compared to the student model without KD, which is a similar performance of the teacher model using only 8% (100k) parameters. Furthermore, with a student model of 2% (30k) parameters, the proposed model using the GTAN improved the dice coefficient about 2% compared to the student model without KD, and the inference time is 13 ms per a 3-D image. Therefore, the proposed method has a great potential for intervention in liver surgery as well as in many real-time applications.
Collapse
Affiliation(s)
- Pengcheng Xu
- College of Optical Science and Engineering, Zhejiang University, Hangzhou, People's Republic of China.,Massachusetts General Hospital and Harvard Medical School, Radiology Department, 55 Fruit Street, Boston, MA 02114, United States of America
| | - Kyungsang Kim
- Massachusetts General Hospital and Harvard Medical School, Radiology Department, 55 Fruit Street, Boston, MA 02114, United States of America
| | - Jeongwan Koh
- Massachusetts General Hospital and Harvard Medical School, Radiology Department, 55 Fruit Street, Boston, MA 02114, United States of America
| | - Dufan Wu
- Massachusetts General Hospital and Harvard Medical School, Radiology Department, 55 Fruit Street, Boston, MA 02114, United States of America
| | - Yu Rim Lee
- Department of Internal Medicine, School of Medicine, Kyungpook National University, Daegu, Republic of Korea
| | - Soo Young Park
- Department of Internal Medicine, School of Medicine, Kyungpook National University, Daegu, Republic of Korea
| | - Won Young Tak
- Department of Internal Medicine, School of Medicine, Kyungpook National University, Daegu, Republic of Korea
| | - Huafeng Liu
- College of Optical Science and Engineering, Zhejiang University, Hangzhou, People's Republic of China
| | - Quanzheng Li
- Massachusetts General Hospital and Harvard Medical School, Radiology Department, 55 Fruit Street, Boston, MA 02114, United States of America
| |
Collapse
|
20
|
Vo VTT, Yang HJ, Lee GS, Kang SR, Kim SH. Effects of Multiple Filters on Liver Tumor Segmentation From CT Images. Front Oncol 2021; 11:697178. [PMID: 34660267 PMCID: PMC8517527 DOI: 10.3389/fonc.2021.697178] [Citation(s) in RCA: 1] [Impact Index Per Article: 0.3] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Download PDF] [Figures] [Journal Information] [Subscribe] [Scholar Register] [Received: 04/19/2021] [Accepted: 09/06/2021] [Indexed: 12/29/2022] Open
Abstract
Segmentation of liver tumors from Computerized Tomography (CT) images remains a challenge due to the natural variation in tumor shape and structure as well as the noise in CT images. A key assumption is that the performance of liver tumor segmentation depends on the characteristics of multiple features extracted from multiple filters. In this paper, we design an enhanced approach based on a two-class (liver, tumor) convolutional neural network that discriminates tumor as well as liver from CT images. First, the contrast and intensity values in CT images are adjusted and high frequencies are removed using Hounsfield units (HU) filtering and standardization. Then, the liver tumor is segmented from entire images with multiple filter U-net (MFU-net). Finally, a quantitative analysis is carried out to evaluate the segmentation results using three different methods: boundary-distance-based metrics, size-based metrics, and overlap-based metrics. The proposed method is validated on CT images from the 3Dircadb and LiTS dataset. The results demonstrate that the multiple filters are useful for extracting local and global feature simultaneously, minimizing the boundary distance errors, and our approach demonstrates better performance in heterogeneous tumor regions of CT images.
Collapse
Affiliation(s)
- Vi Thi-Tuong Vo
- Department of Artificial Intelligence Convergence, Chonnam National University, Gwangju, South Korea
| | - Hyung-Jeong Yang
- Department of Artificial Intelligence Convergence, Chonnam National University, Gwangju, South Korea
| | - Guee-Sang Lee
- Department of Artificial Intelligence Convergence, Chonnam National University, Gwangju, South Korea
| | - Sae-Ryung Kang
- Department of Nuclear Medicine, Chonnam National University Hwasun Hospital, Gwangju, South Korea
| | - Soo-Hyung Kim
- Department of Artificial Intelligence Convergence, Chonnam National University, Gwangju, South Korea
| |
Collapse
|
21
|
DGFAU-Net: Global feature attention upsampling network for medical image segmentation. Neural Comput Appl 2021. [DOI: 10.1007/s00521-021-05908-9] [Citation(s) in RCA: 0] [Impact Index Per Article: 0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 10/21/2022]
|
22
|
The measurement of Cobb angle based on spine X-ray images using multi-scale convolutional neural network. Phys Eng Sci Med 2021; 44:809-821. [PMID: 34251603 DOI: 10.1007/s13246-021-01032-z] [Citation(s) in RCA: 18] [Impact Index Per Article: 4.5] [Reference Citation Analysis] [Abstract] [Key Words] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Received: 01/08/2021] [Accepted: 07/08/2021] [Indexed: 12/29/2022]
Abstract
Adolescent idiopathic scoliosis (AIS) is a structural spinal deformity mainly in the coronal plane and is among the most frequent deformities in children, adolescents, and young adults, with an overall prevalence of 0.47-5.2%. The Cobb angle is an objective measure to determine the progression of deformity and plays a critical role in the planning of surgical treatment. However, existing studies suggested that Cobb angle measurement is susceptible to inter- and intra-observer variability, as well as a high variability in the definition of the end vertebra. In this study, we proposed an automatic method for the spine vertebrae segmentation using Deeplab V3+, a powerful tool that has shown success in the image segmentation of other anatomical regions but spine, and Cobb angle measurement. The segmentation performance was compared to existing mainstay neural networks. Compared to U-Net, Residual U-Net and Dilated U-Net, our method using Deeplab V3+ showed the best performance in the Dice Similarity Coefficient (DSC), accuracy, sensitivity and Jaccard Index. An excellent correlation in the final Cobb angle calculation was achieved between the smallest distance point (SDP) method and two experts (> 0.95), with a small error in the angle estimation compared (MAE < 3°). The proposed method could provide a potential tool for the automatic estimation of the Cobb angle to improve the efficiency and accuracy of the treatment workflow for AIS.
Collapse
|
23
|
Hu Y, Su F, Dong K, Wang X, Zhao X, Jiang Y, Li J, Ji J, Sun Y. Deep learning system for lymph node quantification and metastatic cancer identification from whole-slide pathology images. Gastric Cancer 2021; 24:868-877. [PMID: 33484355 DOI: 10.1007/s10120-021-01158-9] [Citation(s) in RCA: 32] [Impact Index Per Article: 8.0] [Reference Citation Analysis] [Abstract] [Key Words] [MESH Headings] [Track Full Text] [Journal Information] [Submit a Manuscript] [Subscribe] [Scholar Register] [Received: 10/16/2020] [Accepted: 01/07/2021] [Indexed: 02/07/2023]
Abstract
BACKGROUND Traditional diagnosis methods for lymph node metastases are labor-intensive and time-consuming. As a result, diagnostic systems based on deep learning (DL) algorithms have become a hot topic. However, current research lacks testing with sufficient data to verify performance. The aim of this study was to develop and test a deep learning system capable of identifying lymph node metastases. METHODS 921 whole-slide images of lymph nodes were divided into two cohorts: training and testing. For lymph node quantification, we combined Faster RCNN and DeepLab as a cascade DL algorithm to detect regions of interest. For metastatic cancer identification, we fused Xception and DenseNet-121 models and extracted features. Prospective testing to verify the performance of the diagnostic system was performed using 327 unlabeled images. We further validated the proposed system using Positive Predictive Value (PPV) and Negative Predictive Value (NPV) criteria. RESULTS We developed a DL-based system capable of automated quantification and identification of metastatic lymph nodes. The accuracy of lymph node quantification was shown to be 97.13%. The PPV of the combined Xception and DenseNet-121 model was 93.53%, and the NPV was 97.99%. Our experimental results show that the differentiation level of metastatic cancer affects the recognition performance. CONCLUSIONS The diagnostic system we established reached a high level of efficiency and accuracy of lymph node diagnosis. This system could potentially be implemented into clinical workflow to assist pathologists in making a preliminary screening for lymph node metastases in gastric cancer patients.
Collapse
Affiliation(s)
- Yajie Hu
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education), Department of Pathology, Peking University Cancer Hospital and Institute, 52 Fucheng Road, Haidian District, Beijing, 100142, China
| | - Feng Su
- Peking-Tsinghua Center for Life Sciences, Academy for Advanced Interdisciplinary Studies, Peking University, Beijing, 100871, China
| | - Kun Dong
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education), Department of Pathology, Peking University Cancer Hospital and Institute, 52 Fucheng Road, Haidian District, Beijing, 100142, China
| | - Xinyu Wang
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education), Department of Pathology, Peking University Cancer Hospital and Institute, 52 Fucheng Road, Haidian District, Beijing, 100142, China
| | - Xinya Zhao
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education), Department of Pathology, Peking University Cancer Hospital and Institute, 52 Fucheng Road, Haidian District, Beijing, 100142, China
| | - Yumeng Jiang
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education), Department of Pathology, Peking University Cancer Hospital and Institute, 52 Fucheng Road, Haidian District, Beijing, 100142, China
| | - Jianming Li
- Institute for Artificial Intelligence, The State Key Laboratory of Intelligence Technology and Systems, Beijing National Research Center for Information Science and Technology, Department of Computer Science and Technology, Tsinghua University, Beijing, 100084, China
| | - Jiafu Ji
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education), Gastrointestinal Cancer Center, Peking University Cancer Hospital and Institute, Beijing, 100142, China
| | - Yu Sun
- Key Laboratory of Carcinogenesis and Translational Research (Ministry of Education), Department of Pathology, Peking University Cancer Hospital and Institute, 52 Fucheng Road, Haidian District, Beijing, 100142, China.
| |
Collapse
|
24
|
Ieracitano C, Mammone N, Hussain A, Morabito FC. A novel explainable machine learning approach for EEG-based brain-computer interface systems. Neural Comput Appl 2021. [DOI: 10.1007/s00521-020-05624-w] [Citation(s) in RCA: 20] [Impact Index Per Article: 5.0] [Reference Citation Analysis] [Track Full Text] [Journal Information] [Subscribe] [Scholar Register] [Indexed: 01/15/2023]
|
25
|
Semantic Segmentation Framework for Glomeruli Detection and Classification in Kidney Histological Sections. ELECTRONICS 2020. [DOI: 10.3390/electronics9030503] [Citation(s) in RCA: 33] [Impact Index Per Article: 6.6] [Reference Citation Analysis] [Abstract] [Track Full Text] [Subscribe] [Scholar Register] [Indexed: 12/16/2022]
Abstract
The evaluation of kidney biopsies performed by expert pathologists is a crucial process for assessing if a kidney is eligible for transplantation. In this evaluation process, an important step consists of the quantification of global glomerulosclerosis, which is the ratio between sclerotic glomeruli and the overall number of glomeruli. Since there is a shortage of organs available for transplantation, a quick and accurate assessment of global glomerulosclerosis is essential for retaining the largest number of eligible kidneys. In the present paper, the authors introduce a Computer-Aided Diagnosis (CAD) system to assess global glomerulosclerosis. The proposed tool is based on Convolutional Neural Networks (CNNs). In particular, the authors considered approaches based on Semantic Segmentation networks, such as SegNet and DeepLab v3+. The dataset has been provided by the Department of Emergency and Organ Transplantations (DETO) of Bari University Hospital, and it is composed of 26 kidney biopsies coming from 19 donors. The dataset contains 2344 non-sclerotic glomeruli and 428 sclerotic glomeruli. The proposed model consents to achieve promising results in the task of automatically detecting and classifying glomeruli, thus easing the burden of pathologists. We get high performance both at pixel-level, achieving mean F-score higher than 0.81, and Weighted Intersection over Union (IoU) higher than 0.97 for both SegNet and Deeplab v3+ approaches, and at object detection level, achieving 0.924 as best F-score for non-sclerotic glomeruli and 0.730 as best F-score for sclerotic glomeruli.
Collapse
|